Categories
en Uncategorized

Why AI like ChatGPT still quotes retracted papers?

retracted studies

AI models like ChatGPT are trained on massive datasets collected at specific moments in time, which means they lack awareness of papers retracted after their training cutoff. When a scientific paper gets retracted, whether due to errors, fraud, or ethical violations, most AI systems continue referencing it as if nothing happened. This creates a troubling scenario where researchers using AI assistants might unknowingly build their work on discredited foundations.

In other words: retracted papers are the academic world’s way of saying “we got this wrong, please disregard.” Yet the AI tools designed to help us navigate research faster often can’t tell the difference between solid science and work that’s been officially debunked.

ChatGPT and other assistants tested

Recent studies examined how popular AI research tools handle retracted papers, and the results were concerning. Researchers tested ChatGPT, Google’s Gemini, and similar language models by asking them about known retracted papers. In many cases, they not only failed to flag the retractions but actively praised the withdrawn studies.

One investigation found that ChatGPT referenced retracted cancer imaging research without any warning to users, presenting the flawed findings as credible. The problem extends beyond chatbots to AI-powered literature review tools that researchers increasingly rely on for efficiency.

Common failure scenarios

The risks show up across different domains, each with its own consequences:

  • Medical guidance: Healthcare professionals consulting AI for clinical information might receive recommendations based on studies withdrawn for data fabrication or patient safety concerns
  • Literature reviews: Academic researchers face citation issues when AI assistants suggest retracted papers, damaging credibility and delaying peer review
  • Policy decisions: Institutional leaders making evidence-based choices might rely on AI-summarised research without realising the underlying studies have been retracted

A doctor asking about treatment protocols could unknowingly follow advice rooted in discredited research. Meanwhile, detecting retracted citations manually across hundreds of references proves nearly impossible for most researchers.

How Often Retractions Slip Into AI Training Data

The scale of retracted papers entering AI systems is larger than most people realise. Crossref, the scholarly metadata registry that tracks digital object identifiers (DOIs) for academic publications, reports thousands of retraction notices annually. Yet many AI models were trained on datasets harvested years ago, capturing papers before retraction notices appeared.

Here’s where timing becomes critical. A paper published in 2020 and included in an AI training dataset that same year might get retracted in 2023. If the model hasn’t been retrained with updated data, it remains oblivious to the retraction. Some popular language models go years between major training updates, meaning their knowledge of the research landscape grows increasingly outdated.

Lag between retraction and model update

Training Large Language Models requires enormous computational resources and time, which explains why most AI companies don’t continuously update their systems. Even when retraining occurs, the process of identifying and removing retracted papers from massive datasets presents technical challenges that many organisations haven’t prioritised solving.

The result is a growing gap between the current state of scientific knowledge and what AI assistants “know.” You might think AI systems could simply check retraction databases in real-time before responding, but most don’t. Instead, they generate responses based solely on their static training data, unaware that some information has been invalidated.

Risks of Citing Retracted Papers in Practice

The consequences of AI-recommended retracted papers extend beyond embarrassment. When flawed research influences decisions, the ripple effects can be substantial and long-lasting.

Clinical decision errors

Healthcare providers increasingly turn to AI tools for quick access to medical literature, especially when facing unfamiliar conditions or emerging treatments. If an AI assistant recommends a retracted study on drug efficacy or surgical techniques, clinicians might implement approaches that have been proven harmful or ineffective. The 2020 hydroxychloroquine controversy illustrated how quickly questionable research spreads. Imagine that dynamic accelerated by AI systems that can’t distinguish between valid and retracted papers.

Policy and funding implications

Government agencies and research institutions often use AI tools to synthesise large bodies of literature when making funding decisions or setting research priorities. Basing these high-stakes choices on retracted work wastes resources and potentially misdirects entire fields of inquiry. A withdrawn climate study or economic analysis could influence policy for years before anyone discovers the AI-assisted review included discredited research.

Academic reputation damage

For individual researchers, citing retracted papers carries professional consequences. Journals may reject manuscripts, tenure committees question research rigour, and collaborators lose confidence. While honest mistakes happen, the frequency of such errors increases when researchers rely on AI tools that lack retraction awareness, and the responsibility still falls on the researcher, not the AI.

Why Language Models Miss Retraction Signals

The technical architecture of most AI research assistants makes them inherently vulnerable to the retraction problem. Understanding why helps explain what solutions might actually work.

Corpus quality controls lacking

AI models learn from their training corpus, the massive collection of text they analyse during development. Most organisations building these models prioritise breadth over curation, scraping academic databases, preprint servers, and publisher websites without rigorous quality checks.

The assumption is that more data produces better models, but this approach treats all papers equally regardless of retraction status. Even when training data includes retraction notices, the AI might not recognise them as signals to discount the paper’s content. A retraction notice is just another piece of text unless the model has been specifically trained to understand its significance.

Sparse or inconsistent metadata

Publishers handle retractions differently, creating inconsistencies that confuse automated systems:

  • Some journals add “RETRACTED” to article titles
  • Others publish separate retraction notices
  • A few quietly remove papers entirely

This lack of standardisation means AI systems trained to recognise one retraction format might miss others completely. Metadata، the structured information describing each paper, often fails to consistently flag retraction status across databases. A paper retracted in PubMed might still appear without warning in other indexes that AI training pipelines access.

Hallucination and overconfidence

AI hallucination occurs when models generate plausible-sounding but false information, and it exacerbates the retraction problem. Even if a model has no information about a topic, it might confidently fabricate citations or misremember details from its training data. This overconfidence means AI assistants rarely express uncertainty about the papers they recommend, leaving users with no indication that additional verification is needed.

Real-Time Retraction Data Sources Researchers Should Trust

While AI tools struggle with retractions, several authoritative databases exist for manual verification. Researchers concerned about citation integrity can cross-reference their sources against these resources.

Retraction Watch Database

Retraction Watch operates as an independent watchdog, tracking retractions across all academic disciplines and publishers. Their freely accessible database includes detailed explanations of why papers were withdrawn, from honest error to fraud. The organisation’s blog also provides context about patterns in retractions and systemic issues in scholarly publishing.

Crossref metadata service

Crossref maintains the infrastructure that assigns DOIs to scholarly works, and publishers report retractions through this system. While coverage depends on publishers properly flagging retractions, Crossref offers a comprehensive view across multiple disciplines and publication types. Their API allows developers to build tools that automatically check retraction status, a capability that forward-thinking platforms are beginning to implement.

PubMed retracted publication tag

For medical and life sciences research, PubMed provides reliable retraction flagging with daily updates. The National Library of Medicine maintains this database with rigorous quality control, ensuring retracted papers receive prominent warning labels. However, this coverage is limited to biomedical literature, leaving researchers in other fields without equivalent resources.

DatabaseCoverageUpdate SpeedAccess
Retraction WatchAll disciplinesReal-timeFree
CrossrefPublisher-reportedVariableFree API
PubMedMedical/life sciencesDailyFree


Responsible AI Starts with Licensing

When AI systems access research papers, articles, or datasets, authors and publishers have legal and ethical rights that need protection. Ignoring these rights can undermine the sustainability of the research ecosystem and diminish trust between researchers and technology providers.

One of the biggest reasons AI tools get it wrong is that they often cite retracted papers as if they’re still valid. When an article is retracted, e.g. due to peer review process not being conducted properly or failing to meet established standards, most AI systems don’t know, it simply remains part of their training data. This is where licensing plays a crucial role. Licensed data ensures that AI systems are connected to the right sources, continuously updated with accurate, publisher-verified information. It’s the foundation for what platforms like Zendy aim to achieve: making sure the content is clean and trustworthy. 

Licensing ensures that content is used responsibly. Proper agreements between AI companies and copyright holders allow AI systems to access material legally while providing attribution and, when appropriate, compensation. This is especially important when AI tools generate insights or summaries that are distributed at scale, potentially creating value for commercial platforms without benefiting the sources of the content.

in conclusion, consent-driven licensing helps build trust. Publishers and authors can choose whether and how their work is incorporated into AI systems, ensuring that content is included only when rights are respected. Advanced AI platforms, such as Zendy, can even track which licensed sources contributed to a particular output, providing accountability and a foundation for equitable revenue sharing.

Categories
en Uncategorized

Zendy Sponsors the 4th Annual Forum for Open Research in MENA

form

This October, American University of Sharjah will host the 4th Annual Forum for Open Research in MENA (FORM), a leading regional event dedicated to advancing open access, collaboration, and community capacity building across the Arab world. Zendy is proud to be among the official sponsors of this year’s Forum, reinforcing our ongoing commitment to making research more accessible and equitable for all.

Organised by the non-profit Forum for Open Research in MENA, the fourth edition will bring together:

  • 26 sessions across four days
  • 88 expert speakers
  • Representation from 60 global institutions
  • Delegates from 27 countries
  • Dozens of formal and informal networking opportunities, connecting thought leaders, practitioners, and advocates throughout the MENA region

Participants will explore strategies to strengthen open science practices, build sustainable infrastructures, and promote shared learning across borders.

A Shared Vision for Accessible Knowledge

Zendy’s mission to make academic knowledge affordable and accessible worldwide aligns closely with FORM’s goal of fostering open, collaborative research communities in the region. As a sponsor, Zendy supports initiatives that not only expand access to scholarly literature but also empower researchers, educators, and students to participate in a more transparent and connected research ecosystem.

Hosted by the American University of Sharjah

This year’s Forum, hosted by the American University of Sharjah, arrives in the UAE under the national theme, “The Year of Community.” The theme underscores the importance of collective progress and shared learning—values that sit at the heart of both open science and Zendy’s approach to research discovery.

Theme Spotlight: “Becoming Open—Capacity Building and Community Collaboration”

The 2025 theme, “Becoming Open,” highlights the human side of open science: community collaboration, capacity building, and sustainable growth. Through workshops, panels, and discussions, the Annual Forum will address how regional institutions can implement open research policies, share resources effectively, and strengthen local research infrastructures.

Join us in Sharjah from October 20–23, 2025, as we celebrate and support the growing movement toward open science across the Arab world.

The registration is now open, visit https://forumforopenresearch.com/registration/

Categories
en Uncategorized

We’re heading back to Frankfurter Buchmesse 2025

We’re excited to return to Frankfurt this October for the world’s largest book fair, and this year, we’re stepping onto the Innovation Stage with a timely conversation on how artificial intelligence can bring ethical, lasting value to scholarly publishing.

Join us on October 15 at 12:30 p.m. (CEST), Hall 4.0 Innovation Stage
Our Co-Founder, Kamran Kardan, will be joined by Josh Nicholson (Scite) and Julie Atkins (Bristol University Press & Policy Press) for a panel discussion on:

“AI and Ethical Innovation in Publishing: Driving Value for Publishers”

The session will explore how AI can strengthen collaboration between publishers, platforms, librarians and researchers, ensuring research is more transparent, trustworthy, and accessible worldwide.

Some of the themes include:

  • How publishers can benefit from Zendy’s Revenue-Share Model, which compensates them whenever their content is referenced by AI assistants.
  • The role of Scite’s smart citation system in helping users evaluate whether a paper supports, contradicts, or builds upon previous work.
  • Why responsible use of technologies like Retrieval-Augmented Generation (RAG) can combat misinformation and provide more reliable insights.
  • How global equity and inclusivity can remain at the center of AI adoption in research.

Together, these discussions will also link to broader goals, from advancing quality education to supporting sustainable innovation and global partnerships.

Speakers

  • Kamran Kardan, Founder & CEO Knowledge E, Co-Founder Zendy
  • Josh Nicholson, Co-founder of Scite, CSO at Research Solutions
  • Moderator: Sara Crowley Vigneau, Partnership Relations Manager, Zendy

And of course, the Zendy team will be around throughout the fair! Find us at Stand G97, Hall 4.0, where Kamran Kardan, Lisette van Kessel (Head of Marketing), and Sara Crowley Vigneau (Partnership Manager) will be happy to talk about our ethical AI tools for research, open access publishing, and higher education librarynesp programs.

Want to set up a one-on-one meeting? Reach out to l.vankessel@knowledgee.com.

For more details about the fair, visit Frankfurter Buchmesse.

Categories
en

5 Tools Every Librarian Should Know in 2025

ai for libraries

The role of librarians has always been about connecting people with knowledge. But in 2025, with so much information floating around online, the challenge isn’t access, it’s sorting through the noise and finding what really matters. This is where AI for libraries is starting to make a difference. Here are five that are worth keeping in your back pocket this year.

1. Zendy

Zendy is a one-stop AI-powered research library that blends open access with subscription-based resources. Instead of juggling multiple platforms, librarians can point students and researchers to one place where they’ll find academic articles, reports, and AI tools to help with research discovery and literature review. With its growing use of AI for libraries, Zendy makes it easier to summarise research, highlight key ideas, and support literature reviews without adding to the librarian’s workload.

2. LibGuides

Still one of the most practical tools for librarians, LibGuides makes it easy to create tailored resource guides for courses, programs, or specific assignments. Whether you’re curating resources for first-year students or putting together a subject guide for advanced research, it helps librarians stay organised while keeping information accessible to learners.

3. OpenRefine

Cleaning up messy data is nobody’s favourite job, but it’s a reality when working with bibliographic records or digital archives. OpenRefine is like a spreadsheet, but with superpowers, it can quickly detect duplicates, fix formatting issues, and make large datasets more manageable. For librarians working in cataloguing or digital collections, it saves hours of tedious work.

4. PressReader

Library patrons aren’t just looking for academic content; they often want newspapers, magazines, and general reading material too. PressReader gives libraries a simple way to provide access to thousands of publications from around the world. It’s especially valuable in public libraries or institutions with international communities.

5. OCLC WorldShare

Managing collections and sharing resources across institutions is a constant task. OCLC WorldShare helps libraries handle cataloguing, interlibrary loans, and metadata management. It’s not flashy, but it makes collaboration between libraries smoother and ensures that resources don’t sit unused when another community could benefit from them.

Final thought

The tools above aren’t just about technology, they’re about making everyday library work more practical. Whether it’s curating resources with Zendy, cleaning data with OpenRefine, or sharing collections through WorldShare, these platforms help librarians do what they do best: guide people toward knowledge that matters.

Categories
en

Balancing AI Efficiency with Human Expertise in Libraries

AI in libraries

AI in libraries is making some tasks quicker and less repetitive. However, even with these advances, there’s something irreplaceable about a librarian’s judgment and care. The real question isn’t whether AI will take over libraries, it’s how both AI and librarians can work side by side.

How AI Helps in Libraries

According to Clarivate Pulse of the Library 2025 survey, among 2,000 academic library professionals globally, many said they don’t have enough time or budget to learn new tools or skills, a challenge made even harder as global digital content is projected to double every two years.

Here’s where AI tools for librarians prove useful:

  • Cataloguing: AI can scan metadata and suggest subject tags in minutes.
  • Search: Smarter search systems help students and researchers find relevant materials without digging through dozens of irrelevant results.
  • Day-to-day tasks: Think overdue notices, compiling basic reading lists, or identifying key sources and trends to support literature reviews. This is where library automation with AI comes in handy.

Instead of replacing people, these tools free up time. A librarian who doesn’t have to spend hours sorting through data can focus on supporting students, curating collections, analysing usage statistics to make informed decisions or tracking resource usage against budgets.

Where Human Expertise Still Matters

AI is fast, but it’s not thoughtful. A student asking, “I’m researching migration patterns in 19th-century Europe, where do I start?” gets much more from a librarian than from a search algorithm. Librarians bring context, empathy, and critical thinking that machines can’t replicate.

This is why human-AI collaboration in libraries makes sense. AI takes care of the routine. Humans bring the nuance. Together, they cover ground neither could manage alone.

Finding the Balance

So how do libraries get this balance right? A few ideas:

  1. Think of AI as a helper – not a replacement for staff.
  2. Invest in training – librarians need to feel confident using AI tools and knowing when not to rely on them.
  3. Keep the focus on people – the goal isn’t efficiency for its own sake, it’s about better service for students, researchers, and communities.

Final Thoughts

By using AI to handle routine administrative tasks like cataloguing, managing records, or tracking resource usage, librarians free up time to focus on the part of the job that drew them to this profession in the first place: supporting researchers and students, curating meaningful collections, and fostering learning. Combining the efficiency of AI in libraries with the expertise of librarians creates a future where technology supports the human side of education.

Categories
en

How AI in Higher Education Is Helping Libraries Support Research

AI in Higher Education

Libraries have always been at the centre of knowledge in higher education. Beyond curating collections, librarians guide researchers and students through complex databases, teach research skills, and help faculty navigate publishing requirements. They also play a key role in managing institutional resources, preserving archives, and ensuring equitable access to information. These days, libraries are facing new challenges: huge amounts of digital content, tighter budgets, and more demand for remote access. In this environment, AI in higher education is starting to make a real difference.

How AI Makes Life Easier for Librarians

Improving Discovery

AI-powered search tools don’t just look for keywords, they can understand the context of a query. That means students and researchers can find related work they might otherwise miss. It’s like having an extra set of eyes to point them toward useful sources.

Helping with Curation

AI can go through thousands of articles and highlight the ones most relevant to a specific course, project, or research topic. For example, a librarian preparing a reading list for a history class can save hours by letting AI suggest the most relevant papers or reports.

Supporting Remote Access

Students, researchers and faculty aren’t always on campus. AI can summarise long articles, translate content, or adjust resources for different reading levels. This makes it easier for people to get the information they need, even from home.

Working Within Budgets

Subscriptions remain a major expense for libraries, and ongoing budget cuts are forcing many academic institutions to make difficult choices about which resources to keep or cancel. For example, recent surveys show that around 73% of UK higher education libraries are making budget cuts this year, sometimes slashing up to 30% of their overall budgets, and collectively spending £51 million less than the previous year. This trend is not limited to the UK, universities in the U.S. and elsewhere are also reducing library funding, which has dropped by nearly 20% per student over recent years. Even top institutions like Princeton have cut library hours and student staffing to save on costs.

Subscriptions can be expensive, and libraries often have to make tough choices. AI tools that work across large collections help libraries give students and researchers more access without adding extra subscriptions.

Trusted Content Still Matters

AI is helpful, but the resources behind it are just as important. Librarians care about trusted, peer-reviewed, and varied sources.

Librarians and AI: A Partnership

AI isn’t replacing librarians. Instead, it supports the work they already do. Librarians are the ones who guide researchers, check the quality of sources, and teach information skills. By using AI tools, librarians can make research easier for students, researchers and faculty, and they can help their institutions make the most of the resources they have.

Final Thoughts

AI in higher education is making it easier for libraries to support students and faculty, but librarians are still at the centre of the process. By using AI tools alongside strong content collections, libraries can save time, offer more resources, and help researchers find exactly what they need. With the right AI support, research becomes easier to navigate and more accessible without overcomplicating the process.