AI Search: Lies and Fabrications

The Illusion of Accuracy

The original promise of search engines was to connect users with credible and reliable sources of information. This fundamental promise is now being severely compromised. AI-powered search tools, in their quest for speed and efficiency, are increasingly prioritizing rapid responses over factual accuracy. They generate answers that appear confident and authoritative, but often lack the crucial backing of verifiable evidence. We are witnessing a concerning shift: from a system designed to guide users to reliable information, to one that actively manufactures responses, frequently disregarding their truthfulness.

This isn’t about occasional, isolated errors. It’s a systemic, deeply ingrained problem. The Columbia Journalism Review (CJR) study demonstrates that AI search engines aren’t just making mistakes; they are actively constructing a reality that is detached from verifiable sources. They scrape content from across the internet, but instead of directing users to the original sources – the websites, news organizations, and researchers who painstakingly produce and publish information – they provide instant, often fabricated, answers.

The Traffic Drain and the Phantom Citations

The consequences of this approach are far-reaching and deeply damaging to the online information ecosystem. The immediate and most visible impact is a significant reduction in traffic to the original sources of information. Websites, news organizations, and researchers who invest time, effort, and resources in creating content are finding themselves bypassed. Users are getting their answers directly from the AI, eliminating the need to visit the sites that originated the information.

A separate study corroborates this alarming trend, revealing that click-through rates from AI-generated search results and chatbots are substantially lower than those from traditional search engines like Google. This signifies that the lifeblood of online content – the ability to reach an audience and gain visibility – is being slowly choked off.

However, the problem extends far beyond reduced traffic. These AI tools are not just failing to properly credit sources; they are often creating phantom citations. They generate links to non-existent webpages, or to URLs that are broken, irrelevant, or lead to completely different content. This is analogous to a student writing a research paper and inventing sources to support their claims. It’s not just sloppy or careless; it’s a fundamental breach of intellectual honesty and academic integrity.

A Deep Dive into the Deception

The CJR study meticulously analyzed the performance of several leading AI search models, providing a detailed and disturbing picture of the extent of the problem. The findings are deeply concerning. More than half of the citations generated by Google’s Gemini and xAI’s Grok 3 – two prominent and widely used players in the AI search landscape – led to fabricated or inaccessible webpages. This is not a minor glitch or an occasional error; it’s a systemic failure that undermines the very foundation of reliable information retrieval.

The problem extends beyond citations. Chatbots, in general, were found to deliver incorrect information in a staggering more than 60% of cases. Among the models evaluated, Grok 3 stood out as the worst offender, with a shocking 94% of its responses containing inaccuracies. Gemini, while performing slightly better, still managed to provide a fully correct answer only once in every ten attempts. Even Perplexity, which emerged as the most accurate of the models tested, still returned incorrect responses 37% of the time.

These numbers are not mere statistics; they represent a fundamental breakdown in the reliability of information provided by these AI-powered tools. They suggest that the very tools designed to help us navigate the complexities of the digital world and access accurate information are, in fact, leading us astray, providing a distorted and often fabricated view of reality.

Ignoring the Rules: The Robot Exclusion Protocol

The study’s authors unearthed another troubling and ethically questionable aspect of this AI-driven deception. Several of the AI models appeared to be deliberately disregarding the Robot Exclusion Protocol. This protocol is a standard, widely adopted mechanism that allows websites to control which parts of their site can be accessed and scraped by automated bots, including search engine crawlers. It’s a way for websites to protect their content, manage how it’s used, and prevent overloading their servers.

The fact that AI search engines are ignoring this protocol raises serious ethical questions about their respect for the rights of content creators and their willingness to exploit online information without permission. This behavior undermines the very foundations of the web, which relies on a delicate balance between access to information and the protection of intellectual property rights. It suggests a disregard for the established norms and rules that govern the online ecosystem.

Echoes of Past Warnings

The CJR study’s findings are not isolated incidents or unexpected revelations. They resonate with and reinforce the findings of a previous study published in November 2024, which focused specifically on ChatGPT’s search capabilities. That earlier investigation revealed a consistent and disturbing pattern of confident but incorrect responses, misleading citations, and unreliable information retrieval. In other words, the problems identified by the CJR are not new; they are persistent, systemic, and deeply ingrained in the way these AI models operate.

The Erosion of Trust and Agency

Experts in the field of information science and AI ethics have been sounding the alarm about the dangers of generative AI for some time. Critics like Chirag Shah and Emily M. Bender have raised serious concerns that AI search engines are eroding user agency, amplifying biases in information access, and frequently presenting misleading or even toxic answers that users may accept without question or critical evaluation.

The core issue lies in the fact that these AI models are designed to sound authoritative and convincing, even when the information they provide is inaccurate, incomplete, or entirely fabricated. They are trained on vast datasets of text and code, and they are capable of generating responses that mimic human language with remarkable fluency and sophistication. However, this fluency can be deceptive. It can mask the fact that the underlying information is flawed, fabricated, or simply incorrect, leading users to believe they are receiving accurate and reliable information when they are not.

The Mechanics of the Misinformation

The CJR study involved a detailed and rigorous analysis of 1,600 queries, specifically designed to compare how different generative AI search models retrieved and presented information. The researchers focused on key elements such as headlines, publishers, publication dates, and URLs, ensuring a comprehensive evaluation of the models’ performance. They tested a range of models, including ChatGPT Search, Microsoft CoPilot, DeepSeek Search, Perplexity (and its Pro version), xAI’s Grok-2 and Grok-3 Search, and Google Gemini.

The testing methodology was meticulously designed to ensure accuracy and reliability. The researchers used direct excerpts from ten randomly selected articles, sourced from 20 different publishers. This approach ensured that the queries were based on real-world content and that the models were being evaluated on their ability to accurately retrieve and represent that content, rather than generating responses based on potentially flawed or outdated information.

The results, as detailed earlier, paint a grim and concerning picture of the current state of AI-driven search. The tools that are increasingly becoming our primary gateways to information and knowledge are demonstrably unreliable, prone to fabrication and hallucination, and often disrespectful of the very sources they rely on for their information.

The Implications for the Future of Information

The implications of this widespread misinformation and the erosion of trust in AI-powered search are profound and far-reaching. If we cannot trust the tools we use to find information, how can we make informed decisions about our lives, our communities, and our world? How can we engage in meaningful and productive debate on important issues? How can we hold power accountable and ensure transparency and integrity in our institutions?

The rise of AI-powered search, with its inherent flaws, biases, and tendency to fabricate information, poses a significant threat to the very fabric of our information ecosystem. It undermines the credibility of news organizations, researchers, and other content creators who strive to provide accurate and reliable information. It erodes public trust in institutions, including the media, government, and academia. And it empowers those who seek to spread disinformation, propaganda, and manipulate public opinion for their own gain.

The challenge before us is not simply to improve the accuracy of AI search engines through technical fixes or algorithmic adjustments. It’s to fundamentally rethink the way we approach the search for information in the digital age. We need to prioritize transparency, accountability, and respect for the sources of information. We need to develop tools and strategies that empower users to critically evaluate the information they encounter online, distinguish between reliable and unreliable sources, and identify potential biases and misinformation.

We need to foster a culture of skepticism and critical thinking, where we are not simply passive recipients of information, but active participants in the pursuit of truth and knowledge. We must demand greater transparency from the companies that develop and deploy these AI models, holding them accountable for the accuracy and reliability of the information they provide. The future of informed discourse, democratic participation, and perhaps even democracy itself, depends on our ability to address this growing crisis of misinformation in the digital age.


The crisis of misinformation in AI-powered search is not just a technical problem to be solved by engineers and developers; it’s a societal problem that demands a multifaceted and collaborative response. It requires the involvement of not just technologists, but also journalists, educators, policymakers, researchers, and the public at large. We must collectively work to build a more reliable, trustworthy, and transparent information ecosystem, one that serves the needs of informed citizens and promotes the public good, rather than the purveyors of falsehoods and manipulation.


The current trajectory is unsustainable and deeply concerning. If AI search continues to prioritize speed and convenience over accuracy and truth, we risk creating a world where misinformation reigns supreme, where the very notion of objective reality becomes increasingly elusive, and where informed decision-making becomes impossible. The stakes are simply too high to allow this to happen. We must act now to address this crisis and ensure a future where reliable information is accessible to all.