Image: The rack of a broken monitor or TV is the texture of a television screen, background for superimposing with the effect of grain noise, ????????? ?????. Adobe Stock 1050902471 Image: The rack of a broken monitor or TV is the texture of a television screen, background for superimposing with the effect of grain noise, ????????? ?????. Adobe Stock 1050902471

AI search has a citation problem

We Compared Eight AI Search Engines. They’re All Bad at Citing News.

AI search tools are rapidly gaining in popularity, with nearly one in four Americans now saying they have used AI in place of traditional search engines. These tools derive their value from crawling the internet for up-to-date, relevant information—content that is often produced by news publishers. 

Yet a troubling imbalance has emerged: while traditional search engines typically operate as an intermediary, guiding users to news websites and other quality content, generative search tools parse and repackage information themselves, cutting off traffic flow to original sources. These chatbots’ conversational outputs often obfuscate serious underlying issues with information quality. There is an urgent need to evaluate how these systems access, present, and cite news content.

Building on our previous research, the Tow Center for Digital Journalism conducted tests on eight generative search tools with live search features to assess their abilities to accurately retrieve and cite news content, as well as how they behave when they cannot.

We found that…

  • Chatbots were generally bad at declining to answer questions they couldn’t answer accurately, offering incorrect or speculative answers instead. 
  • Premium chatbots provided more confidently incorrect answers than their free counterparts.
  • Multiple chatbots seemed to bypass Robot Exclusion Protocol preferences.
  • Generative search tools fabricated links and cited syndicated and copied versions of articles. 
  • Content licensing deals with news sources provided no guarantee of accurate citation in chatbot responses.

Our findings were consistent with our previous study, proving that our observations are not just a ChatGPT problem, but rather recur across all the prominent generative search tools that we tested. 

Methodology

We systematically tested eight generative search tools: OpenAI’s ChatGPT Search, Perplexity, Perplexity Pro, DeepSeek Search, Microsoft’s Copilot, xAI’s Grok-2 and Grok-3 (beta), and Google’s Gemini. 

We chose 20 news publishers with varying stances on AI access that either permit search bots’ web crawlers via robots.txt, or block them. (The Robot Exclusion Protocol, also known as robots.txt, is a web standard that gives website publishers the option to “disallow” web crawlers–automated programs that systematically browse the internet to discover and retrieve content.) Some of the publishers we included are involved in content licensing or revenue share agreements with the AI companies, while others are pursuing lawsuits against them.

We randomly selected ten articles from each publisher, then manually selected direct excerpts from those articles for use in our queries. After providing each chatbot with the selected excerpts, we asked it to identify the corresponding article’s headline, original publisher, publication date, and URL, using the following query:

We deliberately chose excerpts that, if pasted into a traditional Google search, returned the original source within the first three results. We ran sixteen hundred queries (twenty publishers times ten articles times eight chatbots) in total. We manually evaluated the chatbot responses based on three attributes: the retrieval of (1) the correct article, (2) the correct publisher, and (3) the correct URL. According to these parameters, each response was marked with one of the following labels:

  • Correct: All three attributes were correct.
  • Correct but Incomplete: Some attributes were correct, but the answer was missing information.
  • Partially Incorrect: Some attributes were correct while others were incorrect.
  • Completely Incorrect: All three attributes were incorrect and/or missing.
  • Not Provided: No information was provided.
  • Crawler Blocked: The publisher disallows the chatbot’s crawler in its robots.txt.

Chatbots’ responses to our queries were often confidently wrong

Overall, the chatbots often failed to retrieve the correct articles. Collectively, they provided incorrect answers to more than 60 percent of queries. Across different platforms, the level of inaccuracy varied, with Perplexity answering 37 percent of the queries incorrectly, while Grok 3 had a much higher error rate, answering 94 percent of the queries incorrectly.

For the rest of this article, please use source link below

REGISTER NOW

By Klaudia Jaźwińska and Aisvarya Chandrasekar / Journalists

Klaudia Jaźwińska and Aisvarya Chandrasekar are the authors. Klaudia Jaźwińska is a journalist and researcher for the Tow Center who studies the relationship between the journalism and technology industries. Aisvarya Chandrasekar is a computational journalist at the Tow Center for Digital Journalism at Columbia University, where she studies AI use cases in journalism.

(Source: cjr.org; March 6, 2025; https://tinyurl.com/23ut7lnb)
Back to INF

Loading please wait...