A BBC study reveals significant inaccuracies more than half of AI chatbot summaries of news stories.
Research conducted by the BBC has found that four prominent artificial intelligence (AI) chatbots struggled to accurately summarise news stories, raising concerns about the reliability of these emerging technologies in news publishing. The study examined OpenAI’s ChatGPT, Microsoft’s Copilot, Google’s Gemini and Perplexity AI, providing them with content sourced from the BBC website and subsequently querying them about the news.
The results highlighted “significant inaccuracies” and distortions within the chatbot responses, calling into question the validity of AI-generated summaries in the context of current events. Fifty-one percent of the responses studied had significant issues, while 19% introduced errors of some kind or other.
Deborah Turness, CEO of BBC News and Current Affairs, expressed both optimism and caution regarding the role of AI in news media. In a blog post, she noted that AI offers “endless opportunities” for innovation but cautioned against the risks involved. “We live in troubled times, and how long will it be before an AI-distorted headline causes significant real-world harm?” Turness said, pointing towards the potential dangers of misinformation proliferating through AI-generated content.
Apple recently suspended its AI-generated notifications after it introduced errors into news reports from a number of providers, including the BBC.
This inquiry into AI capabilities is timely as the news publishing industry continues to explore integrating AI-driven tools for various applications, such as content automation and predictive analytics. However, these technologies’ limitations, particularly in presenting accurate information, pose challenges that professionals in the sector must navigate.
The discrepancies in the chatbot outputs could lead to discussions about the standards required for AI-based tools in the media and, more broadly, the ethical considerations surrounding their deployment in critical areas like journalism and public communication.
Source: Noah Wire Services
- https://www.techradar.com/computing/artificial-intelligence/chatgpt-and-google-gemini-are-terrible-at-summarizing-news-according-to-a-new-study – This article corroborates the BBC study’s findings that AI chatbots like ChatGPT and Google Gemini struggle with accurately summarizing news stories, highlighting significant inaccuracies in their responses.
- https://www.width.ai/post/revolutionizing-news-summarization-exploring-the-power-of-gpt-in-zero-shot-and-specialized-tasks – This post explores the capabilities and limitations of AI models like GPT-3 in news summarization, which aligns with the discussion on AI’s role in news publishing.
- https://www.noahwire.com – This source is mentioned as the origin of the article discussing the BBC’s research on AI chatbots’ inability to accurately summarize news, though it does not provide additional external validation.
- https://www.bbc.com/news – The BBC’s news section is the source of the content used in the study, providing the context for the AI chatbots’ summarization tasks.
- https://www.apcentral.collegeboard.org/media/pdf/ap20-seminar-task-2-iwa-directions-and-stimulus-materials.pdf – This document does not directly support the article’s claims but provides a framework for analyzing and presenting information, which is relevant to the broader discussion on information accuracy and presentation.
Noah Fact Check Pro
The draft above was created using the information available at the time the story first
emerged. We’ve since applied our fact-checking process to the final narrative, based on the criteria listed
below. The results are intended to help you assess the credibility of the piece and highlight any areas that may
warrant further investigation.
Freshness check
Score:
8
Notes:
The narrative does not contain specific dates or references to outdated events, but it does not provide a clear timestamp for the study. However, the topic of AI integration in news media is current and ongoing.
Quotes check
Score:
9
Notes:
The quote from Deborah Turness, CEO of BBC News and Current Affairs, appears to be original to this context. Without further online sources, it suggests this might be a first-time use of the quote.
Source reliability
Score:
10
Notes:
The narrative originates from the BBC, a well-established and reputable news organisation known for its reliability.
Plausability check
Score:
9
Notes:
The claims about AI chatbots struggling with news summaries are plausible given the current challenges in AI technology. The narrative aligns with ongoing discussions about AI’s role in media.
Overall assessment
Verdict (FAIL, OPEN, PASS): PASS
Confidence (LOW, MEDIUM, HIGH): HIGH
Summary:
The narrative appears to be current and relevant, with a reliable source in the BBC. The quote from Deborah Turness seems original, and the claims about AI chatbots are plausible. Overall, the information presented is trustworthy and aligns with current discussions in the media industry.


