Is Gen AI a Trustworthy News Source?
As generative AI models start to replace traditional search engines, how good are they at reliably reporting news content? A major new study reveals serious shortcomings and widespread user complacency.
As we reported earlier this year, gen AI search engines are gaining popularity, with well over half of U.S. search queries now resulting in zero clicks. The market predicts gen AI search is here to stay, with All About AI reporting gen AI search engines are “projected to capture 62.2% of total search volume by 2030.”
However, an extensive, international study led by the BBC recently revealed a staggering 45% of gen AI search responses “had at least one significant issue.”
What does this mean for the growing number of users who rely on AI-generated responses to gain accurate information? Let’s take a closer look at the BBC’s findings and explore the nature of trust.
The current state of gen AI news
Drawn to the instantaneous results promised by gen AI summaries, an increasing number of people are getting their news from AI assistants like ChatGPT.
Citing Reuters’ 2025 Digital News Report, the BBC reports, “7% of total online news consumers use AI assistants to get their news.”
Unsurprisingly, digital natives are leading the charge. A study conducted by Salesforce found 70% of Gen Z are using gen AI. However, of that 70%, only 52% reported they “trust the technology to help them make informed decisions,” suggesting younger users are aware that gen AI search responses might be too good to be true.
Assessing the accuracy of gen AI search responses
In a coordinated effort with the European Broadcasting Union (EBU), the BBC worked with 22 public media organizations across 18 countries to determine the accuracy of over 3,000 gen AI search responses from the top four engines: ChatGPT, Copilot, Gemini, and Perplexity.
Assessing the gen AI responses “against key criteria, including accuracy, sourcing, distinguishing opinion from fact, and providing context,” the journalists found:
- 45% of all AI answers had at least one significant issue.
- 31% “showed serious sourcing problems – missing, misleading, or incorrect attributions.
- 20% contained major accuracy issues, including hallucinated details and outdated information.
- Gemini performed worst with significant issues in 76% of responses, more than double the other assistants, largely due to its poor sourcing performance.
- Comparison between the BBC’s results earlier this year and this study show some improvements but still high levels of error.
But do readers care?
With this mountain of evidence confirming suspicions that gen AI engines are indeed peddling misinformation, the question then shifted to the users. Do people still care enough about accuracy to give up the convenience and expediency of gen-AI news summaries? The BBC partnered with global market research company Ipsos to find out.
In Audience Use and Perceptions of AI Assistants for News , the BBC reports that 47% of U.K. adults consider AI-generated “news summaries helpful for understanding complex topics,” and over a third “trust AI to produce accurate summaries of information.”
Trust in gen AI news content is tenuous
But they also discovered that for the vast majority of users, trust is a tenuous and revocable thing that can be withdrawn at the first sign of unreliability.
“84% said a factual error would have a major impact on their trust in an AI summary, with 76% saying the same about errors of sourcing and attribution. This was also high for errors where AI presented opinion as fact (81%) and introduced an opinion itself (73%).”
The BBC also affirms that the erosion of trust has an immediate and measurable impact on user behavior. “After being made aware that summaries may contain mistakes, those who instinctively disagreed with ‘I trust Gen AI to summarise news for me’ rose by 11 percentage points. 45% said they’d be less likely to use Gen AI to ask about the news in future, rising to 50% among those aged 35+.”
Can AI-generated news content regain user trust?
In response to these findings, the BBC published the News Integrity in AI Assistants Toolkit , which includes a comprehensive breakdown of the major failures of gen AI search engines and identifies “four key components of a good AI assistant response”:
- Accuracy: is the information provided by the AI assistant correct?
- Context: is the AI assistant providing all relevant and necessary information?
- Distinguishing opinion from fact: is the AI assistant clear whether the information it is providing is fact or opinion?
- Sourcing: is the AI assistant clear and accurate about where the information it provides comes from?
If gen AI engines can produce content that satisfyingly answers the questions above, then perhaps AI-generated news summaries can meet the lofty expectations fueled by the bullish AI market.
But if this state of misinformation persists, and current user attitudes prevail, gen AI-powered search will run the very real risk of reputational collapse. And as more institutions sound the alarm on gen AI’s systemic failures, the window for meaningful reform is closing.
FAQ
To what extent are AI search engines replacing traditional search?
AI-powered search engines are gaining significant market share. Over half of U.S. search queries now result in zero clicks, and projections suggest AI search engines will capture 62.2% of total search volume by 2030.
How many people use AI assistants for news?
According to Reuters' 2025 Digital News Report, approximately 7% of online news consumers currently use AI assistants like ChatGPT to get their news. This number is higher among younger users, with 70% of Gen Z reported to be using generative AI.
What did the BBC study reveal about AI news accuracy?
The BBC's international study, conducted with 22 public media organizations across 18 countries, found that 45% of AI-generated news responses had at least one significant issue when tested across ChatGPT, Copilot, Gemini, and Perplexity.
What types of errors appear in AI news summaries?
The study identified several error types: 31% showed serious sourcing problems including missing or incorrect attributions, 20% contained major accuracy issues such as hallucinated details and outdated information, and many failed to distinguish opinion from fact.
Which AI assistant performed worst in the study?
Gemini performed worst with significant issues in 76% of responses, more than double the error rate of other assistants. Its poor performance was largely attributed to sourcing problems.
Do users trust AI-generated news summaries?
Trust levels vary. About 47% of U.K. adults find AI news summaries helpful for understanding complex topics, and over a third trust AI to produce accurate summaries. However, only 52% of Gen Z users who use AI trust it to help them make informed decisions.
What happens when users discover errors in AI news?
User trust is highly fragile. When made aware that summaries may contain mistakes, 45% of users said they'd be less likely to use AI for news in the future, rising to 50% among those aged 35 and older.
What are the four key components of a good AI news response?
According to the BBC's News Integrity in AI Assistants Toolkit, the four essential components are: accuracy of information, proper context, clear distinction between opinion and fact, and transparent sourcing.
Has AI news accuracy improved over time?
The BBC study shows some improvements compared to earlier results from the same year, but error rates remain high, with 45% of responses still containing at least one significant issue.
Why are people drawn to AI news summaries?
Users are attracted to the convenience and speed of AI-generated summaries, which provide instantaneous results without requiring them to read multiple articles or sources.
Can AI-generated news regain user trust?
Trust recovery is possible if AI engines can consistently meet the four key criteria: accuracy, proper context, clear distinction between fact and opinion, and transparent sourcing. However, the window for meaningful reform is closing as more institutions highlight systemic failures.
What risk does persistent misinformation pose to AI search?
If current issues persist, AI-powered search faces the real risk of reputational collapse. As more organizations sound alarms about systemic failures and user awareness grows, the credibility of AI news sources could severely decline.
