Study: Google’s AI Overviews show millions of wrong answers every hour

With over 4 trillion Google queries every year, the inaccuracies add up.
Google search page with 'Is AI Overview accurate?' -- the thinking emojis are placed over the two o's in Google
Part of the issue comes from using Facebook and Reddit posts as sources. Credit: Google / Popular Science

Aside from the hallucinations, immense energy requirements, and potentially negative mental health effects, generative artificial intelligence still has an issue with accuracy. However, that hasn’t stopped major tech companies like Google from rolling out features like AI search summaries to its users. Most of the results seem fine at first glance and usually feature multiple source citations, but that doesn’t mean the product works perfectly.

A recent study reported by the New York Times found that Google’s “AI Overview” offers correct and reputably sourced summaries 9 out of 10 times. But while 90 percent sounds like a passing grade, the failure rate adds up in a matter of minutes. Given that the company will process over five trillion searches in 2026, the ensuing math means AI Overview is churning out tens of millions of questionable answers each hour. That’s hundreds of thousands of errors every minute.

What’s more, it’s difficult to truly assess AI Overview’s accuracy. An initially wrong response to a search query may transform into a correct summation when a user repeats the search a second time, making it basically impossible to anticipate. Google’s decision to place the AI Overview tab at the top of most search result pages also means more people see it and can assume its reliability.

Part of the issue relates to the specific sites treated as sources. The study’s authors at the open-source AI company Oumi found that Facebook and Reddit were the second- and fourth-most-cited references for AI Overview. Accurate answers cited Facebook 5 percent of the time, while inaccurate responses cited the social media site 7 percent of the time. In other cases, AI Overview seems to misstate a reliable source when giving a wrong answer.

Then there is the problem of bad actors. Anyone who understands these bugs in the system can potentially game AI Overview into giving inaccurate statements. Hypothetically, a person could author a series of blog posts asserting flat-out wrong historical information, then artificially boost traffic to their website. Google’s AI Overview may include the site in its source-scouring, fail to flag its inaccuracies, then generate a wrong answer.

“Our Search AI features are built on the same ranking and safety protections that block the overwhelming majority of spam from appearing in our results,” Google spokesperson Ned Adriance told the NYT. “Most of these examples are unrealistic searches that people wouldn’t actually do.”

At the very least, it’s important to view tools like AI Overview with a heavy dose of skepticism for the conceivable future. One faulty response in every 10 answers may not seem too serious, but think about how many search queries you have already made today. Studies also already indicate that overreliance on this type of tech may not be great for overall cognitive abilities. But if there is one thing you can trust, it’s AI Overview’s fine-print disclaimer:

“AI can make mistakes, so double-check responses.”

 
products on a page that says best of what's new 2025

2025 PopSci Best of What’s New

 
Andrew Paul Avatar

Andrew Paul

Staff Writer

Andrew Paul is a staff writer for Popular Science.