| United Kingdom
| politics
| ✓ Verified - theguardian.com
‘Very dangerous’: a Mind mental health expert on Google’s AI Overviews
#Google AI Overviews#Mind charity#Mental health misinformation#AI search technology#Digital wellbeing#Tech regulation#Information safety#Rosie Weatherley
📌 Key Takeaways
Mind launches year-long commission to examine AI and mental health
Google's AI Overviews present harmful inaccuracies as facts
Testing revealed multiple dangerous mental health misinformation examples
Google's reactive approach to fixing AI problems is insufficient
📖 Full Retelling
Mind, England and Wales' largest mental health charity, has launched a year-long commission to examine AI and mental health on February 20, 2026, following a Guardian investigation revealing that Google's AI Overviews provided 'very dangerous' mental health advice to its 2 billion monthly users. The commission comes after Rosie Weatherley, information content manager at Mind, documented how Google's AI-generated summaries, which appear above search results on the world's most visited website, present harmful inaccuracies as uncontroversial facts. Weatherley explained that over three decades, Google's search engine allowed credible health content to rise to the top of results, but AI Overviews have replaced this richness with clinical-sounding summaries that create an illusion of definitiveness, prematurely ending information-seeking journeys with incomplete answers. In just two minutes of testing with queries commonly used by people with mental health problems, Weatherley and her team encountered multiple instances of harmful misinformation, including assurances that starvation is healthy, claims that mental health problems are caused solely by chemical imbalances, and suggestions that 60% of benefit claims for mental health conditions are malingering.
🏷️ Themes
Mental health, AI technology, Information accuracy, Corporate responsibility
AI Overviews is an artificial intelligence (AI) feature integrated into Google Search that produces AI-generated summaries of search results. The feature has been criticized for its accuracy and for reducing traffic to content websites.
Mind is a mental health charity in England and Wales. It was founded in 1946 as the National Association for Mental Health (NAMH).
Mind offers information and advice to people with mental health problems and lobbies government and local authorities on their behalf.
Researchers from fields like psychology, sociology, anthropology, and medicine have studied the relationship between digital media use and mental health since the mid-1990s, following the rise of the World Wide Web and text messaging. Much research has focused on patterns of excessive use, often cal...
Google's AI Overviews are shown to 2 billion people each month, yet they can present harmful mental health advice as facts. This risks misinforming vulnerable users and undermining trust in online health information. The issue highlights the need for stricter oversight of AI-generated content.
Context & Background
Google's AI Overviews summarize search results
Mind is the largest mental health charity in England and Wales
Guardian investigation exposed inaccuracies
AI Overviews can give false medical advice
Mind has launched a commission to investigate
What Happens Next
Mind's commission will review AI practices and recommend safeguards. Google may face pressure to improve content accuracy and transparency. The outcome could influence policy on AI-generated health information.
Frequently Asked Questions
What are AI Overviews?
Short AI-generated summaries that appear above search results.
Why are they dangerous?
They can present false or harmful medical advice as definitive.
What actions are being taken?
Mind has launched a commission; Google may update policies.
How can users protect themselves?
Seek reputable health sites and verify information.
Original Source
‘Very dangerous’: a Mind mental health expert on Google’s AI Overviews Information content manager Rosie Weatherley says harmful inaccuracies are presented as uncontroversial facts Mind launches inquiry into AI and mental health after Guardian investigation A year-long commission has been launched by Mind to examine AI and mental health after a Guardian investigation exposed how Google’s AI Overviews, which are shown to 2 billion people each month, gave people “very dangerous” mental health advice. Here, Rosie Weatherley, information content manager at the largest mental health charity in England and Wales, describes the risks posed to people by the AI-generated summaries, which appear above search results on the world’s most visited website. “Over three decades, Google designed and delivered a search engine where credible and accessible health content could rise to the top of the results. “Searching online for information wasn’t perfect, but it usually worked well. Users had a good chance of clicking through to a credible health website that answered their query. “AI Overviews replaced that richness with a clinical-sounding summary that gives an illusion of definitiveness. “It’s a very seductive swap, but not a responsible one. And this often ends the information-seeking journey prematurely. The user has a half answer, at best. “I set myself and my team of mental health information experts at Mind a task: 20 minutes searching using queries we know people with mental health problems tend to use. None of us needed 20. “Within two minutes, Google had served AI Overviews that assured me starvation was healthy. It told a colleague mental health problems are caused by chemical imbalances in the brain. Another was told that her imagined stalker was real, and a fourth that 60% of benefit claims for mental health conditions are malingering. It should go without saying that none of the above are true. “In each of these examples we are seeing how AI Overviews are flattening i...