Google Med-Gemini AI Paper Invents ‘Basilar Ganglia’ in Major Error

Google's Med-Gemini AI research paper erred by inventing the "basilar ganglia," conflating the basilar artery and basal ganglia, exposing AI hallucinations and oversight failures. This incident, unnoticed by authors and reviewers, underscores risks in healthcare AI, urging hybrid human-AI systems and rigorous audits for reliable medical applications.
Google Med-Gemini AI Paper Invents ‘Basilar Ganglia’ in Major Error
Written by John Smart

In the rapidly evolving field of artificial intelligence applied to medicine, a recent incident has spotlighted the perils of unchecked errors in AI-generated content. Google’s Med-Gemini, a specialized AI model designed for healthcare tasks, was highlighted in a 2024 research paper for its impressive capabilities in analyzing medical images and data. However, as detailed in a report by The Verge, the paper contained a glaring mistake: it conflated the “basilar artery” with the “basal ganglia,” essentially inventing a nonexistent body part called the “basilar ganglia.” This error, whether a typo or an AI hallucination, went unnoticed by the paper’s authors and peer reviewers, raising alarms about the reliability of AI in critical healthcare applications.

The mix-up occurred in a section describing Med-Gemini’s prowess in interpreting brain scans. The basilar artery is a vital blood vessel at the base of the brain, while the basal ganglia are clusters of neurons involved in movement and cognition. By merging them into “basilar ganglia,” the AI not only demonstrated a factual inaccuracy but also highlighted a broader issue: AI models can “hallucinate” information, generating plausible-sounding but entirely false details. Industry insiders point out that such errors could propagate through medical literature and tools, potentially influencing diagnoses if not caught.

The Human Oversight Gap in AI-Driven Research

Experts interviewed by The Verge emphasized the challenges in human review processes. Dr. Nigam Shah, a professor at Stanford University, noted that even seasoned professionals might miss these slips, especially when AI outputs are voluminous. “The problem with these typos or other hallucinations is I don’t trust our humans to review them,” Shah told the publication, underscoring how errors can cascade, as seen in cases where AI misreads pathology notes and perpetuates inaccuracies in patient records.

This isn’t an isolated incident for Google. A 2020 article in MIT Technology Review examined how Google’s earlier AI for detecting diabetic retinopathy performed exceptionally in controlled lab settings but faltered in real-world clinics due to variables like poor lighting or patient movement. The Med-Gemini error echoes these concerns, suggesting that while AI excels in benchmarks, its deployment in healthcare demands rigorous validation.

Broader Implications for AI Standards in Medicine

Recent updates from Google’s own events, such as the 2025 Check Up detailed on their official blog, tout advancements in generative AI for health, including better integration with medical records. Yet, the basilar ganglia blunder, still uncorrected in the original paper as of August 2025, fuels skepticism. Posts on X (formerly Twitter) reflect public sentiment, with users expressing frustration over AI’s anatomical inaccuracies, from botched image generations to fundamental misunderstandings of human biology.

Regulatory bodies are taking note. The FDA has been scrutinizing AI tools for medical use, requiring evidence of safety and efficacy. Insiders argue that incidents like this could delay widespread adoption, prompting calls for mandatory “AI audits” where outputs are cross-verified by multiple experts before publication or clinical use.

Lessons from Past AI Controversies and Future Safeguards

Google’s history with AI mishaps extends beyond healthcare. In 2024, its Gemini image generator faced backlash for historical inaccuracies, as covered in various media, leading to admissions from co-founder Sergey Brin that the company “messed up.” Applied to medicine, such flaws carry higher stakes, potentially endangering lives if AI influences treatment decisions.

To mitigate risks, researchers advocate for hybrid systems combining AI with human expertise, alongside transparent error-reporting mechanisms. Google’s March 2025 announcements, reported by Fierce Healthcare, include open models for collaborative AI development, which could foster community-driven corrections. Discussions on platforms like Reddit’s r/technology thread on this topic reveal a consensus among tech enthusiasts that while AI promises transformative healthcare tools, unchecked hallucinations demand a reevaluation of trust in these systems.

As AI integrates deeper into medicine, the Med-Gemini incident serves as a cautionary tale. It underscores the need for robust safeguards to ensure that innovation doesn’t come at the cost of accuracy, pushing the industry toward more accountable practices.

Subscribe for Updates

HealthcareITPro Newsletter

By signing up for our newsletter you agree to receive content related to ientry.com / webpronews.com and our affiliate partners. For additional information refer to our terms of service.

Notice an error?

Help us improve our content by reporting any issues you find.

Get the WebProNews newsletter delivered to your inbox

Get the free daily newsletter read by decision makers

Subscribe
Advertise with Us

Ready to get started?

Get our media kit

Advertise with Us