ChatGPT-5 Slashes Hallucinations to 9.6%, Tops Grok in AI Benchmarks

OpenAI's ChatGPT-5 significantly reduces hallucinations to 9.6% from GPT-4o's 12.9%, outperforming in factual recall, coding, and medical tasks per recent benchmarks. However, xAI's Grok lags with 4.8% errors, prioritizing creativity over accuracy. Enterprises must balance reliability and innovation as AI evolves.
ChatGPT-5 Slashes Hallucinations to 9.6%, Tops Grok in AI Benchmarks
Written by Tim Toole

In the rapidly evolving world of artificial intelligence, the persistent issue of hallucinations—where models generate plausible but inaccurate information—continues to challenge developers and users alike. OpenAI’s latest flagship, ChatGPT-5, has made headlines for its purported reductions in these errors, drawing comparisons to predecessors like GPT-4o and rivals such as xAI’s Grok. Recent benchmarks suggest a marked improvement, but the nuances reveal a complex picture for enterprises relying on AI for decision-making.

Independent tests, including those detailed in a TechRadar analysis published just hours ago, indicate that ChatGPT-5 hallucinates at a rate significantly lower than GPT-4o. According to OpenAI’s internal evaluations shared in the report, the new model errs in factual claims about 9.6% of the time, down from 12.9% for GPT-4o—a 26% relative reduction. This progress stems from enhanced reasoning capabilities and a larger training dataset, allowing the AI to better cross-reference information before responding.

Benchmarking the Breakthrough

Industry insiders point to broader studies for context. A Deloitte survey cited in AIMultiple’s 2025 AI Hallucination report notes that 77% of businesses express concern over these inaccuracies, underscoring the stakes. In controlled experiments with 60 questions across 16 large language models, ChatGPT-5 emerged as a leader in reliability, outperforming GPT-4o in categories like factual recall and logical consistency.

Yet, this isn’t uniform across all tasks. Mashable India’s recent coverage highlights OpenAI’s claims of a 45% overall drop in hallucinations, particularly in coding and medical advice, where error rates fell to as low as 1.6% on tough cases. Posts on X from AI researchers, including those analyzing Vectara’s hallucination leaderboard, corroborate this, showing ChatGPT-5 at a mere 1.4% error rate—far below competitors in precision domains.

Grok’s Persistent Pitfalls

xAI’s Grok, however, tells a different story. The same TechRadar tests crown it “the king of making stuff up,” with hallucination rates hovering around 4.8% on Vectara benchmarks, nearly triple that of ChatGPT-5. Grok’s design prioritizes creative, humorous responses, which critics argue amplifies fabrications, especially in creative thinking tasks where it excels but often veers into inaccuracy.

Comparisons in McNeece’s showdown article reveal Grok 4 edging out in imaginative scenarios but lagging in reliability for business applications. X users, including AI enthusiasts posting real-time evaluations, note Grok’s zero-hallucination performance in niche areas like immunology, yet it stumbles on broader prompts, requiring more user fact-checking.

Implications for Enterprise Adoption

For industry leaders, these disparities influence deployment strategies. OpenAI’s rollout of GPT-5, free to all ChatGPT users with usage limits as reported by AInvest, democratizes access but raises questions about scalability. NetValuator’s comprehensive comparison emphasizes GPT-5’s adaptive reasoning, reducing sycophancy—overly agreeable but wrong answers—by integrating multimodal understanding.

Still, experts warn of underlying mysteries. A New York Times piece from May details how increased model complexity can paradoxically boost hallucinations, a trend OpenAI reversed here through undisclosed architectural tweaks. PC Gamer’s analysis echoes this, noting that while GPT-5 thinks more like a human, it isn’t immune to “robot dreams” in edge cases.

Looking Ahead: Reliability vs. Innovation

As AI integrates deeper into sectors like healthcare and finance, minimizing hallucinations could tip the scales in market dominance. Grok’s higher error rates, per India Today’s coverage of user backlash, highlight the trade-offs between wit and trustworthiness. OpenAI’s pacification efforts, including interface updates, aim to smooth adoption, but rivals like Google’s Gemini, with a 2.6% rate on Vectara, keep the competition fierce.

Ultimately, while ChatGPT-5 sets a new bar, the quest for hallucination-free AI remains ongoing. Insiders speculate that future iterations, informed by these benchmarks, may blend Grok’s creativity with OpenAI’s precision, but for now, enterprises must weigh risks carefully in their AI strategies.

Subscribe for Updates

GenAIPro Newsletter

News, updates and trends in generative AI for the Tech and AI leaders and architects.

By signing up for our newsletter you agree to receive content related to ientry.com / webpronews.com and our affiliate partners. For additional information refer to our terms of service.

Notice an error?

Help us improve our content by reporting any issues you find.

Get the WebProNews newsletter delivered to your inbox

Get the free daily newsletter read by decision makers

Subscribe
Advertise with Us

Ready to get started?

Get our media kit

Advertise with Us