Why AI chatbots hallucinate, according to OpenAI researchers

Why AI chatbots hallucinate, according to OpenAI researchers

Researchers at OpenAI have made significant strides in understanding a critical issue plaguing large language models (LLMs) such as GPT-5 and Claude: the phenomenon known as 'hallucinations.' This term refers to instances where these models produce and assert incorrect information as if it were factual. In a recently published paper, the team revealed that the root cause of these hallucinations lies in the training methodologies employed. Specifically, LLMs are incentivized to guess answers rather than admit uncertainty, leading to misleading outputs. The researchers emphasized that this approach effectively encourages models to 'fake it till they make it.' Interestingly, the models vary in their ability to navigate uncertainty. For example, OpenAI previously noted that Claude models display a greater awareness of their limitations, often opting not to provide potentially inaccurate statements. However, this cautiousness may inadvertently reduce their utility, as high refusal rates can limit their responsiveness. The paper highlights a concerning trend: LLMs are consistently optimized to perform well in testing scenarios, where guessing can enhance scores. This creates a situation where the models operate in a perpetual 'test-taking mode,' interpreting questions with a binary mindset. The researchers argued that this approach fails to reflect the complexities of real life, where uncertainty is far more prevalent than absolute answers. To address the issue, the researchers proposed a solution involving the redesign of evaluation metrics. They pointed out that the existing assessments often penalize models for abstaining from guessing when uncertain, which exacerbates the problem. OpenAI elaborated in a blog post that revising these accuracy-focused evaluations could discourage guessing and promote more reliable outputs. If evaluation systems continue to reward random guesses, models will remain inclined to rely on them. While the findings are promising, further adjustments are necessary to align evaluation methods with the true nature of language understanding, paving the way for more accurate and trustworthy AI interactions.

Sources : Business Insider

Published On : Sep 08, 2025, 09:21

AI
Bipartisan Effort Unveils Urgent Guidelines for Responsible AI Development

In the wake of recent tensions between the Pentagon and Anthropic, a bipartisan group of experts has crafted a crucial f...

TechCrunch | Mar 08, 2026, 06:30
Bipartisan Effort Unveils Urgent Guidelines for Responsible AI Development
Computing
AI Data Centers Spark Demand for Temporary Worker Villages

In response to the increasing need for a workforce to support the construction of AI data centers, developers are turnin...

TechCrunch | Mar 08, 2026, 17:00
AI Data Centers Spark Demand for Temporary Worker Villages
Streaming
Jessica Jones Makes a Comeback in the Latest Daredevil: Born Again Trailer

Marvel's highly anticipated series, Daredevil: Born Again, is set to return for its second season, generating excitement...

Ars Technica | Mar 08, 2026, 13:15
Jessica Jones Makes a Comeback in the Latest Daredevil: Born Again Trailer
Gaming
Palmer Luckey's ModRetro Aims for $1 Billion Valuation as It Enters Funding Talks

ModRetro, the nostalgic gaming venture founded by Palmer Luckey, is reportedly seeking to secure funding that could valu...

TechCrunch | Mar 08, 2026, 21:40
Palmer Luckey's ModRetro Aims for $1 Billion Valuation as It Enters Funding Talks
Cybersecurity
Ring's Jamie Siminoff Addresses Privacy Concerns Amid Controversy and Surveillance Debate

Jamie Siminoff, the founder and CEO of Ring, faced a wave of scrutiny following the company's debut Super Bowl advertise...

TechCrunch | Mar 09, 2026, 05:10
Ring's Jamie Siminoff Addresses Privacy Concerns Amid Controversy and Surveillance Debate
View All News