• Advertise
  • Contact Us
  • Daily The Business
  • Privacy Policy
Thursday, February 5, 2026
Daily The Business
  • Login
No Result
View All Result
DTB
No Result
View All Result
DTB

Why AI chatbots hallucinate, according to OpenAI researchers

September 6, 2025
in AI, chatbots, hallucinations, openai, Tech
Why AI chatbots hallucinate, according to OpenAI researchers
Share on FacebookShare on TwitterWhatsapp

Matthias Balk/picture alliance via Getty Images

  • OpenAI researchers say they've found a reason large language models hallucinate.
  • Hallucinations occur when models confidently generate inaccurate information as facts.
  • Redesigning evaluation metrics can prevent models from guessing, OpenAI researchers said in a paper.

OpenAI researchers claim they've cracked one of the biggest obstacles to large language model performance — hallucinations.

Hallucinations occur when a large language model generates inaccurate information that it presents as fact. They plague the most popular LLMs, from OpenAI's GPT-5 to Anthropic's Claude.

OpenAI's baseline finding, which it made public in a paper released on Thursday, is that large language models hallucinate because the methods they're trained under reward guessing more than admitting uncertainty.

In other words, LLMs are being told to fake it till they make it. Some are better than others, however. In a blog post last month, OpenAI said that Claude models are more "aware of their uncertainty and often avoid making statements that are inaccurate." It also noted that Claude's high refusal rates risked limiting its utility.

"Hallucinations persist due to the way most evaluations are graded — language models are optimized to be good test-takers, and guessing when uncertain improves test performance," the researchers wrote in the paper.

Large language models are essentially always in "test-taking mode," answering questions as if everything in life were binary — right or wrong, black or white.

In many ways, they're not equipped for the realities of life, where uncertainty is more common than certainty, and true accuracy is not a given.

"Humans learn the value of expressing uncertainty outside of school, in the school of hard knocks. On the other hand, language models are primarily evaluated using exams that penalize uncertainty," the researchers wrote.

The good news is that there is a fix, and it has to do with redesigning evaluation metrics.

"The root problem is the abundance of evaluations that are not aligned," they wrote. "The numerous primary evaluations must be adjusted to stop penalizing abstentions when uncertain."

In a blog post about the paper, OpenAI elaborated on what this type of adjustment would entail.

"The widely used, accuracy-based evals need to be updated so that their scoring discourages guessing. If the main scoreboards keep rewarding lucky guesses, models will keep learning to guess," OpenAI said.

OpenAI did not immediately respond to a request for comment from Business Insider.

Read the original article on Business Insider
Share15Tweet10Send
Previous Post

Saudi Arabia to Open World’s Tallest, Fastest, and Longest Roller Coaster

Next Post

Blood moon returns to Pakistan this weekend

Related Posts

Blame Jeff Bezos! And then stop hoping billionaires bail out the news.
amazon

Blame Jeff Bezos! And then stop hoping billionaires bail out the news.

February 5, 2026
What Mark Zuckerberg, Jeff Bezos, Sam Altman, and other tech execs have said about parenting
business-visual-features

What Mark Zuckerberg, Jeff Bezos, Sam Altman, and other tech execs have said about parenting

February 4, 2026
New poll shows the shifting conversation around blue-collar work in the age of AI
AI

New poll shows the shifting conversation around blue-collar work in the age of AI

February 3, 2026
All the ways Elon Musk's companies are already intertwined, from a Tesla 'collab' with SpaceX to Grok in vehicles
AI

All the ways Elon Musk’s companies are already intertwined, from a Tesla ‘collab’ with SpaceX to Grok in vehicles

February 3, 2026
The 'father of the iPod' says Apple should bring back a 'nostalgic version' of the music device
apple

The ‘father of the iPod’ says Apple should bring back a ‘nostalgic version’ of the music device

February 2, 2026
'Unsubscribe' and 'opt out': A new Big Tech boycott to protest ICE starts February 1
AI

‘Unsubscribe’ and ‘opt out’: A new Big Tech boycott to protest ICE starts February 1

February 1, 2026

Popular Post

  • FRSHAR Mail

    FRSHAR Mail set to redefine secure communication, data privacy

    127 shares
    Share 51 Tweet 32
  • How to avoid buyer’s remorse when raising venture capital

    33 shares
    Share 337 Tweet 211
  • Microsoft to pay off cloud industry group to end EU antitrust complaint

    55 shares
    Share 22 Tweet 14
  • Capacity utilisation of Pakistan’s cement industry drops to lowest on record

    48 shares
    Share 19 Tweet 12
  • SingTel annual profit more than halves on $2.3bn impairment charge

    48 shares
    Share 19 Tweet 12
American Dollar Exchange Rate
  • Advertise
  • Contact Us
  • Daily The Business
  • Privacy Policy
Write us: info@dailythebusiness.com

© 2021 Daily The Business

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In

Add New Playlist

No Result
View All Result
  • Advertise
  • Contact Us
  • Daily The Business
  • Privacy Policy

© 2021 Daily The Business

This website uses cookies. By continuing to use this website you are giving consent to cookies being used. Visit our Privacy and Cookie Policy.