Techoreon

  • Home
  • AI
  • Tips & Tricks
  • Info
    • Privacy Policy
    • DMCA & Copyright Notice
    • Contact Us
    • About Us
    • Terms and Conditions
Notification Show More
Latest News
ChatGPT Logo | Users Suffer Rare Delusions After Intensive Use of ChatGPT
Users Suffer Rare Delusions After Intensive Use of ChatGPT
AI OpenAI
Tesla Cars on Road
Man Asks His Tesla to Take Him to a New Place and Ends Up Getting Roasted
Tesla AI
Tesla Cybertrucks Piled Up | Techoreon
Man Returns Tesla Cybertruck Same Day After Finding It Sat in a Vacant Lot for Four Months
Tesla EV
Elon Musk's AI, Grok, Accused of Undressing Women in Public on X
Elon Musk’s AI, Grok, Accused of Undressing Women in Public on X
AI Social Media
U.S. Jury Orders NSO Group to Pay $168 Million for Illegally Targeting WhatsApp Users with Spyware
U.S. Jury Orders NSO Group to Pay $168 Million in WhatsApp Spyware Case
Cybersecurity Meta
Aa

Techoreon

Aa
  • Home
  • AI
  • Tips & Tricks
Search
  • Home
  • Privacy Policy
  • Contact Us
  • About Us
  • Terms and Conditions
  • DMCA & Copyright Notice
Follow US
Techoreon > AI > OpenAI’s New Reasoning Models Hallucinate Up to 48% of the Time, But No One Knows Why
AIOpenAI

OpenAI’s New Reasoning Models Hallucinate Up to 48% of the Time, But No One Knows Why

Piyush Gupta
Last updated: 2025/04/19 at 9:21 PM
Piyush Gupta
Share
6 Min Read
ChatGPT on Phone's Screen
OpenAI says more research is needed to explain why hallucinations worsen as reasoning models expand. | Flickr
SHARE

Virtually every month we see announcements of new AI models that surpass previous ones, the latest ones even having reasoning functions, but these advancements are having a worrying negative aspect: AI hallucinations.

Hallucinations in artificial intelligence, especially LLMs, are one of serious problems that big AI tech giants are facing. This occurs when these models offer type of information or reasoning that makes a firm statement, but in reality it is not accurate or true.

For example, an AI might confidently generate a fake historical quote or invent a non-existent scientific study, presenting it as fact. It might claim that Albert Einstein won a Nobel Prize for his theory of relativity — when in fact, he received it for his work on the photoelectric effect.

The worrying thing is that the new OpenAI models, both the o3 and the o4-mini, generate more hallucinations than ever compared to previous models.

According to OpenAI’s internal testing, both o3 and o4-mini, which feature reasoning capabilities, hallucinate more frequently than the company’s previous reasoning models, even compared to traditional models like GPT-4o.

But the worrying thing is that even OpenAI doesn’t know exactly why this is happening, and if things continue this way, its future models might not be reliable.

In the latest white paper on o3 and o4-mini, OpenAI notes that more research is needed to understand why hallucinations worsen as reasoning models expand.

They point out that o3 amazed by exhibiting a 33% hallucination rate in answering the questions in PersonQA, an internal company benchmark for measuring the accuracy of a model on people. This represents approximately twice the rate of hallucinations in previous reasoning models.

On the other hand, o4-mini performed even worse on PersonQA with hallucinations 48% of the time.

Likewise, external tests like those conducted by Transluce, a nonprofit AI research lab, have also found evidence that these new models tend to invent more than ever: “Our hypothesis is that the ty[e of reinforcement learning used for the ‘o’ series models may amplify problems that are typically mitigated (but not entirely eliminated) by standard post-training processes,” said Neil Chowdhury, a researcher at Transluce.

While these hallucinations can sometimes help models be more imaginative and creative, if the goal is clear, concise, and accurate information, they should have a virtually nonexistent hallucination rate.

One way to increase the accuracy of models is with web search functions, but sometimes not everything found on the web is accurate or trustworthy.

“Addressing hallucinations in all of our models is an area of ​​ongoing research, and we’re constantly working to improve their accuracy and reliability,” OpenAI spokesperson Niko Felix told Techcrunch.


Also Read Loading title…
TAGGED: AI, ChatGPT, LLMs, OpenAI
Share this Article
Facebook Twitter Copy Link Print
Leave a comment Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Latest Posts

ChatGPT Logo | Users Suffer Rare Delusions After Intensive Use of ChatGPT
AIOpenAI

Users Suffer Rare Delusions After Intensive Use of ChatGPT

Dev Mehta Dev Mehta May 13, 2025
Tesla Cars on Road
TeslaAI

Man Asks His Tesla to Take Him to a New Place and Ends Up Getting Roasted

Dev Mehta Dev Mehta May 12, 2025
Tesla Cybertrucks Piled Up | Techoreon
TeslaEV

Man Returns Tesla Cybertruck Same Day After Finding It Sat in a Vacant Lot for Four Months

Dev Mehta Dev Mehta May 11, 2025
Elon Musk's AI, Grok, Accused of Undressing Women in Public on X
AISocial Media

Elon Musk’s AI, Grok, Accused of Undressing Women in Public on X

Dev Mehta Dev Mehta May 8, 2025
U.S. Jury Orders NSO Group to Pay $168 Million for Illegally Targeting WhatsApp Users with Spyware
CybersecurityMeta

U.S. Jury Orders NSO Group to Pay $168 Million in WhatsApp Spyware Case

Dev Mehta Dev Mehta May 7, 2025
Elon Musk seated in a Tesla electric car wearing a black suit, looking to the side. | Elon Musk Threatened to Deport Engineers Who Exposed Flaws in Tesla Cars: "He's Pure Evil"
TeslaEV

Elon Musk Threatened to Deport Engineers Who Exposed Flaws in Tesla Cars: “He’s Pure Evil”

Piyush Gupta Piyush Gupta May 4, 2025
Show More

© 2025 Techoreon. All rights reserved.

  • Home
  • Privacy Policy
  • Contact Us
  • About Us
  • Terms and Conditions
  • DMCA & Copyright Notice

Removed from reading list

Undo
Welcome Back!

Sign in to your account

Lost your password?