OpenAI's new AI models hallucinate more frequently

setn.com (Chinese)

OpenAI's new AI models, o3 and o4-mini, exhibit a higher frequency of generating false or fabricated information compared to previous models. This "hallucination" rate is a concerning development. Internal tests show o3 and o4-mini produce hallucinations more often, with o4-mini generating false information in 48% of its responses. This is despite improved performance in programming and math. OpenAI acknowledges the issue, attributing it to the models' output style. They are researching the cause, while external tests and ChatGPT itself confirm the problem, highlighting the need for accuracy improvements.


With a significance score of 3.6, this news ranks in the top 8.4% of today's 32564 analyzed articles.

Get summaries of news with significance over 5.5 (usually ~10 stories per week). Read by 10,000+ subscribers:


OpenAI's new AI models hallucinate more frequently | News Minimalist