According to OpenAI, chatbot hallucinations are unavoidable. The culprit is not technology, but simple mathematical calculation.
AI hallucinations pose a major issue since we can never predict when it will invent the information it delivers. The problem is that a lot of ChatGPT users use writing assistance tools, and if they don't verify the generated content, they might accidentally pass on errors. Scientists from OpenAI stated that this cannot be avoided (see Computer World).
Would you like to get more news like this or read interesting facts from the gaming world? Join our community on Google News and follow us there! Your support helps us grow and bring you even more content!
In the published document "Why Language Models Hallucinate," a team of four researchers presented their conclusions, and one of the main culprits is the AI benchmark system, which rates any answer, even a wrong one, higher than admitting ignorance. That's why artificial intelligence will try to guess any solution.
The behavior was compared to students who would rather write something, anything, on an exam question than leave the page blank:
Like students facing hard exam questions, large language models sometimes guess when uncertain, producing plausible yet incorrect statements instead of admitting uncertainty. Such 'hallucinations' persist even in state-of-the-art systems and undermine trust.
Models competing with ChatGPT were subjected to an experiment, which revealed that AI systems tend to provide incorrect answers. So the question was asked how many letters "d" are in the word "deepseek." DeepSeek-V3 in ten independent tests gave values such as "2" or "3." Claude 3.7 Sonnet responded even with "6" and "7."
ChatGPT-5 is also prone to hallucinations, although according to scientists, to a lesser extent. The model already showed it in August when it responded "I don't know" to a question from an internet user, which impressed many, including Elon Musk, because it was seen as a very human reaction. Interestingly, in the experiment, less errors were made by the more primitive models than the more advanced ones (o1 with 16% hallucinations, o3 with 33% hallucinations, and o4-mini with 48% hallucinations).
Researchers have found that it's impossible to avoid hallucinations, so we need to learn to control them. They also suggest that we need to make changes to benchmark systems so that they stop rewarding guessing and start penalizing for admitting ignorance. However, this cannot be achieved without appropriate regulations and industry requirements.
1

Author: Zbigniew Woznicki
He began his adventure with journalism and writing on the Allegro website, where he published news related to games, technology, and social media. He soon appeared on Gamepressure and Filmomaniak, writing about news related to the film industry. Despite being a huge fan of various TV series, his heart belongs to games of all kinds. He isn't afraid of any genre, and the adventure with Tibia taught him that sky and music in games are completely unnecessary. Years ago, he shared his experiences, moderating the forum of mmorpg.org.pl. Loves to complain, but of course constructively and in moderation.
Brainteaser answers in Dispatch. Let’s solve his riddles
Which Mecha Man contains the bomb in Dispatch. Here’s how to complete Comically Yours
Which movie theater should we search first in Dispatch. Complete Theatrically Yours easily
Where to find Hollow Abode and Aureate Pavilion in Where Winds Meet. Navigation through Mistveil Forest has never been so easy
How to “Visit Unique Stashes” in BF6. Complete BR REDSEC challenge easily