OpenAI sheds light on the persistence of hallucinations in language models due to evaluation methods favoring guessing over honesty, requiring a shift towards rewarding uncertainty acknowledgment. High model accuracy does not equate to the eradication of hallucinations, as some questions are inherently unanswerable, driving the necessity for a broader adoption of hallucination-reduction techniques through updated scoring mechanisms.