Hallucinations are accidentally created by evals
They come from post-training. Reasoning models hallucinate more because we do more rigorous post-training on them
The problem is we reward them for being confident
cdn.openai.com/pdf/d04913be...
Hallucinations are accidentally created by evals
View original thread
65
7
e.g. MMLU tests knowledge, it's one of the core benchmarks that people care about
if it says "idk", it's score goes down
if it confidently guesses regardless of it's own confidence, it's score goes up (and so do hallucinations)
if it says "idk", it's score goes down
if it confidently guesses regardless of it's own confidence, it's score goes up (and so do hallucinations)
21
3