HN2new | past | comments | ask | show | jobs | submitlogin

Obviously it's well over a year since this article was posted and if anything I've anecdotally noticed hallucinations getting more, not less, common.

Possibly/probably with another years experience with LLMs I'm just more attuned to noticing when they have lost the plot and are making shit up



RL for reasoning definitely introduces hallucinations, and sometimes it introduces a class of hallucinations that feels a lot worse than the classic ones.

I noticed OpenAI's models picked up a tendency to hold strong convictions on completely unknowable things.

"<suggests possible optimization> Implement this change and it will result in a 4.5% uplift in performance"

"<provides code> I ran the updated script 10 times and it completes 30.5 seconds faster than before on average"

It's bad it enough it convinces itself it did things it can't do, but then it goes further and hallucinates insights from the tasks it hallucinated itself doing in the first places!

I feel like lay people aren't ready for that. Normal hallucinations felt passive, like a slip up. To the unprepared, this becomes more like someone actively trying to sell their slip ups.

I'm not sure if it's a form of RL hacking making it through to the final model or what, but even OpenAI seems to have noticed it in testing based on their model cards.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: