The AI Engineer's Diary

The Reasoning Model Paradox: Why Smarter AI Hallucinates More


Listen Later

OpenAI's latest reasoning models can solve complex physics problems and write sophisticated code better than some humans – but they're wrong about basic facts up to 79% of the time. How is that possible?

In this deep dive, we unpack one of the most concerning trends in AI development: the reasoning model paradox. Despite being more capable than ever, models are hallucinating at unprecedented rates, with error rates that have doubled or even tripled compared to previous generations.

...more
View all episodesView all episodes
Download on the App Store

The AI Engineer's DiaryBy Skyward