In the race to build machines that think, a growing body of research suggests we may be mistaking verbosity for intelligence.
While developers of Large Reasoning Models (LRMs), the next generation of language models engineered to mimic step-by-step reasoning, claim dramatic improvements in accuracy and logic, a new study reveals those claims are both overstated and fundamentally flawed.
Published by a team of researchers at Apple, “The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity” offers one of the most sobering evaluations yet of AI’s much-hyped cognitive potential.
Authored by Parshin Shojaee, Iman Mirzadeh, Keivan Alizadeh, Maxwell Horton, Samy Bengio, and Mehrdad Farajtabar, the paper doesn’t simply challeng