Rethinking the Illusion of Thinking

📅 2025-07-01
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the fundamental debate over whether large reasoning models (LRMs) possess genuine reasoning capability, specifically distinguishing whether their failures on complex tasks—such as the Tower of Hanoi and river-crossing puzzles—stem from output-length constraints or intrinsic cognitive limitations. Method: We propose three key methodological innovations: (i) incremental stepwise prompting, (ii) multi-agent collaborative reasoning, and (iii) a controllable benchmarking framework, complemented by a fine-grained ablation analysis to precisely isolate failure sources. Contribution/Results: Empirical evaluation reveals that LRMs encounter a sharp cognitive bottleneck at ~8-disk Tower of Hanoi instances, yet solve river-crossing problems with >100 entity pairs reliably. These findings demonstrate that LRMs function as stochastic search-based reasoners—not mere “hallucinatory” systems—and provide both a novel evaluation paradigm and empirical grounding for assessing LRM reasoning competence.

Technology Category

Application Category

📝 Abstract
Earlier this year, Apple ignited controversy by publishing "The Illusion of Thinking," prompting heated debate within the AI community. Critics seized upon the findings as conclusive evidence that Large Reasoning Models (LRMs) lack genuine reasoning capabilities, branding them as mere stochastic parrots. Meanwhile, defenders-spearheaded by Lawsen et al. (2025)-fired back, condemning the experimental setup as flawed and the conclusions overstated. We clarify this debate by replicating and refining two of the original study's most contentious benchmarks: Towers of Hanoi and River Crossing. By introducing incremental stepwise prompting and agentic collaborative dialogue, we show that previously reported failures solving the Towers of Hanoi were not purely result of output constraints, but also partly a result of cognition limitations: LRMs still stumble when complexity rises moderately (around 8 disks). Moreover, the River Crossing results initially heralded as catastrophic failures turn out to hinge upon testing unsolvable configurations. Once we limit tests strictly to solvable problems-LRMs effortlessly solve large instances involving over 100 agent pairs. Our findings ultimately defy simplistic narratives: today's LRMs are stochastic, RL-tuned searchers in a discrete state space we barely understand. Real progress in symbolic, long-horizon reasoning demands mapping that terrain through fine-grained ablations like those introduced here.
Problem

Research questions and friction points this paper is trying to address.

Clarify debate on LRMs' reasoning capabilities via benchmarks
Address failures in solving Towers of Hanoi with LRMs
Correct misinterpretations of River Crossing test results
Innovation

Methods, ideas, or system contributions that make the work stand out.

Incremental stepwise prompting enhances LRM performance
Agentic collaborative dialogue improves problem-solving accuracy
Fine-grained ablations map symbolic reasoning terrain
🔎 Similar Papers
No similar papers found.
I
Iñaki Dellibarda Varela
Center for Automation and Robotics, Spanish National Research Council (CSIC-UPM), Madrid, Spain
P
Pablo Romero-Sorozabal
Center for Automation and Robotics, Spanish National Research Council (CSIC-UPM), Madrid, Spain
E
Eduardo Rocon
Center for Automation and Robotics, Spanish National Research Council (CSIC-UPM), Madrid, Spain
Manuel Cebrian
Manuel Cebrian
Spanish National Research Council
Computational Social ScienceArtificial Intelligence