🤖 AI Summary
To address the parameter-efficiency and generalization challenges posed by complex reasoning tasks—such as Sudoku, maze solving, and ARC-AGI—for small-scale models, this work proposes the Tiny Recursive Model (TRM): a minimalist recursive neural network with only two layers and 7 million parameters. TRM departs from large-model paradigms by leveraging high-frequency internal recursion to emulate hierarchical reasoning, enabling effective logical inference training on extremely limited data. Its architecture is simpler and more generalizable than that of Hierarchical Recursive Models (HRMs). On the ARC-AGI benchmark, TRM achieves 45% test accuracy on ARC-AGI-1 and 8% on ARC-AGI-2—substantially outperforming state-of-the-art large models like DeepSeek R1 despite using less than 0.01% of their parameters. This work demonstrates the efficacy of lightweight recursive mechanisms for complex reasoning under severe resource constraints, establishing a novel paradigm for efficient AI inference.
📝 Abstract
Hierarchical Reasoning Model (HRM) is a novel approach using two small neural networks recursing at different frequencies. This biologically inspired method beats Large Language models (LLMs) on hard puzzle tasks such as Sudoku, Maze, and ARC-AGI while trained with small models (27M parameters) on small data (around 1000 examples). HRM holds great promise for solving hard problems with small networks, but it is not yet well understood and may be suboptimal. We propose Tiny Recursive Model (TRM), a much simpler recursive reasoning approach that achieves significantly higher generalization than HRM, while using a single tiny network with only 2 layers. With only 7M parameters, TRM obtains 45% test-accuracy on ARC-AGI-1 and 8% on ARC-AGI-2, higher than most LLMs (e.g., Deepseek R1, o3-mini, Gemini 2.5 Pro) with less than 0.01% of the parameters.