From System 1 to System 2: A Survey of Reasoning Large Language Models

📅 2025-02-24
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the paradigm shift in large language models (LLMs) from intuitive, fast System 1 reasoning to deliberate, logical System 2 reasoning. Methodologically, it introduces the first cognitive-science–inspired dual-system framework for LLMs, establishing a unified taxonomy and evolutionary trajectory for reasoning-oriented LLMs; integrates key techniques—including Chain-of-Thought variants, verifier-guided inference, process-supervised fine-tuning, multi-stage architectures, and introspective reinforcement learning; and develops a dynamic open-source tracking ecosystem (Awesome-Slow-Reason-System). Empirically, it conducts systematic benchmarking of models—including o1, o3, and R1—on mathematical and code-reasoning tasks. The contributions include: (1) a theoretically grounded framework for mechanistic understanding of LLM reasoning, (2) principled design principles for controllable model evolution, and (3) reproducible evaluation protocols and open tools to advance research on reasoning-capable LLMs.

Technology Category

Application Category

📝 Abstract
Achieving human-level intelligence requires refining the transition from the fast, intuitive System 1 to the slower, more deliberate System 2 reasoning. While System 1 excels in quick, heuristic decisions, System 2 relies on logical reasoning for more accurate judgments and reduced biases. Foundational Large Language Models (LLMs) excel at fast decision-making but lack the depth for complex reasoning, as they have not yet fully embraced the step-by-step analysis characteristic of true System 2 thinking. Recently, reasoning LLMs like OpenAI's o1/o3 and DeepSeek's R1 have demonstrated expert-level performance in fields such as mathematics and coding, closely mimicking the deliberate reasoning of System 2 and showcasing human-like cognitive abilities. This survey begins with a brief overview of the progress in foundational LLMs and the early development of System 2 technologies, exploring how their combination has paved the way for reasoning LLMs. Next, we discuss how to construct reasoning LLMs, analyzing their features, the core methods enabling advanced reasoning, and the evolution of various reasoning LLMs. Additionally, we provide an overview of reasoning benchmarks, offering an in-depth comparison of the performance of representative reasoning LLMs. Finally, we explore promising directions for advancing reasoning LLMs and maintain a real-time href{https://github.com/zzli2022/Awesome-Slow-Reason-System}{GitHub Repository} to track the latest developments. We hope this survey will serve as a valuable resource to inspire innovation and drive progress in this rapidly evolving field.
Problem

Research questions and friction points this paper is trying to address.

Transition from intuitive to deliberate reasoning in AI.
Enhance large language models for complex logical analysis.
Develop reasoning benchmarks for advanced AI performance evaluation.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Transitioning System 1 to System 2
Enhancing logical reasoning in LLMs
Developing advanced reasoning benchmarks
🔎 Similar Papers
No similar papers found.
Zhongzhi Li
Zhongzhi Li
Institute of Automation, Chinese Academy of Sciences
LLMNLPMath Reason
Duzhen Zhang
Duzhen Zhang
Institute of Automation, Chinese Academy of Sciences
Natural Language ProcessingMultimodalLarge Language ModelsContinual LearningAI4Science
Ming-Liang Zhang
Ming-Liang Zhang
PhD, Senior Algorithm Engineer at Alibaba Beijing
Multimodal ReasoningMath Problem SolvingScene Parsing
J
Jiaxin Zhang
University of Strathclyde, Glasgow, UK
Z
Zengyan Liu
City University of Hong Kong and the Hong Kong University of Science and Technology (Guangzhou), China
Yuxuan Yao
Yuxuan Yao
City University of Hong Kong
LLMDecodingReasoningModel Merging
H
Haotian Xu
Xiaohongshu Inc, Beijing, China
Junhao Zheng
Junhao Zheng
South China University of Technology, Qwen Team
Large Language ModelsPretrainingContinual Learning
P
Pei-Jie Wang
Institute of Automation, Chinese Academy of Sciences, Beijing, China
Xiuyi Chen
Xiuyi Chen
Baidu <<< CASIA
RAGMultiModalDialogue
Y
Yingying Zhang
East China Normal University, Shanghai, China
F
Fei Yin
Institute of Automation, Chinese Academy of Sciences, Beijing, China
J
Jiahua Dong
Mohamed bin Zayed University of Artificial Intelligence, Abu Dhabi, UAE
Zhijiang Guo
Zhijiang Guo
HKUST (GZ) | HKUST
Natural Language ProcessingMachine LearningLarge Language Models
Le Song
Le Song
CTO, GenBio AI; Professor, MBZUAI
AIAI for ScienceMachine Learning
C
Cheng-Lin Liu
Institute of Automation, Chinese Academy of Sciences, Beijing, China