Reasoning Beyond Language: A Comprehensive Survey on Latent Chain-of-Thought Reasoning

📅 2025-05-22
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Traditional chain-of-thought (CoT) prompting in language models relies on explicit natural-language reasoning, which is computationally inefficient and ill-suited for abstract, non-linguistic inference. Method: This work systematically investigates the latent CoT paradigm—where large language models perform implicit, language-free, and efficient abstract reasoning within their internal latent spaces. We propose a novel four-dimensional unified taxonomy covering token-level strategies, architectural mechanisms, analytical methodologies, and application scenarios. Contribution/Results: Our study establishes the first structured knowledge framework for latent CoT, comprehensively surveys over 100 state-of-the-art works—including advances in training paradigms, model architecture innovations, latent-space interpretability analysis, and multi-task empirical validation—and clarifies technical trajectories, recurring design patterns, and fundamental challenges. To foster community progress, we open-source implementation code and a curated resource repository.

Technology Category

Application Category

📝 Abstract
Large Language Models (LLMs) have achieved impressive performance on complex reasoning tasks with Chain-of-Thought (CoT) prompting. However, conventional CoT relies on reasoning steps explicitly verbalized in natural language, introducing inefficiencies and limiting its applicability to abstract reasoning. To address this, there has been growing research interest in latent CoT reasoning, where inference occurs within latent spaces. By decoupling reasoning from language, latent reasoning promises richer cognitive representations and more flexible, faster inference. Researchers have explored various directions in this promising field, including training methodologies, structural innovations, and internal reasoning mechanisms. This paper presents a comprehensive overview and analysis of this reasoning paradigm. We begin by proposing a unified taxonomy from four perspectives: token-wise strategies, internal mechanisms, analysis, and applications. We then provide in-depth discussions and comparative analyses of representative methods, highlighting their design patterns, strengths, and open challenges. We aim to provide a structured foundation for advancing this emerging direction in LLM reasoning. The relevant papers will be regularly updated at https://github.com/EIT-NLP/Awesome-Latent-CoT.
Problem

Research questions and friction points this paper is trying to address.

Explores latent Chain-of-Thought reasoning beyond language constraints
Analyzes inefficiencies in conventional verbalized reasoning steps
Surveys methods for abstract reasoning in latent spaces
Innovation

Methods, ideas, or system contributions that make the work stand out.

Latent CoT reasoning in hidden spaces
Decoupling reasoning from language constraints
Taxonomy covering token strategies and mechanisms
🔎 Similar Papers
No similar papers found.
X
Xinghao Chen
Department of Computing, The Hong Kong Polytechnic University
A
Anhao Zhao
Ningbo Digital Twin Institute, Eastern Institute of Technology, Ningbo, China
Heming Xia
Heming Xia
Natural Language Processing Group, The Hong Kong Polytechnic University
Natural Language ProcessingLarge Language Models
Xuan Lu
Xuan Lu
Assistant Professor, University of Arizona
Human-centered Data ScienceHuman-AI CollaborationCausal InferenceFuture of WorkEmoji
Hanlin Wang
Hanlin Wang
HKUST
Computer visionVideo understanding
Yanjun Chen
Yanjun Chen
University of Illinois Urbana-Champaign
Human Computer InteractionHaptics
W
Wei Zhang
Ningbo Digital Twin Institute, Eastern Institute of Technology, Ningbo, China
J
Jian Wang
Department of Computing, The Hong Kong Polytechnic University
W
Wenjie Li
Department of Computing, The Hong Kong Polytechnic University
Xiaoyu Shen
Xiaoyu Shen
Eastern Institute of Technology, Ningbo
language modelmulti-modal learningreasoning