ReGuLaR: Variational Latent Reasoning Guided by Rendered Chain-of-Thought

📅 2026-01-30
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the performance degradation of existing implicit reasoning methods caused by the absence of effective compression guidance, while explicit chain-of-thought (CoT) approaches suffer from redundant computation. To overcome these limitations, the authors propose a novel implicit reasoning framework based on variational autoencoding, which— for the first time—leverages rendered CoT images as multimodal supervision signals. By extracting visual semantic representations from these images to regularize the posterior distribution, the method achieves highly efficient reasoning with minimal information loss. Extensive experiments demonstrate that the proposed approach significantly outperforms current implicit reasoning techniques across multiple tasks, offering not only superior computational efficiency but also reasoning performance that surpasses that of the original CoT.

Technology Category

Application Category

📝 Abstract
While Chain-of-Thought (CoT) significantly enhances the performance of Large Language Models (LLMs), explicit reasoning chains introduce substantial computational redundancy. Recent latent reasoning methods attempt to mitigate this by compressing reasoning processes into latent space, but often suffer from severe performance degradation due to the lack of appropriate compression guidance. In this study, we propose Rendered CoT-Guided variational Latent Reasoning (ReGuLaR), a simple yet novel latent learning paradigm resolving this issue. Fundamentally, we formulate latent reasoning within the Variational Auto-Encoding (VAE) framework, sampling the current latent reasoning state from the posterior distribution conditioned on previous ones. Specifically, when learning this variational latent reasoning model, we render explicit reasoning chains as images, from which we extract dense visual-semantic representations to regularize the posterior distribution, thereby achieving efficient compression with minimal information loss. Extensive experiments demonstrate that ReGuLaR significantly outperforms existing latent reasoning methods across both computational efficiency and reasoning effectiveness, and even surpasses CoT through multi-modal reasoning, providing a new and insightful solution to latent reasoning. Code: https://github.com/FanmengWang/ReGuLaR.
Problem

Research questions and friction points this paper is trying to address.

latent reasoning
Chain-of-Thought
computational redundancy
performance degradation
reasoning compression
Innovation

Methods, ideas, or system contributions that make the work stand out.

latent reasoning
Chain-of-Thought
Variational Auto-Encoding
multi-modal regularization
computational efficiency
🔎 Similar Papers
No similar papers found.
Fanmeng Wang
Fanmeng Wang
Renmin University of China
geometric deep learninggraph representation learningAI4Science
H
Haotian Liu
Gaoling School of Artificial Intelligence, Renmin University of China
Guojiang Zhao
Guojiang Zhao
DP Technology; Carnegie Mellon University;
LLMAI For Science
H
Hongteng Xu
Gaoling School of Artificial Intelligence, Renmin University of China; Beijing Key Laboratory of Research on Large Models and Intelligent Governance; Engineering Research Center of Next-Generation Intelligent Search and Recommendation, MOE
Zhifeng Gao
Zhifeng Gao
DP Technology
Data MiningMachine LearningAI for ScienceAI for Industry