Agentic Reasoning for Large Language Models

📅 2026-01-18
📈 Citations: 7
Influential: 1
📄 PDF
🤖 AI Summary
This work addresses the limited capacity of large language models (LLMs) to plan, act, and learn through sustained interaction in open, dynamic environments. To overcome this, the authors propose a three-tiered reasoning framework that treats LLMs as autonomous agents, unifying single-agent foundational reasoning, self-evolution, and multi-agent collaboration within a coherent paradigm. The framework orchestrates structured interactions, incorporates memory mechanisms, enables tool use, and integrates both reinforcement learning and supervised fine-tuning. It explicitly distinguishes between in-context reasoning and post-training optimization pathways, systematically coupling cognition with action. Empirical validation across diverse domains—including scientific discovery, robotics, healthcare, autonomous research, and mathematics—demonstrates its effectiveness and highlights promising future directions such as personalized interaction, long-horizon engagement, world modeling, and scalable multi-agent training.

Technology Category

Application Category

📝 Abstract
Reasoning is a fundamental cognitive process underlying inference, problem-solving, and decision-making. While large language models (LLMs) demonstrate strong reasoning capabilities in closed-world settings, they struggle in open-ended and dynamic environments. Agentic reasoning marks a paradigm shift by reframing LLMs as autonomous agents that plan, act, and learn through continual interaction. In this survey, we organize agentic reasoning along three complementary dimensions. First, we characterize environmental dynamics through three layers: foundational agentic reasoning, which establishes core single-agent capabilities including planning, tool use, and search in stable environments; self-evolving agentic reasoning, which studies how agents refine these capabilities through feedback, memory, and adaptation; and collective multi-agent reasoning, which extends intelligence to collaborative settings involving coordination, knowledge sharing, and shared goals. Across these layers, we distinguish in-context reasoning, which scales test-time interaction through structured orchestration, from post-training reasoning, which optimizes behaviors via reinforcement learning and supervised fine-tuning. We further review representative agentic reasoning frameworks across real-world applications and benchmarks, including science, robotics, healthcare, autonomous research, and mathematics. This survey synthesizes agentic reasoning methods into a unified roadmap bridging thought and action, and outlines open challenges and future directions, including personalization, long-horizon interaction, world modeling, scalable multi-agent training, and governance for real-world deployment.
Problem

Research questions and friction points this paper is trying to address.

agentic reasoning
large language models
open-ended environments
multi-agent reasoning
dynamic interaction
Innovation

Methods, ideas, or system contributions that make the work stand out.

Agentic Reasoning
Autonomous Agents
Multi-Agent Collaboration
In-Context Reasoning
Self-Evolving LLMs
🔎 Similar Papers
No similar papers found.
Tianxin Wei
Tianxin Wei
University of Illinois Urbana Champaign
Trustworthy Machine LearningLLMInformation Retrieval
Ting-Wei Li
Ting-Wei Li
University of Illinois Urbana-Champaign
graph machine learningdata-centric machine learning
Zhining Liu
Zhining Liu
Ph.D. Candidate, UIUC
LLMData-centric AIResponsible AIImbalanced LearningGraph Mining
Xuying Ning
Xuying Ning
University of Illinois Urbana-Champaign
Graph LearningLarge Language ModelsRecommender Systems
Z
Ze Yang
Meta
Jiaru Zou
Jiaru Zou
University of Illinois Urbana-Champaign
LLM ReasoningAgentsReinforcement Learning
Zhichen Zeng
Zhichen Zeng
University of Illinois at Urbana-Champaign
Ruizhong Qiu
Ruizhong Qiu
University of Illinois Urbana-Champaign
Large Language ModelsOptimizationGraph Neural Networks
Xiao Lin
Xiao Lin
Ph.D., University of Illinois Urbana-Champaign
Machine LearningGraph LearningTime Series Analysis
Dongqi Fu
Dongqi Fu
Research Scientist, Meta AI
Geometric Deep LearningSequence ModelingProbabilistic Graphical Models
Zihao Li
Zihao Li
Univerisity of Illinois at Urbana-Champaign
musicsportsvideo gamescooking
Mengting Ai
Mengting Ai
PhD Student, University of Illinois, Urabana-Champaign
efficient machine learninglarge language model
Duo Zhou
Duo Zhou
Computer Science, University of Illinois Urbana Champaign
OptimizationMachine LearningTrustworthy AI
Wenxuan Bao
Wenxuan Bao
University of Illinois Urbana-Champaign
federated learningtransfer learning
Yunzhe Li
Yunzhe Li
University of Illinois Urbana-Champaign
RecommedationData MiningInformation RetrievalLLMs
G
Gaotang Li
University of Illinois Urbana-Champaign
Cheng Qian
Cheng Qian
University of Illinois, Urbana-Champaign
Tool LearningAgent
Yu Wang
Yu Wang
PhD student, University of California, San Diego
Natural Language ProcessingMulti Modality
X
Xiangru Tang
Yale
Y
Yin Xiao
University of Illinois Urbana-Champaign
L
Liri Fang
University of Illinois Urbana-Champaign
Hui Liu
Hui Liu
Amazon
Natural Language ProcessingLarge Language ModelsArtificial Intelligence
Xianfeng Tang
Xianfeng Tang
Amazon
Machine LearningLarge Language Models
Y
Yu-jing Zhang
University of Illinois Urbana-Champaign
Chi Wang
Chi Wang
Google DeepMind
AILarge Language ModelsAI AgentsMachine LearningData Mining
Jiaxuan You
Jiaxuan You
Assistant Professor, UIUC CS
Foundation ModelsGNNLarge Language Models
Heng Ji
Heng Ji
Professor of Computer Science, AICE Director, ASKS Director, UIUC, Amazon Scholar
Natural Language ProcessingLarge Language Models
Hanghang Tong
Hanghang Tong
University of Illinois at Urbana-Champaign
Large Scale Data MiningGraph MiningSocial NetworksHealthcareMultimedia
Jingrui He
Jingrui He
University of Illinois at Urbana-Champaign
Machine LearningData MiningSocial NetworksMedical InformaticsSemiconductor Manufacturing