LLM-based Agents Suffer from Hallucinations: A Survey of Taxonomy, Methods, and Directions

📅 2025-09-23
📈 Citations: 0
✨ Influential: 0
📄 PDF
🤖 AI Summary
Large language model (LLM)-based agents frequently suffer from hallucination—generating factually incorrect, inconsistent, or unsubstantiated content during cognition, reasoning, and interaction—severely compromising task reliability. Method: This work introduces the first holistic hallucination taxonomy spanning the agent’s entire lifecycle, identifying multi-stage hallucination types and systematically categorizing eighteen underlying causal mechanisms. Through process analysis, systematic literature review, and attribution modeling, we synthesize over one hundred existing studies to distill prevailing detection and mitigation strategies. Contribution/Results: The study establishes a unified theoretical framework for hallucination research, clarifying critical directions—including enhanced interpretability, multi-source verification, and dynamic feedback integration—that collectively advance the development of robust, trustworthy LLM-based agent systems.

Technology Category

Application Category

📝 Abstract
Driven by the rapid advancements of Large Language Models (LLMs), LLM-based agents have emerged as powerful intelligent systems capable of human-like cognition, reasoning, and interaction. These agents are increasingly being deployed across diverse real-world applications, including student education, scientific research, and financial analysis. However, despite their remarkable potential, LLM-based agents remain vulnerable to hallucination issues, which can result in erroneous task execution and undermine the reliability of the overall system design. Addressing this critical challenge requires a deep understanding and a systematic consolidation of recent advances on LLM-based agents. To this end, we present the first comprehensive survey of hallucinations in LLM-based agents. By carefully analyzing the complete workflow of agents, we propose a new taxonomy that identifies different types of agent hallucinations occurring at different stages. Furthermore, we conduct an in-depth examination of eighteen triggering causes underlying the emergence of agent hallucinations. Through a detailed review of a large number of existing studies, we summarize approaches for hallucination mitigation and detection, and highlight promising directions for future research. We hope this survey will inspire further efforts toward addressing hallucinations in LLM-based agents, ultimately contributing to the development of more robust and reliable agent systems.
Problem

Research questions and friction points this paper is trying to address.

Surveying hallucinations in LLM-based agents across different workflow stages
Analyzing eighteen triggering causes for agent hallucination emergence
Summarizing mitigation approaches and future research directions
Innovation

Methods, ideas, or system contributions that make the work stand out.

Surveyed hallucination taxonomy across agent workflow stages
Analyzed eighteen triggering causes for agent hallucinations
Summarized mitigation and detection methods from existing studies
🔎 Similar Papers
Xixun Lin
Xixun Lin
Institute of Information Engineering, Chinese Academy of Sciences
Data miningGraph representation learningLarge language model
Y
Yucheng Ning
Institute of Information Engineering, Chinese Academy of Sciences, School of Cyber Security, University of Chinese Academy of Sciences, China
J
Jingwen Zhang
Institute of Information Engineering, Chinese Academy of Sciences, School of Cyber Security, University of Chinese Academy of Sciences, China
Yan Dong
Yan Dong
Harbin Engineering University
Structural safetyfatigue strengthfatigue reliabilitycrack initiation and propgationcorrosion
Y
Yilong Liu
Institute of Information Engineering, Chinese Academy of Sciences, School of Cyber Security, University of Chinese Academy of Sciences, China
Y
Yongxuan Wu
Institute of Information Engineering, Chinese Academy of Sciences, School of Cyber Security, University of Chinese Academy of Sciences, China
X
Xiaohua Qi
Institute of Information Engineering, Chinese Academy of Sciences, School of Cyber Security, University of Chinese Academy of Sciences, China
Nan Sun
Nan Sun
University of New South Wales
CybersecurityArtificial Intelligence Applications
Y
Yanmin Shang
Institute of Information Engineering, Chinese Academy of Sciences, School of Cyber Security, University of Chinese Academy of Sciences, China
P
Pengfei Cao
Institute of Automation, Chinese Academy of Sciences, China
Lixin Zou
Lixin Zou
Wuhan University
Information RetrievalRecommender SystemReinforcement LearningLarge Language Model
X
Xu Chen
Gaoling School of Artificial Intelligence, Renmin University of China, China
C
Chuan Zhou
Academy of Mathematics and Systems Science, Chinese Academy of Sciences, School of Cyber Security, University of Chinese Academy of Sciences, China
J
Jia Wu
School of Computing, Faculty of Science and Engineering, Macquarie University, Australia
Shirui Pan
Shirui Pan
Professor, ARC Future Fellow, FQA, Director of TrustAGI Lab, Griffith University
Data MiningMachine LearningGraph Neural NetworksTrustworthy AITime Series
B
Bin Wang
Xiaomi Inc., China
Yanan Cao
Yanan Cao
Institute of Information Engineering, Chinese Academy of Sciences
K
Kai Chen
Institute of Information Engineering, Chinese Academy of Sciences, School of Cyber Security, University of Chinese Academy of Sciences, China
S
Songlin Hu
Institute of Information Engineering, Chinese Academy of Sciences, School of Cyber Security, University of Chinese Academy of Sciences, China
L
Li Guo
Institute of Information Engineering, Chinese Academy of Sciences, School of Cyber Security, University of Chinese Academy of Sciences, China