Lyapunov Probes for Hallucination Detection in Large Foundation Models

๐Ÿ“… 2026-03-06
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
This work addresses the hallucination problem in large language models and multimodal large models by conceptualizing the model as a dynamical system, wherein factual knowledge is modeled as stable equilibrium points in the representation space. Leveraging Lyapunov stability theory, the authors propose a novel probing mechanism that integrates derivative-constrained training, systematic input perturbations, and a two-stage training procedure to effectively capture the modelโ€™s confidence decay under perturbations. This approach enables the identification of boundaries in knowledge transition regions, thereby facilitating hallucination detection. Experimental results across multiple datasets and models demonstrate that the proposed method significantly outperforms existing baselines, achieving more stable and reliable hallucination identification performance.

Technology Category

Application Category

๐Ÿ“ Abstract
We address hallucination detection in Large Language Models (LLMs) and Multimodal Large Language Models (MLLMs) by framing the problem through the lens of dynamical systems stability theory. Rather than treating hallucination as a straightforward classification task, we conceptualize (M)LLMs as dynamical systems, where factual knowledge is represented by stable equilibrium points within the representation space. Our main insight is that hallucinations tend to arise at the boundaries of knowledge-transition regions separating stable and unstable zones. To capture this phenomenon, we propose Lyapunov Probes: lightweight networks trained with derivative-based stability constraints that enforce a monotonic decay in confidence under input perturbations. By performing systematic perturbation analysis and applying a two-stage training process, these probes reliably distinguish between stable factual regions and unstable, hallucination-prone regions. Experiments on diverse datasets and models demonstrate consistent improvements over existing baselines.
Problem

Research questions and friction points this paper is trying to address.

hallucination detection
Large Language Models
Multimodal Large Language Models
dynamical systems
stability
Innovation

Methods, ideas, or system contributions that make the work stand out.

Lyapunov Probes
hallucination detection
dynamical systems
stability theory
representation space
๐Ÿ”Ž Similar Papers
No similar papers found.
Bozhi Luan
Bozhi Luan
University of Science and Technology of China
LVLMsCVDiffusion models
G
Gen Li
Beijing Advanced Innovation Center for Future Blockchain and Privacy Computing, School of Artificial Intelligence, Beihang University
Y
Yalan Qin
Shanghai University
J
Jifeng Guo
School of Electronic and Information Engineering, State Key Laboratory of CNS/ATM, Beihang University
Y
Yun Zhou
National Key Laboratory of Information Systems Engineering, National University of Defense Technology
F
Faguo Wu
Beijing Advanced Innovation Center for Future Blockchain and Privacy Computing, School of Artificial Intelligence, Beihang University
Hongwei Zheng
Hongwei Zheng
Shanghai Jiao Tong University
่ฎก็ฎ—ๆœบ่ง†่ง‰ใ€่”้‚ฆๅญฆไน 
W
Wenjun Wu
Beijing Advanced Innovation Center for Future Blockchain and Privacy Computing, School of Artificial Intelligence, Beihang University
Z
Zhaoxin Fan
Beijing Advanced Innovation Center for Future Blockchain and Privacy Computing, School of Artificial Intelligence, Beihang University