Enhancing LLM Reliability via Explicit Knowledge Boundary Modeling

📅 2025-03-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Large language models (LLMs) frequently generate hallucinations beyond their knowledge boundaries, while existing uncertainty estimation or query rejection methods suffer from either low efficiency or compromised task utility. To address this, we propose the Explicit Knowledge Boundary Modeling (EKBM) framework—the first to explicitly model knowledge boundaries by coupling dual cognitive systems: “fast thinking” (efficient response generation) and “slow thinking” (deliberate refinement). EKBM introduces three core components: confidence-annotated data generation, a goal-directed slow-thinking refinement module, and an uncertainty-aware hybrid supervised training mechanism—jointly optimizing self-awareness alignment and task performance. Evaluated on dialogue state tracking, EKBM substantially outperforms uncertainty-based baselines: refinement-stage accuracy improves significantly, while computational overhead increases by less than 15%. Thus, EKBM achieves a favorable trade-off between reliability and practicality.

Technology Category

Application Category

📝 Abstract
Large language models (LLMs) frequently hallucinate due to misaligned self-awareness, generating erroneous outputs when addressing queries beyond their knowledge boundaries. While existing approaches mitigate hallucinations via uncertainty estimation or query rejection, they suffer from computational inefficiency or sacrificed helpfulness. To address these issues, we propose the Explicit Knowledge Boundary Modeling (EKBM) framework, integrating fast and slow reasoning systems to harmonize reliability and usability. The framework first employs a fast-thinking model to generate confidence-labeled responses, enabling immediate use of high-confidence outputs. For uncertain predictions, a slow refinement model conducts targeted reasoning to improve accuracy. To align model behavior with our proposed object, we propose a hybrid training pipeline, enhancing self-awareness without degrading task performance. Evaluations on dialogue state tracking tasks demonstrate that EKBM achieves superior model reliability over uncertainty-based baselines. Further analysis reveals that refinement substantially boosts accuracy while maintaining low computational overhead. Our work establishes a scalable paradigm for advancing LLM reliability and balancing accuracy and practical utility in error-sensitive applications.
Problem

Research questions and friction points this paper is trying to address.

Mitigate LLM hallucinations via knowledge boundary modeling
Balance reliability and usability in LLM responses
Enhance accuracy with low computational overhead
Innovation

Methods, ideas, or system contributions that make the work stand out.

Integrates fast and slow reasoning systems
Employs confidence-labeled responses for reliability
Uses hybrid training to enhance self-awareness
🔎 Similar Papers
No similar papers found.
Hang Zheng
Hang Zheng
Zhejiang University
array signal processingDOA estimationbeamformingtensor signal processingmachine learning
Hongshen Xu
Hongshen Xu
Shanghai Jiao Tong University
Natural Language ProcessingLarge Language ModelLLM Alignment
Y
Yuncong Liu
X-LANCE Lab, Department of Computer Science and Engineering, Shanghai Jiao Tong University
L
Lu Chen
X-LANCE Lab, Department of Computer Science and Engineering, Shanghai Jiao Tong University
Pascale Fung
Pascale Fung
Dept. of Electronic & Computer Engineering, the Hong Kong University of Science & Technology
artificial intelligenceconversational AIspeech recognitionnatural language processingAI
K
Kai Yu
X-LANCE Lab, Department of Computer Science and Engineering, Shanghai Jiao Tong University