ADVICE: Answer-Dependent Verbalized Confidence Estimation

📅 2025-10-12
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Large language models (LLMs) commonly exhibit overconfidence and answer-agnostic confidence estimation in natural language confidence expression, leading to poor calibration and diminished reliability. This work identifies the lack of semantic dependency between confidence statements and generated answers as the root cause. To address this, we propose ADVICE—a novel Answer-Dependent Confidence Estimation framework—that enforces strong semantic coupling between natural language confidence expressions and answer content. ADVICE achieves this through fine-grained supervised fine-tuning and a dynamic confidence alignment mechanism. Extensive experiments across multiple benchmarks demonstrate that ADVICE significantly improves confidence calibration—reducing Expected Calibration Error (ECE) by 32–58%—while preserving original task accuracy. Unlike prior methods, ADVICE yields more balanced, interpretable, and reliable self-assessment without compromising performance, establishing a new paradigm for answer-aware confidence modeling in LLMs.

Technology Category

Application Category

📝 Abstract
Recent progress in large language models (LLMs) has enabled them to express their confidence in natural language, enhancing transparency and reliability. However, their confidence often exhibits overconfidence, the cause of which remains poorly understood. In this work, we conduct a detailed analysis of the dynamics underlying verbalized confidence and identify answer-independence as a key factor, defined as the model's failure to condition confidence on its own answer. To address this, we propose ADVICE (Answer-Dependent Verbalized Confidence Estimation), a fine-tuning framework that facilitates answer-grounded confidence estimation. Extensive experiments show that ADVICE substantially improves confidence calibration while preserving task performance. Further analyses confirm that ADVICE strengthens answer-groundedness, leading to more balanced and well-calibrated confidence distributions. Our findings shed light on the origin of overconfidence and establish a framework for more trustworthy confidence verbalization.
Problem

Research questions and friction points this paper is trying to address.

Addresses overconfidence in LLMs' verbalized confidence estimates
Identifies answer-independence as key cause of poor calibration
Proposes fine-tuning framework for answer-dependent confidence estimation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Fine-tuning framework for answer-grounded confidence estimation
Improves confidence calibration while preserving task performance
Strengthens answer-groundedness for balanced confidence distributions
🔎 Similar Papers
No similar papers found.
K
Ki Jung Seo
Department of Computer Science, Hanyang University, Seoul, Republic of Korea
S
Sehun Lim
Department of Computer Science, Hanyang University, Seoul, Republic of Korea
Taeuk Kim
Taeuk Kim
Assistant Professor, Hanyang University.
Natural Language ProcessingLarge Language ModelsMachine Learning