On the Role of Unobserved Sequences on Sample-based Uncertainty Quantification for LLMs

πŸ“… 2025-10-05
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work addresses a long-overlooked challenge in uncertainty quantification for large language models (LLMs): the substantial bias introduced by unobserved output sequencesβ€”i.e., valid continuations not sampled during estimation. Existing sampling-based methods (e.g., entropy estimation) routinely neglect the probability mass assigned to such sequences, severely limiting hallucination detection. To remedy this, we formally model the latent output distribution and explicitly incorporate the probability mass of unobserved sequences into uncertainty estimation. We propose a principled framework for improved uncertainty quantification that accounts for this missing mass. Experiments reveal that ignoring unobserved sequences underestimates entropy by an average of 38%, degrading high-risk output identification. Our method improves hallucination detection F1-score by up to 12.6% across multiple benchmarks, offering both theoretical grounding and practical tools for deploying trustworthy LLMs in safety-critical applications.

Technology Category

Application Category

πŸ“ Abstract
Quantifying uncertainty in large language models (LLMs) is important for safety-critical applications because it helps spot incorrect answers, known as hallucinations. One major trend of uncertainty quantification methods is based on estimating the entropy of the distribution of the LLM's potential output sequences. This estimation is based on a set of output sequences and associated probabilities obtained by querying the LLM several times. In this paper, we advocate and experimentally show that the probability of unobserved sequences plays a crucial role, and we recommend future research to integrate it to enhance such LLM uncertainty quantification methods.
Problem

Research questions and friction points this paper is trying to address.

Quantifying uncertainty in LLMs to detect hallucinations
Estimating entropy of potential output sequences distribution
Incorporating unobserved sequences probability for uncertainty quantification
Innovation

Methods, ideas, or system contributions that make the work stand out.

Incorporates unobserved sequences probability in uncertainty quantification
Enhances entropy estimation for LLM output distributions
Improves sample-based uncertainty methods through unobserved sequences
πŸ”Ž Similar Papers
No similar papers found.
L
Lucie Kunitomo-Jacquin
National Institute of Advanced Industrial Science and Technology (AIST), Tokyo, Japan
Edison Marrese-Taylor
Edison Marrese-Taylor
National Institute of Advanced Industrial Science and Technology (AIST)
Natural Language Processing - Machine Learning
K
Ken Fukuda
National Institute of Advanced Industrial Science and Technology (AIST), Tokyo, Japan