π€ AI Summary
This work addresses a long-overlooked challenge in uncertainty quantification for large language models (LLMs): the substantial bias introduced by unobserved output sequencesβi.e., valid continuations not sampled during estimation. Existing sampling-based methods (e.g., entropy estimation) routinely neglect the probability mass assigned to such sequences, severely limiting hallucination detection. To remedy this, we formally model the latent output distribution and explicitly incorporate the probability mass of unobserved sequences into uncertainty estimation. We propose a principled framework for improved uncertainty quantification that accounts for this missing mass. Experiments reveal that ignoring unobserved sequences underestimates entropy by an average of 38%, degrading high-risk output identification. Our method improves hallucination detection F1-score by up to 12.6% across multiple benchmarks, offering both theoretical grounding and practical tools for deploying trustworthy LLMs in safety-critical applications.
π Abstract
Quantifying uncertainty in large language models (LLMs) is important for safety-critical applications because it helps spot incorrect answers, known as hallucinations. One major trend of uncertainty quantification methods is based on estimating the entropy of the distribution of the LLM's potential output sequences. This estimation is based on a set of output sequences and associated probabilities obtained by querying the LLM several times. In this paper, we advocate and experimentally show that the probability of unobserved sequences plays a crucial role, and we recommend future research to integrate it to enhance such LLM uncertainty quantification methods.