đ¤ AI Summary
Large language models (LLMs) suffer from hallucinations due to either internal uncertaintiesâsuch as knowledge gaps or contradictionsâor external uncertaintiesâsuch as ambiguous user queries. Existing methods model these two uncertainty sources separately. Method: We propose Semantic Volume, a general, unsupervised, black-box approach that requires no model access. It quantifies both internal and external uncertainty jointly by perturbing query and response embeddings in semantic space, constructing their Gram matrix, and computing its determinantâthe âsemantic volume.â Contribution/Results: This is the first method to jointly model input ambiguity (external) and knowledge uncertainty (internal). Geometrically grounded, semantic volume admits a theoretical interpretation via differential entropy, unifying and extending sampling-based semantic entropy methods. Experiments demonstrate consistent superiority over baselines across both internal and external uncertainty detection tasks, with strong robustness, black-box compatibility, and interpretabilityâsignificantly enhancing LLM output reliability.
đ Abstract
Large language models (LLMs) have demonstrated remarkable performance across diverse tasks by encoding vast amounts of factual knowledge. However, they are still prone to hallucinations, generating incorrect or misleading information, often accompanied by high uncertainty. Existing methods for hallucination detection primarily focus on quantifying internal uncertainty, which arises from missing or conflicting knowledge within the model. However, hallucinations can also stem from external uncertainty, where ambiguous user queries lead to multiple possible interpretations. In this work, we introduce Semantic Volume, a novel mathematical measure for quantifying both external and internal uncertainty in LLMs. Our approach perturbs queries and responses, embeds them in a semantic space, and computes the determinant of the Gram matrix of the embedding vectors, capturing their dispersion as a measure of uncertainty. Our framework provides a generalizable and unsupervised uncertainty detection method without requiring white-box access to LLMs. We conduct extensive experiments on both external and internal uncertainty detection, demonstrating that our Semantic Volume method consistently outperforms existing baselines in both tasks. Additionally, we provide theoretical insights linking our measure to differential entropy, unifying and extending previous sampling-based uncertainty measures such as the semantic entropy. Semantic Volume is shown to be a robust and interpretable approach to improving the reliability of LLMs by systematically detecting uncertainty in both user queries and model responses.