🤖 AI Summary
Large language models (LLMs) frequently omit critical information or inadequately cover key perspectives—especially concerning sensitive topics—posing significant reliability and safety risks. Method: This paper introduces three automated approaches for evaluating factual completeness: (1) sentence decomposition grounded in natural language inference (NLI), (2) question-answer pair extraction and alignment, and (3) end-to-end missing-information detection using large models. Contribution/Results: Experimental results demonstrate that the end-to-end method substantially outperforms conventional, complex strategies in identifying omissions, validating the efficacy of simple, discriminative LLM-based paradigms. We systematically benchmark multiple leading open-source LLMs across multi-source query tasks, revealing—through the first empirical evidence—a pervasive, systematic tendency toward information omission. This work establishes a novel, comprehensive evaluation benchmark and provides an extensible technical framework for assessing informational coverage and factual completeness in LLM-generated text.
📝 Abstract
Despite demonstrating remarkable performance across a wide range of tasks, large language models (LLMs) have also been found to frequently produce outputs that are incomplete or selectively omit key information. In sensitive domains, such omissions can result in significant harm comparable to that posed by factual inaccuracies, including hallucinations. In this study, we address the challenge of evaluating the comprehensiveness of LLM-generated texts, focusing on the detection of missing information or underrepresented viewpoints. We investigate three automated evaluation strategies: (1) an NLI-based method that decomposes texts into atomic statements and uses natural language inference (NLI) to identify missing links, (2) a Q&A-based approach that extracts question-answer pairs and compares responses across sources, and (3) an end-to-end method that directly identifies missing content using LLMs. Our experiments demonstrate the surprising effectiveness of the simple end-to-end approach compared to more complex methods, though at the cost of reduced robustness, interpretability and result granularity. We further assess the comprehensiveness of responses from several popular open-weight LLMs when answering user queries based on multiple sources.