🤖 AI Summary
This work addresses the challenge of hallucination detection in open-ended long-text generation by large language models (LLMs), particularly in zero-resource settings—where no external knowledge is available and complex inter-factual dependencies and alignments must be modeled implicitly within the text. To this end, we propose a novel graph-structured detection framework. Methodologically, it introduces context-dependency graph modeling based on knowledge triples, integrating response segmentation, relational graph convolutional network (RGCN)-based message passing, LLM-driven inverse triple reconstruction, and consistency alignment. Crucially, our approach enables end-to-end modeling of implicit factual associations within long texts—without external retrieval. Evaluated across multiple long-text hallucination benchmarks, it significantly outperforms state-of-the-art methods in both accuracy and robustness. This establishes a scalable, zero-resource paradigm for hallucination detection in long-text LLM generation.
📝 Abstract
LLMs obtain remarkable performance but suffer from hallucinations. Most research on detecting hallucination focuses on the questions with short and concrete correct answers that are easy to check the faithfulness. Hallucination detections for text generation with open-ended answers are more challenging. Some researchers use external knowledge to detect hallucinations in generated texts, but external resources for specific scenarios are hard to access. Recent studies on detecting hallucinations in long text without external resources conduct consistency comparison among multiple sampled outputs. To handle long texts, researchers split long texts into multiple facts and individually compare the consistency of each pairs of facts. However, these methods (1) hardly achieve alignment among multiple facts; (2) overlook dependencies between multiple contextual facts. In this paper, we propose a graph-based context-aware (GCA) hallucination detection for text generations, which aligns knowledge facts and considers the dependencies between contextual knowledge triples in consistency comparison. Particularly, to align multiple facts, we conduct a triple-oriented response segmentation to extract multiple knowledge triples. To model dependencies among contextual knowledge triple (facts), we construct contextual triple into a graph and enhance triples' interactions via message passing and aggregating via RGCN. To avoid the omission of knowledge triples in long text, we conduct a LLM-based reverse verification via reconstructing the knowledge triples. Experiments show that our model enhances hallucination detection and excels all baselines.