🤖 AI Summary
To address the proliferation of misinformation on social media and the challenges posed by heterogeneous, contradictory multimodal evidence (textual and visual), this paper proposes CRAVE, a retrieval-augmented large language model framework for fact-checking. CRAVE introduces a novel hierarchical clustering–based retrieval-augmented verification paradigm: it jointly embeds cross-modal evidence via alignment-aware encoding, clusters multi-source evidence to construct coherent narratives, and employs an LLM-based adjudicator to assess veracity. Crucially, it incorporates a proxy-driven iterative refinement mechanism to balance evidential consistency and diversity. The framework automatically generates interpretable, auditable verdicts. Extensive evaluation across multiple benchmarks demonstrates that CRAVE significantly outperforms state-of-the-art methods in retrieval precision, clustering quality, and verdict accuracy—validating its robustness, interpretability, and practical utility as a decision-support tool for automated fact-checking.
📝 Abstract
We propose CRAVE (Cluster-based Retrieval Augmented Verification with Explanation); a novel framework that integrates retrieval-augmented Large Language Models (LLMs) with clustering techniques to address fact-checking challenges on social media. CRAVE automatically retrieves multimodal evidence from diverse, often contradictory, sources. Evidence is clustered into coherent narratives, and evaluated via an LLM-based judge to deliver fact-checking verdicts explained by evidence summaries. By synthesizing evidence from both text and image modalities and incorporating agent-based refinement, CRAVE ensures consistency and diversity in evidence representation. Comprehensive experiments demonstrate CRAVE's efficacy in retrieval precision, clustering quality, and judgment accuracy, showcasing its potential as a robust decision-support tool for fact-checkers.