🤖 AI Summary
Large language models (LLMs) suffer from faithfulness hallucinations in retrieval-augmented generation and summarization tasks. Method: This paper proposes a lightweight joint prediction-and-explanation framework, introducing the first regularized reinforcement learning paradigm that jointly optimizes quality rewards and explanation constraints; it employs LLM-driven data synthesis, multi-dimensional filtering, and supervised fine-tuning for cold-start initialization. Contribution/Results: We present the first end-to-end model achieving both high-accuracy binary detection and natural-language explanation of hallucinations. Evaluated on 12 heterogeneous benchmarks, it significantly outperforms GPT-4.1 and o3. With only 8B parameters, it achieves state-of-the-art performance—delivering high accuracy, high-quality explanations, and low inference overhead—thereby offering an efficient solution for trustworthy LLM deployment.
📝 Abstract
Recognizing whether outputs from large language models (LLMs) contain faithfulness hallucination is crucial for real-world applications, e.g., retrieval-augmented generation and summarization. In this paper, we introduce FaithLens, a cost-efficient and effective faithfulness hallucination detection model that can jointly provide binary predictions and corresponding explanations to improve trustworthiness. To achieve this, we first synthesize training data with explanations via advanced LLMs and apply a well-defined data filtering strategy to ensure label correctness, explanation quality, and data diversity. Subsequently, we fine-tune the model on these well-curated training data as a cold start and further optimize it with rule-based reinforcement learning, using rewards for both prediction correctness and explanation quality. Results on 12 diverse tasks show that the 8B-parameter FaithLens outperforms advanced models such as GPT-4.1 and o3. Also, FaithLens can produce high-quality explanations, delivering a distinctive balance of trustworthiness, efficiency, and effectiveness.