π€ AI Summary
To address the hallucination and low faithfulness issues in retrieval-augmented generation (RAG), along with the reliance on supervised fine-tuning or high inference overhead, this paper proposes SSFOβa self-supervised alignment framework. SSFO requires no human annotations; instead, it constructs preference pairs between context-conditioned and context-free generations, and integrates an enhanced DPO loss with a benign likelihood shift mechanism to explicitly model and strengthen the modelβs dependence on retrieved evidence. Technically, it unifies contextual contrastive generation, self-supervised preference learning, and likelihood shift modeling. Evaluated on multilingual QA benchmarks, SSFO achieves substantial improvements in faithfulness (+12.3% on average) over existing state-of-the-art methods, while preserving instruction-following capability and cross-lingual generalization.
π Abstract
Retrieval-Augmented Generation (RAG) systems require Large Language Models (LLMs) to generate responses that are faithful to the retrieved context. However, faithfulness hallucination remains a critical challenge, as existing methods often require costly supervision and post-training or significant inference burdens. To overcome these limitations, we introduce Self-Supervised Faithfulness Optimization (SSFO), the first self-supervised alignment approach for enhancing RAG faithfulness. SSFO constructs preference data pairs by contrasting the model's outputs generated with and without the context. Leveraging Direct Preference Optimization (DPO), SSFO aligns model faithfulness without incurring labeling costs or additional inference burden. We theoretically and empirically demonstrate that SSFO leverages a benign form of emph{likelihood displacement}, transferring probability mass from parametric-based tokens to context-aligned tokens. Based on this insight, we propose a modified DPO loss function to encourage likelihood displacement. Comprehensive evaluations show that SSFO significantly outperforms existing methods, achieving state-of-the-art faithfulness on multiple context-based question-answering datasets. Notably, SSFO exhibits strong generalization, improving cross-lingual faithfulness and preserving general instruction-following capabilities. We release our code and model at the anonymous link: https://github.com/chkwy/SSFO