🤖 AI Summary
To address the limitations of fine-tuning dependency, model complexity, and neglect of affective signals in cross-domain fake news detection, this paper proposes Affect-RAG—the first retrieval-augmented large language model framework integrating affective information. Affect-RAG embeds affective features throughout the entire pipeline: retrieval, indexing, and few-shot in-context learning (ICL), constructing an affect-aware unified vector index to enable fine-tuning-free cross-domain generalization. Its key contributions are threefold: (1) the first systematic integration of affect into both RAG and ICL paradigms; (2) the first fine-tuning-free approach achieving robust cross-domain fake news detection; and (3) substantial performance gains—up to 31.18% absolute improvement over state-of-the-art few-shot methods across three benchmark datasets—while significantly reducing computational overhead and annotation costs.
📝 Abstract
Misinformation is prevalent in various fields such as education, politics, health, etc., causing significant harm to society. However, current methods for cross-domain misinformation detection rely on effort- and resource-intensive fine-tuning and complex model structures. With the outstanding performance of LLMs, many studies have employed them for misinformation detection. Unfortunately, they focus on in-domain tasks and do not incorporate significant sentiment and emotion features (which we jointly call {em affect}). In this paper, we propose RAEmoLLM, the first retrieval augmented (RAG) LLMs framework to address cross-domain misinformation detection using in-context learning based on affective information. RAEmoLLM includes three modules. (1) In the index construction module, we apply an emotional LLM to obtain affective embeddings from all domains to construct a retrieval database. (2) The retrieval module uses the database to recommend top K examples (text-label pairs) from source domain data for target domain contents. (3) These examples are adopted as few-shot demonstrations for the inference module to process the target domain content. The RAEmoLLM can effectively enhance the general performance of LLMs in cross-domain misinformation detection tasks through affect-based retrieval, without fine-tuning. We evaluate our framework on three misinformation benchmarks. Results show that RAEmoLLM achieves significant improvements compared to the other few-shot methods on three datasets, with the highest increases of 15.64%, 31.18%, and 15.73% respectively. This project is available at https://github.com/lzw108/RAEmoLLM.