🤖 AI Summary
This work addresses the performance degradation of retrieval-augmented generation (RAG) systems in specialized domains due to distributional shift. To mitigate this issue, the authors propose TTARAG, a novel approach that introduces test-time adaptation into the RAG framework for the first time. During inference, TTARAG leverages the language model to predict relevant retrieved content and dynamically adjusts its parameters accordingly, enabling unsupervised domain adaptation without requiring additional labeled data. By integrating test-time adaptation, retrieval-augmented generation, and dynamic fine-tuning of the language model, TTARAG achieves substantial improvements over existing RAG baselines across six specialized-domain benchmarks, demonstrating both its effectiveness and generalizability.
📝 Abstract
Retrieval-Augmented Generation (RAG) has emerged as a powerful approach for enhancing large language models'question-answering capabilities through the integration of external knowledge. However, when adapting RAG systems to specialized domains, challenges arise from distribution shifts, resulting in suboptimal generalization performance. In this work, we propose TTARAG, a test-time adaptation method that dynamically updates the language model's parameters during inference to improve RAG system performance in specialized domains. Our method introduces a simple yet effective approach where the model learns to predict retrieved content, enabling automatic parameter adjustment to the target domain. Through extensive experiments across six specialized domains, we demonstrate that TTARAG achieves substantial performance improvements over baseline RAG systems. Code available at https://github.com/sunxin000/TTARAG.