🤖 AI Summary
Traditional RAG systems suffer from inefficient exploration, gradient conflicts, and sparse rewards in multi-step reasoning due to static pipelines and coarse-grained, outcome-level reward signals. To address these issues, this paper proposes RAG-ProGuide—the first reinforcement learning framework guided by fine-grained, process-level rewards. Its core contributions include: (i) constructing a high-quality supervision dataset covering three sequential stages—query generation, evidence retrieval, and answer synthesis; (ii) designing a multi-stage decoupled agent architecture; and (iii) introducing a retrieval-reasoning co-finetuning mechanism. Experiments across five benchmarks demonstrate that RAG-ProGuide significantly outperforms baselines such as Search-R1. It achieves state-of-the-art performance with only 5K training samples—reducing data requirements by 94%—while enhancing training stability, accelerating convergence, and improving sample efficiency.
📝 Abstract
Retrieval-augmented generation (RAG) enhances the text generation capabilities of large language models (LLMs) by integrating external knowledge and up-to-date information. However, traditional RAG systems are limited by static workflows and lack the adaptability required for multistep reasoning and complex task management. To address these limitations, agentic RAG systems (e.g., DeepResearch) have been proposed, enabling dynamic retrieval strategies, iterative context refinement, and adaptive workflows for handling complex search queries beyond the capabilities of conventional RAG. Recent advances, such as Search-R1, have demonstrated promising gains using outcome-based reinforcement learning, where the correctness of the final answer serves as the reward signal. Nevertheless, such outcome-supervised agentic RAG methods face challenges including low exploration efficiency, gradient conflict, and sparse reward signals. To overcome these challenges, we propose to utilize fine-grained, process-level rewards to improve training stability, reduce computational costs, and enhance efficiency. Specifically, we introduce a novel method ReasonRAG that automatically constructs RAG-ProGuide, a high-quality dataset providing process-level rewards for (i) query generation, (ii) evidence extraction, and (iii) answer generation, thereby enhancing model inherent capabilities via process-supervised reinforcement learning. With the process-level policy optimization, the proposed framework empowers LLMs to autonomously invoke search, generate queries, extract relevant evidence, and produce final answers. Compared to existing approaches such as Search-R1 and traditional RAG systems, ReasonRAG, leveraging RAG-ProGuide, achieves superior performance on five benchmark datasets using only 5k training instances, significantly fewer than the 90k training instances required by Search-R1.