DeepRAG: Thinking to Retrieval Step by Step for Large Language Models

📅 2025-02-03
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address factual hallucinations in large language models (LLMs) arising from outdated or incomplete parametric knowledge, and to mitigate noise introduced by ambiguous task decomposition and redundant retrieval in conventional retrieval-augmented generation (RAG), this work formulates retrieval-augmented reasoning as a Markov decision process (MDP), enabling dynamic, stepwise retrieval decisions conditioned on the current reasoning state. The proposed method integrates reinforcement learning, iterative query decomposition, and a learnable retrieval gating mechanism to support adaptive switching between retrieval and reasoning. Evaluated on multi-hop reasoning and open-domain question answering, it achieves a 21.99% absolute improvement in answer accuracy over strong baselines. The approach significantly enhances retrieval efficiency and response quality, effectively reducing hallucinations while improving reasoning controllability and interpretability.

Technology Category

Application Category

📝 Abstract
Large Language Models (LLMs) have shown remarkable potential in reasoning while they still suffer from severe factual hallucinations due to timeliness, accuracy, and coverage of parametric knowledge. Meanwhile, integrating reasoning with retrieval-augmented generation (RAG) remains challenging due to ineffective task decomposition and redundant retrieval, which can introduce noise and degrade response quality. In this paper, we propose DeepRAG, a framework that models retrieval-augmented reasoning as a Markov Decision Process (MDP), enabling strategic and adaptive retrieval. By iteratively decomposing queries, DeepRAG dynamically determines whether to retrieve external knowledge or rely on parametric reasoning at each step. Experiments show that DeepRAG improves retrieval efficiency while improving answer accuracy by 21.99%, demonstrating its effectiveness in optimizing retrieval-augmented reasoning.
Problem

Research questions and friction points this paper is trying to address.

Large Language Models
Knowledge Updating
Retrieval Augmentation
Innovation

Methods, ideas, or system contributions that make the work stand out.

DeepRAG
MDP-Controlled Inference
Knowledge Integration
🔎 Similar Papers
No similar papers found.
Xinyan Guan
Xinyan Guan
Institute of Software, Chinese Academy of Sciences
Jiali Zeng
Jiali Zeng
Tencent
Natural Language ProcessingDeep LearningNeural Machine Translation
Fandong Meng
Fandong Meng
WeChat AI, Tencent
Machine TranslationNatural Language Processing
C
Chunlei Xin
Chinese Information Processing Laboratory, Institute of Software, Chinese Academy of Sciences; University of Chinese Academy of Sciences
Yaojie Lu
Yaojie Lu
Institute of Software, Chinese Academy of Sciences
Information ExtractionLarge Language Models
H
Hongyu Lin
Chinese Information Processing Laboratory, Institute of Software, Chinese Academy of Sciences
X
Xianpei Han
Chinese Information Processing Laboratory, Institute of Software, Chinese Academy of Sciences
Le Sun
Le Sun
Institute of Software, CAS
information_retrievalnatural_language_processing
J
Jie Zhou
Pattern Recognition Center, WeChat AI, Tencent Inc, China