🤖 AI Summary
This work proposes a hierarchical global-to-local three-stage retrieval framework to address the trade-off between comprehensiveness in global search and efficiency in local search within graph-based retrieval-augmented generation, as well as the challenges of path optimization in large-scale hierarchical graph retrieval and insufficient multi-stage re-ranking. The approach integrates inter- and intra-community contextual relationships, introduces a beam search–based dynamic re-ranking module, and employs a lightweight language model combined with a GRPO reinforcement learning strategy featuring dynamic weight rewards for effective knowledge fusion. Evaluated on the Natural Questions and HotpotQA datasets, the method significantly outperforms existing graph retrieval baselines, achieving notable improvements in both accuracy and efficiency—enabling a 1.5B-parameter model to approach the performance of a 70B-parameter model on knowledge-intensive tasks.
📝 Abstract
Graph-based Retrieval-Augmented Generation (GraphRAG) frameworks face a trade-off between the comprehensiveness of global search and the efficiency of local search. Existing methods are often challenged by navigating large-scale hierarchical graphs, optimizing retrieval paths, and balancing exploration-exploitation dynamics, frequently lacking robust multi-stage re-ranking. To overcome these deficits, we propose Deep GraphRAG, a framework designed for a balanced approach to hierarchical retrieval and adaptive integration. It introduces a hierarchical global-to-local retrieval strategy that integrates macroscopic inter-community and microscopic intra-community contextual relations. This strategy employs a three-stage process: (1) inter-community filtering, which prunes the search space using local context; (2) community-level refinement, which prioritizes relevant subgraphs via entity-interaction analysis; and (3) entity-level fine-grained search within target communities. A beam search-optimized dynamic re-ranking module guides this process, continuously filtering candidates to balance efficiency and global comprehensiveness. Deep GraphRAG also features a Knowledge Integration Module leveraging a compact LLM, trained with Dynamic Weighting Reward GRPO (DW-GRPO). This novel reinforcement learning approach dynamically adjusts reward weights to balance three key objectives: relevance, faithfulness, and conciseness. This training enables compact models (1.5B) to approach the performance of large models (70B) in the integration task. Evaluations on Natural Questions and HotpotQA demonstrate that Deep GraphRAG significantly outperforms baseline graph retrieval methods in both accuracy and efficiency.