🤖 AI Summary
Existing knowledge graph–based retrieval-augmented generation (GRAG) methods suffer from domain misalignment and insufficient graph-structural awareness. To address these challenges, we propose GPR—the first end-to-end pre-trained graph retriever operating directly over knowledge graphs. Our approach introduces three key innovations: (1) LLM-guided graph data augmentation to synthesize high-quality subgraph queries; (2) a structure-aware contrastive learning objective enabling fine-grained subgraph matching; and (3) joint multi-hop subgraph encoding with cross-modal alignment. Evaluated across two GRAG benchmarks, three mainstream large language models, and five baseline retrievers, GPR consistently achieves significant improvements in retrieval accuracy and downstream generation quality. Results demonstrate that GPR serves as a robust, structure-aware GRAG retrieval component with strong generalization capability across domains and model architectures.
📝 Abstract
Graph retrieval-augmented generation (GRAG) places high demands on graph-specific retrievers. However, existing retrievers often rely on language models pretrained on plain text, limiting their effectiveness due to domain misalignment and structure ignorance. To address these challenges, we propose GPR, a graph-based retriever pretrained directly on knowledge graphs. GPR aligns natural language questions with relevant subgraphs through LLM-guided graph augmentation and employs a structure-aware objective to learn fine-grained retrieval strategies. Experiments on two datasets, three LLM backbones, and five baselines show that GPR consistently improves both retrieval quality and downstream generation, demonstrating its effectiveness as a robust retrieval solution for GRAG.