RAGTrace: Understanding and Refining Retrieval-Generation Dynamics in Retrieval-Augmented Generation

📅 2025-08-08
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
In Retrieval-Augmented Generation (RAG) systems, opaque interactions between retrieval and generation components hinder knowledge provenance, factual fidelity, and interpretability. Method: We propose the first evaluation framework supporting interactive, fine-grained cross-component analysis—integrating retrieval quality assessment, generation fidelity detection, and knowledge path tracing into a multi-level visual analytics system. Our architecture is co-designed via systematic literature review and expert interviews, moving beyond traditional isolated evaluation paradigms to enable traceable, end-to-end and component-level assessment. Contribution/Results: Evaluated on real-world RAG deployments and through expert validation, our framework accurately identifies failure modes (e.g., retrieval–generation misalignment, hallucination propagation), facilitates domain-specific optimization, and significantly enhances RAG system reliability and explainability.

Technology Category

Application Category

📝 Abstract
Retrieval-Augmented Generation (RAG) systems have emerged as a promising solution to enhance large language models (LLMs) by integrating external knowledge retrieval with generative capabilities. While significant advancements have been made in improving retrieval accuracy and response quality, a critical challenge remains that the internal knowledge integration and retrieval-generation interactions in RAG workflows are largely opaque. This paper introduces RAGTrace, an interactive evaluation system designed to analyze retrieval and generation dynamics in RAG-based workflows. Informed by a comprehensive literature review and expert interviews, the system supports a multi-level analysis approach, ranging from high-level performance evaluation to fine-grained examination of retrieval relevance, generation fidelity, and cross-component interactions. Unlike conventional evaluation practices that focus on isolated retrieval or generation quality assessments, RAGTrace enables an integrated exploration of retrieval-generation relationships, allowing users to trace knowledge sources and identify potential failure cases. The system's workflow allows users to build, evaluate, and iterate on retrieval processes tailored to their specific domains of interest. The effectiveness of the system is demonstrated through case studies and expert evaluations on real-world RAG applications.
Problem

Research questions and friction points this paper is trying to address.

Analyzing opaque retrieval-generation interactions in RAG systems
Evaluating retrieval relevance and generation fidelity in workflows
Tracing knowledge sources to identify RAG failure cases
Innovation

Methods, ideas, or system contributions that make the work stand out.

Interactive system analyzes retrieval-generation dynamics
Multi-level evaluation from performance to fine-grained
Traces knowledge sources and identifies failures
🔎 Similar Papers
No similar papers found.
Sizhe Cheng
Sizhe Cheng
Nanyang Technological University
Human-computer InteractionHuman-AI CollaborationVisual Analysis
J
Jiaping Li
Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, Guangdong, China
Huanchen Wang
Huanchen Wang
City University of Hong Kong; Southern University of Science and Technology
HCIVisualizationHuman-AI CollaborationIntangible Cultural HeritageGenerative AI
Y
Yuxin Ma
Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, Guangdong, China