ChunkRAG: Novel LLM-Chunk Filtering Method for RAG Systems

๐Ÿ“… 2024-10-25
๐Ÿ›๏ธ arXiv.org
๐Ÿ“ˆ Citations: 7
โœจ Influential: 1
๐Ÿ“„ PDF
๐Ÿค– AI Summary
RAG systems frequently suffer from hallucinations and factual inaccuracies due to retrieval of irrelevant or weakly relevant passages; existing approaches are limited to coarse-grained, document-level filtering, resulting in low precision. This paper proposes the first LLM-driven, fine-grained block-level filtering framework, integrating semantic chunking, query-block alignment scoring, and LLM-based re-ranking to enable precise noise suppression prior to generation. Its core innovations are: (1) dynamic relevance assessment and filtering at the semantic block levelโ€”departing from conventional document-level paradigms; and (2) a lightweight LLM agent for efficient, interpretable block re-ranking and adaptive threshold-based pruning. Experiments demonstrate substantial hallucination reduction and consistent superiority over mainstream RAG baselines across multi-hop reasoning and fact-checking tasks, with an average 12.7% improvement in factual accuracy.

Technology Category

Application Category

๐Ÿ“ Abstract
Retrieval-Augmented Generation (RAG) systems using large language models (LLMs) often generate inaccurate responses due to the retrieval of irrelevant or loosely related information. Existing methods, which operate at the document level, fail to effectively filter out such content. We propose LLM-driven chunk filtering, ChunkRAG, a framework that enhances RAG systems by evaluating and filtering retrieved information at the chunk level. Our approach employs semantic chunking to divide documents into coherent sections and utilizes LLM-based relevance scoring to assess each chunk's alignment with the user's query. By filtering out less pertinent chunks before the generation phase, we significantly reduce hallucinations and improve factual accuracy. Experiments show that our method outperforms existing RAG models, achieving higher accuracy on tasks requiring precise information retrieval. This advancement enhances the reliability of RAG systems, making them particularly beneficial for applications like fact-checking and multi-hop reasoning.
Problem

Research questions and friction points this paper is trying to address.

Filters irrelevant chunks in RAG systems to reduce inaccuracies
Improves factual accuracy by semantic chunk-level relevance scoring
Enhances reliability for fact-checking and multi-hop reasoning tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

LLM-driven chunk filtering for RAG
Semantic chunking for coherent sections
LLM-based relevance scoring per chunk
๐Ÿ”Ž Similar Papers
No similar papers found.