🤖 AI Summary
Dense retrieval models exhibit multiple shallow biases—including short-document preference, first-paragraph bias, literal matching, and repeated-entity overreliance—causing severe neglect of answer relevance; these biases interact synergistically, leading to “catastrophic” retrieval failure and significantly misleading LLMs in RAG.
Method: We introduce the first controllable bias evaluation framework, built upon relation extraction datasets (e.g., Re-DocRED), enabling quantitative measurement of compound bias effects; propose a novel robustness-oriented evaluation paradigm; and conduct systematic mechanistic analysis of state-of-the-art models (e.g., Dragon+, Contriever).
Contribution/Results: Experiments reveal that documents containing correct answers are retrieved with <3% probability; RAG performance drops by 34% relative to a no-retrieval baseline—demonstrating that bias harms more than absence of retrieval. Our work establishes theoretical foundations and practical guidelines for building factually reliable, robust, and trustworthy retrieval systems.
📝 Abstract
Dense retrieval models are commonly used in Information Retrieval (IR) applications, such as Retrieval-Augmented Generation (RAG). Since they often serve as the first step in these systems, their robustness is critical to avoid failures. In this work, by repurposing a relation extraction dataset (e.g. Re-DocRED), we design controlled experiments to quantify the impact of heuristic biases, such as favoring shorter documents, in retrievers like Dragon+ and Contriever. Our findings reveal significant vulnerabilities: retrievers often rely on superficial patterns like over-prioritizing document beginnings, shorter documents, repeated entities, and literal matches. Additionally, they tend to overlook whether the document contains the query's answer, lacking deep semantic understanding. Notably, when multiple biases combine, models exhibit catastrophic performance degradation, selecting the answer-containing document in less than 3% of cases over a biased document without the answer. Furthermore, we show that these biases have direct consequences for downstream applications like RAG, where retrieval-preferred documents can mislead LLMs, resulting in a 34% performance drop than not providing any documents at all.