Large Language Model Can Be a Foundation for Hidden Rationale-Based Retrieval

📅 2024-12-21
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses implicit-reasoning retrieval—a challenging task where queries and relevant documents lack direct semantic similarity and instead require inference via logical chains, causal relationships, or experiential knowledge. We propose the first framework leveraging large language models (LLMs) as implicit-reasoning retrievers. Methodologically, we design a generative binary-classification instruction template and integrate it with a cross-encoder architecture, fine-tuned efficiently via Direct Preference Optimization (DPO). The approach achieves strong zero-shot generalization and supervised fine-tuning performance. Evaluated on the Emotional Support Conversation (ESC) benchmark, our model significantly outperforms existing retrieval methods, demonstrating both reasoning effectiveness and robustness. To foster reproducibility and community advancement, we publicly release all code, model weights, and annotated datasets—establishing a foundational resource for implicit-reasoning retrieval research.

Technology Category

Application Category

📝 Abstract
Despite the recent advancement in Retrieval-Augmented Generation (RAG) systems, most retrieval methodologies are often developed for factual retrieval, which assumes query and positive documents are semantically similar. In this paper, we instead propose and study a more challenging type of retrieval task, called hidden rationale retrieval, in which query and document are not similar but can be inferred by reasoning chains, logic relationships, or empirical experiences. To address such problems, an instruction-tuned Large language model (LLM) with a cross-encoder architecture could be a reasonable choice. To further strengthen pioneering LLM-based retrievers, we design a special instruction that transforms the retrieval task into a generative task by prompting LLM to answer a binary-choice question. The model can be fine-tuned with direct preference optimization (DPO). The framework is also optimized for computational efficiency with no performance degradation. We name this retrieval framework by RaHoRe and verify its zero-shot and fine-tuned performance superiority on Emotional Support Conversation (ESC), compared with previous retrieval works. Our study suggests the potential to employ LLM as a foundation for a wider scope of retrieval tasks. Our codes, models, and datasets are available on https://github.com/flyfree5/LaHoRe.
Problem

Research questions and friction points this paper is trying to address.

Retrieving documents with hidden rationales, not semantic similarity
Using LLMs for reasoning-based retrieval tasks
Enhancing retrieval efficiency without performance loss
Innovation

Methods, ideas, or system contributions that make the work stand out.

Instruction-tuned LLM with cross-encoder architecture
Transforms retrieval into generative binary-choice task
Optimized with DPO for computational efficiency
🔎 Similar Papers
No similar papers found.
Luo Ji
Luo Ji
Alibaba Group
Reinforcement LearningAutomatic Control
F
Feixiang Guo
Geely Automobile Research Institute (Ningbo) Co., Ltd
T
Teng Chen
Geely Automobile Research Institute (Ningbo) Co., Ltd
Q
Qingqing Gu
Geely Automobile Research Institute (Ningbo) Co., Ltd
X
Xiaoyu Wang
Beijing Institute of Technology
Ningyuan Xi
Ningyuan Xi
Beihang University
LLMNature Language ProcessingMachine Learning
Y
Yihong Wang
Geely Automobile Research Institute (Ningbo) Co., Ltd
P
Peng Yu
Geely Automobile Research Institute (Ningbo) Co., Ltd
Y
Yue Zhao
Geely Automobile Research Institute (Ningbo) Co., Ltd
Hongyang Lei
Hongyang Lei
Geely Automobile Research Institute (Ningbo) Co., Ltd
Z
Zhonglin Jiang
Geely Automobile Research Institute (Ningbo) Co., Ltd
Y
Yong Chen
Geely Automobile Research Institute (Ningbo) Co., Ltd