Parenting: Optimizing Knowledge Selection of Retrieval-Augmented Language Models with Parameter Decoupling and Tailored Tuning

📅 2024-10-14
🏛️ arXiv.org
📈 Citations: 4
Influential: 0
📄 PDF
🤖 AI Summary
Existing retrieval-augmented generation (RAG) methods struggle to harmonize internal parametric knowledge with external retrieved knowledge, leading to insufficient factual consistency and robustness. To address this, we propose a parameter-decoupled cognitive framework that, for the first time, jointly localizes capability-specific subspaces—where knowledge representations reside—via forward and backward dual-path signals. We further introduce a type-customized low-rank fine-tuning strategy to enable dynamic, interpretable selection of knowledge sources. Our approach comprises three core components: critical parameter identification, capability-oriented subspace decoupling, and task-aware fine-tuning. Extensive evaluation across multiple datasets and architectures (LLaMA, Qwen, ChatGLM) demonstrates substantial improvements: +12.7% in knowledge-source selection accuracy and +9.3 BLEU-Fact points in generation fidelity. The method exhibits strong generalization capability and architecture-agnostic performance.

Technology Category

Application Category

📝 Abstract
Retrieval-Augmented Generation (RAG) offers an effective solution to the issues faced by Large Language Models (LLMs) in hallucination generation and knowledge obsolescence by incorporating externally retrieved knowledge. However, existing methods lack effective control mechanisms for integrating internal and external knowledge. Inspired by human cognitive processes, we propose Parenting, a novel framework that decouples, identifies, and purposefully optimizes parameter subspaces related to adherence and robustness. Specifically, Parenting utilizes a key parameter mining method that combines forward and backward propagation signals to localize subspaces representing different capabilities. Then, Parenting employs a type-tailored tuning strategy, applying specific and appropriate optimizations to different subspaces, aiming to achieve a balanced enhancement of both adherence and robustness. Extensive experiments on various datasets and models validate the effectiveness and generalizability of our method.
Problem

Research questions and friction points this paper is trying to address.

Control mechanisms for integrating internal and external knowledge in RAG
Decoupling and optimizing parameter subspaces for adherence and robustness
Balanced enhancement of knowledge adherence and robustness in LLMs
Innovation

Methods, ideas, or system contributions that make the work stand out.

Decouples parameter subspaces for adherence and robustness
Combines forward and backward propagation signals
Applies type-tailored tuning to different subspaces
🔎 Similar Papers
No similar papers found.
Yongxin Xu
Yongxin Xu
Peking University
Large Language ModelsKnowledge GraphsElectronic Medical Record Analysis
Ruizhe Zhang
Ruizhe Zhang
Purdue University
Quantum computingOptimizationMachine learningComplexity theory
X
Xinke Jiang
School of Computer Science and School of Software & Microelectronics, Peking University, Beijing, China; Key Laboratory of High Confidence Software Technologies, Ministry of Education, Beijing, China
Y
Yujie Feng
Department of Computing, The Hong Kong Polytechnic University, Hong Kong S.A.R.
Y
Yuzhen Xiao
School of Computer Science and School of Software & Microelectronics, Peking University, Beijing, China; Key Laboratory of High Confidence Software Technologies, Ministry of Education, Beijing, China
X
Xinyu Ma
School of Computer Science and School of Software & Microelectronics, Peking University, Beijing, China; Key Laboratory of High Confidence Software Technologies, Ministry of Education, Beijing, China
R
Runchuan Zhu
School of Computer Science and School of Software & Microelectronics, Peking University, Beijing, China; Key Laboratory of High Confidence Software Technologies, Ministry of Education, Beijing, China
X
Xu Chu
School of Computer Science and School of Software & Microelectronics, Peking University, Beijing, China; Key Laboratory of High Confidence Software Technologies, Ministry of Education, Beijing, China; Center on Frontiers of Computing Studies, Peking University, Beijing, China
Junfeng Zhao
Junfeng Zhao
Assistant Professor at Arizona State University, Director of BELIV Lab
Connected & Automated VehicleMotion Planning & ControlsElectric VehiclesAI/ML
Y
Yasha Wang
School of Computer Science and School of Software & Microelectronics, Peking University, Beijing, China; National Engineering Research Center For Software Engineering, Peking University, Beijing, China