Joint Knowledge Base Completion and Question Answering by Combining Large Language Models and Small Language Models

πŸ“… 2026-04-07
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work addresses the limited synergy between knowledge base completion (KBC) and knowledge base question answering (KBQA), as well as the underutilization of large language models’ (LLMs’) reasoning capabilities in existing approaches. The authors propose JCQL, a novel framework that jointly optimizes KBC and KBQA for the first time. JCQL leverages a small language model (SLM) to enhance the accuracy of LLM-generated reasoning paths in KBQA and, in turn, uses these refined reasoning paths to incrementally fine-tune the SLM, thereby improving KBC performance. This bidirectional task enhancement not only mitigates LLM hallucinations and reduces computational overhead but also achieves state-of-the-art results on both tasks across two public benchmarks.
πŸ“ Abstract
Knowledge Bases (KBs) play a key role in various applications. As two representative KB-related tasks, knowledge base completion (KBC) and knowledge base question answering (KBQA) are closely related and inherently complementary with each other. Thus, it will be beneficial to solve the task of joint KBC and KBQA to make them reinforce each other. However, existing studies usually rely on the small language model (SLM) to enhance them jointly, and the large language model (LLM)'s strong reasoning ability is ignored. In this paper, by combining the strengths of the LLM with the SLM, we propose a novel framework JCQL, which can make these two tasks enhance each other in an iterative manner. To make KBC enhance KBQA, we augment the LLM agent-based KBQA model's reasoning paths by incorporating an SLM-trained KBC model as an action of the agent, alleviating the LLM's hallucination and high computational costs issue in KBQA. To make KBQA enhance KBC, we incrementally fine-tune the KBC model by leveraging KBQA's reasoning paths as its supplementary training data, improving the ability of the SLM in KBC. Extensive experiments over two public benchmark data sets demonstrate that JCQL surpasses all baselines for both KBC and KBQA tasks.
Problem

Research questions and friction points this paper is trying to address.

Knowledge Base Completion
Knowledge Base Question Answering
Large Language Models
Small Language Models
Joint Learning
Innovation

Methods, ideas, or system contributions that make the work stand out.

joint learning
large language model
small language model
knowledge base completion
knowledge base question answering
πŸ”Ž Similar Papers
No similar papers found.
Yinan Liu
Yinan Liu
Northeastern University
Artificial IntelligenceData IntegrationKnowledge BaseLarge Language Model
D
Dongying Lin
School of Computer Science and Engineering, Northeastern University, Shenyang, China
S
Sigang Luo
School of Computer Science and Engineering, Northeastern University, Shenyang, China
Xiaochun Yang
Xiaochun Yang
Professor of Computer Science, Northeastern University, China
database
B
Bin Wang
School of Computer Science and Engineering, Northeastern University, Shenyang, China; National Frontiers Science Center for Industrial Intelligence and Systems optimization, Northeastern University, Shenyang, China