Federated Reasoning Distillation Framework with Model Learnability-Aware Data Allocation

📅 2026-02-21
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the inefficiency of knowledge transfer in federated learning when large and small language models collaborate, which stems from disparities in learnability and domain-agnostic reasoning transfer. To tackle this, the authors propose LaDa, a novel framework that introduces, for the first time, a learnability-aware mechanism to dynamically filter high-return samples and employs contrastive distillation to align the joint reasoning-path probabilities of large and small models. This enables adaptive reasoning transfer tailored to local data distributions. LaDa features a plug-in federated distillation architecture that facilitates efficient and lightweight deployment. Experimental results demonstrate that LaDa significantly enhances the step-by-step reasoning capability of small models on local tasks and effectively bridges the bidirectional knowledge gap.

Technology Category

Application Category

📝 Abstract
Data allocation plays a critical role in federated large language model (LLM) and small language models (SLMs) reasoning collaboration. Nevertheless, existing data allocation methods fail to address an under-explored challenge in collaboration: bidirectional model learnability gap, where client-side SLMs cannot identify high-reward samples matching their learnability constraints for effective knowledge transfer from LLMs, while LLMs struggle to select samples contributing novel knowledge beyond their existing data. Furthermore, these collaboration frameworks face another key challenge: domain-agnostic reasoning transfer, where existing reasoning transfer methods fail to flexibly adapt to the local domain data, preventing SLMs from effectively acquiring step-by-step reasoning abilities within from general LLM. To address these challenges, we propose LaDa, a federated reasoning distillation framework with model learnability-aware data allocation. It introduces a model learnability-aware data filter that adaptively allocates high-reward samples based on the learnability gap between each SLM and LLM pair, effectively facilitating bidirectional knowledge transfer. We further design a domain adaptive reasoning distillation method that aligns joint probabilities of reasoning paths on filtered high-reward samples through contrastive distillation learning between SLM and LLM, enabling SLM to capture underlying reasoning patterns under local data distribution. LaDa operates as a plug-in module for existing collaboration frameworks, adapting knowledge transfer based on model learnability gaps.
Problem

Research questions and friction points this paper is trying to address.

federated reasoning
model learnability gap
data allocation
reasoning distillation
domain adaptation
Innovation

Methods, ideas, or system contributions that make the work stand out.

federated reasoning distillation
model learnability-aware data allocation
bidirectional knowledge transfer
domain-adaptive reasoning
contrastive distillation learning
🔎 Similar Papers
No similar papers found.
W
Wei Guo
Beihang University, Beijing 100191, China
Siyuan Lu
Siyuan Lu
Nanjing University
Artificial IntelligenceNatural Language ProcessingAutomatic Speech RecognitionQuantitative InvestmentAI Chip Design
X
Xiangdong Ran
Beijing University of Civil Engineering and Architecture, Beijing 100044, China
Y
Yiqi Tong
Beihang University, Beijing 100191, China
Yikun Ban
Yikun Ban
Beihang University, University of Illinois Urbana-Champaign
Reinforcement LearningEnsemble Learning
Z
Zelong Xu
Yanshan University, Qinhuangdao 066104, China
Jing Fan
Jing Fan
Research Assistant, Vanderbilt University
Human Robot InteractionBrain Computer InterfaceArtificial IntelligenceMachine LearningVirtual Reality
Z
Zixuan Huang
Beihang University, Beijing 100191, China
X
Xiao Zhang
Shandong University, Qingdao 266237, China
Z
Zhaojun Hu
Renmin University of China, Beijing 100872, China
F
Fuzhen Zhuang
Beihang University, Beijing 100191, China