Large-Small Model Collaborative Framework for Federated Continual Learning

πŸ“… 2025-08-13
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
In federated continual learning (FCL), foundational models suffer from weak task-specific adaptability and severe catastrophic forgetting due to inaccessibility of private local data. To address this, we propose the first large–small model collaborative framework, wherein lightweight, heterogeneous local small models dynamically bridge a global foundational model and evolving private task streams. Our approach innovatively integrates continual fine-tuning, one-to-one knowledge distillation, and federated learning to enable personalized small-model training and cross-client knowledge aggregation. Under strict data isolation and communication constraints, the framework significantly mitigates forgetting while improving both forward and backward transfer performance across tasks. Crucially, it exhibits strong robustness to structural heterogeneity among local small models. Extensive experiments on multiple FCL benchmarks demonstrate consistent superiority over state-of-the-art methods.

Technology Category

Application Category

πŸ“ Abstract
Continual learning (CL) for Foundation Models (FMs) is an essential yet underexplored challenge, especially in Federated Continual Learning (FCL), where each client learns from a private, evolving task stream under strict data and communication constraints. Despite their powerful generalization abilities, FMs often exhibit suboptimal performance on local downstream tasks, as they are unable to utilize private local data. Furthermore, enabling FMs to learn new tasks without forgetting prior knowledge is inherently a challenging problem, primarily due to their immense parameter count and high model complexity. In contrast, small models can be trained locally under resource-constrained conditions and benefit from more mature CL techniques. To bridge the gap between small models and FMs, we propose the first collaborative framework in FCL, where lightweight local models act as a dynamic bridge, continually adapting to new tasks while enhancing the utility of the large model. Two novel components are also included: Small Model Continual Fine-tuning is for preventing small models from temporal forgetting; One-by-One Distillation performs personalized fusion of heterogeneous local knowledge on the server. Experimental results demonstrate its superior performance, even when clients utilize heterogeneous small models.
Problem

Research questions and friction points this paper is trying to address.

Addressing suboptimal FM performance on local tasks due to data constraints
Preventing FM forgetting in federated continual learning settings
Bridging small models and FMs for improved task adaptation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Large-Small Model Collaborative Framework
Small Model Continual Fine-tuning
One-by-One Distillation for knowledge fusion
πŸ”Ž Similar Papers
No similar papers found.
H
Hao Yu
Southwestern University of Finance and Economics
X
Xin Yang
Southwestern University of Finance and Economics
B
Boyang Fan
Southwestern University of Finance and Economics
X
Xuemei Cao
Southwestern University of Finance and Economics
Hanlin Gu
Hanlin Gu
Webank
federated learningprivacy and securityLLM
Lixin Fan
Lixin Fan
WeBank
Computer visionmachine learningartificial intelligencefederated learning
Q
Qiang Yang
Hong Kong Polytechnic University