Adaptive Model Ensemble for Continual Learning

📅 2025-09-24
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address task-interference and intra-layer knowledge conflicts in model ensembling—key contributors to catastrophic forgetting in continual learning—this paper proposes a meta-learning-driven adaptive hierarchical ensembling method. Our approach introduces a lightweight mixture coefficient generator that takes task identifiers as input and dynamically produces layer-wise parameter interpolation weights, enabling task- and layer-aware knowledge fusion without architectural modifications to the backbone network. The method is plug-and-play, readily enhancing mainstream continual learning algorithms. Extensive experiments on benchmarks including Split-CIFAR100 and ImageNet-R demonstrate substantial mitigation of forgetting: our method achieves average accuracy improvements of 2.3–4.7 percentage points over state-of-the-art approaches, validating its effectiveness and generalizability across diverse continual learning scenarios.

Technology Category

Application Category

📝 Abstract
Model ensemble is an effective strategy in continual learning, which alleviates catastrophic forgetting by interpolating model parameters, achieving knowledge fusion learned from different tasks. However, existing model ensemble methods usually encounter the knowledge conflict issue at task and layer levels, causing compromised learning performance in both old and new tasks. To solve this issue, we propose meta-weight-ensembler that adaptively fuses knowledge of different tasks for continual learning. Concretely, we employ a mixing coefficient generator trained via meta-learning to generate appropriate mixing coefficients for model ensemble to address the task-level knowledge conflict. The mixing coefficient is individually generated for each layer to address the layer-level knowledge conflict. In this way, we learn the prior knowledge about adaptively accumulating knowledge of different tasks in a fused model, achieving efficient learning in both old and new tasks. Meta-weight-ensembler can be flexibly combined with existing continual learning methods to boost their ability of alleviating catastrophic forgetting. Experiments on multiple continual learning datasets show that meta-weight-ensembler effectively alleviates catastrophic forgetting and achieves state-of-the-art performance.
Problem

Research questions and friction points this paper is trying to address.

Addresses knowledge conflicts in model ensemble methods for continual learning
Solves task-level and layer-level knowledge conflicts using adaptive fusion
Alleviates catastrophic forgetting while learning new tasks efficiently
Innovation

Methods, ideas, or system contributions that make the work stand out.

Meta-learning trains generator for adaptive model ensemble coefficients
Layer-specific coefficients resolve task and layer knowledge conflicts
Flexible integration with existing methods boosts forgetting prevention
🔎 Similar Papers
No similar papers found.
Y
Yuchuan Mao
Beijing Key Laboratory of Intelligent Information Technology, School of Computer Science, Beijing Institute of Technology, China
Z
Zhi Gao
Beijing Key Laboratory of Intelligent Information Technology, School of Computer Science, Beijing Institute of Technology, China
Xiaomeng Fan
Xiaomeng Fan
Beijing Institute of Technology
machine learningcomputer vision
Yuwei Wu
Yuwei Wu
Ph.D. candidate, GRASP Lab, University of Pennsylvania
RoboticsTrajectory OptimizationTask and Motion Planning
Y
Yunde Jia
Guangdong Laboratory of Machine Perception and Intelligent Computing, Shenzhen MSU-BIT University, China
C
Chenchen Jing
Zhejiang University, Hangzhou, China