Dual Consolidation for Pre-Trained Model-Based Domain-Incremental Learning

📅 2024-10-01
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address catastrophic forgetting—where pretrained knowledge is overwritten during domain incremental learning (DIL)—this paper proposes Duct, a dual-path knowledge consolidation framework. Duct jointly preserves historical domain knowledge at both the representation and classifier levels: first, via multi-stage backbone fusion and class-level semantic embedding mapping to achieve unified cross-domain feature modeling; second, through a novel semantic-guided mechanism for estimating legacy classifier weights, enabling joint alignment between the embedding space and classifier parameters. Evaluated on four standard DIL benchmarks, Duct significantly mitigates forgetting while enhancing cross-domain generalization, achieving state-of-the-art performance.

Technology Category

Application Category

📝 Abstract
Domain-Incremental Learning (DIL) involves the progressive adaptation of a model to new concepts across different domains. While recent advances in pre-trained models provide a solid foundation for DIL, learning new concepts often results in the catastrophic forgetting of pre-trained knowledge. Specifically, sequential model updates can overwrite both the representation and the classifier with knowledge from the latest domain. Thus, it is crucial to develop a representation and corresponding classifier that accommodate all seen domains throughout the learning process. To this end, we propose DUal ConsolidaTion (Duct) to unify and consolidate historical knowledge at both the representation and classifier levels. By merging the backbone of different stages, we create a representation space suitable for multiple domains incrementally. The merged representation serves as a balanced intermediary that captures task-specific features from all seen domains. Additionally, to address the mismatch between consolidated embeddings and the classifier, we introduce an extra classifier consolidation process. Leveraging class-wise semantic information, we estimate the classifier weights of old domains within the latest embedding space. By merging historical and estimated classifiers, we align them with the consolidated embedding space, facilitating incremental classification. Extensive experimental results on four benchmark datasets demonstrate Duct's state-of-the-art performance. Code is available at https://github.com/Estrella-fugaz/CVPR25-Duct
Problem

Research questions and friction points this paper is trying to address.

Prevent catastrophic forgetting in domain-incremental learning.
Develop representation and classifier for multiple domains.
Align historical and estimated classifiers with consolidated embeddings.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dual consolidation for representation and classifier
Merged backbone for multi-domain representation
Classifier consolidation using semantic information
🔎 Similar Papers
No similar papers found.
Da-Wei Zhou
Da-Wei Zhou
Associate Researcher, Nanjing University
Incremental LearningContinual LearningOpen-World LearningModel Reuse
Z
Zi-Wen Cai
School of Artificial Intelligence, Nanjing University, National Key Laboratory for Novel Software Technology, Nanjing University
Han-Jia Ye
Han-Jia Ye
Nanjing University
Machine LearningData MiningMetric LearningMeta-Learning
L
Lijun Zhang
School of Artificial Intelligence, Nanjing University, National Key Laboratory for Novel Software Technology, Nanjing University
De-Chuan Zhan
De-Chuan Zhan
Nanjing University, China
Machine LearningData Mining