🤖 AI Summary
This paper addresses catastrophic forgetting and impaired cross-modal coordination in multimodal large language models (MLLMs) during continual instruction tuning. To tackle these challenges, we propose MCL-Instruction—the first open-source framework specifically designed for multimodal continual instruction tuning. It uniformly integrates eight state-of-the-art continual learning algorithms and establishes a standardized benchmark comprising two high-quality multimodal continual instruction datasets. The framework supports cross-modal alignment, inter-modal knowledge transfer, and progressive instruction adaptation. Comprehensive experiments demonstrate substantial mitigation of forgetting and significant improvements in cross-task generalization. All components—including source code, datasets, and evaluation tools—are publicly released to ensure full reproducibility and extensibility. MCL-Instruction thus provides foundational technical infrastructure and a community-standard benchmark for advancing research in multimodal continual learning.
📝 Abstract
Continual learning aims to equip AI systems with the ability to continuously acquire and adapt to new knowledge without forgetting previously learned information, similar to human learning. While traditional continual learning methods focusing on unimodal tasks have achieved notable success, the emergence of Multimodal Large Language Models has brought increasing attention to Multimodal Continual Learning tasks involving multiple modalities, such as vision and language. In this setting, models are expected to not only mitigate catastrophic forgetting but also handle the challenges posed by cross-modal interactions and coordination. To facilitate research in this direction, we introduce MCITlib, a comprehensive and constantly evolving code library for continual instruction tuning of Multimodal Large Language Models. In MCITlib, we have currently implemented 8 representative algorithms for Multimodal Continual Instruction Tuning and systematically evaluated them on 2 carefully selected benchmarks. MCITlib will be continuously updated to reflect advances in the Multimodal Continual Learning field. The codebase is released at https://github.com/Ghy0501/MCITlib.