Learning the Mechanism of Catastrophic Forgetting: A Perspective from Gradient Similarity

📅 2026-01-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses catastrophic forgetting in large language models during continual knowledge injection. The authors provide the first theoretical explanation from the perspective of gradient similarity, revealing that strong negative gradient similarity is the root cause of forgetting. Based on this insight, they categorize neurons into conflicting and synergistic types. They propose Synergistic Neuron Learning (CNL), a method that updates only synergistic neurons while freezing conflicting ones, thereby theoretically eliminating forgetting. Extensive experiments across five large language models, four datasets, and four optimizers demonstrate that CNL achieves zero forgetting in in-distribution scenarios and reduces forgetting by 59.1%–81.7% in out-of-distribution settings.

Technology Category

Application Category

📝 Abstract
Catastrophic forgetting during knowledge injection severely undermines the continual learning capability of large language models (LLMs). Although existing methods attempt to mitigate this issue, they often lack a foundational theoretical explanation. We establish a gradient-based theoretical framework to explain catastrophic forgetting. We first prove that strongly negative gradient similarity is a fundamental cause of forgetting. We then use gradient similarity to identify two types of neurons: conflicting neurons that induce forgetting and account for 50%-75% of neurons, and collaborative neurons that mitigate forgetting and account for 25%-50%. Based on this analysis, we propose a knowledge injection method, Collaborative Neural Learning (CNL). By freezing conflicting neurons and updating only collaborative neurons, CNL theoretically eliminates catastrophic forgetting under an infinitesimal learning rate eta and an exactly known mastered set. Experiments on five LLMs, four datasets, and four optimizers show that CNL achieves zero forgetting in in-set settings and reduces forgetting by 59.1%-81.7% in out-of-set settings.
Problem

Research questions and friction points this paper is trying to address.

catastrophic forgetting
continual learning
large language models
knowledge injection
Innovation

Methods, ideas, or system contributions that make the work stand out.

catastrophic forgetting
gradient similarity
collaborative neurons
conflicting neurons
continual learning
🔎 Similar Papers
No similar papers found.
M
Mutian Yang
Department of Electronic Engineering, Tsinghua University, Beijing, China
Z
Zisen Zhan
Institute of Medical Technology, Peking University Health Science Center, Peking University, Beijing, China
Yutong Chen
Yutong Chen
ETH Zurich
Computer Vision Natural Language Processing
H
Haolin Li
Department of Electronic Engineering, Tsinghua University, Beijing, China
K
Kaiwen Wang
Department of Electronic Engineering, Tsinghua University, Beijing, China
K
Kaili Zheng
Department of Electronic Engineering, Tsinghua University, Beijing, China
Y
Yuguang Wang
Institute of Medical Technology, Peking University Health Science Center, Peking University, Beijing, China
Q
Qi Wang
College of Information Science and Engineering, Northeastern University, Shenyang, China
J
Jiandong Gao
Department of Electronic Engineering, Tsinghua University, Beijing, China
Ji Wu
Ji Wu
Tsinghua University
Artificial Intelligence,smart healthcaremachine learningpattern recognitionspeech recognition