Sparsity-Aware Unlearning for Large Language Models

πŸ“… 2026-01-31
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Existing machine unlearning methods suffer significant performance degradation when applied to sparse large language models, struggling to effectively remove sensitive information. This work proposes a novel unlearning paradigm tailored for sparse large language models, which decouples the unlearning objective from the model’s sparsification goal for the first time. The approach employs gradient masking to steer parameter updates toward retained weights and incorporates an importance-aware redistribution strategy to compensate for the impact of pruned parameters. By integrating gradient modulation, parameter redistribution, and pruning structure, the method achieves efficient removal of sensitive data while substantially outperforming existing unlearning techniques and preserving overall model performance.

Technology Category

Application Category

πŸ“ Abstract
Large Language Models (LLMs) inevitably memorize sensitive information during training, posing significant privacy risks. Machine unlearning has emerged as a promising solution to selectively remove such information without full retraining. However, existing methods are designed for dense models and overlook model sparsification-an essential technique for efficient LLM deployment. We find that unlearning effectiveness degrades substantially on sparse models. Through empirical analysis, we reveal that this degradation occurs because existing unlearning methods require updating all parameters, yet sparsification prunes substantial weights to zero, fundamentally limiting the model's forgetting capacity. To address this challenge, we propose Sparsity-Aware Unlearning (SAU), which decouples unlearning from sparsification objectives through gradient masking that redirects updates to surviving weights, combined with importance-aware redistribution to compensate for pruned parameters. Extensive experiments demonstrate that SAU significantly outperforms existing methods on sparse LLMs, achieving effective forgetting while preserving model utility.
Problem

Research questions and friction points this paper is trying to address.

machine unlearning
large language models
model sparsification
privacy
parameter pruning
Innovation

Methods, ideas, or system contributions that make the work stand out.

sparsity-aware unlearning
large language models
machine unlearning
model sparsification
gradient masking
πŸ”Ž Similar Papers
No similar papers found.
Yuze Wang
Yuze Wang
Beihang University
3D VisionComputer GraphicNeural Renderingin-the-wild Reconstruction
Yujia Tong
Yujia Tong
Wuhan University of Technology
Machine LearningEfficient Computing
Ke Xu
Ke Xu
Wuhan University
Processor
J
Jingling Yuan
School of Computer Science and Artificial Intelligence, Wuhan University of Technology, Hubei 430070, China
Jiawei Jiang
Jiawei Jiang
Wuhan University
Machine Learning SystemFederated LearningGraph Learning
C
Chuang Hu
School of Computer Science, Wuhan University, Hubei 430072, China