🤖 AI Summary
This work addresses the urgent need for efficient and secure machine unlearning in generative foundation models. To overcome key bottlenecks—namely, reliance on full-parameter updates, high computational cost, and dependence on original training data—we propose a low-overhead, data-free unlearning method. For the first time, we integrate Singular Value Decomposition (SVD) into the machine unlearning framework, combining low-dimensional projection with sparse parameter updates to selectively erase specific harmful knowledge. Our approach modifies only ~8% of model parameters—reducing parameter updates by 92% on average—while eliminating the need for original training data entirely. It achieves forgetting accuracy comparable to state-of-the-art methods, significantly improves training stability, and enhances computational efficiency. The method provides a scalable, verifiable technical pathway compliant with emerging AI safety and regulatory requirements.
📝 Abstract
While the capabilities of generative foundational models have advanced rapidly in recent years, methods to prevent harmful and unsafe behaviors remain underdeveloped. Among the pressing challenges in AI safety, machine unlearning (MU) has become increasingly critical to meet upcoming safety regulations. Most existing MU approaches focus on altering the most significant parameters of the model. However, these methods often require fine-tuning substantial portions of the model, resulting in high computational costs and training instabilities, which are typically mitigated by access to the original training dataset. In this work, we address these limitations by leveraging Singular Value Decomposition (SVD) to create a compact, low-dimensional projection that enables the selective forgetting of specific data points. We propose Singular Value Decomposition for Efficient Machine Unlearning (SEMU), a novel approach designed to optimize MU in two key aspects. First, SEMU minimizes the number of model parameters that need to be modified, effectively removing unwanted knowledge while making only minimal changes to the model's weights. Second, SEMU eliminates the dependency on the original training dataset, preserving the model's previously acquired knowledge without additional data requirements. Extensive experiments demonstrate that SEMU achieves competitive performance while significantly improving efficiency in terms of both data usage and the number of modified parameters.