🤖 AI Summary
To address the cumulative utility degradation of large language models (LLMs) under sequential machine unlearning requests—caused by reliance on retained data—we propose a novel data-free cognitive rotation space unlearning paradigm. Our method models parameter updates as orthogonal rotations in a learned cognitive space: (i) an antisymmetric loss constrains rotation direction; (ii) rotation significance weights govern fine-grained forgetting granularity; and (iii) orthogonal rotation-axis regularization minimizes interference across multiple unlearning rounds. This enables angular, quantifiable, and controllable unlearning directly in parameter space. Experiments across multiple benchmarks demonstrate state-of-the-art data-free unlearning performance, significantly mitigating catastrophic utility loss while preserving both security guarantees and practical model utility.
📝 Abstract
As Large Language Models (LLMs) become increasingly prevalent, their security vulnerabilities have already drawn attention. Machine unlearning is introduced to seek to mitigate these risks by removing the influence of undesirable data. However, existing methods not only rely on the retained dataset to preserve model utility, but also suffer from cumulative catastrophic utility loss under continuous unlearning requests. To solve this dilemma, we propose a novel method, called Rotation Control Unlearning (RCU), which leverages the rotational salience weight of RCU to quantify and control the unlearning degree in the continuous unlearning process. The skew symmetric loss is designed to construct the existence of the cognitive rotation space, where the changes of rotational angle can simulate the continuous unlearning process. Furthermore, we design an orthogonal rotation axes regularization to enforce mutually perpendicular rotation directions for continuous unlearning requests, effectively minimizing interference and addressing cumulative catastrophic utility loss. Experiments on multiple datasets confirm that our method without retained dataset achieves SOTA performance.