Leveraging Per-Instance Privacy for Machine Unlearning

📅 2025-05-24
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the key challenge in machine unlearning—quantifying the difficulty of single-sample unlearning—by proposing a fine-grained difficulty assessment method based on per-sample Rényi privacy loss, moving beyond traditional worst-case analysis. Innovatively, it links the Rényi divergence-defined privacy loss to intrinsic data hardness, revealing its strong correlation with loss barriers, and establishes the first interpretable and computationally tractable framework for unlearning difficulty evaluation. Through theoretical analysis grounded in noisy gradient descent and stochastic gradient Langevin dynamics (SGLD), the metric is rigorously validated to accurately identify hard-to-forget samples under both standard fine-tuning and privacy-preserving training. Experimental results demonstrate that the proposed method provides a principled foundation for adaptive unlearning strategies, ensuring thorough forgetting while significantly improving model utility—thereby achieving a superior utility–unlearning trade-off.

Technology Category

Application Category

📝 Abstract
We present a principled, per-instance approach to quantifying the difficulty of unlearning via fine-tuning. We begin by sharpening an analysis of noisy gradient descent for unlearning (Chien et al., 2024), obtaining a better utility-unlearning tradeoff by replacing worst-case privacy loss bounds with per-instance privacy losses (Thudi et al., 2024), each of which bounds the (Renyi) divergence to retraining without an individual data point. To demonstrate the practical applicability of our theory, we present empirical results showing that our theoretical predictions are born out both for Stochastic Gradient Langevin Dynamics (SGLD) as well as for standard fine-tuning without explicit noise. We further demonstrate that per-instance privacy losses correlate well with several existing data difficulty metrics, while also identifying harder groups of data points, and introduce novel evaluation methods based on loss barriers. All together, our findings provide a foundation for more efficient and adaptive unlearning strategies tailored to the unique properties of individual data points.
Problem

Research questions and friction points this paper is trying to address.

Quantify difficulty of unlearning via per-instance privacy losses
Improve utility-unlearning tradeoff with instance-specific privacy bounds
Develop adaptive unlearning strategies for individual data points
Innovation

Methods, ideas, or system contributions that make the work stand out.

Per-instance privacy losses for unlearning
Improved utility-unlearning tradeoff via fine-tuning
Novel evaluation methods using loss barriers
🔎 Similar Papers
No similar papers found.
N
N. Sepahvand
Department of Electrical and Computer Engineering, McGill University, Montreal, Canada; Mila, Montreal, Canada
Anvith Thudi
Anvith Thudi
CS PhD, UofT
Machine LearningComputer Security
B
Berivan Isik
Google, Mountain View, US
A
Ashmita Bhattacharyya
Department of Computer Science, University of Toronto, Toronto, Canada; Vector Institute, Toronto, Canada
Nicolas Papernot
Nicolas Papernot
University of Toronto and Vector Institute
Computer SecurityDeep LearningData Privacy
Eleni Triantafillou
Eleni Triantafillou
Google DeepMind
Machine LearningFew-shot LearningMeta-Learning
Daniel M. Roy
Daniel M. Roy
Research Director, Vector Institute; Prof., U. Toronto (Statistics, CS)
Machine learningTrustworthy AIMathematical StatisticsLearning TheoryTheoretical CS
G
G. Dziugaite
Department of Computer Science, McGill University, Montreal, Canada; Google DeepMind, Toronto, Canada