🤖 AI Summary
Existing evaluation metrics for model unlearning rely on reference responses or classifier outputs, measuring only superficial consistency while neglecting the core objective: whether the unlearned model’s behavior is statistically indistinguishable from a model never exposed to sensitive data—resulting in systemic assessment blind spots. This paper proposes FADE (Forget Alignment via Distribution Equivalence), the first metric that directly assesses output distribution equivalence between an unlearned model and a “from-scratch” baseline using bidirectional likelihood ratios, without requiring predefined classifiers or reference samples. FADE enables functional unlearning verification through distributional alignment. Evaluated on the TOFU and UnlearnCanvas benchmarks, FADE reveals that state-of-the-art high-scoring unlearning methods fail to achieve distribution-level unlearning and significantly deviate from the gold-standard baseline—exposing a fundamental flaw in current evaluation paradigms.
📝 Abstract
Current unlearning metrics for generative models evaluate success based on reference responses or classifier outputs rather than assessing the core objective: whether the unlearned model behaves indistinguishably from a model that never saw the unwanted data. This reference-specific approach creates systematic blind spots, allowing models to appear successful while retaining unwanted knowledge accessible through alternative prompts or attacks. We address these limitations by proposing Functional Alignment for Distributional Equivalence (FADE), a novel metric that measures distributional similarity between unlearned and reference models by comparing bidirectional likelihood assignments over generated samples. Unlike existing approaches that rely on predetermined references, FADE captures functional alignment across the entire output distribution, providing a principled assessment of genuine unlearning. Our experiments on the TOFU benchmark for LLM unlearning and the UnlearnCanvas benchmark for text-to-image diffusion model unlearning reveal that methods achieving near-optimal scores on traditional metrics fail to achieve distributional equivalence, with many becoming more distant from the gold standard than before unlearning. These findings expose fundamental gaps in current evaluation practices and demonstrate that FADE provides a more robust foundation for developing and assessing truly effective unlearning methods.