AUVIC: Adversarial Unlearning of Visual Concepts for Multi-modal Large Language Models

📅 2025-11-14
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Multimodal large language models (MLLMs) face significant challenges in selectively forgetting fine-grained visual concepts—e.g., removing specific sensitive categories (e.g., “person” or “license plate”)—while preserving model performance on semantically similar yet non-target concepts, all under regulatory constraints such as the “right to be forgotten.” Method: We propose AUVIC, the first framework systematically addressing fine-grained visual concept unlearning in MLLMs. It employs adversarial perturbation and gradient-based optimization to isolate and erase target concepts without full model retraining. Contribution/Results: We introduce VCUBench—the first benchmark for group-level visual concept unlearning—and demonstrate that AUVIC achieves state-of-the-art target-concept forgetting rates while inducing minimal degradation (<1.5% average accuracy drop) on non-target concepts. AUVIC significantly outperforms existing unlearning methods in both efficacy and generalization preservation.

Technology Category

Application Category

📝 Abstract
Multimodal Large Language Models (MLLMs) achieve impressive performance once optimized on massive datasets. Such datasets often contain sensitive or copyrighted content, raising significant data privacy concerns. Regulatory frameworks mandating the'right to be forgotten'drive the need for machine unlearning. This technique allows for the removal of target data without resource-consuming retraining. However, while well-studied for text, visual concept unlearning in MLLMs remains underexplored. A primary challenge is precisely removing a target visual concept without disrupting model performance on related entities. To address this, we introduce AUVIC, a novel visual concept unlearning framework for MLLMs. AUVIC applies adversarial perturbations to enable precise forgetting. This approach effectively isolates the target concept while avoiding unintended effects on similar entities. To evaluate our method, we construct VCUBench. It is the first benchmark designed to assess visual concept unlearning in group contexts. Experimental results demonstrate that AUVIC achieves state-of-the-art target forgetting rates while incurs minimal performance degradation on non-target concepts.
Problem

Research questions and friction points this paper is trying to address.

Removing sensitive visual concepts from MLLMs without retraining
Precisely forgetting target concepts while preserving related entities
Addressing data privacy concerns through visual concept unlearning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Adversarial perturbations enable precise visual concept forgetting
Isolates target concepts without affecting similar entities
Benchmark VCUBench assesses unlearning in group contexts
🔎 Similar Papers
No similar papers found.
H
Haokun Chen
LMU Munich, Munich, Germany
J
Jianing Li
Technical University of Munich, Munich, Germany
Y
Yao Zhang
LMU Munich, Munich, Germany
Jinhe Bi
Jinhe Bi
LMU Munich
Efficient AIM/LLM
Y
Yan Xia
University of Science and Technology of China (USTC), Hefei, China
Jindong Gu
Jindong Gu
Google Research & DeepMind, University of Oxford
Trustworthy AIAI SafetyMultimodal AI
Volker Tresp
Volker Tresp
Ludwig-Maximilians-Universität München (LMU Munich)
Machine LearningArtificial IntelligenceComputational Cognitive NeuroscienceKnowledge Graphs