🤖 AI Summary
This work exposes a fundamental robustness deficiency in current proactive deepfake defenses—particularly adversarial image perturbations. To address this, we propose the first LoRA-based *aggressive patching* framework: it employs learnable low-rank adaptation modules to bypass mainstream defenses and introduces a multi-modal feature alignment (MMFA) loss to enhance cross-modal attack transferability. Furthermore, we design an adaptive gating mechanism coupled with defensive LoRA patches to enable real-time, interpretable attack visualization and early warning. With only 1,000 face samples and a single fine-tuning round, our method efficiently breaks multiple state-of-the-art defense systems. Beyond revealing critical security vulnerabilities in existing defense paradigms, this work pioneers the “interpretable co-evolution of attack and defense” principle—establishing both theoretical foundations and practical pathways toward building robust, transparent deepfake defense architectures.
📝 Abstract
Deepfakes pose significant societal risks, motivating the development of proactive defenses that embed adversarial perturbations in facial images to prevent manipulation. However, in this paper, we show that these preemptive defenses often lack robustness and reliability. We propose a novel approach, Low-Rank Adaptation (LoRA) patching, which injects a plug-and-play LoRA patch into Deepfake generators to bypass state-of-the-art defenses. A learnable gating mechanism adaptively controls the effect of the LoRA patch and prevents gradient explosions during fine-tuning. We also introduce a Multi-Modal Feature Alignment (MMFA) loss, encouraging the features of adversarial outputs to align with those of the desired outputs at the semantic level. Beyond bypassing, we present defensive LoRA patching, embedding visible warnings in the outputs as a complementary solution to mitigate this newly identified security vulnerability. With only 1,000 facial examples and a single epoch of fine-tuning, LoRA patching successfully defeats multiple proactive defenses. These results reveal a critical weakness in current paradigms and underscore the need for more robust Deepfake defense strategies. Our code is available at https://github.com/ZOMIN28/LoRA-Patching.