FaceShield: Defending Facial Image against Deepfake Threats

📅 2024-12-13
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
The escalating misuse of deepfakes—particularly diffusion models (DMs)—poses severe security threats, yet existing passive detection methods and GAN-specific active defenses lack generalization across both DMs and GANs. Method: We propose the first unified active defense framework for face images targeting both DMs and GANs, featuring three innovations: (i) a novel attention-mechanism interference strategy tailored to diffusion models; (ii) a cross-architecture (DM+GAN) transferable feature extractor to counter manipulation; and (iii) joint optimization of feature-level perturbations and signal-domain low-pass filtering, balancing imperceptibility, JPEG robustness (QF=50), and generalization. Results: Our method achieves state-of-the-art performance on CelebA-HQ and VGGFace2-HQ: <3.2% success rate against mainstream DM-based attacks and >87% cross-model defense efficacy against unseen GANs.

Technology Category

Application Category

📝 Abstract
The rising use of deepfakes in criminal activities presents a significant issue, inciting widespread controversy. While numerous studies have tackled this problem, most primarily focus on deepfake detection. These reactive solutions are insufficient as a fundamental approach for crimes where authenticity is disregarded. Existing proactive defenses also have limitations, as they are effective only for deepfake models based on specific Generative Adversarial Networks (GANs), making them less applicable in light of recent advancements in diffusion-based models. In this paper, we propose a proactive defense method named FaceShield, which introduces novel defense strategies targeting deepfakes generated by Diffusion Models (DMs) and facilitates defenses on various existing GAN-based deepfake models through facial feature extractor manipulations. Our approach consists of three main components: (i) manipulating the attention mechanism of DMs to exclude protected facial features during the denoising process, (ii) targeting prominent facial feature extraction models to enhance the robustness of our adversarial perturbation, and (iii) employing Gaussian blur and low-pass filtering techniques to improve imperceptibility while enhancing robustness against JPEG compression. Experimental results on the CelebA-HQ and VGGFace2-HQ datasets demonstrate that our method achieves state-of-the-art performance against the latest deepfake models based on DMs, while also exhibiting transferability to GANs and showcasing greater imperceptibility of noise along with enhanced robustness.
Problem

Research questions and friction points this paper is trying to address.

Addresses deepfake threats in facial images
Proposes proactive defense against diffusion-based deepfakes
Enhances robustness and imperceptibility of adversarial perturbations
Innovation

Methods, ideas, or system contributions that make the work stand out.

Manipulates DM attention to exclude facial features
Targets facial feature extraction for robust perturbations
Uses Gaussian blur and filtering for imperceptible noise
🔎 Similar Papers
No similar papers found.
Jaehwan Jeong
Jaehwan Jeong
Samsung Electronics
AI SecurityComputer Security
S
Sumin In
Korea University
S
Sieun Kim
Korea University
H
Hannie Shin
Korea University
J
Jongheon Jeong
Korea University
Sang Ho Yoon
Sang Ho Yoon
Associate Professor, KAIST
HCIInteraction TechniquesHapticsWearable ComputingAR/VR
J
Jaewook Chung
Samsung Research
Sangpil Kim
Sangpil Kim
Korea University
Computer Vision