PUREVQ-GAN: Defending Data Poisoning Attacks through Vector-Quantized Bottlenecks

📅 2025-09-30
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the challenge of fine-grained backdoor triggers in data poisoning attacks—whose removal is difficult and often compromises semantic fidelity—this paper proposes a defense method based on a vector-quantized bottleneck. Specifically, it jointly models a Vector-Quantized Variational Autoencoder (VQ-VAE) with a GAN discriminator to perform semantics-preserving quantization-based purification of poisoned images. This approach disrupts trigger patterns while enforcing output adherence to the natural image distribution. Evaluated on CIFAR-10, the method achieves a zero percent poisoning success rate (PSR), maintains clean-accuracy between 91% and 95%, and accelerates inference by over 50× compared to diffusion-based defenses. The framework thus simultaneously delivers strong robustness against backdoor attacks, high-fidelity reconstruction, and superior computational efficiency.

Technology Category

Application Category

📝 Abstract
We introduce PureVQ-GAN, a defense against data poisoning that forces backdoor triggers through a discrete bottleneck using Vector-Quantized VAE with GAN discriminator. By quantizing poisoned images through a learned codebook, PureVQ-GAN destroys fine-grained trigger patterns while preserving semantic content. A GAN discriminator ensures outputs match the natural image distribution, preventing reconstruction of out-of-distribution perturbations. On CIFAR-10, PureVQ-GAN achieves 0% poison success rate (PSR) against Gradient Matching and Bullseye Polytope attacks, and 1.64% against Narcissus while maintaining 91-95% clean accuracy. Unlike diffusion-based defenses requiring hundreds of iterative refinement steps, PureVQ-GAN is over 50x faster, making it practical for real training pipelines.
Problem

Research questions and friction points this paper is trying to address.

Defending against data poisoning attacks using vector-quantized bottlenecks
Destroying backdoor triggers while preserving image semantic content
Achieving high clean accuracy with significantly faster processing speed
Innovation

Methods, ideas, or system contributions that make the work stand out.

Vector-Quantized VAE with GAN discriminator creates discrete bottleneck
Learned codebook destroys triggers while preserving semantic content
GAN discriminator ensures outputs match natural image distribution
🔎 Similar Papers
No similar papers found.
A
Alexander Branch
University of California, Los Angeles
Omead Pooladzandi
Omead Pooladzandi
ECE Researcher Deep Learning
OptimizationGenerative ModelsAdversarial DefenseDeep Learning
R
Radin Khosraviani
S
Sunay Gajanan Bhat
University of California, Los Angeles
Jeffrey Jiang
Jeffrey Jiang
UCLA
machine learningcausal inference
G
Gregory Pottie
University of California, Los Angeles