Sharpness-Aware Data Poisoning Attack

📅 2023-05-24
🏛️ International Conference on Learning Representations
📈 Citations: 7
Influential: 0
📄 PDF
🤖 AI Summary
Deep neural networks (DNNs) exhibit unstable backdoor and misclassification attack performance under data poisoning due to retraining uncertainty—e.g., variations in weight initialization and optimization algorithms. Method: This paper proposes a robust poisoning method grounded in loss surface sharpness awareness. It is the first to incorporate sharpness-aware optimization into poisoning frameworks, integrating Sharpness-Aware Minimization (SAM), adversarial gradient optimization, and worst-case modeling to ensure poisoned samples remain effective across diverse retraining configurations. Contribution/Results: The method significantly improves attack success rates and cross-model/cross-optimizer generalizability on both backdoor and misclassification attacks. It achieves strong universality—requiring no knowledge of the victim’s training setup—while maintaining theoretical consistency with recent advances in flat-minima optimization and adversarial robustness. Experimental results demonstrate consistent efficacy across architectures (e.g., ResNet, ViT), optimizers (SGD, Adam), and initialization schemes, validating its robustness to real-world retraining variability.
📝 Abstract
Recent research has highlighted the vulnerability of Deep Neural Networks (DNNs) against data poisoning attacks. These attacks aim to inject poisoning samples into the models' training dataset such that the trained models have inference failures. While previous studies have executed different types of attacks, one major challenge that greatly limits their effectiveness is the uncertainty of the re-training process after the injection of poisoning samples, including the re-training initialization or algorithms. To address this challenge, we propose a novel attack method called ''Sharpness-Aware Data Poisoning Attack (SAPA)''. In particular, it leverages the concept of DNNs' loss landscape sharpness to optimize the poisoning effect on the worst re-trained model. It helps enhance the preservation of the poisoning effect, regardless of the specific retraining procedure employed. Extensive experiments demonstrate that SAPA offers a general and principled strategy that significantly enhances various types of poisoning attacks.
Problem

Research questions and friction points this paper is trying to address.

DNNs vulnerable to data poisoning attacks
Uncertainty limits effectiveness of retraining post-attack
SAPA optimizes poisoning via loss landscape sharpness
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses loss landscape sharpness concept
Optimizes poisoning for worst retrained model
Enhances attack effectiveness universally
P
P. He
Michigan State University
H
Han Xu
Michigan State University
J
J. Ren
Michigan State University
Yingqian Cui
Yingqian Cui
Michigan State University
Trustworthy AI
H
Hui Liu
Michigan State University
C
C. Aggarwal
IBM T. J. Watson Research Center
Jiliang Tang
Jiliang Tang
University Foundation Professor of Computer Science and Engineering, Michigan State University
Trustworthy AIGraph Neural NetworksFeature SelectionRecommendations