Adversarial Example Soups: Improving Transferability and Stealthiness for Free

📅 2024-02-27
🏛️ IEEE Transactions on Information Forensics and Security
📈 Citations: 1
Influential: 0
📄 PDF
🤖 AI Summary
To address the limited transferability and stealthiness of adversarial examples in black-box transfer attacks, this paper proposes the Adversarial Example Soup (AES) framework—breaking the conventional paradigm of retaining only the optimal adversarial example by systematically reusing suboptimal candidates discarded during optimization. We introduce two complementary strategies—AES-tune and AES-rand—that jointly enhance transferability and stealthiness via parameter-space and sample-space averaging, respectively, and extend AES to multi-source adversarial example fusion. Inspired by Model Soups, AES integrates gradient-based optimization, stability evaluation, and hyperparameter tuning. Evaluated on ten defended models, AES boosts the success rates of ten state-of-the-art transfer attacks by up to 13% on average, while significantly reducing perturbation variance—thereby improving both visual naturalness and quantitative stealthiness.

Technology Category

Application Category

📝 Abstract
Transferable adversarial examples cause practical security risks since they can mislead a target model without knowing its internal knowledge. A conventional recipe for maximizing transferability is to keep only the optimal adversarial example from all those obtained in the optimization pipeline. In this paper, for the first time, we revisit this convention and demonstrate that those discarded, sub-optimal adversarial examples can be reused to boost transferability. Specifically, we propose “Adversarial Example Soups” (AES), with AES-tune for averaging discarded adversarial examples in hyperparameter tuning and AES-rand for stability testing. In addition, our AES is inspired by “model soups”, which averages weights of multiple fine-tuned models for improved accuracy without increasing inference time. Extensive experiments validate the global effectiveness of our AES, boosting 10 state-of-the-art transfer attacks and their combinations by up to 13% against 10 diverse (defensive) target models. We also show the possibility of generalizing AES to other types, e.g., directly averaging multiple in-the-wild adversarial examples that yield comparable success. A promising byproduct of AES is the improved stealthiness of adversarial examples since the perturbation variances are naturally reduced.
Problem

Research questions and friction points this paper is trying to address.

Enhancing transferability of adversarial examples by reusing sub-optimal ones
Improving stealthiness of adversarial examples through reduced perturbation variances
Generalizing adversarial example averaging to boost attack success rates
Innovation

Methods, ideas, or system contributions that make the work stand out.

Averages discarded adversarial examples for transferability
Inspired by model soups for improved accuracy
Enhances stealthiness by reducing perturbation variances
🔎 Similar Papers
No similar papers found.
B
Bo Yang
PLA Strategic Support Force Information Engineering University
H
Hengwei Zhang
PLA Strategic Support Force Information Engineering University
J
Jin-dong Wang
PLA Strategic Support Force Information Engineering University
Yulong Yang
Yulong Yang
Princeton University
Dynamics and ControlPhysics Guided Deep Learning
C
Chenhao Lin
C
Chao Shen
Zhengyu Zhao
Zhengyu Zhao
Xi'an Jiaotong University, China
Adversarial Machine LearningComputer Vision