Two Heads are Better than One: Towards Better Adversarial Robustness by Combining Transduction and Rejection

📅 2023-05-27
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the insufficient robustness of deep models under adversarial attacks, this paper proposes a synergistic defense paradigm integrating transductive learning with selective rejection. Methodologically, we extend Tramèr’s classifier-detector transformation framework—originally designed for supervised settings—to the semi-supervised transductive setting, designing an end-to-end trainable selective transductive model. Guided by theoretical analysis, we reduce the sample complexity required for robust generalization and introduce rigorous evaluation using AutoAttack and the stronger GMSA attack. Extensive experiments on multiple benchmark datasets demonstrate significant improvements in robust accuracy; notably, under GMSA, our method substantially outperforms existing approaches. These results validate the effectiveness and practicality of the “transduction + rejection” synergy as a principled defense strategy against adaptive adversarial threats.
📝 Abstract
Both transduction and rejection have emerged as important techniques for defending against adversarial perturbations. A recent work by Tram`er showed that, in the rejection-only case (no transduction), a strong rejection-solution can be turned into a strong (but computationally inefficient) non-rejection solution. This detector-to-classifier reduction has been mostly applied to give evidence that certain claims of strong selective-model solutions are susceptible, leaving the benefits of rejection unclear. On the other hand, a recent work by Goldwasser et al. showed that rejection combined with transduction can give provable guarantees (for certain problems) that cannot be achieved otherwise. Nevertheless, under recent strong adversarial attacks (GMSA, which has been shown to be much more effective than AutoAttack against transduction), Goldwasser et al.'s work was shown to have low performance in a practical deep-learning setting. In this paper, we take a step towards realizing the promise of transduction+rejection in more realistic scenarios. Theoretically, we show that a novel application of Tram`er's classifier-to-detector technique in the transductive setting can give significantly improved sample-complexity for robust generalization. While our theoretical construction is computationally inefficient, it guides us to identify an efficient transductive algorithm to learn a selective model. Extensive experiments using state of the art attacks (AutoAttack, GMSA) show that our solutions provide significantly better robust accuracy.
Problem

Research questions and friction points this paper is trying to address.

Improving adversarial robustness via transduction and rejection
Enhancing robust generalization with reduced sample complexity
Achieving higher robust accuracy against strong adversarial attacks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Combines transduction and rejection for robustness
Applies reduction technique to construct defenses
Improves robust accuracy with selective model
🔎 Similar Papers
No similar papers found.
Nils Palumbo
Nils Palumbo
PhD Student in Computer Science, UW-Madison
Y
Yang Guo
University of Wisconsin-Madison
X
Xi Wu
Google
J
Jiefeng Chen
University of Wisconsin-Madison
Yingyu Liang
Yingyu Liang
The University of Hong Kong
machine learning
S
S. Jha
University of Wisconsin-Madison