On Transfer-based Universal Attacks in Pure Black-box Setting

📅 2025-04-11
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing black-box adversarial attack methods implicitly rely on prior knowledge—such as the target model’s training dataset and number of classes—violating the pure black-box assumption and inflating estimates of transferability. Method: This paper introduces the first systematic framework for prior-agnostic transfer attack evaluation, explicitly identifying and eliminating such priors. It proposes a novel image fusion augmentation strategy to enhance query-based surrogate model training and establishes a rigorous, reproducible, and interpretable pure black-box evaluation paradigm. Contribution/Results: Experiments demonstrate that prior knowledge substantially overestimates transfer success rates. Under strictly zero-knowledge conditions—i.e., without any access to target model priors—the proposed framework enables robust, verifiable, and query-compatible attack performance assessment. It provides a reliable benchmark for evaluating black-box adversarial robustness, advancing methodological rigor in this domain.

Technology Category

Application Category

📝 Abstract
Despite their impressive performance, deep visual models are susceptible to transferable black-box adversarial attacks. Principally, these attacks craft perturbations in a target model-agnostic manner. However, surprisingly, we find that existing methods in this domain inadvertently take help from various priors that violate the black-box assumption such as the availability of the dataset used to train the target model, and the knowledge of the number of classes in the target model. Consequently, the literature fails to articulate the true potency of transferable black-box attacks. We provide an empirical study of these biases and propose a framework that aids in a prior-free transparent study of this paradigm. Using our framework, we analyze the role of prior knowledge of the target model data and number of classes in attack performance. We also provide several interesting insights based on our analysis, and demonstrate that priors cause overestimation in transferability scores. Finally, we extend our framework to query-based attacks. This extension inspires a novel image-blending technique to prepare data for effective surrogate model training.
Problem

Research questions and friction points this paper is trying to address.

Studying biases in transferable black-box adversarial attacks
Analyzing impact of prior knowledge on attack performance
Proposing prior-free framework for transparent attack evaluation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Framework for prior-free black-box attack study
Analyzes impact of target model data priors
Image-blending technique for surrogate training
🔎 Similar Papers
2024-03-22arXiv.orgCitations: 8
M
M. Jalwana
Department of Computer Science and Software Engineering, The University of Western Australia, Perth, WA 6009, Australia
N
Naveed Akhtar
Department of Computer Science and Software Engineering, The University of Western Australia, Perth, WA 6009, Australia
A
Ajmal Mian
Department of Computer Science and Software Engineering, The University of Western Australia, Perth, WA 6009, Australia
Nazanin Rahnavard
Nazanin Rahnavard
University of Central Florida
Deep LearningCompressive SensingCognitive Radio Networksmm-wave Communications and Localization
Mubarak Shah
Mubarak Shah
Trustee Chair Professor of Computer Science, University of Central Florida
Computer Vision