3S-Attack: Spatial, Spectral and Semantic Invisible Backdoor Attack Against DNN Models

📅 2025-07-14
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing backdoor attacks are vulnerable to defenses operating in the spatial, spectral, and semantic domains. To address this, we propose the first triple-domain stealthy backdoor attack. Our method leverages Grad-CAM to identify semantically salient regions from benign samples as natural triggers, embeds them in the spectral domain, and applies pixel-level constraints to ensure visual imperceptibility, semantic plausibility, and minimal spectral perturbation. This achieves synergistic stealth across spatial, spectral, and semantic domains, significantly enhancing trigger concealment and cross-model transferability. Extensive experiments on multiple benchmark datasets demonstrate that our attack maintains high clean accuracy (>98%) while achieving >95% attack success rate. Moreover, it exhibits strong robustness against state-of-the-art detection and defense methods—including STRIP, Neural Cleanse (NC), and SPECTRE—outperforming prior approaches in stealth and resilience.

Technology Category

Application Category

📝 Abstract
Backdoor attacks involve either poisoning the training data or directly modifying the model in order to implant a hidden behavior, that causes the model to misclassify inputs when a specific trigger is present. During inference, the model maintains high accuracy on benign samples but misclassifies poisoned samples into an attacker-specified target class. Existing research on backdoor attacks has explored developing triggers in the spatial, spectral (frequency), and semantic (feature) domains, aiming to make them stealthy. While some approaches have considered designing triggers that are imperceptible in both spatial and spectral domains, few have incorporated the semantic domain. In this paper, we propose a novel backdoor attack, termed 3S-attack, which is stealthy across the spatial, spectral, and semantic domains. The key idea is to exploit the semantic features of benign samples as triggers, using Gradient-weighted Class Activation Mapping (Grad-CAM) and a preliminary model for extraction. The trigger is then embedded in the spectral domain, followed by pixel-level restrictions after converting the samples back to the spatial domain. This process minimizes the distance between poisoned and benign samples, making the attack harder to detect by existing defenses and human inspection. Extensive experiments on various datasets, along with theoretical analysis, demonstrate the stealthiness of 3S-attack and highlight the need for stronger defenses to ensure AI security. Our code is available at: https://anonymous.4open.science/r/anon-project-3776/
Problem

Research questions and friction points this paper is trying to address.

Develops stealthy backdoor attack across spatial, spectral, semantic domains
Uses semantic features as triggers to evade detection
Minimizes distance between poisoned and benign samples
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses semantic features as stealthy triggers
Embeds triggers in spectral domain
Applies pixel-level spatial restrictions
🔎 Similar Papers
No similar papers found.
J
Jianyao Yin
University of Birmingham, Birmingham, UK
L
Luca Arnaboldi
University of Birmingham, Birmingham, UK
H
Honglong Chen
China University of Petroleum (East China), Qingdao, Shandong, China
Pascal Berrang
Pascal Berrang
University of Birmingham, UK
Computer ScienceSecurityPrivacy