🤖 AI Summary
This work exposes the severe vulnerability of automatic speech recognition (ASR) systems to adversarial attacks. To address both white-box and non-transferable black-box settings, we propose an efficient adversarial example generation method that integrates the Fast Gradient Sign Method (FGSM) with zeroth-order optimization, enabling low-perturbation, highly imperceptible adversarial speech at signal-to-noise ratios up to 35 dB and generation times under 60 seconds. Additionally, we design a novel data poisoning strategy that significantly degrades recognition accuracy—and induces semantic misclassifications—in mainstream open-source ASR models, including Whisper and Wav2Vec 2.0. Experimental results demonstrate high attack success rates under minimal perturbations, providing the first systematic empirical validation of real-world security risks for deployed ASR systems. Our findings establish critical empirical foundations and concrete technical pathways for advancing robustness research in speech AI.
📝 Abstract
Recent studies have demonstrated the vulnerability of Automatic Speech Recognition systems to adversarial examples, which can deceive these systems into misinterpreting input speech commands. While previous research has primarily focused on white-box attacks with constrained optimizations, and transferability based black-box attacks against commercial Automatic Speech Recognition devices, this paper explores cost efficient white-box attack and non transferability black-box adversarial attacks on Automatic Speech Recognition systems, drawing insights from approaches such as Fast Gradient Sign Method and Zeroth-Order Optimization. Further, the novelty of the paper includes how poisoning attack can degrade the performances of state-of-the-art models leading to misinterpretation of audio signals. Through experimentation and analysis, we illustrate how hybrid models can generate subtle yet impactful adversarial examples with very little perturbation having Signal Noise Ratio of 35dB that can be generated within a minute. These vulnerabilities of state-of-the-art open source model have practical security implications, and emphasize the need for adversarial security.