Vulnerabilities of Audio-Based Biometric Authentication Systems Against Deepfake Speech Synthesis

📅 2026-01-06
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses the critical threat posed by deepfake audio to speaker verification systems, particularly in high-stakes security scenarios where such systems can be readily bypassed. It presents the first systematic evaluation of leading commercial speaker verification platforms under realistic deepfake attacks, leveraging state-of-the-art voice cloning models and cross-domain synthetic datasets. The findings reveal two fundamental vulnerabilities: attackers can achieve highly effective spoofing with only a minimal number of genuine voice samples, and current anti-spoofing detectors exhibit markedly poor generalization when confronted with deepfakes generated by unseen synthesis methods. These results underscore the fragility of existing authentication architectures and provide empirical grounding and strategic direction for developing more robust defense mechanisms.

Technology Category

Application Category

📝 Abstract
As audio deepfakes transition from research artifacts to widely available commercial tools, robust biometric authentication faces pressing security threats in high-stakes industries. This paper presents a systematic empirical evaluation of state-of-the-art speaker authentication systems based on a large-scale speech synthesis dataset, revealing two major security vulnerabilities: 1) modern voice cloning models trained on very small samples can easily bypass commercial speaker verification systems; and 2) anti-spoofing detectors struggle to generalize across different methods of audio synthesis, leading to a significant gap between in-domain performance and real-world robustness. These findings call for a reconsideration of security measures and stress the need for architectural innovations, adaptive defenses, and the transition towards multi-factor authentication.
Problem

Research questions and friction points this paper is trying to address.

audio deepfakes
biometric authentication
speaker verification
voice cloning
anti-spoofing
Innovation

Methods, ideas, or system contributions that make the work stand out.

deepfake speech synthesis
speaker verification
anti-spoofing
voice cloning
biometric security
🔎 Similar Papers
No similar papers found.
M
Mengze Hong
Hong Kong Polytechnic University
D
Di Jiang
Hong Kong Polytechnic University
Z
Zeying Xie
AI Group, WeBank Co., Ltd
W
Weiwei Zhao
AI Group, WeBank Co., Ltd
G
Guan Wang
Hong Kong Polytechnic University
Chen Jason Zhang
Chen Jason Zhang
Hong Kong Polytechnic University
Human-Centered ComputingAI for ScienceAI for Hospitality Management