Sparse deepfake detection promotes better disentanglement

📅 2025-10-07
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the trade-off between model interpretability and detection performance in deepfake speech detection, this paper introduces a Top-K sparse activation mechanism into the embedding layer of the AASIST architecture, proposing a latent-space sparse disentangled representation method. Inspired by sparse autoencoders, the approach enforces controllable sparsity to compel the model to focus exclusively on discriminative spoofing features, thereby achieving disentangled encoding of attack patterns in the latent space. To quantitatively evaluate disentanglement quality, we design integrity and modularity metrics based on mutual information. Experimental results on the ASVSpoof2021 LA and DF evaluation sets demonstrate an EER of 23.36% with 95% sparsity—significantly improving both detection accuracy and interpretability while preserving model lightweightness.

Technology Category

Application Category

📝 Abstract
Due to the rapid progress of speech synthesis, deepfake detection has become a major concern in the speech processing community. Because it is a critical task, systems must not only be efficient and robust, but also provide interpretable explanations. Among the different approaches for explainability, we focus on the interpretation of latent representations. In such paper, we focus on the last layer of embeddings of AASIST, a deepfake detection architecture. We use a TopK activation inspired by SAEs on this layer to obtain sparse representations which are used in the decision process. We demonstrate that sparse deepfake detection can improve detection performance, with an EER of 23.36% on ASVSpoof5 test set, with 95% of sparsity. We then show that these representations provide better disentanglement, using completeness and modularity metrics based on mutual information. Notably, some attacks are directly encoded in the latent space.
Problem

Research questions and friction points this paper is trying to address.

Developing sparse deepfake detection for improved interpretability and performance
Enhancing latent representation disentanglement using TopK activation sparsity
Analyzing attack encoding in latent space via mutual information metrics
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses TopK activation for sparse representations
Applies sparse detection on AASIST embeddings layer
Improves disentanglement via mutual information metrics
A
Antoine Teissier
LIUM, Le Mans University
M
Marie Tahon
LIUM, Le Mans University
Nicolas Dugué
Nicolas Dugué
Associate professor, University of Le Mans
InterpretabilityComplex networksComputational linguistics
A
Aghilas Sini
LIUM, Le Mans University