đ€ AI Summary
This study addresses the limited interpretability of deep learning models in audio anomaly detection, which often undermines user trust. By systematically comparing standard autoencoders with masked autoencoders and integrating attribution methodsânamely SmoothGrad, Integrated Gradients, and Grad-CAMâthe work evaluates the faithfulness and temporal precision of generated explanations. The authors propose a perturbation-based metric to assess explanation faithfulness and demonstrate that masked training significantly enhances explanation quality without compromising detection performance. Experiments on real-world industrial data show that masked autoencoders produce more faithful and temporally accurate explanations of anomalies, thereby substantially improving model trustworthiness.
đ Abstract
Interpretability is essential for user trust in real-world anomaly detection applications. However, deep learning models, despite their strong performance, often lack transparency. In this work, we study the interpretability of autoencoder-based models for audio anomaly detection, by comparing a standard autoencoder (AE) with a mask autoencoder (MAE) in terms of detection performance and interpretability. We applied several attribution methods, including error maps, saliency maps, SmoothGrad, Integrated Gradients, GradSHAP, and Grad-CAM. Although MAE shows a slightly lower detection, it consistently provides more faithful and temporally precise explanations, suggesting a better alignment with true anomalies. To assess the relevance of the regions highlighted by the explanation method, we propose a perturbation-based faithfulness metric that replaces them with their reconstructions to simulate normal input. Our findings, based on experiments in a real industrial scenario, highlight the importance of incorporating interpretability into anomaly detection pipelines and show that masked training improves explanation quality without compromising performance.