🤖 AI Summary
This work addresses the challenge of detecting locally manipulated deepfake speech—audio in which only specific segments are altered via neural speech editing techniques. We introduce PartialEdit, the first dedicated benchmark dataset supporting both forgery detection and tampering localization. We formally define “partially edited deepfake speech” as a novel threat model and propose a unified benchmark framework jointly optimizing detection and localization. Crucially, we identify and model learnable forensic traces introduced by neural audio codecs, and design a feature extraction module integrating contrastive learning with spectral residual analysis. On PartialEdit, state-of-the-art detection models suffer a sharp accuracy drop to 52%, whereas our method achieves 98.3% detection accuracy and 86.7% localization IoU—demonstrating substantial improvements in robustness and practical applicability for real-world partial-editing scenarios.
📝 Abstract
Neural speech editing enables seamless partial edits to speech utterances, allowing modifications to selected content while preserving the rest of the audio unchanged. This useful technique, however, also poses new risks of deepfakes. To encourage research on detecting such partially edited deepfake speech, we introduce PartialEdit, a deepfake speech dataset curated using advanced neural editing techniques. We explore both detection and localization tasks on PartialEdit. Our experiments reveal that models trained on the existing PartialSpoof dataset fail to detect partially edited speech generated by neural speech editing models. As recent speech editing models almost all involve neural audio codecs, we also provide insights into the artifacts the model learned on detecting these deepfakes. Further information about the PartialEdit dataset and audio samples can be found on the project page: https://yzyouzhang.com/PartialEdit/index.html.