🤖 AI Summary
This study addresses the challenge in neural target speech extraction (TSE) systems of accurately modeling fine-grained speech distortions. We propose the first human-feedback-driven iterative optimization framework: users annotate only erroneous speech segments, and the system generates an editing mask to perform localized refinement while freezing unannotated regions to preserve semantic integrity and audio fidelity. To avoid costly real-world human annotations, we design an automatic mask generation strategy based on noise power (dBFS) and probability thresholds, and simulate annotation errors using synthetically corrupted data for efficient weakly supervised training. Experiments demonstrate strong agreement between predicted and manual masks (IoU = 0.82); a user study with 22 participants shows statistically significant preference for optimized outputs (p < 0.01); and objective evaluations confirm substantial improvements in speech intelligibility and target speaker fidelity.
📝 Abstract
We present the first neural target speech extraction (TSE) system that uses human feedback for iterative refinement. Our approach allows users to mark specific segments of the TSE output, generating an edit mask. The refinement system then improves the marked sections while preserving unmarked regions. Since large-scale datasets of human-marked errors are difficult to collect, we generate synthetic datasets using various automated masking functions and train models on each. Evaluations show that models trained with noise power-based masking (in dBFS) and probabilistic thresholding perform best, aligning with human annotations. In a study with 22 participants, users showed a preference for refined outputs over baseline TSE. Our findings demonstrate that human-in-the-loop refinement is a promising approach for improving the performance of neural speech extraction.