🤖 AI Summary
Addressing the challenge of detecting overlapping multi-keyword utterances in mixed speech under few-shot conditions, this paper proposes MT-HuBERT, a self-supervised pretraining framework. Departing from reliance on labeled data, MT-HuBERT jointly models contextual cues and mixture components via a novel Mix-Training objective, learning separable acoustic unit representations from unlabeled mixed-speech data to enable precise localization of overlapping keywords. Its core innovation lies in the first integration of mixture decomposition modeling into the HuBERT architecture, overcoming the dual bottlenecks of speech overlap handling and label scarcity that limit conventional approaches. Evaluated on the Google Speech Commands v2 benchmark, MT-HuBERT achieves significant improvements over existing state-of-the-art methods under both clean and mixed-speech few-shot settings, demonstrating markedly enhanced generalization capability and robustness to acoustic interference.
📝 Abstract
Few-shot keyword spotting aims to detect previously unseen keywords with very limited labeled samples. A pre-training and adaptation paradigm is typically adopted for this task. While effective in clean conditions, most existing approaches struggle with mixed keyword spotting--detecting multiple overlapping keywords within a single utterance--a capability essential for real-world applications. We have previously proposed a pre-training approach based on Mix-Training (MT) to tackle the mixed keyword detection problem and demonstrated its efficiency. However, this approach is fully supervised, unable to utilize vast unlabeled data. To this end, we propose Mix-Training HuBERT (MT-HuBERT), a self-supervised learning (SSL) pre-training framework that implements the MT criterion during pre-training. MT-HuBERT predicts, in a self-supervised manner, the clean acoustic units of each constituent signal from contextual cues, in contrast to predicting compositional patterns of mixed speech. Experiments conducted on the Google Speech Commands (GSC v2) corpus demonstrate that our proposed MT-HuBERT consistently outperforms several state-of-the-art baselines in few-shot KWS tasks under both mixed and clean conditions.