🤖 AI Summary
Long-standing limitations in speech pretraining stem from the need for task-specific models—discriminative (e.g., ASR) and generative (e.g., TTS, speech tokenization)—which hinder unified representation learning. This paper introduces UniSpeech, the first unified speech pretraining framework built upon a shared encoder-decoder architecture. It jointly optimizes a representation encoder and a waveform-generation decoder through waveform-level autoregressive and non-autoregressive modeling, coupled with multi-task collaborative objectives to simultaneously learn discriminative and generative capabilities. Evaluated on ASR, TTS, and speech tokenization, UniSpeech achieves state-of-the-art performance across all three tasks. These results demonstrate the feasibility of replacing multiple specialized models with a single foundational model, substantially reducing pretraining cost and deployment complexity. UniSpeech thus establishes a new paradigm for speech foundation models.
📝 Abstract
Pre-training and representation learning have been playing an increasingly important role in modern speech processing. Nevertheless, different applications have been relying on different foundation models, since predominant pre-training techniques are either designed for discriminative tasks or generative tasks. In this work, we make the first attempt at building a unified pre-training framework for both types of tasks in speech. We show that with the appropriate design choices for pre-training, one can jointly learn a representation encoder and generative audio decoder that can be applied to both types of tasks. We propose UniWav, an encoder-decoder framework designed to unify pre-training representation learning and generative tasks. On speech recognition, text-to-speech, and speech tokenization, UniWav achieves comparable performance to different existing foundation models, each trained on a specific task. Our findings suggest that a single general-purpose foundation model for speech can be built to replace different foundation models, reducing the overhead and cost of pre-training.