π€ AI Summary
Large-scale zero-shot text-to-speech (TTS) models achieve high speech quality but suffer from excessive parameter counts and slow inference. This paper proposes ZipVoice, a lightweight zero-shot TTS model based on flow matching. Our method introduces three key innovations: (1) the Zipformerβa novel flow-matching decoder architecture designed for efficiency; (2) an average upsampling alignment mechanism that replaces autoregressive or attention-based duration modeling; and (3) flow distillation, which compresses the teacher model without classifier-free guidance, eliminating its computational overhead. Trained on 100K hours of multilingual speech data, ZipVoice achieves state-of-the-art speech quality while reducing model size by 3Γ and accelerating end-to-end inference by up to 30Γ over prior zero-shot TTS systems. The approach thus bridges the longstanding trade-off between synthesis fidelity and real-time efficiency.
π Abstract
Existing large-scale zero-shot text-to-speech (TTS) models deliver high speech quality but suffer from slow inference speeds due to massive parameters. To address this issue, this paper introduces ZipVoice, a high-quality flow-matching-based zero-shot TTS model with a compact model size and fast inference speed. Key designs include: 1) a Zipformer-based flow-matching decoder to maintain adequate modeling capabilities under constrained size; 2) Average upsampling-based initial speech-text alignment and Zipformer-based text encoder to improve speech intelligibility; 3) A flow distillation method to reduce sampling steps and eliminate the inference overhead associated with classifier-free guidance. Experiments on 100k hours multilingual datasets show that ZipVoice matches state-of-the-art models in speech quality, while being 3 times smaller and up to 30 times faster than a DiT-based flow-matching baseline. Codes, model checkpoints and demo samples are publicly available.