🤖 AI Summary
This work addresses the challenges of alignment modeling and high inference cost in non-autoregressive text-to-speech (TTS) synthesis by proposing a flow-matching-based TTS model. The approach introduces a self-supervised semantic aligner to effectively enhance temporal and semantic consistency between text and audio, while incorporating an encoder feature reuse mechanism to accelerate inference. Additionally, a CTC auxiliary loss is integrated to strengthen semantic modeling. Experimental results demonstrate that the proposed model achieves competitive performance, attaining a word error rate (WER) of 1.98% on LibriSpeech-PC test-clean and 1.47% and 1.42% on SeedTTS test-en and test-zh, respectively. The system exhibits both high inference efficiency and superior performance compared to current state-of-the-art methods.
📝 Abstract
Although diffusion-based, non-autoregressive text-to-speech (TTS) systems have demonstrated impressive zero-shot synthesis capabilities, their efficacy is still hindered by two key challenges: the difficulty of text-speech alignment modeling and the high computational overhead of the iterative denoising process. To address these limitations, we propose ARCHI-TTS that features a dedicated semantic aligner to ensure robust temporal and semantic consistency between text and audio. To overcome high computational inference costs, ARCHI-TTS employs an efficient inference strategy that reuses encoder features across denoising steps, drastically accelerating synthesis without performance degradation. An auxiliary CTC loss applied to the condition encoder further enhances the semantic understanding. Experimental results demonstrate that ARCHI-TTS achieves a WER of 1.98% on LibriSpeech-PC test-clean, and 1.47%/1.42% on SeedTTS test-en/test-zh with a high inference efficiency, consistently outperforming recent state-of-the-art TTS systems.