🤖 AI Summary
To address data scarcity in Arabic multi-dialect automatic speech recognition (ASR), this work conducts a systematic fine-tuning study of the Whisper architecture across five major dialects—Gulf, Levantine, Iraqi, Egyptian, and Maghrebi—as well as Modern Standard Arabic (MSA). We propose three strategies: (1) balanced mixed-dialect training, (2) lightweight pre-fine-tuning on MSA, and (3) empirical analysis of cross-variant feature transferability. Results show that the mixed-dialect model achieves performance comparable to dialect-specific models; fine-tuning Whisper Tiny on only a small amount of MSA data surpasses the zero-shot performance of the larger Whisper Base; and empirical analysis reveals limited representational overlap between MSA and dialects, challenging prevailing assumptions about cross-variant transfer. This study establishes a reproducible, data-efficient paradigm for low-resource multi-dialect ASR, offering practical guidance for leveraging heterogeneous Arabic speech resources.
📝 Abstract
Although commercial Arabic automatic speech recognition (ASR) systems support Modern Standard Arabic (MSA), they struggle with dialectal speech. We investigate the effect of fine-tuning OpenAI's Whisper on five major Arabic dialects (Gulf, Levantine, Iraqi, Egyptian, Maghrebi) using Mozilla Common Voice for MSA and the MASC dataset for dialectal speech. We evaluate MSA training size effects, benefits of pre-training on MSA data, and dialect-specific versus dialect-pooled models. We find that small amounts of MSA fine-tuning data yield substantial improvements for smaller models, matching larger non-fine-tuned models. While MSA pre-training shows minimal benefit, suggesting limited shared features between MSA and dialects, our dialect-pooled models perform comparably to dialect-specific ones. This indicates that pooling dialectal data, when properly balanced, can help address data scarcity in low-resource ASR without significant performance loss.