🤖 AI Summary
Traditional sign language translation (SLT) relies on monolingual text supervision, resulting in poor generalization and limited cross-lingual scalability. To address these limitations, this work proposes a multilingual SLT framework. First, it introduces language-agnostic multimodal sentence embeddings—jointly encoding multilingual text and speech—as a unified supervisory signal, enabling direct cross-lingual translation without intermediate pivoting. Second, it designs a coupling enhancement strategy that jointly applies multilingual back-translation and video-level perturbations to mitigate data scarcity in low-resource settings. Implemented within an end-to-end neural machine translation architecture, the method achieves significant improvements over monolingual embedding baselines on the BLEURT metric—particularly for low-resource language pairs—demonstrating superior generalization capability, semantic robustness, and cross-lingual extensibility.
📝 Abstract
Sign language translation (SLT) is typically trained with text in a single spoken language, which limits scalability and cross-language generalization. Earlier approaches have replaced gloss supervision with text-based sentence embeddings, but up to now, these remain tied to a specific language and modality. In contrast, here we employ language-agnostic, multimodal embeddings trained on text and speech from multiple languages to supervise SLT, enabling direct multilingual translation. To address data scarcity, we propose a coupled augmentation method that combines multilingual target augmentations (i.e. translations into many languages) with video-level perturbations, improving model robustness. Experiments show consistent BLEURT gains over text-only sentence embedding supervision, with larger improvements in low-resource settings. Our results demonstrate that language-agnostic embedding supervision, combined with coupled augmentation, provides a scalable and semantically robust alternative to traditional SLT training.