🤖 AI Summary
To address the challenges of data scarcity and inconsistent cross-lingual representations in machine translation for Indian low-resource languages (LRLs), this paper proposes a layer-wise cross-lingual representation alignment method tailored for the decoder-only multilingual large language model Aya-23 (8B). Our approach innovatively integrates centered kernel alignment (CKA) with parameter regularization (REPINA), imposing cross-lingual similarity constraints at specific intermediate decoder layers, and employs QLoRA for efficient fine-tuning. Crucially, it requires no additional parallel data. Empirical results demonstrate substantial improvements in translation quality—from Mundari, Santali, and Bhili to English—across zero-shot, few-shot, and full fine-tuning settings, with particularly strong gains under extremely low-resource conditions. This work establishes a scalable, parameter-efficient paradigm for cross-lingual representation alignment in low-resource machine translation.
📝 Abstract
The 2025 Multimodal Models for Low-Resource Contexts and Social Impact (MMLoSo) Language Challenge addresses one of India's most pressing linguistic gaps: the lack of resources for its diverse low-resource languages (LRLs). In this study, we investigate whether enforcing cross-lingual similarity in specific internal layers of a decoder-only multilingual large language model (LLM) can improve translation quality from LRL to high-resource language (HRL). Specifically, we combine Centered Kernel Alignment (CKA), a similarity metric that encourages representations of different languages to align, with REPINA, a regularization method that constrains parameter updates to remain close to the pretrained model, into a joint method we call TRepLiNa. In this research project, we experiment with zero-shot, few-shot, and fine-tuning settings using Aya-23 8B with QLoRA across MMLoSo shared task language pairs (Mundari, Santali, Bhili) with Hindi/English pivots. Our results show that aligning mid-level layers using TRepLiNa (CKA+REPINA) is a low-cost, practical approach to improving LRL translation, especially in data-scarce settings.