AlignX: Advancing Multilingual Large Language Models with Multilingual Representation Alignment

πŸ“… 2025-09-29
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Multilingual large language models (LLMs) suffer from substantial performance degradation on non-dominant languages and exhibit insufficient cross-lingual representation alignment, hindering effective knowledge transfer. To address this, we propose AlignXβ€”a two-stage representation-level alignment framework. In the first stage, fine-grained multilingual semantic alignment is achieved via contrastive learning; in the second stage, language-specific features are integrated with multilingual instruction tuning to jointly optimize cross-lingual understanding and generation capabilities. Unlike conventional output-layer alignment or single-stage fine-tuning, AlignX jointly models semantic commonalities and linguistic idiosyncrasies at the representation level. Extensive experiments across 12 languages and multiple pretrained LLMs (e.g., mBERT, XGLM) demonstrate that AlignX significantly narrows the multilingual performance gap: it yields average improvements of 4.2–7.8 percentage points on cross-lingual understanding benchmarks (XNLI, XCOPA) and the XGen multilingual generation task, while markedly enhancing representation alignment quality and generalization capacity.

Technology Category

Application Category

πŸ“ Abstract
Multilingual large language models (LLMs) possess impressive multilingual understanding and generation capabilities. However, their performance and cross-lingual alignment often lag for non-dominant languages. A common solution is to fine-tune LLMs on large-scale and more balanced multilingual corpus, but such approaches often lead to imprecise alignment and suboptimal knowledge transfer, struggling with limited improvements across languages. In this paper, we propose AlignX to bridge the multilingual performance gap, which is a two-stage representation-level framework for enhancing multilingual performance of pre-trained LLMs. In the first stage, we align multilingual representations with multilingual semantic alignment and language feature integration. In the second stage, we stimulate the multilingual capability of LLMs via multilingual instruction fine-tuning. Experimental results on several pre-trained LLMs demonstrate that our approach enhances LLMs' multilingual general and cross-lingual generation capability. Further analysis indicates that AlignX brings the multilingual representations closer and improves the cross-lingual alignment.
Problem

Research questions and friction points this paper is trying to address.

Improving multilingual performance for non-dominant languages
Addressing imprecise alignment in multilingual representation transfer
Enhancing cross-lingual generation capabilities through semantic alignment
Innovation

Methods, ideas, or system contributions that make the work stand out.

Two-stage framework aligns multilingual representations semantically
Integrates language features to enhance cross-lingual alignment
Uses multilingual instruction fine-tuning to stimulate capabilities
πŸ”Ž Similar Papers
No similar papers found.
Mengyu Bu
Mengyu Bu
Institute of Computing Technology, Chinese Academy of Sciences
Large Language ModelMultilingualityMachine Translation
Shaolei Zhang
Shaolei Zhang
Institute of Computing Technology, Chinese Academy of Sciences (ICT/CAS)
Natural Language ProcessingLarge Language ModelMultimodal LLMsSimultaneous Translation
Z
Zhongjun He
Baidu Inc.
H
Hua Wu
Baidu Inc.
Y
Yang Feng
Key Laboratory of Intelligent Information Processing, Institute of Computing Technology, Chinese Academy of Sciences (ICT/CAS); Key Laboratory of AI Safety, Chinese Academy of Sciences; University of Chinese Academy of Sciences, Beijing, China