Cross-Lingual Activation Steering for Multilingual Language Models

πŸ“… 2026-01-23
πŸ“ˆ Citations: 1
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work addresses the significant performance gap between high-resource and low-resource languages in multilingual large language models, which stems from an imbalance between shared and language-specific neuronal representations. The authors propose a training-free, inference-stage activation modulation method that selectively adjusts neuron activations to dynamically enhance separation among language clusters without altering model weights. Their approach demonstrates that effective cross-lingual transfer arises from functional differentiation rather than strict representational alignment, thereby unlocking the model’s latent multilingual capacity. Experimental results show consistent improvements across diverse tasks, with average gains of 2.3% in accuracy on classification and 3.4% in F1 score on generation tasks, while preserving performance on high-resource languages.

Technology Category

Application Category

πŸ“ Abstract
Large language models exhibit strong multilingual capabilities, yet significant performance gaps persist between dominant and non-dominant languages. Prior work attributes this gap to imbalances between shared and language-specific neurons in multilingual representations. We propose Cross-Lingual Activation Steering (CLAS), a training-free inference-time intervention that selectively modulates neuron activations. We evaluate CLAS on classification and generation benchmarks, achieving average improvements of 2.3% (Acc.) and 3.4% (F1) respectively, while maintaining high-resource language performance. We discover that effective transfer operates through functional divergence rather than strict alignment; performance gains correlate with increased language cluster separation. Our results demonstrate that targeted activation steering can unlock latent multilingual capacity in existing models without modification to model weights.
Problem

Research questions and friction points this paper is trying to address.

multilingual language models
performance gap
cross-lingual transfer
language imbalance
neuron activation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Cross-Lingual Activation Steering
multilingual language models
neuron activation modulation
training-free intervention
language cluster separation
πŸ”Ž Similar Papers
No similar papers found.