Contextual Cues in Machine Translation: Investigating the Potential of Multi-Source Input Strategies in LLMs and NMT Systems

📅 2025-03-10
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study investigates the impact of multi-source input strategies on English/Chinese-to-Portuguese machine translation quality. We propose a context-enhanced paradigm leveraging a high-resource pivot language (e.g., English), wherein pivot translations are injected as prompts into both large language models (GPT-4o) and multilingual neural machine translation (NMT) systems (Transformer-based). Methodologically, we integrate multi-source encoding with shallow fusion techniques and systematically evaluate how language distance and resource availability modulate contextual gains. Key contributions include: (1) the first empirical validation in NMT that using a high-resource pivot language yields substantial improvements (+4.2 BLEU on a domain-specific dataset); (2) identification of language-distance sensitivity and resource dependency in contextual gain; and (3) demonstration that the strategy outperforms single-source baselines for Chinese→Portuguese translation, while gains diminish on general benchmarks—highlighting its particular efficacy in low-resource, typologically distant language pairs.

Technology Category

Application Category

📝 Abstract
We explore the impact of multi-source input strategies on machine translation (MT) quality, comparing GPT-4o, a large language model (LLM), with a traditional multilingual neural machine translation (NMT) system. Using intermediate language translations as contextual cues, we evaluate their effectiveness in enhancing English and Chinese translations into Portuguese. Results suggest that contextual information significantly improves translation quality for domain-specific datasets and potentially for linguistically distant language pairs, with diminishing returns observed in benchmarks with high linguistic variability. Additionally, we demonstrate that shallow fusion, a multi-source approach we apply within the NMT system, shows improved results when using high-resource languages as context for other translation pairs, highlighting the importance of strategic context language selection.
Problem

Research questions and friction points this paper is trying to address.

Impact of multi-source input on translation quality
Effectiveness of contextual cues in domain-specific datasets
Strategic context language selection in NMT systems
Innovation

Methods, ideas, or system contributions that make the work stand out.

Multi-source input strategies enhance translation quality.
Intermediate language translations provide effective contextual cues.
Shallow fusion improves NMT with strategic context language selection.
🔎 Similar Papers
No similar papers found.
L
Lia Shahnazaryan
P
P. Simianer
Joern Wuebker
Joern Wuebker
Lilt
Machine Translation