Semantic Prosody in Machine Translation: the English-Chinese Case of Passive Structures

📅 2025-10-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses the absence of semantic prosody modeling for passive constructions in machine translation, specifically focusing on the negative semantic prosody associated with Chinese *bèi*-passives in English-to-Chinese translation. We introduce the first semantic-prosody-annotated English–Chinese parallel dataset and propose a semantic prosody knowledge injection method, enabling the first explicit modeling of the negative prosody of *bèi*-passives. Fine-tuning on OPUS-MT, NLLB-600M, and mBART50 demonstrates that our approach significantly improves the accuracy of *bèi*-passive generation in negative contexts while effectively suppressing erroneous usage in neutral or positive contexts. Moreover, the injected prosodic knowledge exhibits cross-lingual transferability, generalizing to other multilingual models. This work fills a critical gap in semantic prosody modeling for machine translation and establishes a novel paradigm for prosody-aware neural machine translation.

Technology Category

Application Category

📝 Abstract
Semantic prosody is a collocational meaning formed through the co-occurrence of a linguistic unit and a consistent series of collocates, which should be treated separately from semantic meaning. Since words that are literal translations of each other may have different semantic prosody, more attention should be paid to this linguistic property to generate accurate translations. However, current machine translation models cannot handle this problem. To bridge the gap, we propose an approach to teach machine translation models about semantic prosody of a specific structure. We focus on Chinese BEI passives and create a dataset of English-Chinese sentence pairs with the purpose of demonstrating the negative semantic prosody of BEI passives. Then we fine-tune OPUS-MT, NLLB-600M and mBART50 models with our dataset for the English-Chinese translation task. Our results show that fine-tuned MT models perform better on using BEI passives for translating unfavourable content and avoid using it for neutral and favourable content. Also, in NLLB-600M, which is a multilingual model, this knowledge of semantic prosody can be transferred from English-Chinese translation to other language pairs, such as Spanish-Chinese.
Problem

Research questions and friction points this paper is trying to address.

Machine translation models ignore semantic prosody differences between languages
Literal translations fail to convey negative connotations of Chinese BEI passives
Current MT systems cannot handle collocational meaning in passive structures
Innovation

Methods, ideas, or system contributions that make the work stand out.

Fine-tuned MT models handle semantic prosody
Created dataset for negative BEI passive prosody
Transferred prosody knowledge across language pairs
🔎 Similar Papers
No similar papers found.
X
Xinyue Ma
Centre de Llenguatge i Computacio (CLiC), Universitat de Barcelona, Spain
P
Pol Pastells
Centre de Llenguatge i Computacio (CLiC), Universitat de Barcelona, Spain
Mireia Farrús
Mireia Farrús
Universitat de Barcelona
prosodyspeech technologiesmachine translationbiometricsethics
Mariona Taulé
Mariona Taulé
Full Professor of Linguistics, Universitat de Barcelona
Computational LinguisticsLinguisticsNatural Language Processing