Mitigating Translationese Bias in Multilingual LLM-as-a-Judge via Disentangled Information Bottleneck

📅 2026-03-11
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the susceptibility of multilingual large language models (LLMs) to "translationese" bias when serving as evaluators, particularly their tendency to favor machine-translated text over human-written content in low-resource languages. To mitigate this systematic bias, the authors propose DIBJudge, a novel framework that introduces the disentangled information bottleneck to multilingual evaluation for the first time. By leveraging variational information compression, DIBJudge learns minimal sufficient representations for judgment while explicitly isolating spurious correlations responsible for bias into a separate branch. A cross-covariance penalty further enforces effective disentanglement between robust and biased features. Experimental results demonstrate that DIBJudge significantly outperforms strong baselines on both multilingual reward modeling benchmarks and dedicated translationese bias evaluation suites, effectively alleviating such systemic biases.

Technology Category

Application Category

📝 Abstract
Large language models (LLMs) have become a standard for multilingual evaluation, yet they exhibit a severe systematic translationese bias. In this paper, translationese bias is characterized as LLMs systematically favoring machine-translated text over human-authored references, particularly in low-resource languages. We attribute this bias to spurious correlations with (i) latent manifold alignment with English and (ii) cross-lingual predictability. To mitigate this bias, we propose DIBJudge, a robust fine-tuning framework that learns a minimally sufficient, judgment-critical representation via variational information compression, while explicitly isolating spurious factors into the dedicated bias branch. Furthermore, we incorporate a cross-covariance penalty that explicitly suppresses statistical dependence between robust and bias representations, thereby encouraging effective disentanglement. Extensive evaluations on multilingual reward modeling benchmarks and a dedicated translationese bias evaluation suite demonstrate that the proposed DIBJudge consistently outperforms strong baselines and substantially mitigates translationese bias.
Problem

Research questions and friction points this paper is trying to address.

translationese bias
multilingual LLM-as-a-Judge
spurious correlations
low-resource languages
machine-translated text
Innovation

Methods, ideas, or system contributions that make the work stand out.

translationese bias
disentangled representation
information bottleneck
cross-lingual evaluation
LLM-as-a-Judge
🔎 Similar Papers
No similar papers found.
Hongbin Zhang
Hongbin Zhang
Institute of Materials Science, TU Darmstadt
DFT and beyondinverse design
Kehai Chen
Kehai Chen
Harbin Institute of Technolgy (Shenzhen)
LLMNatural Language ProcessingAgentMulti-model Generation
X
Xuefen Bai
Institute of Computing and Intelligence, Harbin Institute of Technology, Shenzhen, China
Y
Youcheng Pan
Pengcheng Laboratory, Shenzhen, China
Yang Xiang
Yang Xiang
Associate Professor of Peng Cheng Laboratory, China
artificial intelligencepredictive modelingmachine learningnatural language processing
J
Jinpeng Wang
Keeta AI, Meituan, Beijing, China
M
Min Zhang
Institute of Computing and Intelligence, Harbin Institute of Technology, Shenzhen, China; Pengcheng Laboratory, Shenzhen, China