🤖 AI Summary
This work addresses the pervasive biases related to gender, race, and religion in multilingual pre-trained language models by proposing the first end-to-end debiasing framework that integrates both preprocessing and postprocessing strategies. The approach combines multilingual counterfactual data augmentation, multilingual Self-Debias, and parameter-efficient fine-tuning. Furthermore, it extends the CrowS-Pairs benchmark for the first time to German, Spanish, Chinese, and Japanese. Experimental results demonstrate that the proposed method significantly reduces multiple forms of bias across all four languages, validating the effectiveness of cross-lingual debiasing information fusion. The framework consistently outperforms monolingual debiasing approaches, thereby achieving a systematic improvement in fairness for multilingual language models.
📝 Abstract
Multilingual Pre-trained Language Models (MPLMs) have become essential tools for natural language processing. However, they often exhibit biases related to sensitive attributes such as gender, race, and religion. In this paper, we introduce a comprehensive multilingual debiasing method named Multiple-Debias to address these issues across multiple languages. By incorporating multilingual counterfactual data augmentation and multilingual Self-Debias across both pre-processing and post-processing stages, alongside parameter-efficient fine-tuning, we significantly reduced biases in MPLMs across three sensitive attributes in four languages. We also extended CrowS-Pairs to German, Spanish, Chinese, and Japanese, validating our full-process multilingual debiasing method for gender, racial, and religious bias. Our experiments show that (i) multilingual debiasing methods surpass monolingual approaches in effectively mitigating biases, and (ii) integrating debiasing information from different languages notably improves the fairness of MPLMs.