MMBERT: Scaled Mixture-of-Experts Multimodal BERT for Robust Chinese Hate Speech Detection under Cloaking Perturbations

📅 2025-08-01
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
In Chinese social media, hate speech frequently evades detection via multimodal obfuscation—such as lexical camouflage, pitch-altered speech, and metaphorical imagery—rendering unimodal text-based detection methods insufficiently robust. To address this, we propose MoE-BERT, a multimodal mixture-of-experts architecture tailored for Chinese hate speech detection. Our method integrates modality-specific experts with a shared self-attention mechanism and introduces a router-guided dynamic expert assignment strategy to stabilize joint training of MoE and BERT. Furthermore, we adopt a three-stage progressive training paradigm to enhance generalization under adversarial perturbations. Evaluated on multiple Chinese multimodal hate speech benchmarks, our approach consistently outperforms fine-tuned BERT baselines, state-of-the-art large language models, and in-context learning methods, achieving superior accuracy and significantly improved robustness in identifying covert hateful content.

Technology Category

Application Category

📝 Abstract
Hate speech detection on Chinese social networks presents distinct challenges, particularly due to the widespread use of cloaking techniques designed to evade conventional text-based detection systems. Although large language models (LLMs) have recently improved hate speech detection capabilities, the majority of existing work has concentrated on English datasets, with limited attention given to multimodal strategies in the Chinese context. In this study, we propose MMBERT, a novel BERT-based multimodal framework that integrates textual, speech, and visual modalities through a Mixture-of-Experts (MoE) architecture. To address the instability associated with directly integrating MoE into BERT-based models, we develop a progressive three-stage training paradigm. MMBERT incorporates modality-specific experts, a shared self-attention mechanism, and a router-based expert allocation strategy to enhance robustness against adversarial perturbations. Empirical results in several Chinese hate speech datasets show that MMBERT significantly surpasses fine-tuned BERT-based encoder models, fine-tuned LLMs, and LLMs utilizing in-context learning approaches.
Problem

Research questions and friction points this paper is trying to address.

Detect Chinese hate speech under cloaking perturbations
Address lack of multimodal strategies in Chinese context
Improve robustness with MoE-integrated BERT framework
Innovation

Methods, ideas, or system contributions that make the work stand out.

Multimodal BERT with Mixture-of-Experts architecture
Progressive three-stage training paradigm
Router-based expert allocation for robustness
🔎 Similar Papers
No similar papers found.
Q
Qiyao Xue
University of Pittsburgh
Y
Yuchen Dou
University of Pittsburgh
R
Ryan Shi
University of Pittsburgh
Xiang Lorraine Li
Xiang Lorraine Li
Assistant Professor, University of Pittsburgh
Natural Language ProcessingMachine Learning
W
Wei Gao
University of Pittsburgh