Mitigating Judgment Preference Bias in Large Language Models through Group-Based Polling

📅 2025-10-09
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Large language models (LLMs) employed as automated evaluators (“LLM-as-a-Judge”) suffer from self-preference bias—systematically overrating responses generated by themselves—thereby compromising evaluation reliability. To address this, we propose Genii, an unsupervised multi-agent collaborative framework that emulates a client-server interactive group-voting mechanism without requiring human annotations. Methodologically, Genii employs a heterogeneous LLM ensemble architecture, dynamically assigning distinct roles to agents and aggregating judgments via consensus-based voting to mitigate bias. Notably, the server agent can be instantiated with a weaker, more efficient model, enhancing deployment flexibility. Experiments across multiple benchmarks and diverse client LLMs demonstrate that Genii consistently outperforms supervised baselines, yielding evaluations that are significantly more robust and fair.

Technology Category

Application Category

📝 Abstract
Large Language Models (LLMs) as automatic evaluators, commonly referred to as LLM-as-a-Judge, have also attracted growing attention. This approach plays a vital role in aligning LLMs with human judgments, providing accurate and reliable assessments. However, LLM-based judgment models often exhibit judgment preference bias during the evaluation phase, tending to favor responses generated by themselves, undermining the reliability of their judgments. This paper introduces the Group-Based Polling Optimization (Genii), an unsupervised multi-agent collaborative optimization framework that mitigates the inherent judgment preference bias of judgment models. Specifically, Genii integrates various LLM-based judgment models into a multi-agent system and simulates the interactive client-server polling mechanism to optimize each client agent unsupervisedly. Our experiments demonstrate that Genii outperforms supervised models trained on annotated judgment data, while requiring no human-labeled annotations. Genii consistently improves performance across different client agents during the polling, even when weaker models act as server agents. Further analysis reveals that Genii effectively mitigates judgment preference bias of LLM-based judgment models, demonstrating its effectiveness. All codes are available at https://github.com/NEUIR/Genii.
Problem

Research questions and friction points this paper is trying to address.

Mitigating judgment preference bias in LLM evaluators
Reducing self-preference bias during evaluation phase
Improving reliability of unsupervised LLM-based assessments
Innovation

Methods, ideas, or system contributions that make the work stand out.

Unsupervised multi-agent collaborative optimization framework
Simulates interactive client-server polling mechanism
Mitigates judgment preference bias without human annotations
🔎 Similar Papers
No similar papers found.
Shuliang Liu
Shuliang Liu
PhD, HKUST(GZ)
Trustworthy LLMVLMRecommendation System
Zhipeng Xu
Zhipeng Xu
Northeastern University
NLPInformation Retrieval
Zhenghao Liu
Zhenghao Liu
Northeastern University
NLPInformation Retrieval
Yukun Yan
Yukun Yan
Tsinghua University
Large Language Model
M
Minghe Yu
School of Computer Science and Engineering, Northeastern University, China
Y
Yu Gu
School of Computer Science and Engineering, Northeastern University, China
C
Chong Chen
Huawei Cloud Business Unit, China
Huiyuan Xie
Huiyuan Xie
University of Cambridge
Natural Language ProcessingAI and Law
G
Ge Yu
School of Computer Science and Engineering, Northeastern University, China