Key Decision-Makers in Multi-Agent Debates: Who Holds the Power?

📅 2025-11-14
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
The impact of role assignment strategies on reasoning performance in Multi-Agent Debate (MAD) remains systematically unexplored, particularly in real-world scenarios where ground truth is unknown and no effective mechanism exists. Method: We identify the critical influence of positional placement of viewpoints within the debate sequence and propose the “Truth-Last” strategy—deliberately assigning the correct viewpoint to a later position in the debate chain. Furthermore, we introduce Multi-Agent Debate Consistency (MADC), a novel mechanism that automatically identifies high-credibility agents via path-level consistency, thereby addressing the truth-agnostic challenge. Contribution/Results: Extensive experiments across nine mainstream large language models—including a distilled variant of DeepSeek-R1—demonstrate that our approach achieves an average 22% improvement on complex reasoning tasks. MADC consistently outperforms conventional MAD across all models, effectively overcoming key performance bottlenecks in multi-agent collaborative reasoning.

Technology Category

Application Category

📝 Abstract
Recent studies on LLM agent scaling have highlighted the potential of Multi-Agent Debate (MAD) to enhance reasoning abilities. However, the critical aspect of role allocation strategies remains underexplored. In this study, we demonstrate that allocating roles with differing viewpoints to specific positions significantly impacts MAD's performance in reasoning tasks. Specifically, we find a novel role allocation strategy,"Truth Last", which can improve MAD performance by up to 22% in reasoning tasks. To address the issue of unknown truth in practical applications, we propose the Multi-Agent Debate Consistency (MADC) strategy, which systematically simulates and optimizes its core mechanisms. MADC incorporates path consistency to assess agreement among independent roles, simulating the role with the highest consistency score as the truth. We validated MADC across a range of LLMs (9 models), including the DeepSeek-R1 Distilled Models, on challenging reasoning tasks. MADC consistently demonstrated advanced performance, effectively overcoming MAD's performance bottlenecks and providing a crucial pathway for further improvements in LLM agent scaling.
Problem

Research questions and friction points this paper is trying to address.

Optimizing role allocation strategies in multi-agent debates
Improving reasoning performance through novel viewpoint positioning
Addressing unknown truth with consistency-based simulation mechanisms
Innovation

Methods, ideas, or system contributions that make the work stand out.

Truth Last role allocation boosts reasoning performance
MADC strategy simulates and optimizes debate mechanisms
Path consistency assesses agreement among independent roles
🔎 Similar Papers
No similar papers found.
Q
Qian Zhang
School of New Media and Communication, Tianjin University, China
Y
Yan Zheng
College of Intelligence and Computing, Tianjin University, China
Jinyi Liu
Jinyi Liu
Tianjin University
Reinforcement LearningLarge Language Model
Hebin Liang
Hebin Liang
Tianjin University
Reinforcement learningLLMEmbodied AI
L
Lanjun Wang
School of New Media and Communication, Tianjin University, China