ALIGN: Aligned Delegation with Performance Guarantees for Multi-Agent LLM Reasoning

๐Ÿ“… 2026-01-28
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
This work addresses the limitations of large language models (LLMs) in complex reasoning tasks, where reliance on a single generate-and-select pipeline constrains performance and existing ensemble methods lack theoretical guarantees. The authors propose a multi-agent reasoning framework grounded in aligned delegation games: a principal delegates tasks to multiple agents through incentive mechanisms to generate candidate solutions and then selects the final answer, thereby enabling structured interaction and objective alignment. This approach provides the first theoretical guarantee for performance improvement in multi-agent LLM reasoning, relaxing the independence assumption while explicitly accounting for correlations among candidate solutionsโ€”thus overcoming key limitations of conventional ensembles. Extensive experiments demonstrate that the framework consistently outperforms strong single-agent and ensemble baselines across diverse reasoning benchmarks, confirming its effectiveness and robustness.

Technology Category

Application Category

๐Ÿ“ Abstract
LLMs often underperform on complex reasoning tasks when relying on a single generation-and-selection pipeline. Inference-time ensemble methods can improve performance by sampling diverse reasoning paths or aggregating multiple candidate answers, but they typically treat candidates independently and provide no formal guarantees that ensembling improves reasoning quality. We propose a novel method, Aligned Delegation for Multi-Agent LLM Reasoning (ALIGN), which formulates LLM reasoning as an aligned delegation game. In ALIGN, a principal delegates a task to multiple agents that generate candidate solutions under designed incentives, and then selects among their outputs to produce a final answer. This formulation induces structured interaction among agents while preserving alignment between agent and principal objectives. We establish theoretical guarantees showing that, under a fair comparison with equal access to candidate solutions, ALIGN provably improves expected performance over single-agent generation. Our analysis accommodates correlated candidate answers and relaxes independence assumptions that are commonly used in prior work. Empirical results across a broad range of LLM reasoning benchmarks consistently demonstrate that ALIGN outperforms strong single-agent and ensemble baselines.
Problem

Research questions and friction points this paper is trying to address.

multi-agent reasoning
large language models
ensemble methods
reasoning quality
performance guarantees
Innovation

Methods, ideas, or system contributions that make the work stand out.

multi-agent reasoning
aligned delegation
performance guarantees
LLM ensembling
correlated candidates
๐Ÿ”Ž Similar Papers
No similar papers found.
T
Tong Zhu
Department of Biostatistics, UCLA
B
Baiting Chen
Department of Statistics and Data Science, UCLA
J
Jin Zhou
Department of Biostatistics, UCLA
Hua Zhou
Hua Zhou
Advance Photon Source, Argonne National Laboratory
Materials PhysicsSynchrotron RadiationSurface and InterfaceQuantum MaterialsEnergy Materials
Sriram Sankararaman
Sriram Sankararaman
UCLA
Computational BiologyPopulation geneticsBayesian statisticsProbabilistic graphical modelsMachine learning
X
Xiaowu Dai
Departments of Statistics and Data Science, and of Biostatistics, UCLA