π€ AI Summary
The scaling behavior of reasoning compute in multi-agent systems remains poorly understood. This paper addresses data synthesis under multi-agent collaborative sampling and proposes TOAβthe first framework integrating Monte Carlo Tree Search (MCTS) into multilingual model collaboration. TOA formalizes agent coordination as a dynamic tree-search decision process, enabling input-adaptive online workflow orchestration and reward-driven exploration. It unifies MCTS, reward modeling, multi-agent sampling, and synthetic data distillation to significantly improve reasoning compute scalability. Experiments demonstrate state-of-the-art performance on WMT and a 71.8% win rate on AlpacaEval. After fine-tuning with TOA-synthesized data, the resulting models surpass strong preference-learning baselines and achieve leading results on challenging benchmarks including Arena-Hard.
π Abstract
Scaling laws for inference compute in multi-agent systems remain under-explored compared to single-agent scenarios. This work aims to bridge this gap by investigating the problem of data synthesis through multi-agent sampling, where synthetic responses are generated by sampling from multiple distinct language models. Effective model coordination is crucial for successful multi-agent collaboration. Unlike previous approaches that rely on fixed workflows, we treat model coordination as a multi-step decision-making process, optimizing generation structures dynamically for each input question. We introduce Tree Search-based Orchestrated Agents~(TOA), where the workflow evolves iteratively during the sequential sampling process. To achieve this, we leverage Monte Carlo Tree Search (MCTS), integrating a reward model to provide real-time feedback and accelerate exploration. Our experiments on alignment, machine translation, and mathematical reasoning demonstrate that multi-agent sampling significantly outperforms single-agent sampling as inference compute scales. TOA is the most compute-efficient approach, achieving SOTA performance on WMT and a 71.8% LC win rate on AlpacaEval. Moreover, fine-tuning with our synthesized alignment data surpasses strong preference learning methods on challenging benchmarks such as Arena-Hard and AlpacaEval.