ACC-Collab: An Actor-Critic Approach to Multi-Agent LLM Collaboration

📅 2024-10-30
📈 Citations: 1
Influential: 0
📄 PDF
🤖 AI Summary
Existing multi-agent frameworks rely on implicit collaboration capabilities acquired during large language model (LLM) pretraining, leading to poor generalizability and optimizability of collaborative behavior. This paper proposes the first end-to-end trainable two-agent collaborative framework, explicitly modeling collaboration as a learnable process: an Actor-Agent executes tasks, while a Critic-Agent specializes in evaluating collaboration quality and providing feedback—marking the first application of the Actor-Critic paradigm to multi-LLM collaboration. Our method integrates LLM supervised fine-tuning, dialogue policy gradient optimization, and collaboration-aware reward modeling, enabling joint optimization of collaborative policies via multi-turn trajectory sampling. Evaluated on multiple benchmarks, our approach significantly outperforms state-of-the-art methods, achieving consistent improvements in task completion rate, response quality, and collaboration stability.

Technology Category

Application Category

📝 Abstract
Large language models (LLMs) have demonstrated a remarkable ability to serve as general-purpose tools for various language-based tasks. Recent works have demonstrated that the efficacy of such models can be improved through iterative dialog between multiple models. While these paradigms show promise in improving model efficacy, most works in this area treat collaboration as an emergent behavior, rather than a learned behavior. In doing so, current multi-agent frameworks rely on collaborative behaviors to have been sufficiently trained into off-the-shelf models. To address this limitation, we propose ACC-Collab, an Actor-Critic based learning framework to produce a two-agent team (an actor-agent and a critic-agent) specialized in collaboration. We demonstrate that ACC-Collab outperforms SotA multi-agent techniques on a wide array of benchmarks.
Problem

Research questions and friction points this paper is trying to address.

Improves multi-agent LLM collaboration through learned behavior
Addresses reliance on emergent collaboration in current frameworks
Proposes Actor-Critic framework for specialized two-agent teams
Innovation

Methods, ideas, or system contributions that make the work stand out.

Actor-Critic framework for multi-agent collaboration
Specialized two-agent team for improved collaboration
Outperforms state-of-the-art multi-agent techniques
🔎 Similar Papers
No similar papers found.
Andrew Estornell
Andrew Estornell
ByteDance Research
Large Language ModelsMulti-Agent SystemsAlgorithmic Fairness
J
Jean-François Ton
ByteDance Research
Y
Yuanshun Yao
Meta GenAI
Y
Yang Liu
University of California, Santa Cruz