🤖 AI Summary
This work addresses deadlock arising from symmetry in distributed graph coloring tasks—specifically on odd-length cycles (C₃, C₅, C₁₁)—by non-communicating intelligent agents. Method: We introduce LoopBench, the first benchmark systematically evaluating large language models’ (LLMs) joint capabilities in distributed symmetry breaking and metacognitive reasoning. Our approach designs an LLM-based multi-agent coordination mechanism grounded in policy transfer and shared memory, integrating prompt engineering, graph-theoretic constraint modeling, and a metacognitive reasoning framework to enable collaborative decision-making under ultra-low communication. Contribution/Results: We demonstrate that LLM swarms can spontaneously emerge correct distributed algorithms; notably, advanced reasoning models such as O3 achieve zero-deadlock coloring across all tested cycle graphs—surpassing the limitations of conventional deterministic protocols—and establish LLM swarms as a novel emergent substrate for distributed algorithm execution.
📝 Abstract
Large Language Models (LLMs) are increasingly being utilized as autonomous agents, yet their ability to coordinate in distributed systems remains poorly understood. We introduce extbf{LoopBench}, a benchmark to evaluate LLM reasoning in distributed symmetry breaking and meta-cognitive thinking. The benchmark focuses on coloring odd cycle graphs ($C_3, C_5, C_{11}$) with limited colors, where deterministic, non-communicating agents fail in infinite loops. A strategy passing mechanism is implemented as a form of consistent memory. We show that while standard LLMs and classical heuristics struggle, advanced reasoning models (e.g., O3) devise strategies to escape deadlocks. LoopBench allows the study of emergent distributed algorithms based on language-based reasoning, offering a testbed for collective intelligence.