🤖 AI Summary
Large language models (LLMs) face significant challenges in telecommunications—characterized by dense domain-specific terminology, rapid knowledge evolution, and high technical complexity—limiting their practical applicability.
Method: This paper proposes a consensus-driven multi-LLM collaboration framework featuring a novel two-stage consensus mechanism: (i) multiple proposer LLMs generate rationale-augmented responses; (ii) an adjudicator LLM dynamically evaluates response quality and synthesizes final outputs via domain-adapted consensus aggregation. The framework integrates multi-LLM collaborative reasoning, a proposer-adjudicator architecture, and adaptive consensus strategies.
Contribution/Results: It overcomes limitations of single-model and naive ensemble approaches by jointly suppressing bias and enabling cross-model knowledge complementarity. Evaluated on telecom-specific tasks, the framework achieves a 9.7% absolute accuracy gain over strong baselines—including RAG, Mixture-of-Experts (MoE), and fine-tuning—demonstrating superior scalability and robustness. It establishes a new paradigm for deploying LLMs in fast-evolving, knowledge-intensive professional domains.
📝 Abstract
Large language models (LLMs) face significant challenges in specialized domains like telecommunication (Telecom) due to technical complexity, specialized terminology, and rapidly evolving knowledge. Traditional methods, such as scaling model parameters or retraining on domain-specific corpora, are computationally expensive and yield diminishing returns, while existing approaches like retrieval-augmented generation, mixture of experts, and fine-tuning struggle with accuracy, efficiency, and coordination. To address this issue, we propose Telecom mixture of models (TeleMoM), a consensus-driven ensemble framework that integrates multiple LLMs for enhanced decision-making in Telecom. TeleMoM employs a two-stage process: proponent models generate justified responses, and an adjudicator finalizes decisions, supported by a quality-checking mechanism. This approach leverages strengths of diverse models to improve accuracy, reduce biases, and handle domain-specific complexities effectively. Evaluation results demonstrate that TeleMoM achieves a 9.7% increase in answer accuracy, highlighting its effectiveness in Telecom applications.