🤖 AI Summary
Existing FD-SLM evaluation benchmarks are limited to single-turn dialogues, overlooking the complexity of multi-turn interaction and lacking systematic assessment of instruction following and safety. To address challenges in full-duplex multi-turn dialogue—such as ambiguous turn boundaries and contextual inconsistency—this paper introduces MTR-DuplexBench, the first benchmark enabling fine-grained multi-turn evaluation. Our method discretizes continuous speech interaction via dialogue segmentation and establishes a four-dimensional, human-in-the-loop scoring framework covering dialogue quality, interaction dynamics, instruction adherence, and safety. Experimental results reveal significant deficiencies in current state-of-the-art FD-SLMs regarding multi-turn consistency and cross-dimensional coordination, thereby validating both the necessity and effectiveness of MTR-DuplexBench. The benchmark provides quantifiable metrics to guide iterative model improvement, advancing rigorous, holistic evaluation of full-duplex conversational AI systems.
📝 Abstract
Full-Duplex Speech Language Models (FD-SLMs) enable real-time, overlapping conversational interactions, offering a more dynamic user experience compared to traditional half-duplex models. However, existing benchmarks primarily focus on evaluating single-round interactions and conversational features, neglecting the complexities of multi-round communication and critical capabilities such as instruction following and safety. Evaluating FD-SLMs in multi-round settings poses significant challenges, including blurred turn boundaries in communication and context inconsistency during model inference. To address these gaps, we introduce MTR-DuplexBench, a novel benchmark that segments continuous full-duplex dialogues into discrete turns, enabling comprehensive, turn-by-turn evaluation of FD-SLMs across dialogue quality, conversational dynamics, instruction following, and safety. Experimental results reveal that current FD-SLMs face difficulties in maintaining consistent performance across multiple rounds and evaluation dimensions, highlighting the necessity and effectiveness of our proposed benchmark. The benchmark and code will be available in the future.