🤖 AI Summary
Current evaluation methodologies for dialogue systems are constrained by static scoring criteria and fixed scenarios, limiting their ability to capture dynamic behaviors in multi-turn interactions. This work proposes an adaptive co-evolutionary evaluation framework that integrates generation and assessment into an iterative optimization loop through a closed-loop collaboration between a dialogue planner and a reflective analyzer. The framework employs structured templates to guide a user simulator in generating goal-oriented dialogues and automatically refines evaluation rubrics based on behavioral pattern analysis. This approach simultaneously enhances test case complexity and diagnostic precision of scoring, significantly improving coverage and accuracy in evaluating the multi-turn capabilities of advanced dialogue systems while reducing reliance on manual intervention.
📝 Abstract
Evaluating conversational systems in multi-turn settings remains a fundamental challenge. Conventional pipelines typically rely on manually defined rubrics and fixed conversational context$-$a static approach that limits coverage and fails to capture the diverse, emergent behaviors of dialogue models. To address this, we introduce CoReflect (Conversational Evaluation via Co-Evolutionary Simulation and Reflective Rubric Refinement), which unifies dialogue simulation and evaluation into an adaptive, iterative process. CoReflect employs a conversation planner that generates structured templates to guide a user simulator through diverse, goal-directed dialogues. Subsequently, a reflective analyzer processes these dialogues to identify systematic behavioral patterns and automatically refine the evaluation rubrics. Crucially, the insights from the conversation analysis are fed back into the planner to update conversation templates for subsequent iterations. This co-evolution loop ensures that the complexity of test cases and the diagnostic precision of rubrics improve in tandem. By minimizing human intervention, CoReflect provides a scalable and self-refining methodology that allows evaluation protocols to adapt alongside the rapidly advancing capabilities of dialogue models.