🤖 AI Summary
This study addresses the challenge of insufficient role consistency and dialogue coherence exhibited by large language models (LLMs) when simulating client roles in online psychological counseling training. To systematically evaluate model performance, the authors introduce a novel role-consistency evaluation dataset constructed using adversarial examples specifically tailored to counseling scenarios. Leveraging metrics that jointly assess role fidelity and conversational coherence, the work provides a comprehensive benchmark of several open-source LLMs, including Vicuna. The findings not only quantify disparities in role-maintenance capabilities across models but also offer empirical guidance for selecting appropriate LLMs in virtual counseling training systems. Furthermore, this research highlights the innovative value of adversarial evaluation methodologies in role-playing tasks within specialized domains such as mental health counseling.
📝 Abstract
The rise of online counseling services has highlighted the need for effective training methods for future counselors. This paper extends research on VirCo, a Virtual Client for Online Counseling, designed to complement traditional role-playing methods in academic training by simulating realistic client interactions. Building on previous work, we introduce a new dataset incorporating adversarial attacks to test the ability of large language models (LLMs) to maintain their assigned roles (role-consistency). The study focuses on evaluating the role consistency and coherence of the Vicuna model's responses, comparing these findings with earlier research. Additionally, we assess and compare various open-source LLMs for their performance in sustaining role consistency during virtual client interactions. Our contributions include creating an adversarial dataset, evaluating conversation coherence and persona consistency, and providing a comparative analysis of different LLMs.