Evaluating an LLM-Powered Chatbot for Cognitive Restructuring: Insights from Mental Health Professionals

📅 2025-01-26
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study investigates the clinical feasibility and ethical risks of large language models (LLMs) in cognitive restructuring (CR)—a core evidence-based psychotherapeutic technique. To this end, we developed a CR-specific chatbot via prompt engineering and conducted an empirical user study involving 19 real participants; their interaction logs were subjected to qualitative analysis and clinical evaluation by licensed mental health professionals—the first study to integrate authentic user engagement with expert clinical assessment. Results indicate that LLMs can execute fundamental CR procedures (e.g., Socratic questioning, empathic reflection, protocol adherence), yet exhibit persistent structural risks: misinterpretation of client cues, boundary violations in therapeutic advice, power imbalances, and excessive positivity bias. Critically, we propose and substantiate a novel “real-time clinical supervision” design paradigm, wherein human experts actively monitor and intervene during AI-assisted sessions. This framework advances both theoretical understanding and practical implementation pathways for safe, clinically integrated AI interventions.

Technology Category

Application Category

📝 Abstract
Recent advancements in large language models (LLMs) promise to expand mental health interventions by emulating therapeutic techniques, potentially easing barriers to care. Yet there is a lack of real-world empirical evidence evaluating the strengths and limitations of LLM-enabled psychotherapy interventions. In this work, we evaluate an LLM-powered chatbot, designed via prompt engineering to deliver cognitive restructuring (CR), with 19 users. Mental health professionals then examined the resulting conversation logs to uncover potential benefits and pitfalls. Our findings indicate that an LLM-based CR approach has the capability to adhere to core CR protocols, prompt Socratic questioning, and provide empathetic validation. However, issues of power imbalances, advice-giving, misunderstood cues, and excessive positivity reveal deeper challenges, including the potential to erode therapeutic rapport and ethical concerns. We also discuss design implications for leveraging LLMs in psychotherapy and underscore the importance of expert oversight to mitigate these concerns, which are critical steps toward safer, more effective AI-assisted interventions.
Problem

Research questions and friction points this paper is trying to address.

Large Language Models
Psychological Therapy
Ethical Considerations
Innovation

Methods, ideas, or system contributions that make the work stand out.

Large Language Models
Mental Health Therapy
Chatbot Design
🔎 Similar Papers
No similar papers found.
Y
Yinzhou Wang
College of William & Mary, USA
Y
Yimeng Wang
College of William & Mary, USA
Y
Ye Xiao
College of William & Mary, USA
L
Liabette Escamilla
College of William & Mary, USA
B
Bianca Augustine
College of William & Mary, USA
K
Kelly Crace
University of Virginia, USA
Gang Zhou
Gang Zhou
Class of 2027 Professor, IEEE Fellow, EiC of ACM HEALTH, William & Mary
Sensor SystemsAI for HealthMobile and Ubiquitous ComputingInternet of ThingsNetworking
Y
Yixuan Zhang
College of William & Mary, USA