Position: Thematic Analysis of Unstructured Clinical Transcripts with Large Language Models

📅 2025-09-18
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Current large language models (LLMs) face four critical fragmentation challenges in thematic analysis of unstructured clinical text: heterogeneous analytical objectives, inconsistent benchmark datasets, nonstandardized prompting strategies, and absent evaluation criteria—severely impeding cross-study comparability and clinical deployment. Through a systematic literature review, in-depth interviews with practicing clinicians, and multi-strategy prompting experiments, this study identifies *evaluation heterogeneity* as the primary bottleneck. We thus propose a novel three-dimensional standardized evaluation framework centered on *effectiveness*, *reliability*, and *interpretability*, integrating automated semantic similarity metrics with qualitative expert validation. To our knowledge, this is the first reproducible, comparable, and interpretable evaluation paradigm specifically designed for LLM-based clinical thematic analysis. The framework substantially enhances methodological rigor, enables robust cross-model comparison, and strengthens translational relevance for real-world clinical applications.

Technology Category

Application Category

📝 Abstract
This position paper examines how large language models (LLMs) can support thematic analysis of unstructured clinical transcripts, a widely used but resource-intensive method for uncovering patterns in patient and provider narratives. We conducted a systematic review of recent studies applying LLMs to thematic analysis, complemented by an interview with a practicing clinician. Our findings reveal that current approaches remain fragmented across multiple dimensions including types of thematic analysis, datasets, prompting strategies and models used, most notably in evaluation. Existing evaluation methods vary widely (from qualitative expert review to automatic similarity metrics), hindering progress and preventing meaningful benchmarking across studies. We argue that establishing standardized evaluation practices is critical for advancing the field. To this end, we propose an evaluation framework centered on three dimensions: validity, reliability, and interpretability.
Problem

Research questions and friction points this paper is trying to address.

LLMs support thematic analysis of clinical transcripts
Current evaluation methods are fragmented and inconsistent
Standardized evaluation framework is needed for validity and reliability
Innovation

Methods, ideas, or system contributions that make the work stand out.

Proposing evaluation framework for validity reliability
Systematic review of LLMs thematic analysis
Standardized evaluation practices critical advancement
🔎 Similar Papers
No similar papers found.
S
Seungjun Yi
Department of Biomedical Engineering, The University of Texas at Austin
J
Joakim Nguyen
Department of Computer Science, The University of Texas at Austin
T
Terence Lim
College of Natural Sciences, The University of Texas at Austin
Andrew Well
Andrew Well
Research Assistant Professor of Cardiac Surgery
J
Joseph Skrovan
School of Information, The University of Texas at Austin
M
Mehak Beri
School of Information, The University of Texas at Austin
Y
YongGeon Lee
School of Information, The University of Texas at Austin
Kavita Radhakrishnan
Kavita Radhakrishnan
University of Texas - Austin
telehealthheart failureserious gamesself-managementprecision health
Liu Leqi
Liu Leqi
UT Austin
Artificial IntelligenceMachine Learning
M
Mia Markey
Department of Biomedical Engineering, The University of Texas at Austin
Ying Ding
Ying Ding
Bill & Lewis Suit Professor, School of Information, Dell Med, University of Texas at Austin
AI in HealthKnowledge GraphScience of Science