🤖 AI Summary
This work proposes a collaborative framework in which large language models (LLMs) serve as “empathy editors” to refine clinician-authored messages, enhancing empathetic expression while preserving medical accuracy. Recognizing that clinicians under pressure often struggle to balance factual precision with emotional attunement—leading to perceived empathy deficits in patient communication—the approach emphasizes human–AI collaboration rather than end-to-end generation. To evaluate performance, the study introduces two novel metrics: the Empathy Ranking Score, which assesses perceived empathy, and the MedFactChecking Score, which verifies clinical fidelity. Experimental results demonstrate that LLM-edited texts significantly improve patients’ perception of empathy without compromising medical correctness, outperforming fully LLM-generated responses.
📝 Abstract
Clinical empathy is essential for patient care, but physicians need continually balance emotional warmth with factual precision under the cognitive and emotional constraints of clinical practice. This study investigates how large language models (LLMs) can function as empathy editors, refining physicians'written responses to enhance empathetic tone while preserving underlying medical information. More importantly, we introduce novel quantitative metrics, an Empathy Ranking Score and a MedFactChecking Score to systematically assess both emotional and factual quality of the responses. Experimental results show that LLM edited responses significantly increase perceived empathy while preserving factual accuracy compared with fully LLM generated outputs. These findings suggest that using LLMs as editorial assistants, rather than autonomous generators, offers a safer, more effective pathway to empathetic and trustworthy AI-assisted healthcare communication.