Who's Sorry Now: User Preferences Among Rote, Empathic, and Explanatory Apologies from LLM Chatbots

📅 2025-07-03
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study investigates how large language model (LLM) chatbots can restore user trust after errors—specifically bias, factual inaccuracies, and hallucinations. Using a pre-registered within-subjects experiment (N = 162), we compared the efficacy of three apology strategies: mechanical, explanatory, and empathetic. Results show that explanatory apologies yield the highest overall trust restoration; empathetic apologies significantly outperform others in bias-related errors; and no significant preference emerges for hallucination errors—highlighting strong contextual dependence in trust repair. This work provides the first systematic empirical evidence of interaction effects between error type and apology strategy. We propose two design frameworks—“personalized apology” and “error–apology calibration”—to guide context-sensitive apology generation. These findings offer empirically grounded principles for developing trustworthy, human-centered AI dialogue systems.

Technology Category

Application Category

📝 Abstract
As chatbots driven by large language models (LLMs) are increasingly deployed in everyday contexts, their ability to recover from errors through effective apologies is critical to maintaining user trust and satisfaction. In a preregistered study with Prolific workers (N=162), we examine user preferences for three types of apologies (rote, explanatory, and empathic) issued in response to three categories of common LLM mistakes (bias, unfounded fabrication, and factual errors). We designed a pairwise experiment in which participants evaluated chatbot responses consisting of an initial error, a subsequent apology, and a resolution. Explanatory apologies were generally preferred, but this varied by context and user. In the bias scenario, empathic apologies were favored for acknowledging emotional impact, while hallucinations, though seen as serious, elicited no clear preference, reflecting user uncertainty. Our findings show the complexity of effective apology in AI systems. We discuss key insights such as personalization and calibration that future systems must navigate to meaningfully repair trust.
Problem

Research questions and friction points this paper is trying to address.

Examines user preferences for chatbot apology types
Investigates effectiveness of apologies for LLM error categories
Explores trust repair strategies in AI systems
Innovation

Methods, ideas, or system contributions that make the work stand out.

Pairwise experiment design for user evaluation
Explanatory apologies generally preferred by users
Empathic apologies favored in bias scenarios
🔎 Similar Papers
No similar papers found.
Zahra Ashktorab
Zahra Ashktorab
IBM Research
Human Computer Interaction
A
Alessandra Buccella
University at Albany, Department of Philosophy, USA
Jason D'Cruz
Jason D'Cruz
Associate Professor of Philosophy, University at Albany, SUNY
EthicstrustpromisesrationalizationAI Ethics
Zoe Fowler
Zoe Fowler
University at Albany, Department of Psychology, USA
A
Andrew Gill
University at Albany, Department of Philosophy, USA
K
Kei Yan Leung
University at Albany, Department of Philosophy, USA
P
P. D. Magnus
University at Albany, Department of Philosophy, USA
John Richards
John Richards
IBM Research, USA
Kush R. Varshney
Kush R. Varshney
IBM Research
Statistical Signal ProcessingMachine LearningData MiningImage ProcessingSocial Good