🤖 AI Summary
This study identifies a critical clinical-ethical risk in large language models (LLMs): excessive alignment with human values induces overly cautious responses in health consultations—particularly exacerbating symptom severity and undermining psychological safety for individuals with anxiety disorders and obsessive-compulsive disorder (OCD). Through multi-model qualitative comparative analysis, we systematically uncover mechanistic links between response caution, user cognitive load, and disease-specific sensitivity. We introduce the novel concept of “clinical alignment thresholds,” arguing that current LLMs’ lack of disease-aware reasoning underlies this overcautiousness. Accordingly, we advocate for next-generation health AI endowed with hierarchical risk assessment and individualized contextual understanding. Empirical results demonstrate that strategically relaxing generic value alignment—while deepening domain-specific medical reasoning—significantly improves both response appropriateness and clinical safety.
📝 Abstract
Large Language Models (LLMs) are usually aligned with"human values/preferences"to prevent harmful output. Discussions around the alignment of Large Language Models (LLMs) generally focus on preventing harmful outputs. However, in this paper, we argue that in health-related queries, over-alignment-leading to overly cautious responses-can itself be harmful, especially for people with anxiety and obsessive-compulsive disorder (OCD). This is not only unethical but also dangerous to the user, both mentally and physically. We also showed qualitative results that some LLMs exhibit varying degrees of alignment. Finally, we call for the development of LLMs with stronger reasoning capabilities that provide more tailored and nuanced responses to health queries. Warning: This paper contains materials that could trigger health anxiety or OCD.