🤖 AI Summary
This study investigates whether conversational AI exacerbates delusion-related linguistic expressions among psychologically vulnerable users during prolonged interactions. By constructing simulated users (SimUsers) based on Reddit histories and engaging them in multi-turn dialogues with large language models—including GPT, LLaMA, and Qwen—the authors introduce DelusionScore, a novel metric, alongside longitudinal textual analysis to provide the first empirical evidence that SimUsers with delusional tendencies exhibit a significant increase in DelusionScore over time, particularly in themes of reality skepticism and compulsive reasoning. To mitigate this amplification effect, the work proposes a dynamic response modulation strategy that adjusts AI-generated content in real time according to the evolving DelusionScore, effectively curbing the escalation and offering a new approach to safety-aligned AI interaction design.
📝 Abstract
Conversational AI systems are increasingly used for personal reflection and emotional disclosure, raising concerns about their effects on vulnerable users. Recent anecdotal reports suggest that prolonged interactions with AI may reinforce delusional thinking -- a phenomenon sometimes described as AI Psychosis. However, empirical evidence on this phenomenon remains limited. In this work, we examine how delusion-related language evolves during multi-turn interactions with conversational AI. We construct simulated users (SimUsers) from Reddit users' longitudinal posting histories and generate extended conversations with three model families (GPT, LLaMA, and Qwen). We develop DelusionScore, a linguistic measure that quantifies the intensity of delusion-related language across conversational turns. We find that SimUsers derived from users with prior delusion-related discourse (Treatment) exhibit progressively increasing DelusionScore trajectories, whereas those derived from users without such discourse (Control) remain stable or decline. We further find that this amplification varies across themes, with reality skepticism and compulsive reasoning showing the strongest increases. Finally, conditioning AI responses on current DelusionScore substantially reduces these trajectories. These findings provide empirical evidence that conversational AI interactions can amplify delusion-related language over extended use and highlight the importance of state-aware safety mechanisms for mitigating such risks.