🤖 AI Summary
This paper addresses the critical problem of human overreliance on large language models (LLMs) in high-stakes domains such as healthcare and personalized advice. Identifying three key gaps in measuring overreliance, the study proposes a novel, interdisciplinary framework integrating human factors engineering, cognitive science, and AI system design. Methodologically, it analyzes overreliance through three synergistic lenses: systemic design flaws, user cognitive biases, and model uncertainty—thereby enabling both quantitative measurement and tiered intervention strategies. The contributions include: (1) clarifying severe consequences—high-risk errors, governance failures, and cognitive atrophy; (2) reframing LLM development toward human capability augmentation rather than automation; and (3) establishing theoretical foundations and actionable pathways for safe, trustworthy human-AI collaboration. This work advances responsible AI deployment by bridging technical design, cognitive understanding, and socio-technical governance.
📝 Abstract
Large language models (LLMs) distinguish themselves from previous technologies by functioning as collaborative "thought partners," capable of engaging more fluidly in natural language. As LLMs increasingly influence consequential decisions across diverse domains from healthcare to personal advice, the risk of overreliance - relying on LLMs beyond their capabilities - grows. This position paper argues that measuring and mitigating overreliance must become central to LLM research and deployment. First, we consolidate risks from overreliance at both the individual and societal levels, including high-stakes errors, governance challenges, and cognitive deskilling. Then, we explore LLM characteristics, system design features, and user cognitive biases that - together - raise serious and unique concerns about overreliance in practice. We also examine historical approaches for measuring overreliance, identifying three important gaps and proposing three promising directions to improve measurement. Finally, we propose mitigation strategies that the AI research community can pursue to ensure LLMs augment rather than undermine human capabilities.