"Are You Sure?": An Empirical Study of Human Perception Vulnerability in LLM-Driven Agentic Systems

πŸ“… 2026-02-24
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This study investigates human users’ perceptual vulnerability to Agent-Mediated Deception (AMD) attacks in large language model (LLM)-driven agent systems. Leveraging a high-fidelity experimental platform, HAT-Lab, the authors conducted a large-scale empirical study with 303 participants across nine scenarios spanning everyday and professional contexts. The research reveals, for the first time, critical cognitive failure modes among users confronting compromised agents: only 8.6% detected the attack, and experts were paradoxically more susceptible in certain scenarios. To address this vulnerability, the work proposes a novel defense mechanism centered on low-cost verification, workflow-interruptive warnings, and experiential learning. Experimental results demonstrate that over 90% of users significantly enhanced their vigilance after experiential training, confirming the effectiveness and feasibility of the proposed defensive approach.

Technology Category

Application Category

πŸ“ Abstract
Large language model (LLM) agents are rapidly becoming trusted copilots in high-stakes domains like software development and healthcare. However, this deepening trust introduces a novel attack surface: Agent-Mediated Deception (AMD), where compromised agents are weaponized against their human users. While extensive research focuses on agent-centric threats, human susceptibility to deception by a compromised agent remains unexplored. We present the first large-scale empirical study with 303 participants to measure human susceptibility to AMD. This is based on HAT-Lab (Human-Agent Trust Laboratory), a high-fidelity research platform we develop, featuring nine carefully crafted scenarios spanning everyday and professional domains (e.g., healthcare, software development, human resources). Our 10 key findings reveal significant vulnerabilities and provide future defense perspectives. Specifically, only 8.6% of participants perceive AMD attacks, while domain experts show increased susceptibility in certain scenarios. We identify six cognitive failure modes in users and find that their risk awareness often fails to translate to protective behavior. The defense analysis reveals that effective warnings should interrupt workflows with low verification costs. With experiential learning based on HAT-Lab, over 90% of users who perceive risks report increased caution against AMD. This work provides empirical evidence and a platform for human-centric agent security research.
Problem

Research questions and friction points this paper is trying to address.

Agent-Mediated Deception
Human Perception Vulnerability
LLM Agents
Trust
Deception
Innovation

Methods, ideas, or system contributions that make the work stand out.

Agent-Mediated Deception
Human-Agent Trust
LLM Agents
Cognitive Failure Modes
Experiential Learning
πŸ”Ž Similar Papers
No similar papers found.
X
Xinfeng Li
Nanyang Technological University
S
Shenyu Dai
KTH
K
Kelong Zheng
Nanyang Technological University
Y
Yue Xiao
William & Mary
Gelei Deng
Gelei Deng
Nanyang Technological University
CybersecuritySystem securityRobotics SecurityAI SecuritySoftware Testing
W
Wei Dong
Nanyang Technological University
X
Xiaofeng Wang
Nanyang Technological University