"I know it's not right, but that's what it said to do": Investigating Trust in AI Chatbots for Cybersecurity Policy

📅 2025-10-09
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study investigates whether users, when performing cybersecurity policy tasks, can be misled by adversarially manipulated AI chatbots into adopting recommendations that undermine system defenses. Method: A controlled experiment (N=15) combined behavioral observation and semi-structured interviews to simulate attacker-controlled AI outputs, quantifying user trust and its modulation by task familiarity and self-assessed confidence. Contribution/Results: Despite expressed skepticism, nearly half of participants executed high-risk instructions. Trust decisions were significantly moderated by task familiarity (β = 0.42, p < 0.05) and subjective confidence (β = −0.38, p < 0.05). This work provides the first empirical characterization of the dynamic “human-factor trust vulnerability” in AI security—where users’ cognitive biases interact with manipulated AI outputs—and offers evidence-based design principles for resilient human-AI collaborative defense frameworks resistant to adversarial manipulation.

Technology Category

Application Category

📝 Abstract
AI chatbots are an emerging security attack vector, vulnerable to threats such as prompt injection, and rogue chatbot creation. When deployed in domains such as corporate security policy, they could be weaponized to deliver guidance that intentionally undermines system defenses. We investigate whether users can be tricked by a compromised AI chatbot in this scenario. A controlled study (N=15) asked participants to use a chatbot to complete security-related tasks. Without their knowledge, the chatbot was manipulated to give incorrect advice for some tasks. The results show how trust in AI chatbots is related to task familiarity, and confidence in their ownn judgment. Additionally, we discuss possible reasons why people do or do not trust AI chatbots in different scenarios.
Problem

Research questions and friction points this paper is trying to address.

Investigating user trust in compromised AI chatbots for cybersecurity policy guidance
Examining whether users follow incorrect security advice from manipulated AI chatbots
Analyzing how task familiarity affects trust in AI chatbot security recommendations
Innovation

Methods, ideas, or system contributions that make the work stand out.

Manipulated chatbot to give incorrect security advice
Studied user trust through controlled security tasks
Analyzed trust factors like task familiarity and confidence
🔎 Similar Papers
No similar papers found.
B
Brandon Lit
Cheriton School of Computer Science, University of Waterloo, Canada
E
Edward Crowder
Computer Science, University of Guelph, Canada
Daniel Vogel
Daniel Vogel
University of Waterloo
Human-Computer InteractionInput DevicesInteraction TechniquesLarge Displays
H
Hassan Khan
Computer Science, University of Guelph, Canada