GhostEI-Bench: Do Mobile Agents Resilience to Environmental Injection in Dynamic On-Device Environments?

📅 2025-10-23
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the vulnerability of mobile vision-language model (VLM)-based agents to environmental injection attacks—such as deceptive notifications and overlay windows—in dynamic GUI environments, which cause perceptual inaccuracies and erroneous actions. Methodologically, we construct the first Android-specific benchmark for evaluating environmental injection robustness: adversarial UI elements are dynamically injected into a real Android emulator; automated assessment is performed via screenshot sequences, action trajectories, and multimodal playback; and a fine-grained failure analysis protocol—leveraging a judge LLM—is introduced to precisely localize failures in perception, recognition, or reasoning. Our contributions include: (1) establishing environmental injection as a novel threat paradigm for mobile VLM agents; (2) developing an executable, reproducible dynamic evaluation framework; and (3) empirically demonstrating severe robustness deficiencies across state-of-the-art VLM-based agents—thereby providing a foundational benchmark and empirical basis for future defense research.

Technology Category

Application Category

📝 Abstract
Vision-Language Models (VLMs) are increasingly deployed as autonomous agents to navigate mobile graphical user interfaces (GUIs). Operating in dynamic on-device ecosystems, which include notifications, pop-ups, and inter-app interactions, exposes them to a unique and underexplored threat vector: environmental injection. Unlike prompt-based attacks that manipulate textual instructions, environmental injection corrupts an agent's visual perception by inserting adversarial UI elements (for example, deceptive overlays or spoofed notifications) directly into the GUI. This bypasses textual safeguards and can derail execution, causing privacy leakage, financial loss, or irreversible device compromise. To systematically evaluate this threat, we introduce GhostEI-Bench, the first benchmark for assessing mobile agents under environmental injection attacks within dynamic, executable environments. Moving beyond static image-based assessments, GhostEI-Bench injects adversarial events into realistic application workflows inside fully operational Android emulators and evaluates performance across critical risk scenarios. We further propose a judge-LLM protocol that conducts fine-grained failure analysis by reviewing the agent's action trajectory alongside the corresponding screenshot sequence, pinpointing failure in perception, recognition, or reasoning. Comprehensive experiments on state-of-the-art agents reveal pronounced vulnerability to deceptive environmental cues: current models systematically fail to perceive and reason about manipulated UIs. GhostEI-Bench provides a framework for quantifying and mitigating this emerging threat, paving the way toward more robust and secure embodied agents.
Problem

Research questions and friction points this paper is trying to address.

Evaluating mobile agents' resilience to adversarial UI injections in dynamic environments
Assessing how environmental injection attacks bypass textual safeguards on devices
Measuring agent vulnerability to deceptive visual elements causing execution failures
Innovation

Methods, ideas, or system contributions that make the work stand out.

Introduces benchmark for environmental injection attacks
Uses Android emulators for dynamic workflow evaluation
Proposes judge-LLM protocol for failure analysis
🔎 Similar Papers
No similar papers found.
C
Chiyu Chen
Shanghai Jiao Tong University, School of Computer Science
X
Xinhao Song
Shanghai Jiao Tong University, School of Computer Science
Y
Yunkai Chai
Shanghai Artificial Intelligence Laboratory
Y
Yang Yao
The University of Hong Kong
Haodong Zhao
Haodong Zhao
Shanghai Jiao Tong University
Federated LearningLLM
L
Lijun Li
Shanghai Artificial Intelligence Laboratory
J
Jie Li
Shanghai Artificial Intelligence Laboratory
Y
Yan Teng
Shanghai Artificial Intelligence Laboratory
G
Gongshen Liu
Shanghai Jiao Tong University, School of Computer Science
Y
Yingchun Wang
Shanghai Artificial Intelligence Laboratory