Attention Hijackers: Detect and Disentangle Attention Hijacking in LVLMs for Hallucination Mitigation

📅 2025-03-11
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work identifies a novel cause of hallucination in large vision-language models (LVLMs): certain instruction tokens hijack visual attention during decoding, suppressing perception of discriminative image regions—a phenomenon termed “attention hijacking.” To address this, we propose Attention Interference Decoupling (AID), a training-free method that (i) models instruction-driven visual saliency, (ii) applies dynamic attention masking to suppress hijacking tokens, and (iii) rebalances saliency from dual sources (instruction and image) to restore faithful visual grounding. AID is the first attention-decoupling framework for hallucination mitigation that requires no fine-tuning. Evaluated across multiple benchmarks, it significantly reduces hallucination rates while improving answer accuracy and visual faithfulness. Crucially, AID exhibits strong model-agnostic compatibility and incurs zero training overhead.

Technology Category

Application Category

📝 Abstract
Despite their success, Large Vision-Language Models (LVLMs) remain vulnerable to hallucinations. While existing studies attribute the cause of hallucinations to insufficient visual attention to image tokens, our findings indicate that hallucinations also arise from interference from instruction tokens during decoding. Intuitively, certain instruction tokens continuously distort LVLMs' visual perception during decoding, hijacking their visual attention toward less discriminative visual regions. This distortion prevents them integrating broader contextual information from images, ultimately leading to hallucinations. We term this phenomenon 'Attention Hijacking', where disruptive instruction tokens act as 'Attention Hijackers'. To address this, we propose a novel, training-free strategy namely Attention HIjackers Detection and Disentanglement (AID), designed to isolate the influence of Hijackers, enabling LVLMs to rely on their context-aware intrinsic attention map. Specifically, AID consists of three components: First, Attention Hijackers Detection identifies Attention Hijackers by calculating instruction-driven visual salience. Next, Attention Disentanglement mechanism is proposed to mask the visual attention of these identified Hijackers, and thereby mitigate their disruptive influence on subsequent tokens. Finally, Re-Disentanglement recalculates the balance between instruction-driven and image-driven visual salience to avoid over-masking effects. Extensive experiments demonstrate that AID significantly reduces hallucination across various LVLMs on several benchmarks.
Problem

Research questions and friction points this paper is trying to address.

Detects and mitigates attention hijacking in LVLMs
Addresses hallucinations caused by disruptive instruction tokens
Proposes a training-free strategy to disentangle attention hijackers
Innovation

Methods, ideas, or system contributions that make the work stand out.

Detects Attention Hijackers via instruction-driven salience
Disentangles attention by masking disruptive token influences
Recalculates salience balance to prevent over-masking effects
🔎 Similar Papers
2024-10-06Conference on Empirical Methods in Natural Language ProcessingCitations: 33
B
Beitao Chen
Center for Future Media, University of Electronic Science and Technology of China
Xinyu Lyu
Xinyu Lyu
Southwestern University of Finance and Economics
AI safetyMultimedia Learning
Lianli Gao
Lianli Gao
UESTC
Vision and Language
J
Jingkuan Song
Center for Future Media, University of Electronic Science and Technology of China; Tongji University
H
Heng Tao Shen
Center for Future Media, University of Electronic Science and Technology of China; Tongji University