HII-DPO: Eliminate Hallucination via Accurate Hallucination-Inducing Counterfactual Images

πŸ“… 2026-02-11
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work addresses the persistent challenge of hallucinations in large vision-language models, which often stem from linguistic biases and lack identifiable, stable patterns. To this end, we systematically construct Hallucination-Inducing Images (HII) for the first time, uncovering reproducible hallucination mechanisms under specific scene conditions. Building upon these insights, we introduce the Masked-Object-Hallucination (MOH) benchmark for targeted evaluation. Leveraging synthetic HII data and fine-grained preference annotations, we employ Direct Preference Optimization (DPO) for alignment training, achieving effective hallucination mitigation. Our approach improves performance by up to 38% over current state-of-the-art methods on standard hallucination benchmarks while preserving the model’s general capabilities.

Technology Category

Application Category

πŸ“ Abstract
Large Vision-Language Models (VLMs) have achieved remarkable success across diverse multimodal tasks but remain vulnerable to hallucinations rooted in inherent language bias. Despite recent progress, existing hallucination mitigation methods often overlook the underlying hallucination patterns driven by language bias. In this work, we design a novel pipeline to accurately synthesize Hallucination-Inducing Images (HIIs). Using synthesized HIIs, we reveal a consistent scene-conditioned hallucination pattern: models tend to mention objects that are highly typical of the scene even when visual evidence is removed. To quantify the susceptibility of VLMs to this hallucination pattern, we establish the Masked-Object-Hallucination (MOH) benchmark to rigorously evaluate existing state-of-the-art alignment frameworks. Finally, we leverage HIIs to construct high-quality preference datasets for fine-grained alignment. Experimental results demonstrate that our approach effectively mitigates hallucinations while preserving general model capabilities. Specifically, our method achieves up to a 38% improvement over the current state-of-the-art on standard hallucination benchmarks.
Problem

Research questions and friction points this paper is trying to address.

hallucination
vision-language models
language bias
counterfactual images
object hallucination
Innovation

Methods, ideas, or system contributions that make the work stand out.

Hallucination-Inducing Images
Language Bias
Masked-Object-Hallucination Benchmark
Fine-grained Alignment
Visual-Language Models
πŸ”Ž Similar Papers
No similar papers found.