🤖 AI Summary
The increasing photorealism of AI-generated images blurs the boundary between authentic and synthetic content, while existing vision-language model (VLM)-based detection methods lack robustness in identifying subtle artificial artifacts. To address this, we propose ZoomIn, a two-stage interpretable digital forensics framework: (1) coarse-grained localization of suspicious regions, followed by (2) fine-grained magnification and analysis of local anomalies—mimicking human visual inspection to jointly optimize detection accuracy and interpretability. To support training and evaluation, we introduce MagniFake, the first high-quality forgery dataset with pixel-level annotations, and design a VLM-driven automated pipeline for synthetic data generation. Experiments demonstrate that ZoomIn achieves 96.39% accuracy across diverse multi-source test sets, exhibits strong generalization, and generates natural-language explanations grounded in visual evidence.
📝 Abstract
The rapid growth of AI-generated imagery has blurred the boundary between real and synthetic content, raising critical concerns for digital integrity. Vision-language models (VLMs) offer interpretability through explanations but often fail to detect subtle artifacts in high-quality synthetic images. We propose ZoomIn, a two-stage forensic framework that improves both accuracy and interpretability. Mimicking human visual inspection, ZoomIn first scans an image to locate suspicious regions and then performs a focused analysis on these zoomed-in areas to deliver a grounded verdict. To support training, we introduce MagniFake, a dataset of 20,000 real and high-quality synthetic images annotated with bounding boxes and forensic explanations, generated through an automated VLM-based pipeline. Our method achieves 96.39% accuracy with robust generalization, while providing human-understandable explanations grounded in visual evidence.