🤖 AI Summary
To address the challenge that multimodal large language models (MLLMs) struggle to parse fine-grained details in high-resolution images, this paper proposes a training-free, two-stage method. First, it leverages the MLLM’s native zero-shot spatial localization capability to generate bounding boxes that pinpoint salient regions. Second, it crops the corresponding high-resolution sub-image and jointly processes it with the original input across multiple rounds of inference and response regeneration, enabling self-correction. This work introduces the first “local focusing–self-correction” closed-loop mechanism, relying solely on the MLLM’s intrinsic spatial understanding, contextual reasoning, and comparative analysis—without fine-tuning, external annotations, or auxiliary parameters. The method achieves significant performance gains over state-of-the-art approaches on two major high-resolution multimodal benchmarks. Code is publicly available.
📝 Abstract
Multimodal Large Language Models (MLLM) often struggle to interpret high-resolution images accurately, where fine-grained details are crucial for complex visual understanding. We introduce Zoom-Refine, a novel training-free method that enhances MLLM capabilities to address this issue. Zoom-Refine operates through a synergistic process of extit{Localized Zoom} and extit{Self-Refinement}. In the extit{Localized Zoom} step, Zoom-Refine leverages the MLLM to provide a preliminary response to an input query and identifies the most task-relevant image region by predicting its bounding box coordinates. During the extit{Self-Refinement} step, Zoom-Refine then integrates fine-grained details from the high-resolution crop (identified by extit{Localized Zoom}) with its initial reasoning to re-evaluate and refine its preliminary response. Our method harnesses the MLLM's inherent capabilities for spatial localization, contextual reasoning and comparative analysis without requiring additional training or external experts. Comprehensive experiments demonstrate the efficacy of Zoom-Refine on two challenging high-resolution multimodal benchmarks. Code is available at href{https://github.com/xavier-yu114/Zoom-Refine}{color{magenta}github.com/xavier-yu114/Zoom-Refine}