🤖 AI Summary
This work addresses the limited effectiveness of multimodal large language models (MLLMs) as general-purpose visual backbones for dense prediction tasks such as semantic segmentation and depth estimation, primarily due to deficiencies in their visual encoders’ pixel-level feature representations. To overcome this limitation, the authors propose VersaViT, a Vision Transformer–based framework that incorporates lightweight multi-task heads and multi-granularity supervision signals to enable task-guided collaborative post-training of MLLM visual encoders. This approach systematically identifies and effectively mitigates the representational shortcomings of MLLMs at the pixel level, yielding substantial performance gains across diverse downstream dense prediction benchmarks. VersaViT thus establishes a unified visual backbone capable of simultaneously supporting high-level language reasoning and fine-grained pixel-level understanding.
📝 Abstract
Multimodal Large Language Models (MLLMs) have recently achieved remarkable success in visual-language understanding, demonstrating superior high-level semantic alignment within their vision encoders. An important question thus arises: Can these encoders serve as versatile vision backbones, capable of reliably performing classic vision-centric tasks as well? To address the question, we make the following contributions: (i) we identify that the vision encoders within MLLMs exhibit deficiencies in their dense feature representations, as evidenced by their suboptimal performance on dense prediction tasks (e.g., semantic segmentation, depth estimation); (ii) we propose VersaViT, a well-rounded vision transformer that instantiates a novel multi-task framework for collaborative post-training. This framework facilitates the optimization of the vision backbone via lightweight task heads with multi-granularity supervision; (iii) extensive experiments across various downstream tasks demonstrate the effectiveness of our method, yielding a versatile vision backbone suited for both language-mediated reasoning and pixel-level understanding.