🤖 AI Summary
This study investigates how visual information is propagated into the textual domain within vision-language models (VLMs). To this end, we employ residual stream analysis, token-level ablation, cross-modal attention visualization, and targeted token editing. Our analysis reveals that multimodal generative models rely on a single “bottleneck token” to achieve localized visual-semantic transmission, whereas purely text-output models utilize distributed communication across tokens. Crucially, we identify— for the first time—the bottleneck token’s decisive role in image understanding: targeted intervention on this token enables global, fine-grained control over image semantics, facilitating precise semantic steering. This finding establishes a novel paradigm for controllable multimodal understanding, substantially enhancing model controllability and interpretability. It overcomes the limitations of conventional black-box multimodal modeling by exposing and leveraging interpretable, localized cross-modal coupling mechanisms.
📝 Abstract
Recent advances in multimodal training have significantly improved the integration of image understanding and generation within a unified model. This study investigates how vision-language models (VLMs) handle image-understanding tasks, specifically focusing on how visual information is processed and transferred to the textual domain. We compare VLMs that generate both images and text with those that output only text, highlighting key differences in information flow. We find that in models with multimodal outputs, image and text embeddings are more separated within the residual stream. Additionally, models vary in how information is exchanged from visual to textual tokens. VLMs that only output text exhibit a distributed communication pattern, where information is exchanged through multiple image tokens. In contrast, models trained for image and text generation rely on a single token that acts as a narrow gate for the visual information. We demonstrate that ablating this single token significantly deteriorates performance on image understanding tasks. Furthermore, modifying this token enables effective steering of the image semantics, showing that targeted, local interventions can reliably control the model's global behavior.