đ€ AI Summary
Vision-language models (VLMs) consistently fail at feature bindingâaccurately associating multimodal attributes of the same object. This work identifies, for the first time, an internally emergent, content-agnostic spatial indexing mechanism within VLMs that implements feature binding in a symbol-like manner; we empirically establish a causal link between its failure and binding errors. Using interpretability analysis, neuron activation tracing, and structured intervention experimentsâcomplemented by contrastive task diagnostics and mechanistic attributionâwe precisely localize the neural computational units underpinning binding. Our findings reveal a symbolic foundation for visual reasoning in VLMs and provide an interpretable, intervenable theoretical framework and practical pathway to enhance their compositional generalization and structured understanding capabilities.
đ Abstract
To accurately process a visual scene, observers must bind features together to represent individual objects. This capacity is necessary, for instance, to distinguish an image containing a red square and a blue circle from an image containing a blue square and a red circle. Recent work has found that language models solve this'binding problem'via a set of symbol-like, content-independent indices, but it is unclear whether similar mechanisms are employed by vision language models (VLMs). This question is especially relevant, given the persistent failures of VLMs on tasks that require binding. Here, we identify a set of emergent symbolic mechanisms that support binding in VLMs via a content-independent, spatial indexing scheme. Moreover, we find that binding errors can be traced directly to failures in these mechanisms. Taken together, these results shed light on the mechanisms that support symbol-like processing in VLMs, and suggest possible avenues for addressing the persistent binding failures exhibited by these models.