🤖 AI Summary
Transformer-based multimodal large language models face dual bottlenecks: diminishing returns in performance gains and sharply increasing training/inference costs. To address this, we propose *Platonic Grounding*—the first method to model the implicit deep cross-modal alignment observed in pretrained models as a plug-and-play, zero-parameter, incremental alignment module. Our approach freezes the backbone, projects modality-specific features into a shared, modality-agnostic latent space, and enforces alignment via contrastive learning—integrating mechanistic interpretability insights with lightweight fine-tuning. On multimodal understanding and generation benchmarks, it matches or exceeds baseline performance while reducing training computation by 37% and inference latency by 52%, enabling composable, plug-and-play integration of large models. The core contribution is explicit, transferable, and interpretable deep cross-modal representation alignment—achieved without introducing any new parameters.
📝 Abstract
The hyperscaling of data and parameter count in Transformer-based models is yielding diminishing performance improvement, especially when weighed against training costs. Such plateauing indicates the importance of methods for more efficient finetuning and inference, while retaining similar performance. This is especially relevant for multimodal learning paradigms, where inference costs of processing multimodal tokens can determine the model's practical viability. At the same time, research on representations and mechanistic interpretability has improved our understanding of the inner workings of Transformer-based models; one such line of work reveals an implicit alignment in the deeper layers of pretrained models, across modalities. Taking inspiration from this, we motivate and propose a simple modification to existing multimodal frameworks that rely on aligning pretrained models. We demonstrate that our approach maintains and, in some cases, even improves performance of baseline methods while achieving significant gains in both training and inference-time compute. Our work also has implications for combining pretrained models into larger systems efficiently.