🤖 AI Summary
Existing multimodal large language models (MLLMs) commonly rely on post-encoder visual token compression, limiting efficiency gains. To address this, this paper proposes a novel layer-wise compression framework embedded within the visual encoder’s intermediate layers. Our approach innovatively integrates pixel-reordering–driven spatial-channel transformation with parameter-free residual shortcuts, enabling dynamic token compression during encoding while preserving both accuracy and efficiency. By avoiding auxiliary modules, the method eliminates associated computational overhead: training efficiency improves by over 20%, and inference throughput increases by more than 15%. Extensive experiments across diverse downstream tasks and model scales demonstrate consistent and significant superiority over state-of-the-art external compression methods. The proposed framework establishes a scalable architectural paradigm for efficient multimodal understanding.
📝 Abstract
Existing visual token compression methods for Multimodal Large Language Models (MLLMs) predominantly operate as post-encoder modules, limiting their potential for efficiency gains. To address this limitation, we propose LaCo (Layer-wise Visual Token Compression), a novel framework that enables effective token compression within the intermediate layers of the vision encoder. LaCo introduces two core components: 1) a layer-wise pixel-shuffle mechanism that systematically merges adjacent tokens through space-to-channel transformations, and 2) a residual learning architecture with non-parametric shortcuts that preserves critical visual information during compression. Extensive experiments indicate that our LaCo outperforms all existing methods when compressing tokens in the intermediate layers of the vision encoder, demonstrating superior effectiveness. In addition, compared to external compression, our method improves training efficiency beyond 20% and inference throughput over 15% while maintaining strong performance.