π€ AI Summary
This work addresses the excessive computational overhead in large spoken language models caused by high sampling rates, which result in unnecessarily long input sequences far exceeding semantic requirements. Through inter-layer oracle intervention analysis, the study reveals that while shallow layers retain fine-grained acoustic details, deeper layers exhibit highly structured redundancy. Leveraging this insight, the authors propose Affinity Poolingβa training-free method that dynamically merges tokens based on similarity, both at the input and within deep layers. This approach preserves semantic accuracy while significantly improving efficiency: it reduces prefill FLOPs by 27.48%, decreases memory consumption by approximately 1.7Γ, and accelerates first-token latency by about 1.1Γ.
π Abstract
Large Speech Language Models (LSLMs) typically operate at high token rates (tokens/s) to ensure acoustic fidelity, yet this results in sequence lengths that far exceed the underlying semantic content, incurring prohibitive inference costs. In this paper, we empirically revisit the necessity of such granular token-level processing. Through layer-wise oracle interventions, we unveil a structured redundancy hierarchy: while shallow layers encode essential acoustic details, deep layers exhibit extreme redundancy, allowing for aggressive compression. Motivated by these findings, we introduce Affinity Pooling, a training-free, similarity-based token merging mechanism. By strategically applying this method at both input and deep layers, we effectively compress speech representations without compromising semantic information. Extensive evaluations across three tasks demonstrate that our approach reduces prefilling FLOPs by 27.48\% while maintaining competitive accuracy. Practical deployment further confirms significant efficiency gains, yielding up to $\sim$1.7$\times$ memory savings and $\sim$1.1$\times$ faster time-to-first-token on long utterances. Our results challenge the necessity of fully distinct token representations, providing new perspectives on LSLM efficiency.