🤖 AI Summary
This work addresses the poor transferability of Transformers in time-series modeling, attributable to modality-specific characteristics. We first identify rapid spectral decay in time-series embeddings as the root cause of low-rank attention matrices in time-series Transformers, and introduce the novel concept of “flowing rank” to theoretically characterize the depth-dependent, nonlinear growth of matrix rank through network layers. Building on this insight, we propose a singular-value-analysis-based low-rank approximation method for Q/K/V projections and an attention-layer compression scheme. Empirical evaluation on the Chronos model demonstrates a 65% reduction in inference latency and an 81% decrease in memory footprint, with zero accuracy degradation. Our work provides both interpretable, generalizable theoretical foundations—linking embedding spectra, rank dynamics, and attention geometry—and practical techniques for efficient architecture design, theoretical analysis, and deployment of time-series Transformers.
📝 Abstract
Transformers are widely used across data modalities, and yet the principles distilled from text models often transfer imperfectly to models trained to other modalities. In this paper, we analyze Transformers through the lens of rank structure. Our focus is on the time series setting, where the structural properties of the data differ remarkably from those of text or vision. We show that time-series embeddings, unlike text or vision, exhibit sharply decaying singular value spectra: small patch sizes and smooth continuous mappings concentrate the data into low-rank subspaces. From this, we prove that the associated $Q/K/V$ projections admit accurate low-rank approximations, and that attention layers become compressible in proportion to the decay of the embedding spectrum. We introduce the concept of flow-of-ranks, a phenomenon by which nonlinear mixing across depth inflates the rank, explaining why early layers are most amenable to compression and why ranks grow with depth. Guided by these theoretical and empirical results, we use these insights to compress Chronos, a large time series foundation model, achieving a reduction of $65%$ in inference time and $81%$ in memory, without loss of accuracy. Our findings provide principled guidance for allocating width, depth, and heads in time series foundation models, and for exploiting their inherent compressibility.