🤖 AI Summary
Long-context language models suffer from low inference efficiency on consumer-grade hardware. Method: This work presents the first systematic, end-to-end performance evaluation of Transformer, State Space Model (SSM), and hybrid architectures on embedded and consumer GPUs (e.g., 24 GB VRAM), leveraging operator-level fine-grained analysis and real hardware benchmarking. Contribution/Results: We find SSMs significantly outperform Transformers for ultra-long sequences (up to 220K tokens), with performance crossover occurring at ~57K tokens and peak speedups of 4×. Moreover, SSMs support sequence lengths four times longer than Transformers on a 24 GB GPU. Crucially, we identify that over 55% of SSM latency stems from custom operators—revealing the primary bottleneck for edge deployment. This finding provides empirical grounding and a clear optimization pathway for efficient SSM implementation and hardware-software co-design on resource-constrained devices.
📝 Abstract
The demand for machine intelligence capable of processing continuous, long-context inputs on local devices is growing rapidly. However, the quadratic complexity and memory requirements of traditional Transformer architectures make them inefficient and often unusable for these tasks. This has spurred a paradigm shift towards new architectures like State Space Models (SSMs) and hybrids, which promise near-linear scaling. While most current research focuses on the accuracy and theoretical throughput of these models, a systematic performance characterization on practical consumer hardware is critically needed to guide system-level optimization and unlock new applications.
To address this gap, we present a comprehensive, comparative benchmarking of carefully selected Transformer, SSM, and hybrid models specifically for long-context inference on consumer and embedded GPUs. Our analysis reveals that SSMs are not only viable but superior for this domain, capable of processing sequences up to 220K tokens on a 24GB consumer GPU-approximately 4x longer than comparable Transformers. While Transformers may be up to 1.8x faster at short sequences, SSMs demonstrate a dramatic performance inversion, becoming up to 4x faster at very long contexts (~57K tokens). Our operator-level analysis reveals that custom, hardware-aware SSM kernels dominate the inference runtime, accounting for over 55% of latency on edge platforms, identifying them as a primary target for future hardware acceleration. We also provide detailed, device-specific characterization results to guide system co-design for the edge. To foster further research, we will open-source our characterization framework.