π€ AI Summary
To address the challenge of efficiently deploying large language model (LLM) inference on resource-constrained edge CPU platforms, this paper proposes a full-stack co-designed ternary LLM inference framework. The method innovatively leverages SIMD register files to dynamically construct on-chip lookup tables, enabling register-level in-situ computation and circumventing memory-access bottlenecks. It integrates ternary quantization, register reorganization, ALU restructuring, and data-level parallelism to significantly accelerate GEMM and GEMVβcore operators in LLM inference. Evaluated on an NVIDIA Jetson AGX Orin, the framework achieves 5.6β24.5Γ lower GEMM latency and 1.1β86.2Γ higher GEMV throughput, while improving energy efficiency by 2.5β4.9Γ. Critically, it incurs only 3.2% additional power consumption and 1.4% area overhead. This work establishes a scalable, low-power LLM deployment paradigm tailored for pure-CPU edge devices.
π Abstract
Recent advances in LLMs have outpaced the computational and memory capacities of edge platforms that primarily employ CPUs, thereby challenging efficient and scalable deployment. While ternary quantization enables significant resource savings, existing CPU solutions rely heavily on memory-based lookup tables (LUTs) which limit scalability, and FPGA or GPU accelerators remain impractical for edge use. This paper presents T-SAR, the first framework to achieve scalable ternary LLM inference on CPUs by repurposing the SIMD register file for dynamic, in-register LUT generation with minimal hardware modifications. T-SAR eliminates memory bottlenecks and maximizes data-level parallelism, delivering 5.6-24.5x and 1.1-86.2x improvements in GEMM latency and GEMV throughput, respectively, with only 3.2% power and 1.4% area overheads in SIMD units. T-SAR achieves up to 2.5-4.9x the energy efficiency of an NVIDIA Jetson AGX Orin, establishing a practical approach for efficient LLM inference on edge platforms.