Temporal-adaptive Weight Quantization for Spiking Neural Networks

📅 2025-11-14
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the significant accuracy degradation in Spiking Neural Networks (SNNs) caused by ultra-low-bit weight quantization, this paper proposes Temporal-adaptive Weight Quantization (TaWQ). Inspired by astrocyte-mediated regulation of synaptic plasticity in biological neural systems, TaWQ is the first method to incorporate temporal dynamics into weight quantization—adaptively allocating ultra-low-bit weights along the time dimension and jointly optimizing spiking dynamics and quantization error during training. The approach enables end-to-end quantized training on both static (ImageNet) and neuromorphic datasets. Experiments demonstrate that, on ImageNet, TaWQ incurs only a 0.22% Top-1 accuracy drop while compressing model parameters to 4.12M and reducing single-inference energy consumption to 0.63 mJ. This achieves an unprecedented balance between high accuracy and extreme energy efficiency, establishing a novel paradigm for brain-inspired computing at the edge.

Technology Category

Application Category

📝 Abstract
Weight quantization in spiking neural networks (SNNs) could further reduce energy consumption. However, quantizing weights without sacrificing accuracy remains challenging. In this study, inspired by astrocyte-mediated synaptic modulation in the biological nervous systems, we propose Temporal-adaptive Weight Quantization (TaWQ), which incorporates weight quantization with temporal dynamics to adaptively allocate ultra-low-bit weights along the temporal dimension. Extensive experiments on static (e.g., ImageNet) and neuromorphic (e.g., CIFAR10-DVS) datasets demonstrate that our TaWQ maintains high energy efficiency (4.12M, 0.63mJ) while incurring a negligible quantization loss of only 0.22% on ImageNet.
Problem

Research questions and friction points this paper is trying to address.

Quantizing SNN weights without accuracy loss
Adaptive ultra-low-bit allocation across time
Maintaining energy efficiency with minimal quantization error
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses temporal-adaptive weight quantization for SNNs
Adaptively allocates ultra-low-bit weights temporally
Maintains high energy efficiency with minimal accuracy loss
🔎 Similar Papers
No similar papers found.
H
Han Zhang
Faculty of Computing, Harbin Institute of Technology, Harbin 150001, China, and Pengcheng Laboratory, Nanshan, Shenzhen 518000, China
Qingyan Meng
Qingyan Meng
Ph.D. student, The Chinese University of Hong Kong, Shenzhen
Machine Learning
J
Jiaqi Wang
Institute of Computing and Intelligence (ICI), Harbin Institute of Technology, Shenzhen 518055, China, and Pengcheng Laboratory, Nanshan, Shenzhen 518000, China
B
Baiyu Chen
Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China, and Pengcheng Laboratory, Nanshan, Shenzhen 518000, China
Zhengyu Ma
Zhengyu Ma
Pengcheng Laboratory
NeuroscienceNeural Network DynamicsComputational Physics
Xiaopeng Fan
Xiaopeng Fan
Professor, Harbin Institute of Technology
Video/ImageWireless