Efficient Training of Spiking Neural Networks by Spike-aware Data Pruning

📅 2025-10-05
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Training large-scale Spiking Neural Networks (SNNs) incurs prohibitive computational overhead, and existing data pruning methods—designed for Artificial Neural Networks (ANNs)—exhibit biased importance estimation and high gradient variance in SNNs due to their neglect of spiking dynamics. Method: This work introduces data pruning into SNN training for the first time, proposing a spike-aware importance score mechanism that jointly incorporates gradient norm and spiking activity characteristics. It enables efficient, low-variance sample importance estimation without real-time gradient computation and is tightly integrated with SNN backpropagation for co-optimization. Contribution/Results: On ImageNet, our method achieves 35% training speedup with zero accuracy degradation, approaching the theoretical optimal acceleration ratio. Extensive generalization experiments confirm its effectiveness across diverse SNN architectures and datasets. This work establishes a novel paradigm for resource-efficient SNN training.

Technology Category

Application Category

📝 Abstract
Spiking neural networks (SNNs), recognized as an energy-efficient alternative to traditional artificial neural networks (ANNs), have advanced rapidly through the scaling of models and datasets. However, such scaling incurs considerable training overhead, posing challenges for researchers with limited computational resources and hindering the sustained development of SNNs. Data pruning is a promising strategy for accelerating training by retaining the most informative examples and discarding redundant ones, but it remains largely unexplored in SNNs. Directly applying ANN-based data pruning methods to SNNs fails to capture the intrinsic importance of examples and suffers from high gradient variance. To address these challenges, we propose a novel spike-aware data pruning (SADP) method. SADP reduces gradient variance by determining each example's selection probability to be proportional to its gradient norm, while avoiding the high cost of direct gradient computation through an efficient upper bound, termed spike-aware importance score. This score accounts for the influence of all-or-nothing spikes on the gradient norm and can be computed with negligible overhead. Extensive experiments across diverse datasets and architectures demonstrate that SADP consistently outperforms data pruning baselines and achieves training speedups close to the theoretical maxima at different pruning ratios. Notably, SADP reduces training time by 35% on ImageNet while maintaining accuracy comparable to that of full-data training. This work, therefore, establishes a data-centric paradigm for efficient SNN training and paves the way for scaling SNNs to larger models and datasets. The source code will be released publicly after the review process.
Problem

Research questions and friction points this paper is trying to address.

Reducing computational training overhead for spiking neural networks
Addressing inefficiency of ANN pruning methods for SNN training
Minimizing gradient variance while maintaining training accuracy
Innovation

Methods, ideas, or system contributions that make the work stand out.

Spike-aware data pruning method reduces gradient variance
Uses efficient upper bound for spike-aware importance score
Achieves training speedups while maintaining comparable accuracy
🔎 Similar Papers
No similar papers found.
C
Chenxiang Ma
Department of Data Science and Artificial Intelligence, The Hong Kong Polytechnic University, Hong Kong, SAR
X
Xinyi Chen
Department of Data Science and Artificial Intelligence, The Hong Kong Polytechnic University, Hong Kong, SAR
Yujie Wu
Yujie Wu
Assistant Professor, The Hong Kong Polytechnic University
Brain-inspired AIComputational neuroscienceNeuromorphic computing
K
Kay Chen Tan
Department of Data Science and Artificial Intelligence and the Research Center of Data Science and Artificial Intelligence, The Hong Kong Polytechnic University, Hong Kong, SAR
Jibin Wu
Jibin Wu
The Hong Kong Polytechnic University
Spiking Neural NetworkNeuromorphic ComputingSpeech ProcessingCognitive Modelling