๐ค AI Summary
Spiking neural networks (SNNs) offer significant energy efficiency advantages but face a fundamental training challenge: their non-differentiable, sparse spiking dynamics impede gradient-based end-to-end optimization. To address this, we propose S-SWIMโa gradient-free, interpretable, random-feature-driven training algorithm that systematically incorporates random feature mechanisms into spike response modeling for the first time, eliminating reliance on surrogate gradients or differentiable approximations of the spiking function. Unlike conventional random weight initialization, S-SWIM enables data-adaptive, rapid, and stable training. Experimental results demonstrate that S-SWIM achieves high accuracy on time-series forecasting tasks. It functions effectively both as a standalone training method and as a principled initialization strategy for gradient-based SNN training. By decoupling learning from spike differentiability constraints, S-SWIM establishes a novel, efficient training paradigm for SNNs.
๐ Abstract
Spiking Neural Networks (SNNs) as Machine Learning (ML) models have recently received a lot of attention as a potentially more energy-efficient alternative to conventional Artificial Neural Networks. The non-differentiability and sparsity of the spiking mechanism can make these models very difficult to train with algorithms based on propagating gradients through the spiking non-linearity. We address this problem by adapting the paradigm of Random Feature Methods (RFMs) from Artificial Neural Networks (ANNs) to Spike Response Model (SRM) SNNs. This approach allows training of SNNs without approximation of the spike function gradient. Concretely, we propose a novel data-driven, fast, high-performance, and interpretable algorithm for end-to-end training of SNNs inspired by the SWIM algorithm for RFM-ANNs, which we coin S-SWIM. We provide a thorough theoretical discussion and supplementary numerical experiments showing that S-SWIM can reach high accuracies on time series forecasting as a standalone strategy and serve as an effective initialisation strategy before gradient-based training. Additional ablation studies show that our proposed method performs better than random sampling of network weights.