🤖 AI Summary
This work addresses temporal redundancy and low hardware efficiency in artificial neural network (ANN) to spiking neural network (SNN) conversion. We propose a novel conversion method based on Sigma-Delta (ΣΔ) modulation—the first application of ΣΔ principles to SNN spike encoding design. Leveraging Loihi 2’s hierarchical spike encoding capability, our approach replaces conventional binary rate coding with multi-level quantized spikes, preserving spatiotemporal sparsity while substantially reducing the number of inference time steps. Experiments demonstrate that deploying our method on Loihi 2 achieves zero-accuracy-loss inference with significantly lower energy consumption and latency compared to Jetson Xavier—an edge AI platform. The core contribution is the introduction of a ΣΔ-driven hierarchical spike conversion paradigm, establishing a scalable, co-designed software–hardware optimization pathway for efficient neuromorphic inference.
📝 Abstract
Neuromorphic computing aims to improve the efficiency of artificial neural networks by taking inspiration from biological neurons and leveraging temporal sparsity, spatial sparsity, and compute near/in memory. Although these approaches have shown efficiency gains, training these spiking neural networks (SNN) remains difficult. The original attempts at converting trained conventional analog neural networks (ANN) to SNNs used the rate of binary spikes to represent neuron activations. This required many simulation time steps per inference, which degraded efficiency. Intel's Loihi 2 is a neuromorphic platform that supports graded spikes which can be used to represent changes in neuron activation. In this work, we use Loihi 2's graded spikes to develop a method for converting ANN networks to spiking networks, which take advantage of temporal and spatial sparsity. We evaluated the performance of this network on Loihi 2 and compared it to NVIDIA's Jetson Xavier edge AI platform.