SuperSNN: A Hardware-Aware Framework for Physically Realizable, High-Performance Superconducting Spiking Neural Network Chips

📅 2025-09-05
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the scalability limitations of superconducting spiking neural networks (SNNs)—constrained by chip area (3.4 × 3.9 mm²), I/O pin count (40 pins), and routing density—this work proposes the first hardware-aware, full-chip implementation framework. The framework integrates hardware-constraint-driven training, high-fan-in single-flux-quantum (SFQ) neuron design, custom superconducting logic cells, and a LAGS (Local-Global Synchronization) clock architecture, alongside off-chip pruning, weight quantization, and SFQ circuit co-optimization. Evaluated on MNIST, the framework achieves 96.47% training accuracy and 86.2% measured on-chip inference accuracy, operating at 3.02 GHz with 2.15 mW power consumption and only 6.55 fJ per inference. This represents a significant advancement in balancing physical realizability, energy efficiency, and accuracy for superconducting SNNs.

Technology Category

Application Category

📝 Abstract
Despite numerous proposed designs for superconducting neural networks (SNNs), most have overlooked practical fabrication constraints, leading to implementations limited to only a few neurons or synapses. Current superconducting technologies, such as MIT LL SFQ5ee, impose severe limitations on chip area, routing, and input/output pin counts (e.g., 5x5 mm^2 chip with 40 pins), drastically restricting network size and complexity. These hardware constraints necessitate a comprehensive framework to tailor network designs for physical realizability while minimizing accuracy loss. This paper introduces SuperSNN, a comprehensive framework for the implementation of full superconducting SNNs on a chip within these constraints. The key technical contributions include: (1) A hardware-aware training methodology for SNNs, utilizing off-chip pruning and weight quantization for energy-efficient superconducting implementations. (2) Design and layout of an inference SNN chip that incorporates novel high fan-in neurons and custom superconducting cells. (3) An optimized locally synchronous, globally synchronous (LAGS) clock distribution scheme for robust circuit implementation and management of data transfer delays in SFQ SNNs. The main results and findings demonstrate the effectiveness of the framework: (1) The complete network achieved 96.47% accuracy on the full MNIST dataset after quantization and pruning. (2) The fabricated SuperSNN chip successfully classified a reduced set of digits (2, 3, and 4) with 80.07% accuracy, reaching a maximum of 86.2% accuracy for digits 0, 1, and 2. (3) The chip operates at an ultra-high 3.02 GHz clock frequency. (4) It occupies a compact area of 3.4 x 3.9 mm^2, incorporates 5,822 Josephson Junctions, consumes 2.15 mW static power, and has an exceptionally low energy cost of 6.55 fJ (or 1.31e-6 nJ) per inference.
Problem

Research questions and friction points this paper is trying to address.

Addressing fabrication constraints in superconducting neural networks
Overcoming hardware limitations for scalable SNN chip implementation
Ensuring physical realizability while minimizing accuracy loss
Innovation

Methods, ideas, or system contributions that make the work stand out.

Hardware-aware training with pruning and quantization
High fan-in neurons and custom superconducting cells
LAGS clock distribution for SFQ SNNs
🔎 Similar Papers
No similar papers found.
C
Changxu Song
University of Southern California, Los Angeles, USA
A
Arda Caliskan
University of Southern California, Los Angeles, USA
B
Beyza Zeynep Ucpinar
University of Southern California, Los Angeles, USA
Y
Yasemin Kopur
University of Southern California, Los Angeles, USA
M
Mustafa Altay Karamuftuoglu
University of Southern California, Los Angeles, USA
S
Sasan Razmkhah
University of Southern California, Los Angeles, USA
Shahin Nazarian
Shahin Nazarian
University of Southern California, Professor of Electrical and Computer Engineering Practice
VerificationSoftware/hardware efficiencyHardware accelerationComputer aided designMachine
Massoud Pedram
Massoud Pedram
Charles Lee Powell Professor of Electrical Engineering & Computer Science, Univ of Southern
Energy Efficient ElectronicsPower-aware computingEnergy Harvesting and Storage SystemsComputer Aided Design of VLSI Circui