๐ค AI Summary
Current semi-supervised learning (SSL) methods for spiking neural networks (SNNs) are scarce and prone to confirmation bias due to inadequate modeling of temporal dynamics. Method: We propose SpikeMatchโthe first SSL framework explicitly incorporating SNN temporal dynamics. It leverages the leak factor of leaky integrate-and-fire neurons to modulate temporal responses, generating multi-view pseudo-labels for the same input; integrates strong/weak data augmentations with multi-prediction consistency constraints; and enables discriminative learning on weakly labeled samples via single-model co-training. Contribution/Results: SpikeMatch effectively mitigates confirmation bias inherent in conventional pseudo-labeling approaches and significantly improves few-shot generalization. On multiple standard benchmarks, it substantially outperforms existing SNN-adapted SSL methods, empirically validating the critical role of explicit temporal dynamic modeling in SNN semi-supervised learning.
๐ Abstract
Spiking neural networks (SNNs) have recently been attracting significant attention for their biological plausibility and energy efficiency, but semi-supervised learning (SSL) methods for SNN-based models remain underexplored compared to those for artificial neural networks (ANNs). In this paper, we introduce SpikeMatch, the first SSL framework for SNNs that leverages the temporal dynamics through the leakage factor of SNNs for diverse pseudo-labeling within a co-training framework. By utilizing agreement among multiple predictions from a single SNN, SpikeMatch generates reliable pseudo-labels from weakly-augmented unlabeled samples to train on strongly-augmented ones, effectively mitigating confirmation bias by capturing discriminative features with limited labels. Experiments show that SpikeMatch outperforms existing SSL methods adapted to SNN backbones across various standard benchmarks.