4.7 Article

Temporal Backpropagation for Spiking Neural Networks with One Spike per Neuron

Journal

INTERNATIONAL JOURNAL OF NEURAL SYSTEMS
Volume 30, Issue 6, Pages -

Publisher

WORLD SCIENTIFIC PUBL CO PTE LTD
DOI: 10.1142/S0129065720500276

Keywords

Spiking neural network; supervised learning; temporal backpropagation; single spike coding

Funding

  1. French Agence Nationale de la Recherche [ANR-16-CE28-0017-01]
  2. Agence Nationale de la Recherche (ANR) [ANR-16-CE28-0017] Funding Source: Agence Nationale de la Recherche (ANR)

Ask authors/readers for more resources

We propose a new supervised learning rule for multilayer spiking neural networks (SNNs) that use a form of temporal coding known as rank-order-coding. With this coding scheme, all neurons fire exactly one spike per stimulus, but the firing order carries information. In particular, in the readout layer, the first neuron to fire determines the class of the stimulus. We derive a new learning rule for this sort of network, named S4NN, akin to traditional error backpropagation, yet based on latencies. We show how approximated error gradients can be computed backward in a feedforward network with any number of layers. This approach reaches state-of-the-art performance with supervised multi-fully connected layer SNNs: test accuracy of 97.4% for the MNIST dataset, and 99.2% for the ('altech Face/Motorbike dataset. Yet, the neuron model that we use, nonleaky integrate-and-fire, is much simpler than the one used in all previous works. The source codes of the proposed S4NN are publicly available at https://github.com/SRKH/S4NN.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available