4.6 Article

STT-BSNN: An In-Memory Deep Binary Spiking Neural Network Based on STT-MRAM

Journal

IEEE ACCESS
Volume 9, Issue -, Pages 151373-151385

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/ACCESS.2021.3125685

Keywords

Binary spiking neural network; emerging memory technology; in-memory computing; neuromorphic computing; process variation; STT-MRAM

Funding

  1. Vietnam National Foundation for Science and Technology Development (NAFOSTED) [102.01-2018.310]

Ask authors/readers for more resources

This paper introduces an in-memory BSNN based on STT-MRAM, utilising surrogate gradient learning to shorten time steps while maintaining accuracy. At the circuit level, presynaptic spikes are input into memory units via differential bit lines, with binarized weights stored in nonvolatile STT-MRAM, enabling high parallelism.
This paper proposes an in-memory binary spiking neural network (BSNN) based on spin-transfer-torque magnetoresistive RAM (STT-MRAM). We propose residual BSNN learning using a surrogate gradient that shortens the time steps in the BSNN while maintaining sufficient accuracy. At the circuit level, presynaptic spikes are fed to memory units through differential bit lines (BLs), while binarized weights are stored in a subarray of nonvolatile STT-MRAM. When the common inputs are fed through BLs, vector-to-matrix multiplication can be performed in a single memory sensing phase, hence achieving massive parallelism with low power and low latency. We further introduce the concept of a dynamic threshold to reduce the implementation complexity of synapses and neuron circuitry. This adjustable threshold also permits a nonlinear batch normalization (BN) function to be incorporated into the integrate-and-fire (IF) neuron circuit. The circuitry greatly improves the overall performance and enables high regularity in circuit layouts. Our proposed netlist circuits are built on a 65-nm CMOS with a fitted magnetic tunnel junction (MTJ) model for performance evaluation. The hardware/software co-simulation results indicate that the proposed design can deliver a performance of 176.6 TOPS/W for an in-memory computing (IMC) subarray size of 1 x 288. The classification accuracy reaches 97.92% (83.85%) on the MNIST (CIFAR-10) dataset. The impacts of the device non-idealities and process variations are also thoroughly covered in the analysis.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available