38K
views
29
authors
6
articles
Editors
5
Impact
Loading...
6,822 views
28 citations
Methods
04 November 2021

Spiking Neural Networks (SNNs) are a pathway that could potentially empower low-power event-driven neuromorphic hardware due to their spatio-temporal information processing capability and high biological plausibility. Although SNNs are currently more efficient than artificial neural networks (ANNs), they are not as accurate as ANNs. Error backpropagation is the most common method for directly training neural networks, promoting the prosperity of ANNs in various deep learning fields. However, since the signals transmitted in the SNN are non-differentiable discrete binary spike events, the activation function in the form of spikes presents difficulties for the gradient-based optimization algorithms to be directly applied in SNNs, leading to a performance gap (i.e., accuracy and latency) between SNNs and ANNs. This paper introduces a new learning algorithm, called SSTDP, which bridges the gap between backpropagation (BP)-based learning and spike-time-dependent plasticity (STDP)-based learning to train SNNs efficiently. The scheme incorporates the global optimization process from BP and the efficient weight update derived from STDP. It not only avoids the non-differentiable derivation in the BP process but also utilizes the local feature extraction property of STDP. Consequently, our method can lower the possibility of vanishing spikes in BP training and reduce the number of time steps to reduce network latency. In SSTDP, we employ temporal-based coding and use Integrate-and-Fire (IF) neuron as the neuron model to provide considerable computational benefits. Our experiments show the effectiveness of the proposed SSTDP learning algorithm on the SNN by achieving the best classification accuracy 99.3% on the Caltech 101 dataset, 98.1% on the MNIST dataset, and 91.3% on the CIFAR-10 dataset compared to other SNNs trained with other learning methods. It also surpasses the best inference accuracy of the directly trained SNN with 25~32 × less inference latency. Moreover, we analyze event-based computations to demonstrate the efficacy of the SNN for inference operation in the spiking domain, and SSTDP methods can achieve 1.3~37.7 × fewer addition operations per inference. The code is available at: https://github.com/MXHX7199/SNN-SSTDP.

15,215 views
49 citations
Recommended Research Topics
Frontiers Logo

Frontiers in Neuroscience

Spiking Neural Network Learning, Benchmarking, Programming and Executing
Edited by Guoqi Li, Yam Song (Yansong) Chua, Haizhou Li, Peng Li, Emre O Neftci, Lei Deng
168.6K
views
75
authors
16
articles
Frontiers Logo

Frontiers in Neuroscience

Neuro-inspired Computing for Next-gen AI: Computing Model, Architectures and Learning Algorithms
Edited by Angeliki Pantazi, Emre O Neftci, Bipin Rajendran, Osvaldo Simeone
58.6K
views
41
authors
11
articles
Frontiers Logo

Frontiers in Neuroscience

Recent Advances of Deep Spiking Neural Networks and Neuromorphic Computation for Computer Vision
Edited by Nikola K Kasabov, Maryam Doborjeh, Zeng-Guang Hou, WEIQI YAN, Jie Yang
40.2K
views
14
authors
6
articles
Frontiers Logo

Frontiers in Neuroscience

Digital Neuromorphic System with Online Learning: Algorithms, Implementation, and Large-scale Design
Edited by Shuangming Yang, Mostafa Rahimi Azghadi, Jungong Han
18.2K
views
36
authors
5
articles
Frontiers Logo

Frontiers in Neuroscience

Deep Spiking Neural Networks: Models, Algorithms and Applications
Edited by Yonghong Tian, Timothée Masquelier, Guoqi Li, Huihui Zhou, Zhengyu Ma
36.6K
views
33
authors
9
articles