Training High-Performance Low-Latency Spiking Neural Networks by Differentiation on Spike Representation

被引:65
|
作者
Meng, Qingyan [1 ,2 ]
Xiao, Mingqing [3 ]
Yan, Shen [4 ]
Wang, Yisen [3 ,5 ]
Lin, Zhouchen [3 ,5 ,6 ]
Luo, Zhi-Quan [1 ,2 ]
机构
[1] Chinese Univ Hong Kong, Shenzhen, Peoples R China
[2] Shenzhen Res Inst Big Data, Shenzhen, Peoples R China
[3] Peking Univ, Sch Artificial Intelligence, Key Lab Machine Percept MoE, Beijing, Peoples R China
[4] Peking Univ, Ctr Data Sci, Beijing, Peoples R China
[5] Peking Univ, Inst Artificial Intelligence, Beijing, Peoples R China
[6] Peng Cheng Lab, Shenzhen, Guangdong, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52688.2022.01212
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Spiking Neural Network (SNN) is a promising energy-efficient AI model when implemented on neuromorphic hardware. However, it is a challenge to efficiently train SNNs due to their non-differentiability. Most existing methods either suffer from high latency (i.e., long simulation time steps), or cannot achieve as high performance as Artificial Neural Networks (ANNs). In this paper, we propose the Differentiation on Spike Representation (DSR) method, which could achieve high performance that is competitive to ANNs yet with low latency. First, we encode the spike trains into spike representation using (weighted) firing rate coding. Based on the spike representation, we systematically derive that the spiking dynamics with common neural models can be represented as some sub-differentiable mapping. With this viewpoint, our proposed DSR method trains SNNs through gradients of the mapping and avoids the common non-differentiability problem in SNN training. Then we analyze the error when representing the specific mapping with the forward computation of the SNN. To reduce such error, we propose to train the spike threshold in each layer, and to introduce a new hyperparameter for the neural models. With these components, the DSR method can achieve state-of-the-art SNN performance with low latency on both static and neuromorphic datasets, including CIFAR-10, CIFAR-100, ImageNet, and DVS-CIFAR10.
引用
收藏
页码:12434 / 12443
页数:10
相关论文
共 50 条
  • [1] Training Low-Latency Spiking Neural Network with Orthogonal Spiking Neurons
    Yao, Yunpeng
    Wu, Man
    Zhang, Renyuan
    2023 21ST IEEE INTERREGIONAL NEWCAS CONFERENCE, NEWCAS, 2023,
  • [2] Highway Connection for Low-Latency and High-Accuracy Spiking Neural Networks
    Zhang, Anguo
    Wu, Junyi
    Li, Xiumin
    Li, Hung Chun
    Gao, Yueming
    Pun, Sio Hang
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2023, 70 (12) : 4579 - 4583
  • [3] Optimized Potential Initialization for Low-Latency Spiking Neural Networks
    Bu, Tong
    Ding, Jianhao
    Yu, Zhaofei
    Huang, Tiejun
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 11 - 20
  • [4] Analytical performance assessment and high-throughput low-latency spike routing algorithm for spiking neural network systems
    Huy-The Vu
    Okuyama, Yuichi
    Ben Abdallah, Abderazek
    JOURNAL OF SUPERCOMPUTING, 2019, 75 (08): : 5367 - 5397
  • [5] Analytical performance assessment and high-throughput low-latency spike routing algorithm for spiking neural network systems
    Huy-The Vu
    Yuichi Okuyama
    Abderazek Ben Abdallah
    The Journal of Supercomputing, 2019, 75 : 5367 - 5397
  • [6] Trainable Spiking-YOLO for low-latency and high-performance object detection
    Yuan, Mengwen
    Zhang, Chengjun
    Wang, Ziming
    Liu, Huixiang
    Pan, Gang
    Tang, Huajin
    NEURAL NETWORKS, 2024, 172
  • [7] Revisiting Batch Normalization for Training Low-Latency Deep Spiking Neural Networks From Scratch
    Kim, Youngeun
    Panda, Priyadarshini
    FRONTIERS IN NEUROSCIENCE, 2021, 15
  • [8] Can Deep Neural Networks be Converted to Ultra Low-Latency Spiking Neural Networks?
    Datta, Gourav
    Beerel, Peter A.
    PROCEEDINGS OF THE 2022 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2022), 2022, : 718 - 723
  • [9] Training Low-Latency Spiking Neural Network through Knowledge Distillation
    Takuya, Sugahara
    Zhang, Renyuan
    Nakashima, Yasuhiko
    2021 IEEE COOL CHIPS 24: IEEE SYMPOSIUM IN LOW-POWER AND HIGH-SPEED CHIPS, 2021,
  • [10] Constrain Bias Addition to Train Low-Latency Spiking Neural Networks
    Lin, Ranxi
    Dai, Benzhe
    Zhao, Yingkai
    Chen, Gang
    Lu, Huaxiang
    BRAIN SCIENCES, 2023, 13 (02)