Efficient Hardware Acceleration of Sparsely Active Convolutional Spiking Neural Networks

被引:12
|
作者
Sommer, Jan [1 ]
Ozkan, M. Akif [1 ]
Keszocze, Oliver [2 ]
Teich, Juergen [2 ]
机构
[1] Friedrich Alexander Univ Erlangen Nurnberg, Chair Hardware Software Codesign, D-91058 Erlangen, Germany
[2] Max Planck Inst Sci Light, D-91058 Erlangen, Germany
关键词
Event-based processing; field-programmable gate array (FPGA); hardware acceleration; spiking convolutional neural networks (SNNs);
D O I
10.1109/TCAD.2022.3197512
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Spiking neural networks (SNNs) compute in an event-based manner to achieve a more efficient computation than standard neural networks. In SNNs, neuronal outputs are not encoded as real-valued activations but as sequences of binary spikes. The motivation of using SNNs over conventional neural networks is rooted in the special computational aspects of spike-based processing, especially the high degree of sparsity of spikes. Well-established implementations of convolutional neural networks (CNNs) feature large spatial arrays of processing elements (PEs) that remain highly underutilized in the face of activation sparsity. We propose a novel architecture optimized for the processing of convolutional SNNs (CSNNs) featuring a high degree of sparsity. The proposed architecture consists of an array of PEs of the size of the kernel of a convolution and an intelligent spike queue that provides a high PE utilization. A constant flow of spikes is ensured by compressing the feature maps into queues that can then be processed spike-by-spike. This compression is performed at run-time, leading to a self-timed schedule. This allows the processing time to scale with the number of spikes. Also, a novel memory organization scheme is introduced to efficiently store and retrieve the membrane potentials of the individual neurons using multiple small parallel on-chip RAMs. Each RAM is hardwired to its PE, reducing switching circuitry. We implemented the proposed architecture on an FPGA and achieved a significant speedup compared to previously proposed SNN implementations (similar to 10 times) while needing less hardware resources and maintaining a higher energy efficiency (similar to 15 times).
引用
收藏
页码:3767 / 3778
页数:12
相关论文
共 50 条
  • [41] Compiling Spiking Neural Networks to Neuromorphic Hardware
    Song, Shihao
    Balaji, Adarsha
    Das, Anup
    Kandasamy, Nagarajan
    Shackleford, James
    21ST ACM SIGPLAN/SIGBED CONFERENCE ON LANGUAGES, COMPILERS, AND TOOLS FOR EMBEDDED SYSTEMS (LCTES '20), 2020, : 38 - 50
  • [42] Spiking Deep Convolutional Neural Networks for Energy-Efficient Object Recognition
    Cao, Yongqiang
    Chen, Yang
    Khosla, Deepak
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2015, 113 (01) : 54 - 66
  • [43] An adaptive threshold mechanism for accurate and efficient deep spiking convolutional neural networks
    Chen, Yunhua
    Mai, Yingchao
    Feng, Ren
    Xiao, Jinsheng
    NEUROCOMPUTING, 2022, 469 : 189 - 197
  • [44] An Efficient Hardware Accelerator for Block Sparse Convolutional Neural Networks on FPGA
    Yin, Xiaodi
    Wu, Zhipeng
    Li, Dejian
    Shen, Chongfei
    Liu, Yu
    IEEE EMBEDDED SYSTEMS LETTERS, 2024, 16 (02) : 158 - 161
  • [45] EFFICIENT CONVOLUTIONAL NEURAL NETWORKS FOR PIXELWISE CLASSIFICATION ON HETEROGENEOUS HARDWARE SYSTEMS
    Tschopp, Fabian
    Martel, Julien N. P.
    Turaga, Srinivas C.
    Cook, Matthew
    Funke, Jan
    2016 IEEE 13TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI), 2016, : 1225 - 1228
  • [46] Efficient Hardware Design of Convolutional Neural Networks for Accelerated Deep Learning
    Khalil, Kasem
    Khan, Md Rahat
    Bayoumi, Magdy
    Sherif, Ahmed
    2024 IEEE 67TH INTERNATIONAL MIDWEST SYMPOSIUM ON CIRCUITS AND SYSTEMS, MWSCAS 2024, 2024, : 1075 - 1079
  • [47] An Efficient Hardware Accelerator for Structured Sparse Convolutional Neural Networks on FPGAs
    Zhu, Chaoyang
    Huang, Kejie
    Yang, Shuyuan
    Zhu, Ziqi
    Zhang, Hejia
    Shen, Haibin
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2020, 28 (09) : 1953 - 1965
  • [48] Hardware Acceleration of Graph Neural Networks
    Auten, Adam
    Tomei, Matthew
    Kumar, Rakesh
    PROCEEDINGS OF THE 2020 57TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2020,
  • [49] HFP: Hardware-Aware Filter Pruning for Deep Convolutional Neural Networks Acceleration
    Yu, Fang
    Han, Chuanqi
    Wang, Pengcheng
    Huang, Ruoran
    Huang, Xi
    Cui, Li
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 255 - 262
  • [50] HILP: hardware-in-loop pruning of convolutional neural networks towards inference acceleration
    Dong Li
    Qianqian Ye
    Xiaoyue Guo
    Yunda Sun
    Li Zhang
    Neural Computing and Applications, 2024, 36 : 8825 - 8842