SpikeSim: An End-to-End Compute-in-Memory Hardware Evaluation Tool for Benchmarking Spiking Neural Networks

被引:9
|
作者
Moitra, Abhishek [1 ]
Bhattacharjee, Abhiroop [1 ]
Kuang, Runcong [2 ]
Krishnan, Gokul [3 ]
Cao, Yu [2 ]
Panda, Priyadarshini [1 ]
机构
[1] Yale Univ, Dept Elect Engn, New Haven, CT 06520 USA
[2] Arizona State Univ, Sch Elect Comp & Energy Engn, Tempe, AZ 85287 USA
[3] Meta Real Labs, Redmond, WA USA
基金
美国国家科学基金会;
关键词
Analog crossbars; emerging devices; in-memory computing (IMC); spiking neural networks (SNNs);
D O I
10.1109/TCAD.2023.3274918
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Spiking neural networks (SNNs) are an active research domain toward energy-efficient machine intelligence. Compared to conventional artificial neural networks (ANNs), SNNs use temporal spike data and bio-plausible neuronal activation functions such as leaky-integrate fire/integrate fire (LIF/IF) for data processing. However, SNNs incur significant dot-product operations causing high memory and computation overhead in standard von-Neumann computing platforms. To this end, in memory computing (IMC) architectures have been proposed to alleviate the "memory-wall bottleneck" prevalent in von Neumann architectures. Although recent works have proposed IMC-based SNN hardware accelerators, the following key implementation aspects have been overlooked: 1) the adverse effects of crossbar nonideality on SNN performance due to repeated analog dot-product operations over multiple time-steps and 2) hardware overheads of essential SNN-specific components, such as the LIF/IF and data communication modules. To this end, we propose SpikeSim, a tool that can perform realistic performance, energy, latency and area evaluation of IMC-mapped SNNs. SpikeSim consists of a practical monolithic IMC architecture called SpikeFlow for mapping SNNs. Additionally, the nonideality computation engine (NICE) and energy-latency-area (ELA) engine performs hardware-realistic evaluation of SpikeFlow-mapped SNNs. Based on 65nm CMOS implementation and experiments on CIFAR10, CIFAR100 and TinyImagenet datasets, we find that the LIF/IF neuronal module has significant area contribution (> 11% of the total hardware area). To this end, we propose SNN topological modifications that leads to 1.24x and 10x reduction in the neuronal module's area and the overall energy-delay-product value, respectively. Furthermore, in this work, we perform a holistic comparison between IMC implemented ANN and SNNs and conclude that lower number of time-steps are the key to achieve higher throughput and energy-efficiency for SNNs compared to 4-bit ANNs. The code repository for the SpikeSim tool is available at Github link.
引用
收藏
页码:3815 / 3828
页数:14
相关论文
共 50 条
  • [31] End-to-End Contextualized Document Indexing and Retrieval with Neural Networks
    Hofstaetter, Sebastian
    PROCEEDINGS OF THE 43RD INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '20), 2020, : 2481 - 2481
  • [32] Towards End-to-End Speech Recognition with Recurrent Neural Networks
    Graves, Alex
    Jaitly, Navdeep
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 32 (CYCLE 2), 2014, 32 : 1764 - 1772
  • [33] End-to-end risk budgeting portfolio optimization with neural networks
    Uysal, A. Sinem
    Li, Xiaoyue
    Mulvey, John M.
    ANNALS OF OPERATIONS RESEARCH, 2024, 339 (1-2) : 397 - 426
  • [34] END-TO-END OPTIMIZED SPEECH CODING WITH DEEP NEURAL NETWORKS
    Kankanahalli, Srihari
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 2521 - 2525
  • [35] AutoQNN: An End-to-End Framework for Automatically Quantizing Neural Networks
    Gong, Cheng
    Lu, Ye
    Dai, Su-Rong
    Deng, Qian
    Du, Cheng-Kun
    Li, Tao
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2024, 39 (02) : 401 - 420
  • [36] End-to-end heterogeneous graph neural networks for traffic assignment
    Liu, Tong
    Meidani, Hadi
    TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2024, 165
  • [37] Sequential neural networks for noetic end-to-end response selection
    Chen, Qian
    Wang, Wen
    COMPUTER SPEECH AND LANGUAGE, 2020, 62
  • [38] An End-to-End Compression Framework Based on Convolutional Neural Networks
    Tao, Wen
    Jiang, Feng
    Zhang, Shengping
    Ren, Jie
    Shi, Wuzhen
    Zuo, Wangmeng
    Guo, Xun
    Zhao, Debin
    2017 DATA COMPRESSION CONFERENCE (DCC), 2017, : 463 - 463
  • [39] End-to-End Training of Deep Neural Networks in the Fourier Domain
    Fulop, Andras
    Horvath, Andras
    MATHEMATICS, 2022, 10 (12)
  • [40] CiMComp: An Energy Efficient Compute-in-Memory based Comparator for Convolutional Neural Networks
    Kavitha, S.
    Kailath, Binsu J.
    Reniwal, B. S.
    2024 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION, DATE, 2024,