Saving Energy of RRAM-Based Neural Accelerator Through State-Aware Computing

被引:1
|
作者
He, Yintao [1 ,2 ]
Wang, Ying [1 ,2 ]
Li, Huawei [1 ,2 ,3 ]
Li, Xiaowei [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Comp Technol, State Key Lab Comp Architecture, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Beijing 100190, Peoples R China
[3] Peng Cheng Lab, Shenzhen 518066, Peoples R China
基金
中国国家自然科学基金;
关键词
Computer architecture; Microprocessors; Resistance; Power demand; Training; Biological neural networks; Optimization; Low power (LP); neural networks; processing-in-memory; resistive random-access memory (RRAM);
D O I
10.1109/TCAD.2021.3103147
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In-memory computing (IMC) is recognized as one of the most promising architecture solution to realize energy-efficient neural network inference. Amongst many memory technology, resistive RAM (RRAM) is a very attractive device to implement the IMC-based neural network accelerator architecture, which is particularly suitable for power-constrained IoT systems. Due to the nature of low leakage and in-situ computing, the dynamic power consumption of dot-production operations in RRAM crossbars dominates the chip power, especially when applied to low-precision neural networks. This work investigates the correlation between the cell resistance state and the crossbar operation power, and proposes a state-aware RRAM accelerator (SARA) architecture for energy-efficient low-precision neural networks. With the proposed state-aware network training and mapping strategy, crossbars in the RRAM accelerator can perform in a lower power state. Furthermore, we also leverage the proposed RRAM accelerator architecture to reduce the power consumption of high-precision network inference with both single-level or multilevel RRAM. The evaluation results show that for binary neural networks, our design saves 40.53% RRAM computing energy on average over the baseline. For high precision neural networks, the proposed method reduces 11.67% computing energy on average without any accuracy loss.
引用
收藏
页码:2115 / 2127
页数:13
相关论文
共 50 条
  • [1] SubMac: Exploiting the subword-based computation in RRAM-based CNN accelerator for energy saving and speedup
    Chen, Xizi
    Jiang, Jingbo
    Zhu, Jingyang
    Tsui, Chi-Ying
    INTEGRATION-THE VLSI JOURNAL, 2019, 69 : 356 - 368
  • [2] A RRAM-based FPGA for Energy-efficient Edge Computing
    Tang, Xifan
    Giacomin, Edouard
    Cadareanu, Patsy
    Gore, Ganesh
    Gaillardon, Pierre-Emmanuel
    PROCEEDINGS OF THE 2020 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2020), 2020,
  • [3] RRAM-Based In-Memory Computing for Embedded Deep Neural Networks
    Bankman, D.
    Messner, J.
    Gural, A.
    Murmann, B.
    CONFERENCE RECORD OF THE 2019 FIFTY-THIRD ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, 2019, : 1511 - 1515
  • [4] CompRRAE: RRAM-based Convolutional Neural Network Accelerator with Reduced Computations through a Runtime Activation Estimation
    Chen, Xizi
    Zhu, Jingyang
    Jiang, Jingbo
    Tsui, Chi-Ying
    24TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC 2019), 2019, : 133 - 139
  • [5] RRAM-Based Analog Approximate Computing
    Li, Boxun
    Gu, Peng
    Shan, Yi
    Wang, Yu
    Chen, Yiran
    Yang, Huazhong
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2015, 34 (12) : 1905 - 1917
  • [6] Robustness of Neuromorphic Computing with RRAM-based Crossbars and Optical Neural Networks
    Zhang, Grace Li
    Li, Bing
    Zhu, Ying
    Wang, Tianchen
    Shi, Yiyu
    Yin, Xunzhao
    Zhuo, Cheng
    Gu, Huaxi
    Ho, Tsung-Yi
    Schlichtmann, Ulf
    2021 26TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2021, : 853 - 858
  • [7] An RRAM-Based Oscillatory Neural Network
    Jackson, Thomas C.
    Sharma, Abhishek A.
    Bain, James A.
    Weldon, Jeffrey A.
    Pileggi, Lawrence
    2015 IEEE 6TH LATIN AMERICAN SYMPOSIUM ON CIRCUITS & SYSTEMS (LASCAS), 2015,
  • [8] On the Reliability of RRAM-Based Neural Networks
    Aziza, Hassen
    Zambelli, Cristian
    Hamdioui, Said
    Diware, Sumit
    Bishnoi, Rajendra
    Gebregiorgis, Anteneh
    2023 IFIP/IEEE 31ST INTERNATIONAL CONFERENCE ON VERY LARGE SCALE INTEGRATION, VLSI-SOC, 2023, : 13 - 20
  • [9] R-Accelerator: An RRAM-Based CGRA Accelerator With Logic Contraction
    Chen, Zhengyu
    Zhou, Hai
    Gu, Jie
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2019, 27 (11) : 2655 - 2667
  • [10] RRAM-based Analog In-Memory Computing
    Chen, Xiaoming
    Song, Tao
    Han, Yinhe
    2021 IEEE/ACM INTERNATIONAL SYMPOSIUM ON NANOSCALE ARCHITECTURES (NANOARCH), 2021,