Analog Weights in ReRAM DNN Accelerators

被引:0
|
作者
Eshraghian, Jason K. [1 ]
Kang, Sung-Mo [2 ]
Baek, Seungbum [3 ]
Orchard, Garrick [4 ,5 ]
Iu, Herbert Ho-Ching [1 ]
Lei, Wen [1 ]
机构
[1] Univ Western Australia, Sch Elect Elect & Comp Engn, Crawley, WA 6009, Australia
[2] Univ Calif Santa Cruz, Baskin Sch Engn, Santa Cruz, CA 95064 USA
[3] Chungbuk Natl Univ, Coll Elect & Comp Engn, Cheongju 362763, South Korea
[4] Natl Univ Singapore, Temasek Labs, Singapore 117411, Singapore
[5] Natl Univ Singapore, Singapore Inst Neurotechnol, Singapore 117411, Singapore
关键词
accelerator; analog; memristor; neural network; ReRAM;
D O I
10.1109/aicas.2019.8771550
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Artificial neural networks have become ubiquitous in modern life, which has triggered the emergence of a new class of application specific integrated circuits for their acceleration. ReRAM-based accelerators have gained significant traction due to their ability to leverage in-memory computations. In a crossbar structure, they can perform multiply-and-accumulate operations more efficiently than standard CMOS logic. By virtue of being resistive switches, ReRAM switches can only reliably store one of two states. This is a severe limitation on the range of values in a computational kernel. This paper presents a novel scheme in alleviating the single-bit-per-device restriction by exploiting frequency dependence of v-i plane hysteresis, and assigning kernel information not only to the device conductance but also partially distributing it to the frequency of a time-varying input. We show this approach reduces average power consumption for a single crossbar convolution by up to a factor of x16 for an unsigned 8-bit input image, where each convolutional process consumes a worst-case of 1.1mW, and reduces area by a factor of x8, without reducing accuracy to the level of binarized neural networks. This presents a massive saving in computing cost when there are many simultaneous in-situ multiply-and-accumulate processes occurring across different crossbars.
引用
收藏
页码:267 / 271
页数:5
相关论文
共 50 条
  • [21] Control Variate Approximation for DNN Accelerators
    Zervakis, Georgios
    Spantidi, Ourania
    Anagnostopoulos, Iraklis
    Amrouch, Hussam
    Henkel, Joerg
    2021 58TH ACM/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2021, : 481 - 486
  • [22] Hardware attacks on ReRAM-based AI accelerators
    Heidary, Masoud
    Joardar, Biresh Kumar
    17TH IEEE DALLAS CIRCUITS AND SYSTEMS CONFERENCE, DCAS 2024, 2024,
  • [23] Enabling High-Performance DNN Inference Accelerators using Non-Volatile Analog Memory (Invited)
    Chen, An
    Ambrogio, Stefano
    Narayanan, Pritish
    Tsai, Hsinyu
    Mackin, Charles
    2020 IEEE ELECTRON DEVICES TECHNOLOGY AND MANUFACTURING CONFERENCE (EDTM 2020), 2020,
  • [24] Design of Reliable DNN Accelerator with Un-reliable ReRAM
    Long, Yun
    She, Xueyuan
    Mukhopadhyay, Saibal
    2019 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE), 2019, : 1769 - 1774
  • [25] Soft errors in DNN accelerators: A comprehensive review
    Ibrahim, Younis
    Wang, Haibin
    Liu, Junyang
    Wei, Jinghe
    Chen, Li
    Rech, Paolo
    Adam, Khalid
    Guo, Gang
    MICROELECTRONICS RELIABILITY, 2020, 115 (115)
  • [26] Targeting DNN Inference Via Efficient Utilization of Heterogeneous Precision DNN Accelerators
    Spantidi, Ourania
    Zervakis, Georgios
    Alsalamin, Sami
    Roman-Ballesteros, Isai
    Henkel, Joerg
    Amrouch, Hussam
    Anagnostopoulos, Iraklis
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTING, 2023, 11 (01) : 112 - 125
  • [27] Boosting ReRAM-based DNN by Row Activation Oversubscription
    Guo, Mengyu
    Zhang, Zihan
    Jiang, Jianfei
    Wang, Qin
    Jing, Naifeng
    27TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE, ASP-DAC 2022, 2022, : 604 - 609
  • [28] Increasing Throughput of In-Memory DNN Accelerators by Flexible Layerwise DNN Approximation
    De la Parra, Cecilia
    Soliman, Taha
    Guntoro, Andre
    Kumar, Akash
    Wehn, Norbert
    IEEE MICRO, 2022, 42 (06) : 17 - 24
  • [29] Energy Efficient Computing with Heterogeneous DNN Accelerators
    Hossain, Md Shazzad
    Savidis, Ioannis
    2021 IEEE 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS), 2021,
  • [30] Flexion: A Quantitative Metric for Flexibility in DNN Accelerators
    Kwon, Hyoukjun
    Pellauer, Michael
    Parashar, Angshuman
    Krishna, Tushar
    IEEE COMPUTER ARCHITECTURE LETTERS, 2021, 20 (01) : 1 - 4