Analog Weights in ReRAM DNN Accelerators

被引:0
|
作者
Eshraghian, Jason K. [1 ]
Kang, Sung-Mo [2 ]
Baek, Seungbum [3 ]
Orchard, Garrick [4 ,5 ]
Iu, Herbert Ho-Ching [1 ]
Lei, Wen [1 ]
机构
[1] Univ Western Australia, Sch Elect Elect & Comp Engn, Crawley, WA 6009, Australia
[2] Univ Calif Santa Cruz, Baskin Sch Engn, Santa Cruz, CA 95064 USA
[3] Chungbuk Natl Univ, Coll Elect & Comp Engn, Cheongju 362763, South Korea
[4] Natl Univ Singapore, Temasek Labs, Singapore 117411, Singapore
[5] Natl Univ Singapore, Singapore Inst Neurotechnol, Singapore 117411, Singapore
关键词
accelerator; analog; memristor; neural network; ReRAM;
D O I
10.1109/aicas.2019.8771550
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Artificial neural networks have become ubiquitous in modern life, which has triggered the emergence of a new class of application specific integrated circuits for their acceleration. ReRAM-based accelerators have gained significant traction due to their ability to leverage in-memory computations. In a crossbar structure, they can perform multiply-and-accumulate operations more efficiently than standard CMOS logic. By virtue of being resistive switches, ReRAM switches can only reliably store one of two states. This is a severe limitation on the range of values in a computational kernel. This paper presents a novel scheme in alleviating the single-bit-per-device restriction by exploiting frequency dependence of v-i plane hysteresis, and assigning kernel information not only to the device conductance but also partially distributing it to the frequency of a time-varying input. We show this approach reduces average power consumption for a single crossbar convolution by up to a factor of x16 for an unsigned 8-bit input image, where each convolutional process consumes a worst-case of 1.1mW, and reduces area by a factor of x8, without reducing accuracy to the level of binarized neural networks. This presents a massive saving in computing cost when there are many simultaneous in-situ multiply-and-accumulate processes occurring across different crossbars.
引用
收藏
页码:267 / 271
页数:5
相关论文
共 50 条
  • [1] Mixed Precision Quantization for ReRAM-based DNN Inference Accelerators
    Huang, Sitao
    Ankit, Aayush
    Silveira, Plinio
    Antunes, Rodrigo
    Chalamalasetti, Sai Rahul
    El Hajj, Izzat
    Kim, Dong Eun
    Aguiar, Glaucimar
    Bruel, Pedro
    Serebryakov, Sergey
    Xu, Cong
    Li, Can
    Faraboschi, Paolo
    Strachan, John Paul
    Chen, Deming
    Roy, Kaushik
    Hwu, Wen-mei
    Milojicic, Dejan
    2021 26TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2021, : 372 - 377
  • [2] APQ: Automated DNN Pruning and Quantization for ReRAM-Based Accelerators
    Yang, Siling
    He, Shuibing
    Duan, Hexiao
    Chen, Weijian
    Zhang, Xuechen
    Wu, Tong
    Yin, Yanlong
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2023, 34 (09) : 2498 - 2511
  • [3] Effective Zero Compression on ReRAM-based Sparse DNN Accelerators
    Shin, Hoon
    Park, Rihae
    Lee, Seung Yul
    Park, Yeonhong
    Lee, Hyunseung
    Lee, Jae W.
    PROCEEDINGS OF THE 59TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC 2022, 2022, : 949 - 954
  • [4] Design Framework for ReRAM-Based DNN Accelerators with Accuracy and Hardware Evaluation
    Kao, Hsu-Yu
    Huang, Shih-Hsu
    Cheng, Wei-Kai
    ELECTRONICS, 2022, 11 (13)
  • [5] ReHarvest: An ADC Resource-Harvesting Crossbar Architecture for ReRAM-Based DNN Accelerators
    Xu, Jiahong
    Li, Haikun
    Duan, Zhuohui
    Liao, Xiaofei
    Jin, Hai
    Yang, Xiaokang
    Li, Huize
    Liu, Cong
    Mao, Fubing
    Zhang, Yu
    ACM TRANSACTIONS ON ARCHITECTURE AND CODE OPTIMIZATION, 2024, 21 (03)
  • [6] Weight Programming in DNN Analog Hardware Accelerators in the Presence of NVM Variability
    Mackin, Charles
    Tsai, Hsinyu
    Ambrogio, Stefano
    Narayanan, Pritish
    Chen, An
    Burr, Geoffrey W.
    ADVANCED ELECTRONIC MATERIALS, 2019, 5 (09):
  • [7] AutoWS: Automate Weights Streaming in Layer-wise Pipelined DNN Accelerators
    Yu, Zhewen
    Bouganis, Christos-Savvas
    2024 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION, DATE, 2024,
  • [8] Re2fresh: A Framework for Mitigating Read Disturbance in ReRAM-based DNN Accelerators
    Shin, Hyein
    Kang, Myeonggu
    Kim, Lee-Sup
    2022 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2022,
  • [9] Runtime Row/Column Activation Pruning for ReRAM-based Processing-in-Memory DNN Accelerators
    Jiang, Xikun
    Shen, Zhaoyan
    Sun, Siqing
    Yin, Ping
    Jia, Zhiping
    Ju, Lei
    Zhang, Zhiyong
    Yu, Dongxiao
    2023 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2023,
  • [10] CRPIM: An efficient compute-reuse scheme for ReRAM-based Processing-in-Memory DNN accelerators
    Hong, Shihao
    Chung, Yeh-Ching
    JOURNAL OF SYSTEMS ARCHITECTURE, 2024, 153