Partial Sum Quantization for Reducing ADC Size in ReRAM-Based Neural Network Accelerators

被引:0
|
作者
Azamat, Azat [1 ]
Asim, Faaiz [2 ]
Kim, Jintae [3 ]
Lee, Jongeun [2 ]
机构
[1] Ulsan Natl Inst Sci & Technol, Dept Comp Sci & Engn, Ulsan 44919, South Korea
[2] Ulsan Natl Inst Sci & Technol, Dept Elect Engn, Ulsan 44919, South Korea
[3] Konkuk Univ, Dept Elect & Elect Engn, Seoul 143701, South Korea
关键词
Quantization (signal); Hardware; Artificial neural networks; Convolutional neural networks; Training; Throughput; Costs; AC-DC power converters; Memristors; Analog-to-digital conversion (ADC); convolutional neural network (CNN); in-memory computing accelerator; memristor; quantization;
D O I
10.1109/TCAD.2023.3294461
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
While resistive random-access memory (ReRAM) crossbar arrays have the potential to significantly accelerate deep neural network (DNN) training through fast and low-cost matrix-vector multiplication, peripheral circuits like analog-to-digital converters (ADCs) create a high overhead. These ADCs consume over half of the chip power and a considerable portion of the chip cost. To address this challenge, we propose advanced quantization techniques that can significantly reduce the ADC overhead of ReRAM crossbar arrays (RCAs). Our methodology interprets ADC as a quantization mechanism, allowing us to scale the range of ADC input optimally along with the weight parameters of a DNN, resulting in multiple-bit reduction in ADC precision. This approach reduces ADC size and power consumption by several times, and it is applicable to any DNN type (binarized or multibit) and any RCA size. Additionally, we propose ways to minimize the overhead of the digital scaler, which is an essential part of our scheme and sometimes required. Our experimental results using ResNet-18 on the ImageNet dataset demonstrate that our method can reduce the size of the ADC by 32 times compared to ISAAC with only a minimal accuracy loss degradation of 0.24%. We also present evaluation results in the presence of ReRAM nonideality (such as stuck-at fault).
引用
收藏
页码:4897 / 4908
页数:12
相关论文
共 50 条
  • [41] MAX2: An ReRAM-Based Neural Network Accelerator That Maximizes Data Reuse and Area Utilization
    Mao, Manqing
    Peng, Xiaochen
    Liu, Rui
    Li, Jingtao
    Yu, Shimeng
    Chakrabarti, Chaitali
    IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2019, 9 (02) : 398 - 410
  • [42] Re2fresh: A Framework for Mitigating Read Disturbance in ReRAM-based DNN Accelerators
    Shin, Hyein
    Kang, Myeonggu
    Kim, Lee-Sup
    2022 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2022,
  • [43] PRIME: A Novel Processing-in-memory Architecture for Neural Network Computation in ReRAM-based Main Memory
    Chi, Ping
    Li, Shuangchen
    Xu, Cong
    Zhang, Tao
    Zhao, Jishen
    Liu, Yongpan
    Wang, Yu
    Xie, Yuan
    2016 ACM/IEEE 43RD ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA), 2016, : 27 - 39
  • [44] RRAMedy: Protecting ReRAM-based Neural Network from Permanent and Soft Faults During Its Lifetime
    Li, Wen
    Wang, Ying
    Li, Huawei
    Li, Xiaowei
    2019 IEEE 37TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD 2019), 2019, : 91 - 99
  • [45] DL-RSIM: A Reliability and Deployment Strategy Simulation Framework for ReRAM-based CNN Accelerators
    Lin, Wei-Ting
    Cheng, Hsiang-Yun
    Yang, Chia-Lin
    Lin, Meng-Yao
    Lien, Kai
    Hu, Han-Wen
    Chang, Hung-Sheng
    Li, Hsiang-Pang
    Chang, Meng-Fan
    Tsou, Yen-Ting
    Nien, Chin-Fu
    ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2022, 21 (03)
  • [46] Runtime Row/Column Activation Pruning for ReRAM-based Processing-in-Memory DNN Accelerators
    Jiang, Xikun
    Shen, Zhaoyan
    Sun, Siqing
    Yin, Ping
    Jia, Zhiping
    Ju, Lei
    Zhang, Zhiyong
    Yu, Dongxiao
    2023 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2023,
  • [47] DL-RSIM: A Simulation Framework to Enable Reliable ReRAM-based Accelerators for Deep Learning
    Lin, Meng-Yao
    Cheng, Hsiang-Yun
    Lin, Wei-Ting
    Yang, Tzu-Hsien
    Tseng, I-Ching
    Yang, Chia-Lin
    Hu, Han-Wen
    Chang, Hung-Sheng
    Li, Hsiang-Pang
    Chang, Meng-Fan
    2018 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER-AIDED DESIGN (ICCAD) DIGEST OF TECHNICAL PAPERS, 2018,
  • [48] REC: REtime Convolutional Layers to Fully Exploit Harvested Energy for ReRAM-based CNN Accelerators
    Zhou, Kunyu
    Qiu, Keni
    ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2024, 23 (06) : 33 - 33
  • [49] Accelerating Graph Neural Network Training on ReRAM-Based PIM Architectures via Graph and Model Pruning
    Ogbogu, Chukwufumnanya O.
    Arka, Aqeeb Iqbal
    Pfromm, Lukas
    Joardar, Biresh Kumar
    Doppa, Janardhan Rao
    Chakrabarty, Krishnendu
    Pande, Partha Pratim
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (08) : 2703 - 2716
  • [50] A Lazy Engine for High-utilization and Energy-efficient ReRAM-based Neural Network Accelerator
    Yang, Wei-Yi
    Chen, Ya-Shu
    Xiao, Jin-Wen
    2022 IEEE 20TH INTERNATIONAL CONFERENCE ON INDUSTRIAL INFORMATICS (INDIN), 2022, : 140 - 145