Partial Sum Quantization for Reducing ADC Size in ReRAM-Based Neural Network Accelerators

被引:0
|
作者
Azamat, Azat [1 ]
Asim, Faaiz [2 ]
Kim, Jintae [3 ]
Lee, Jongeun [2 ]
机构
[1] Ulsan Natl Inst Sci & Technol, Dept Comp Sci & Engn, Ulsan 44919, South Korea
[2] Ulsan Natl Inst Sci & Technol, Dept Elect Engn, Ulsan 44919, South Korea
[3] Konkuk Univ, Dept Elect & Elect Engn, Seoul 143701, South Korea
关键词
Quantization (signal); Hardware; Artificial neural networks; Convolutional neural networks; Training; Throughput; Costs; AC-DC power converters; Memristors; Analog-to-digital conversion (ADC); convolutional neural network (CNN); in-memory computing accelerator; memristor; quantization;
D O I
10.1109/TCAD.2023.3294461
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
While resistive random-access memory (ReRAM) crossbar arrays have the potential to significantly accelerate deep neural network (DNN) training through fast and low-cost matrix-vector multiplication, peripheral circuits like analog-to-digital converters (ADCs) create a high overhead. These ADCs consume over half of the chip power and a considerable portion of the chip cost. To address this challenge, we propose advanced quantization techniques that can significantly reduce the ADC overhead of ReRAM crossbar arrays (RCAs). Our methodology interprets ADC as a quantization mechanism, allowing us to scale the range of ADC input optimally along with the weight parameters of a DNN, resulting in multiple-bit reduction in ADC precision. This approach reduces ADC size and power consumption by several times, and it is applicable to any DNN type (binarized or multibit) and any RCA size. Additionally, we propose ways to minimize the overhead of the digital scaler, which is an essential part of our scheme and sometimes required. Our experimental results using ResNet-18 on the ImageNet dataset demonstrate that our method can reduce the size of the ADC by 32 times compared to ISAAC with only a minimal accuracy loss degradation of 0.24%. We also present evaluation results in the presence of ReRAM nonideality (such as stuck-at fault).
引用
收藏
页码:4897 / 4908
页数:12
相关论文
共 50 条
  • [31] Learning to Predict IR Drop with Effective Training for ReRAM-based Neural Network Hardware
    Lee, Sugil
    Jung, Giju
    Fouda, Mohammed E.
    Lee, Jongeun
    Eltawil, Ahmed
    Kurdahi, Fadi
    PROCEEDINGS OF THE 2020 57TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2020,
  • [32] An Energy-Efficient Inference Engine for a Configurable ReRAM-Based Neural Network Accelerator
    Zheng, Yang-Lin
    Yang, Wei-Yi
    Chen, Ya-Shu
    Han, Ding-Hung
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (03) : 740 - 753
  • [33] FARe: Fault-Aware GNN Training on ReRAM-based PIM Accelerators
    Dhingra, Pratyush
    Ogbogu, Chukwufumnanya
    Joardar, Biresh Kumar
    Doppa, Janardhan Rao
    Kalyanaraman, Ananth
    Pande, Partha Pratim
    2024 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION, DATE, 2024,
  • [34] A Unified Framework for Training, Mapping and Simulation of ReRAM-Based Convolutional Neural Network Acceleration
    Liu, He
    Han, Jianhui
    Zhang, Youhui
    IEEE COMPUTER ARCHITECTURE LETTERS, 2019, 18 (01) : 63 - 66
  • [35] A Thermal-aware Optimization Framework for ReRAM-based Deep Neural Network Acceleration
    Shin, Hyein
    Kang, Myeonggu
    Kim, Lee-Sup
    2020 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED-DESIGN (ICCAD), 2020,
  • [36] ADC-Free ReRAM-Based In-Situ Accelerator for Energy-Efficient Binary Neural Networks
    Kim, Hyeonuk
    Jung, Youngbeom
    Kim, Lee-Sup
    IEEE TRANSACTIONS ON COMPUTERS, 2024, 73 (02) : 353 - 365
  • [37] On-Line Fault Protection for ReRAM-Based Neural Networks
    Li, Wen
    Wang, Ying
    Liu, Cheng
    He, Yintao
    Liu, Lian
    Li, Huawei
    Li, Xiaowei
    IEEE TRANSACTIONS ON COMPUTERS, 2023, 72 (02) : 423 - 437
  • [38] PHANES: ReRAM-based Photonic Accelerator for Deep Neural Networks
    Liu, Yinyi
    Liu, Jiaqi
    Fu, Yuxiang
    Chen, Shixi
    Zhang, Jiaxu
    Xu, Jiang
    PROCEEDINGS OF THE 59TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC 2022, 2022, : 103 - 108
  • [39] On Minimizing Analog Variation Errors to Resolve the Scalability Issue of ReRAM-Based Crossbar Accelerators
    Kang, Yao-Wen
    Wu, Chun-Feng
    Chang, Yuan-Hao
    Kuo, Tei-Wei
    Ho, Shu-Yin
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2020, 39 (11) : 3856 - 3867
  • [40] An Empirical Fault Vulnerability Exploration of ReRAM-Based Process-in-Memory CNN Accelerators
    Dorostkar, Aniseh
    Farbeh, Hamed
    Zarandi, Hamid R.
    IEEE TRANSACTIONS ON RELIABILITY, 2024, : 1 - 15