Partial Sum Quantization for Reducing ADC Size in ReRAM-Based Neural Network Accelerators

被引:0
|
作者
Azamat, Azat [1 ]
Asim, Faaiz [2 ]
Kim, Jintae [3 ]
Lee, Jongeun [2 ]
机构
[1] Ulsan Natl Inst Sci & Technol, Dept Comp Sci & Engn, Ulsan 44919, South Korea
[2] Ulsan Natl Inst Sci & Technol, Dept Elect Engn, Ulsan 44919, South Korea
[3] Konkuk Univ, Dept Elect & Elect Engn, Seoul 143701, South Korea
关键词
Quantization (signal); Hardware; Artificial neural networks; Convolutional neural networks; Training; Throughput; Costs; AC-DC power converters; Memristors; Analog-to-digital conversion (ADC); convolutional neural network (CNN); in-memory computing accelerator; memristor; quantization;
D O I
10.1109/TCAD.2023.3294461
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
While resistive random-access memory (ReRAM) crossbar arrays have the potential to significantly accelerate deep neural network (DNN) training through fast and low-cost matrix-vector multiplication, peripheral circuits like analog-to-digital converters (ADCs) create a high overhead. These ADCs consume over half of the chip power and a considerable portion of the chip cost. To address this challenge, we propose advanced quantization techniques that can significantly reduce the ADC overhead of ReRAM crossbar arrays (RCAs). Our methodology interprets ADC as a quantization mechanism, allowing us to scale the range of ADC input optimally along with the weight parameters of a DNN, resulting in multiple-bit reduction in ADC precision. This approach reduces ADC size and power consumption by several times, and it is applicable to any DNN type (binarized or multibit) and any RCA size. Additionally, we propose ways to minimize the overhead of the digital scaler, which is an essential part of our scheme and sometimes required. Our experimental results using ResNet-18 on the ImageNet dataset demonstrate that our method can reduce the size of the ADC by 32 times compared to ISAAC with only a minimal accuracy loss degradation of 0.24%. We also present evaluation results in the presence of ReRAM nonideality (such as stuck-at fault).
引用
收藏
页码:4897 / 4908
页数:12
相关论文
共 50 条
  • [21] Mathematical Framework for Optimizing Crossbar Allocation for ReRAM-based CNN Accelerators
    Li, Wanqian
    Han, Yinhe
    Chen, Xiaoming
    ACM TRANSACTIONS ON DESIGN AUTOMATION OF ELECTRONIC SYSTEMS, 2024, 29 (01)
  • [22] Design Framework for ReRAM-Based DNN Accelerators with Accuracy and Hardware Evaluation
    Kao, Hsu-Yu
    Huang, Shih-Hsu
    Cheng, Wei-Kai
    ELECTRONICS, 2022, 11 (13)
  • [23] ReRAM-Based In-Memory Computing for Search Engine and Neural Network Applications
    Halawani, Yasmin
    Mohammad, Baker
    Abu Lebdeh, Muath
    Al-Qutayri, Mahmoud
    Al-Sarawi, Said E.
    IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2019, 9 (02) : 388 - 397
  • [24] ReRAM-Sharing: Fine-Grained Weight Sharing for ReRAM-Based Deep Neural Network Accelerator
    Song, Zhuoran
    Li, Dongyue
    He, Zhezhi
    Liang, Xiaoyao
    Jiang, Li
    2021 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2021,
  • [25] ReNEW: Enhancing Lifetime for ReRAM Crossbar based Neural Network Accelerators
    Wen, Wen
    Zhang, Youtao
    Yang, Jun
    2019 IEEE 37TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD 2019), 2019, : 487 - 496
  • [26] Data Pruning-enabled High Performance and Reliable Graph Neural Network Training on ReRAM-based Processing-in-Memory Accelerators
    Ogbogu, Chukwufumnanya
    Joardar, Biresh
    Chakrabarty, Krishnendu
    Doppa, Jana
    Pande, Partha Pratim
    ACM TRANSACTIONS ON DESIGN AUTOMATION OF ELECTRONIC SYSTEMS, 2024, 29 (05)
  • [27] CNNWire: Boosting Convolutional Neural Network with Winograd on ReRAM based Accelerators
    Lin, Jilan
    Li, Shuangchen
    Hu, Xing
    Deng, Lei
    Xie, Yuan
    GLSVLSI '19 - PROCEEDINGS OF THE 2019 ON GREAT LAKES SYMPOSIUM ON VLSI, 2019, : 283 - 286
  • [28] REC: REtime Convolutional layers in energy harvesting ReRAM-based CNN accelerators
    Zhou, Kunyu
    Qiu, Keni
    PROCEEDINGS OF THE 19TH ACM INTERNATIONAL CONFERENCE ON COMPUTING FRONTIERS 2022 (CF 2022), 2022, : 185 - 188
  • [29] A Versatile ReRAM-based Accelerator for Convolutional Neural Networks
    Mao, Manqing
    Sun, Xiao Yu
    Peng, Xiaochen
    Yu, Shimeng
    Chakrabarti, Chaitali
    PROCEEDINGS OF THE 2018 IEEE INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS), 2018, : 211 - 216
  • [30] A ReRAM-Based Convolutional Neural Network Accelerator Using the Analog Layer Normalization Technique
    Gi, Sang-Gyun
    Lee, Hyunkeun
    Jang, Jingon
    Lee, Byung-Geun
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2023, 70 (06) : 6442 - 6451