Exploring Model Stability of Deep Neural Networks for Reliable RRAM-Based In-Memory Acceleration

被引:5
|
作者
Krishnan, Gokul [1 ]
Yang, Li [1 ]
Sun, Jingbo [1 ]
Hazra, Jubin [2 ]
Du, Xiaocong [1 ]
Liehr, Maximilian [2 ]
Li, Zheng [1 ]
Beckmann, Karsten [2 ]
Joshi, Rajiv, V [3 ]
Cady, Nathaniel C. [2 ]
Fan, Deliang [1 ]
Cao, Yu [1 ]
机构
[1] Arizona State Univ, Sch Elect Comp & Energy Engn, Tempe, AZ 85287 USA
[2] State Univ New York Polytech, Albany, NY 12246 USA
[3] IBM Corp, TJ Watson Res Ctr, Yorktown Hts, NY 10598 USA
关键词
Stability analysis; Computational modeling; Quantization (signal); Semiconductor device modeling; Training; Perturbation methods; Neural networks; In-memory computing; RRAM; model stability; deep neural networks; reliability; pruning; quantization;
D O I
10.1109/TC.2022.3174585
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
RRAM-based in-memory computing (IMC) effectively accelerates deep neural networks (DNNs). Furthermore, model compression techniques, such as quantization and pruning, are necessary to improve algorithm mapping and hardware performance. However, in the presence of RRAM device variations, low-precision and sparse DNNs suffer from severe post-mapping accuracy loss. To address this, in this work, we investigate a new metric, model stability, from the loss landscape to help shed light on accuracy loss under variations and model compression, which guides an algorithmic solution to maximize model stability and mitigate accuracy loss. Based on statistical data from a CMOS/RRAM 1T1R test chip at 65nm, we characterize wafer-level RRAM variations and develop a cross-layer benchmark tool that incorporates quantization, pruning, device variations, model stability, and IMC architecture parameters to assess post-mapping accuracy and hardware performance. Leveraging this tool, we show that a loss-landscape-based DNN model selection for stability effectively tolerates device variations and achieves a post-mapping accuracy higher than that with 50% lower RRAM variations. Moreover, we quantitatively interpret why model pruning increases the sensitivity to variations, while a lower-precision model has better tolerance to variations. Finally, we propose a novel variation-aware training method to improve model stability, in which there exists the most stable model for the best post-mapping accuracy of compressed DNNs. Experimental evaluation of the method shows up to 19%, 21%, and 11% post-mapping accuracy improvement for our 65nm RRAM device, across various precision and sparsity, on CIFAR-10, CIFAR-100, and SVHN datasets, respectively.
引用
收藏
页码:2740 / 2752
页数:13
相关论文
共 50 条
  • [1] RRAM-Based In-Memory Computing for Embedded Deep Neural Networks
    Bankman, D.
    Messner, J.
    Gural, A.
    Murmann, B.
    CONFERENCE RECORD OF THE 2019 FIFTY-THIRD ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, 2019, : 1511 - 1515
  • [2] Robust RRAM-based In-Memory Computing in Light of Model Stability
    Krishnan, Gokul
    Sun, Jingbo
    Hazra, Jubin
    Du, Xiaocong
    Liehr, Maximilian
    Li, Zheng
    Beckmann, Karsten
    Joshi, Rajiv, V
    Cady, Nathaniel C.
    Cao, Yu
    2021 IEEE INTERNATIONAL RELIABILITY PHYSICS SYMPOSIUM (IRPS), 2021,
  • [3] Error Correction Scheme for Reliable RRAM-Based In-Memory Computing
    Hu, Yixuan
    Cheng, Kaili
    Zhang, Zuodong
    Wang, Runsheng
    Wang, Yuan
    Huang, Ru
    2021 5TH IEEE ELECTRON DEVICES TECHNOLOGY & MANUFACTURING CONFERENCE (EDTM), 2021,
  • [4] PRIVE: Efficient RRAM Programming with Chip Verification for RRAM-based In-Memory Computing Acceleration
    He, Wangxin
    Meng, Jian
    Gonugondla, Sujan Kumar
    Yu, Shimeng
    Shanbhag, Naresh R.
    Seo, Jae-sun
    2023 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION, DATE, 2023,
  • [5] RRAM-based Analog In-Memory Computing
    Chen, Xiaoming
    Song, Tao
    Han, Yinhe
    2021 IEEE/ACM INTERNATIONAL SYMPOSIUM ON NANOSCALE ARCHITECTURES (NANOARCH), 2021,
  • [6] Structured Pruning of RRAM Crossbars for Efficient In-Memory Computing Acceleration of Deep Neural Networks
    Meng, Jian
    Yang, Li
    Peng, Xiaochen
    Yu, Shimeng
    Fan, Deliang
    Seo, Jae-Sun
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2021, 68 (05) : 1576 - 1580
  • [7] TIME: A Training-in-Memory Architecture for RRAM-Based Deep Neural Networks
    Cheng, Ming
    Xia, Lixue
    Zhu, Zhenhua
    Cai, Yi
    Xie, Yuan
    Wang, Yu
    Yang, Huazhong
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2019, 38 (05) : 834 - 847
  • [8] Logic Synthesis for RRAM-Based In-Memory Computing
    Shirinzadeh, Saeideh
    Soeken, Mathias
    Gaillardon, Pierre-Emmanuel
    Drechsler, Rolf
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2018, 37 (07) : 1422 - 1435
  • [9] A Flexible and Reliable RRAM-Based In-Memory Computing Architecture for Data-Intensive Applications
    Eslami, Nima
    Moaiyeri, Mohammad Hossein
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTING, 2023, 11 (03) : 736 - 748
  • [10] On the Reliability of RRAM-Based Neural Networks
    Aziza, Hassen
    Zambelli, Cristian
    Hamdioui, Said
    Diware, Sumit
    Bishnoi, Rajendra
    Gebregiorgis, Anteneh
    2023 IFIP/IEEE 31ST INTERNATIONAL CONFERENCE ON VERY LARGE SCALE INTEGRATION, VLSI-SOC, 2023, : 13 - 20