Exploring Model Stability of Deep Neural Networks for Reliable RRAM-Based In-Memory Acceleration

被引:5
|
作者
Krishnan, Gokul [1 ]
Yang, Li [1 ]
Sun, Jingbo [1 ]
Hazra, Jubin [2 ]
Du, Xiaocong [1 ]
Liehr, Maximilian [2 ]
Li, Zheng [1 ]
Beckmann, Karsten [2 ]
Joshi, Rajiv, V [3 ]
Cady, Nathaniel C. [2 ]
Fan, Deliang [1 ]
Cao, Yu [1 ]
机构
[1] Arizona State Univ, Sch Elect Comp & Energy Engn, Tempe, AZ 85287 USA
[2] State Univ New York Polytech, Albany, NY 12246 USA
[3] IBM Corp, TJ Watson Res Ctr, Yorktown Hts, NY 10598 USA
关键词
Stability analysis; Computational modeling; Quantization (signal); Semiconductor device modeling; Training; Perturbation methods; Neural networks; In-memory computing; RRAM; model stability; deep neural networks; reliability; pruning; quantization;
D O I
10.1109/TC.2022.3174585
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
RRAM-based in-memory computing (IMC) effectively accelerates deep neural networks (DNNs). Furthermore, model compression techniques, such as quantization and pruning, are necessary to improve algorithm mapping and hardware performance. However, in the presence of RRAM device variations, low-precision and sparse DNNs suffer from severe post-mapping accuracy loss. To address this, in this work, we investigate a new metric, model stability, from the loss landscape to help shed light on accuracy loss under variations and model compression, which guides an algorithmic solution to maximize model stability and mitigate accuracy loss. Based on statistical data from a CMOS/RRAM 1T1R test chip at 65nm, we characterize wafer-level RRAM variations and develop a cross-layer benchmark tool that incorporates quantization, pruning, device variations, model stability, and IMC architecture parameters to assess post-mapping accuracy and hardware performance. Leveraging this tool, we show that a loss-landscape-based DNN model selection for stability effectively tolerates device variations and achieves a post-mapping accuracy higher than that with 50% lower RRAM variations. Moreover, we quantitatively interpret why model pruning increases the sensitivity to variations, while a lower-precision model has better tolerance to variations. Finally, we propose a novel variation-aware training method to improve model stability, in which there exists the most stable model for the best post-mapping accuracy of compressed DNNs. Experimental evaluation of the method shows up to 19%, 21%, and 11% post-mapping accuracy improvement for our 65nm RRAM device, across various precision and sparsity, on CIFAR-10, CIFAR-100, and SVHN datasets, respectively.
引用
收藏
页码:2740 / 2752
页数:13
相关论文
共 50 条
  • [31] Temperature-Dependent Accuracy Analysis and Resistance Temperature Correction in RRAM-Based In-Memory Computing
    Ling, Yaotian
    Wang, Zongwei
    Yu, Zhizhen
    Bao, Shengyu
    Yang, Yuhang
    Bao, Lin
    Sun, Yining
    Cai, Yimao
    Huang, Ru
    IEEE TRANSACTIONS ON ELECTRON DEVICES, 2024, 71 (01) : 294 - 300
  • [32] Fast Logic Synthesis for RRAM-based In-Memory Computing using Majority-Inverter Graphs
    Shirinzadeh, Saeideh
    Soeken, Mathias
    Gaillardon, Pierre-Emmanuel
    Drechsler, Rolf
    PROCEEDINGS OF THE 2016 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE), 2016, : 948 - 953
  • [33] 2-Bit-Per-Cell RRAM-Based In-Memory Computing for Area-/Energy-Efficient Deep Learning
    He, Wangxin
    Yin, Shihui
    Kim, Yulhwa
    Sun, Xiaoyu
    Kim, Jae-Joon
    Yu, Shimeng
    Seo, Jae-Sun
    IEEE SOLID-STATE CIRCUITS LETTERS, 2020, 3 : 194 - 197
  • [34] Robustness of Neuromorphic Computing with RRAM-based Crossbars and Optical Neural Networks
    Zhang, Grace Li
    Li, Bing
    Zhu, Ying
    Wang, Tianchen
    Shi, Yiyu
    Yin, Xunzhao
    Zhuo, Cheng
    Gu, Huaxi
    Ho, Tsung-Yi
    Schlichtmann, Ulf
    2021 26TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2021, : 853 - 858
  • [35] Attar: RRAM-based in-memory attention accelerator with software-hardware co-optimization
    Bing LI
    Ying QI
    Ying WANG
    Yinhe HAN
    Science China(Information Sciences), 2025, 68 (03) : 371 - 387
  • [36] Attar: RRAM-based in-memory attention accelerator with software-hardware co-optimization
    Li, Bing
    Qi, Ying
    Wang, Ying
    Han, Yinhe
    SCIENCE CHINA-INFORMATION SCIENCES, 2025, 68 (03)
  • [37] Experimental Assessment of Multilevel RRAM-Based Vector-Matrix Multiplication Operations for In-Memory Computing
    Quesada, Emilio Perez-Bosch
    Mahadevaiah, Mamathamba Kalishettyhalli
    Rizzi, Tommaso
    Wen, Jianan
    Ulbricht, Markus
    Krstic, Milos
    Wenger, Christian
    Perez, Eduardo
    IEEE TRANSACTIONS ON ELECTRON DEVICES, 2023, 70 (04) : 2009 - 2014
  • [38] In-Memory Computing Based Hardware Accelerator Module for Deep Neural Networks
    Appukuttan, Allen
    Thomas, Emmanuel
    Nair, Harinandan R.
    Hemanth, S.
    Dhanaraj, K. J.
    Azeez, Maleeha Abdul
    2022 IEEE 19TH INDIA COUNCIL INTERNATIONAL CONFERENCE, INDICON, 2022,
  • [39] RRAM-Based Binary Neural Networks Using Back-Propagation Learning
    Jiang, Yuning
    Yang, Yunfan
    Zhou, Zheng
    Xiang, Yachen
    Huang, Peng
    Kang, Jinfeng
    2018 14TH IEEE INTERNATIONAL CONFERENCE ON SOLID-STATE AND INTEGRATED CIRCUIT TECHNOLOGY (ICSICT), 2018, : 1252 - 1254
  • [40] Challenges and Opportunities toward Online Training Acceleration using RRAM-based Hardware Neural Network
    Chang, Chih-Cheng
    Liu, Jen-Chieh
    Shen, Yu-Lin
    Chou, Teyuh
    Chen, Pin-Chun
    Wang, I-Ting
    Su, Chih-Chun
    Wu, Ming-Hong
    Hudec, Boris
    Chang, Che-Chia
    Tsai, Chia-Ming
    Chang, Tian-Sheuan
    Wong, H-S Philip
    Hou, Tuo-Hung
    2017 IEEE INTERNATIONAL ELECTRON DEVICES MEETING (IEDM), 2017,