Exploring Model Stability of Deep Neural Networks for Reliable RRAM-Based In-Memory Acceleration

被引:5
|
作者
Krishnan, Gokul [1 ]
Yang, Li [1 ]
Sun, Jingbo [1 ]
Hazra, Jubin [2 ]
Du, Xiaocong [1 ]
Liehr, Maximilian [2 ]
Li, Zheng [1 ]
Beckmann, Karsten [2 ]
Joshi, Rajiv, V [3 ]
Cady, Nathaniel C. [2 ]
Fan, Deliang [1 ]
Cao, Yu [1 ]
机构
[1] Arizona State Univ, Sch Elect Comp & Energy Engn, Tempe, AZ 85287 USA
[2] State Univ New York Polytech, Albany, NY 12246 USA
[3] IBM Corp, TJ Watson Res Ctr, Yorktown Hts, NY 10598 USA
关键词
Stability analysis; Computational modeling; Quantization (signal); Semiconductor device modeling; Training; Perturbation methods; Neural networks; In-memory computing; RRAM; model stability; deep neural networks; reliability; pruning; quantization;
D O I
10.1109/TC.2022.3174585
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
RRAM-based in-memory computing (IMC) effectively accelerates deep neural networks (DNNs). Furthermore, model compression techniques, such as quantization and pruning, are necessary to improve algorithm mapping and hardware performance. However, in the presence of RRAM device variations, low-precision and sparse DNNs suffer from severe post-mapping accuracy loss. To address this, in this work, we investigate a new metric, model stability, from the loss landscape to help shed light on accuracy loss under variations and model compression, which guides an algorithmic solution to maximize model stability and mitigate accuracy loss. Based on statistical data from a CMOS/RRAM 1T1R test chip at 65nm, we characterize wafer-level RRAM variations and develop a cross-layer benchmark tool that incorporates quantization, pruning, device variations, model stability, and IMC architecture parameters to assess post-mapping accuracy and hardware performance. Leveraging this tool, we show that a loss-landscape-based DNN model selection for stability effectively tolerates device variations and achieves a post-mapping accuracy higher than that with 50% lower RRAM variations. Moreover, we quantitatively interpret why model pruning increases the sensitivity to variations, while a lower-precision model has better tolerance to variations. Finally, we propose a novel variation-aware training method to improve model stability, in which there exists the most stable model for the best post-mapping accuracy of compressed DNNs. Experimental evaluation of the method shows up to 19%, 21%, and 11% post-mapping accuracy improvement for our 65nm RRAM device, across various precision and sparsity, on CIFAR-10, CIFAR-100, and SVHN datasets, respectively.
引用
收藏
页码:2740 / 2752
页数:13
相关论文
共 50 条
  • [21] Impact of On-chip Interconnect on In-memory Acceleration of Deep Neural Networks
    Krishnan, Gokul
    Mandal, Sumit K.
    Chakrabarti, Chaitali
    Seo, Jae-Sun
    Ogras, Umit Y.
    Cao, Yu
    ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2022, 18 (02)
  • [22] In-Memory and Error-Immune Differential RRAM Implementation of Binarized Deep Neural Networks
    Bocquet, M.
    Hirztlin, T.
    Klein, J. -O.
    Nowak, E.
    Vianello, E.
    Portal, J. -M.
    Querlioz, D.
    2018 IEEE INTERNATIONAL ELECTRON DEVICES MEETING (IEDM), 2018,
  • [23] Fully Binarized, Parallel, RRAM-Based Computing Primitive for In-Memory Similarity Search
    Kingra, Sandeep Kaur
    Parmar, Vivek
    Verma, Deepak
    Bricalli, Alessandro
    Piccolboni, Giuseppe
    Molas, Gabriel
    Regev, Amir
    Suri, Manan
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2023, 70 (01) : 46 - 50
  • [24] Parasitic Resistance Effect Analysis in RRAM-based TCAM for Memory Augmented Neural Networks
    Liao, Yan
    Gao, Bin
    Zhang, Wenqiang
    Yao, Peng
    Li, Xinyi
    Tang, Jianshi
    Li, Zhen
    Cui, Shuguang
    Wu, Huaqiang
    Qian, He
    2020 IEEE INTERNATIONAL MEMORY WORKSHOP (IMW 2020), 2020, : 127 - 130
  • [25] Improving the accuracy and robustness of RRAM-based in-memory computing against RRAM hardware noise and adversarial attacks
    Cherupally, Sai Kiran
    Meng, Jian
    Rakin, Adnan Siraj
    Yin, Shihui
    Yeo, Injune
    Yu, Shimeng
    Fan, Deliang
    Seo, Jae-Sun
    SEMICONDUCTOR SCIENCE AND TECHNOLOGY, 2022, 37 (03)
  • [26] Bidirectional Analog Conductance Modulation for RRAM-Based Neural Networks
    Jiang, Zizhen
    Wang, Ziwen
    Zheng, Xin
    Fong, Scott W.
    Qin, Shengjun
    Chen, Hong-Yu
    Ahn, Ethan C.
    Cao, Ji
    Nishi, Yoshio
    Wong, S. Simon
    Wong, H-S. Philip
    IEEE TRANSACTIONS ON ELECTRON DEVICES, 2020, 67 (11) : 4904 - 4910
  • [27] Energy-Efficient SNN Implementation Using RRAM-Based Computation In-Memory (CIM)
    El Arrassi, Asmae
    Gebregiorgis, Anteneh
    El Haddadi, Anass
    Hamdioui, Said
    PROCEEDINGS OF THE 2022 IFIP/IEEE 30TH INTERNATIONAL CONFERENCE ON VERY LARGE SCALE INTEGRATION (VLSI-SOC), 2022,
  • [28] An MRAM-based Deep In-Memory Architecture for Deep Neural Networks
    Patil, Ameya D.
    Hua, Haocheng
    Gonugondla, Sujan
    Kang, Mingu
    Shanbhag, Naresh R.
    2019 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2019,
  • [29] Single RRAM Cell-based In-Memory Accelerator Architecture for Binary Neural Networks
    Oh, Hyunmyung
    Kim, Hyungjun
    Kang, Nameun
    Kim, Yulhwa
    Park, Jihoon
    Kim, Jae-Joon
    2021 IEEE 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS), 2021,
  • [30] Deep learning acceleration based on in-memory computing
    Eleftheriou, E.
    Le Gallo, M.
    Nandakumar, S. R.
    Piveteau, C.
    Boybat, I
    Joshi, V
    Khaddam-Aljameh, R.
    Dazzi, M.
    Giannopoulos, I
    Karunaratne, G.
    Kersting, B.
    Stanisavljevic, M.
    Jonnalagadda, V. P.
    Ioannou, N.
    Kourtis, K.
    Francese, P. A.
    Sebastian, A.
    IBM JOURNAL OF RESEARCH AND DEVELOPMENT, 2019, 63 (06)