MINT: Mixed-precision RRAM-based IN-memory Training Architecture

被引:27
|
作者
Jiang, Hongwu [1 ]
Huang, Shanshi [1 ]
Peng, Xiaochen [1 ]
Yu, Shimeng [1 ]
机构
[1] Georgia Inst Technol, Sch Elect & Comp Engn, Atlanta, GA 30332 USA
来源
2020 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS) | 2020年
关键词
RRAM; compute-in-memory; deep neural network; hardware accelerator;
D O I
10.1109/iscas45731.2020.9181020
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
On-chip training of large-scale deep neural networks (DNNs) is challenging. To solve the memory wall problem, compute-in-memory (CIM) is a promising approach that exploits the analog computation inside the memory array to speed up the vector-matrix multiplication (VMM). Challenges for on-chip CIM training include higher weight precision and higher analog-to-digital converter (ADC) resolution. In this work, we propose a mixed-precision RRAM-based CIM architecture that overcomes these challenges and supports on-chip training. In particular, we split the multi-bit weight into the most significant bits (MSBs) and the least significant bits (LSBs). The forward and backward propagations are performed with CIM transposable arrays for MSBs only, while the weight update is performed in regular memory arrays that store LSBs. Impact of ADC resolution on training accuracy is analyzed. We explore the training performance of a convolutional VGG-like network on the CIFAR-10 dataset using this Mixed-precision IN-memory Training architecture, namely MINT, showing that it can achieve similar to 91% accuracy under hardware constraints and similar to 4.46TOPS/W energy efficiency. Compared with the baseline CIM architectures based on RRAM, it can achieve 1.35x higher energy efficiency and only 31.9% chip size (similar to 98.86 mm(2) at 32 nm node).
引用
收藏
页数:5
相关论文
共 50 条
  • [41] Device Non-Ideality Effects and Architecture-aware Training in RRAM In-Memory Computing Modules
    Wang, Qiwen
    Park, Yongmo
    Lu, Wei D.
    2021 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2021,
  • [42] Neural Network Training Acceleration With RRAM-Based Hybrid Synapses
    Choi, Wooseok
    Kwak, Myonghoon
    Kim, Seyoung
    Hwang, Hyunsang
    FRONTIERS IN NEUROSCIENCE, 2021, 15
  • [43] Compensation Architecture to Alleviate Noise Effects in RRAM-based Computing-in-memory Chips with Residual Resource
    Zhao, Xiaoqing
    Liu, Longjun
    Liu, Yuyi
    Gao, Bin
    Sun, Hongbin
    2024 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, ISCAS 2024, 2024,
  • [44] RRAM-VAC: A Variability-Aware Controller for RRAM-based Memory Architectures
    Tuli, Shikhar
    Rios, Marco
    Levisse, Alexandre
    Atienza, David
    2020 25TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE, ASP-DAC 2020, 2020, : 181 - 186
  • [45] Mixed-Precision Partial Differential Equation Solver Design Based on Nonvolatile Memory
    Yang, Haozhang
    Huang, Peng
    Zhou, Zheng
    Zhang, Yizhou
    Han, Runze
    Liu, Xiaoyan
    Kang, Jinfeng
    IEEE TRANSACTIONS ON ELECTRON DEVICES, 2022, 69 (07) : 3708 - 3715
  • [46] NAS4RRAM: neural network architecture search for inference on RRAM-based accelerators
    Zhihang Yuan
    Jingze Liu
    Xingchen Li
    Longhao Yan
    Haoxiang Chen
    Bingzhe Wu
    Yuchao Yang
    Guangyu Sun
    Science China Information Sciences, 2021, 64
  • [47] Mixed-Precision Continual Learning Based on Computational Resistance Random Access Memory
    Li, Yi
    Zhang, Woyu
    Xu, Xiaoxin
    He, Yifan
    Dong, Danian
    Jiang, Nanjia
    Wang, Fei
    Guo, Zeyu
    Wang, Shaocong
    Dou, Chunmeng
    Liu, Yongpan
    Wang, Zhongrui
    Shang, Dashan
    ADVANCED INTELLIGENT SYSTEMS, 2022, 4 (08)
  • [48] ADC-Less Reprogrammable RRAM Array Architecture for In-Memory Computing
    Dongre, Ashvinikumar
    Boro, Bipul
    Trivedi, Gaurav
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2023, 31 (12) : 2053 - 2060
  • [49] NAS4RRAM: neural network architecture search for inference on RRAM-based accelerators
    Zhihang YUAN
    Jingze LIU
    Xingchen LI
    Longhao YAN
    Haoxiang CHEN
    Bingzhe WU
    Yuchao YANG
    Guangyu SUN
    Science China(Information Sciences), 2021, 64 (06) : 89 - 99
  • [50] NAS4RRAM: neural network architecture search for inference on RRAM-based accelerators
    Yuan, Zhihang
    Liu, Jingze
    Li, Xingchen
    Yan, Longhao
    Chen, Haoxiang
    Wu, Bingzhe
    Yang, Yuchao
    Sun, Guangyu
    SCIENCE CHINA-INFORMATION SCIENCES, 2021, 64 (06)