On-Device Continual Learning With STT-Assisted-SOT MRAM-Based In-Memory Computing

被引:1
|
作者
Zhang, Fan [1 ]
Sridharan, Amitesh [1 ]
Hwang, William [2 ]
Xue, Fen [2 ]
Tsai, Wilman [3 ]
Wang, Shan Xiang [4 ,5 ]
Fan, Deliang [1 ]
机构
[1] Johns Hopkins Univ, Dept Elect & Comp Engn, Baltimore, MD 21218 USA
[2] Stanford Univ, Dept Elect Engn, Stanford, CA 94305 USA
[3] Stanford Univ, Dept Mat Sci & Engn, Stanford, CA 94305 USA
[4] Stanford Univ, Dept Elect Engn, Stanford, CA 94305 USA
[5] Stanford Univ, Dept Mat Sci & Engn, Stanford, CA 94305 USA
基金
美国国家科学基金会;
关键词
Magnetic tunneling; Training; In-memory computing; Task analysis; Quantization (signal); Nonvolatile memory; Resistance; Continual learning; in-memory computing (IMC); MRAM; neural network;
D O I
10.1109/TCAD.2024.3371268
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Due to the separate memory and computation units in traditional von Neumann architecture, massive data transfer dominates the overall computing system's power and latency, known as the "Memory-Wall" issue. Especially with ever-increasing deep learning-based AI model size and computing complexity, it becomes the bottleneck for state-of-the-art AI computing systems. To address this challenge, in-memory computing (IMC)-based Neural Network accelerators have been widely investigated to support AI computing within memory. However, most of those works focus only on inference. The on-device training and continual learning have not been well explored yet. In this work, for the first time, we introduce on-device continual learning with STT-assisted-SOT (SAS) magnetoresistive random-access memory (MRAM)-based IMC system. On the hardware side, we have fabricated a STT-assisted-SOT MRAM (SAS-MRAM) device prototype with 4 magnetic tunnel junctions (MTJs, each at 100 nm x50 nm) sharing a common heavy metal layer, achieving significantly improved memory writing and area efficiency compared to traditional SOT-MRAM. Next, we designed fully digital IMC circuits with our SAS-MRAM to support both neural network inference and on-device learning. To enable efficient on-device continual learning for new task data, we present an 8-bit integer (INT8)-based continual learning algorithm that utilizes our SAS-MRAM IMC-supported bit-serial digital in-memory convolution operations to train a small parallel reprogramming network (Rep-Net) while freezing the major backbone model. Extensive studies have been presented based on our fabricated SAS-MRAM device prototype, cross-layer device-circuit benchmarking and simulation, as well as the on-device continual learning system evaluation.
引用
收藏
页码:2393 / 2404
页数:12
相关论文
共 50 条
  • [41] A Study of STT-RAM-based In-Memory Computing Across the Memory Hierarchy
    Gajaria, Dhruv
    Gomez, Kevin Antony
    Adegbija, Tosiron
    2022 IEEE 40TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD 2022), 2022, : 685 - 692
  • [42] An STT-MRAM Based in Memory Architecture for Low Power Integral Computing
    Zhao, Yinglin
    Ouyang, Peng
    Kang, Wang
    Yin, Shouyi
    Zhang, Youguang
    Wei, Shaojun
    Zhao, Weisheng
    IEEE TRANSACTIONS ON COMPUTERS, 2019, 68 (04) : 617 - 623
  • [43] An STT-MRAM based reconfigurable computing-in-memory architecture for general purpose computing
    Yu Pan
    Xiaotao Jia
    Zhen Cheng
    Peng Ouyang
    Xueyan Wang
    Jianlei Yang
    Weisheng Zhao
    CCF Transactions on High Performance Computing, 2020, 2 : 272 - 281
  • [44] An STT-MRAM based reconfigurable computing-in-memory architecture for general purpose computing
    Pan, Yu
    Jia, Xiaotao
    Cheng, Zhen
    Ouyang, Peng
    Wang, Xueyan
    Yang, Jianlei
    Zhao, Weisheng
    CCF TRANSACTIONS ON HIGH PERFORMANCE COMPUTING, 2020, 2 (03) : 272 - 281
  • [45] A New MRAM-based Process In-Memory Accelerator for Efficient Neural Network Training with Floating Point Precision
    Wang, Hongjie
    Zhao, Yang
    Li, Chaojian
    Wang, Yue
    Lin, Yingyan
    2020 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2020,
  • [46] A 14 nm MRAM-Based Multi-bit Analog In-Memory Computing With Process-Variation Calibration for 72 Macros-Based Accelerator
    Myung, Sungmeen
    Yun, Seok-Ju
    Kim, Minje
    Yi, Wooseok
    Lee, Jaehyuk
    An, Jangho
    Lee, Kyoung-Rog
    Shin, Chang-Woo
    Jung, Seungchul
    Kwon, Soonwan
    IEEE SOLID-STATE CIRCUITS LETTERS, 2024, 7 : 295 - 298
  • [47] A 6T-3M SOT-MRAM for in-memory computing with reconfigurable arith- metic operations
    Jin, Xing
    Yin, Ningyuan
    Chen, Weichong
    Li, Ximing
    Zhao, Guihua
    Yu, Zhiyi
    IEICE ELECTRONICS EXPRESS, 2023, 20 (11):
  • [48] SOT and STT-Based 4-Bit MRAM Cell for High-Density Memory Applications
    Nisar, Arshid
    Dhull, Seema
    Mittal, Sparsh
    Kaushik, Brajesh Kumar
    IEEE TRANSACTIONS ON ELECTRON DEVICES, 2021, 68 (09) : 4384 - 4390
  • [49] High-Performance Computing-in-Memory Architecture Using STT-/SOT-Based Series Triple-Level Cell MRAM
    Nehra, Vikas
    Prajapati, Sanjay
    Kumar, T. Nandha
    Kaushik, Brajesh Kumar
    IEEE TRANSACTIONS ON MAGNETICS, 2021, 57 (08)
  • [50] Design of an Area-Efficient Computing in Memory Platform Based on STT-MRAM
    Wang, Chao
    Wang, Zhaohao
    Wang, Gefei
    Zhang, Youguang
    Zhao, Weisheng
    IEEE TRANSACTIONS ON MAGNETICS, 2021, 57 (02)