On-Device Continual Learning With STT-Assisted-SOT MRAM-Based In-Memory Computing

被引:1
|
作者
Zhang, Fan [1 ]
Sridharan, Amitesh [1 ]
Hwang, William [2 ]
Xue, Fen [2 ]
Tsai, Wilman [3 ]
Wang, Shan Xiang [4 ,5 ]
Fan, Deliang [1 ]
机构
[1] Johns Hopkins Univ, Dept Elect & Comp Engn, Baltimore, MD 21218 USA
[2] Stanford Univ, Dept Elect Engn, Stanford, CA 94305 USA
[3] Stanford Univ, Dept Mat Sci & Engn, Stanford, CA 94305 USA
[4] Stanford Univ, Dept Elect Engn, Stanford, CA 94305 USA
[5] Stanford Univ, Dept Mat Sci & Engn, Stanford, CA 94305 USA
基金
美国国家科学基金会;
关键词
Magnetic tunneling; Training; In-memory computing; Task analysis; Quantization (signal); Nonvolatile memory; Resistance; Continual learning; in-memory computing (IMC); MRAM; neural network;
D O I
10.1109/TCAD.2024.3371268
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Due to the separate memory and computation units in traditional von Neumann architecture, massive data transfer dominates the overall computing system's power and latency, known as the "Memory-Wall" issue. Especially with ever-increasing deep learning-based AI model size and computing complexity, it becomes the bottleneck for state-of-the-art AI computing systems. To address this challenge, in-memory computing (IMC)-based Neural Network accelerators have been widely investigated to support AI computing within memory. However, most of those works focus only on inference. The on-device training and continual learning have not been well explored yet. In this work, for the first time, we introduce on-device continual learning with STT-assisted-SOT (SAS) magnetoresistive random-access memory (MRAM)-based IMC system. On the hardware side, we have fabricated a STT-assisted-SOT MRAM (SAS-MRAM) device prototype with 4 magnetic tunnel junctions (MTJs, each at 100 nm x50 nm) sharing a common heavy metal layer, achieving significantly improved memory writing and area efficiency compared to traditional SOT-MRAM. Next, we designed fully digital IMC circuits with our SAS-MRAM to support both neural network inference and on-device learning. To enable efficient on-device continual learning for new task data, we present an 8-bit integer (INT8)-based continual learning algorithm that utilizes our SAS-MRAM IMC-supported bit-serial digital in-memory convolution operations to train a small parallel reprogramming network (Rep-Net) while freezing the major backbone model. Extensive studies have been presented based on our fabricated SAS-MRAM device prototype, cross-layer device-circuit benchmarking and simulation, as well as the on-device continual learning system evaluation.
引用
收藏
页码:2393 / 2404
页数:12
相关论文
共 50 条
  • [31] An In-Memory Power Efficient Computing Architecture with Emerging VGSOT MRAM Device
    Sarkar, Md Rubel
    Chowdhury, Shirazush Salekin
    Walling, Jeffrey Sean
    Yi, Cindy Yang
    2024 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, ISCAS 2024, 2024,
  • [32] STT-BSNN: An In-Memory Deep Binary Spiking Neural Network Based on STT-MRAM
    Van-Tinh Nguyen
    Quang-Kien Trinh
    Zhang, Renyuan
    Nakashima, Yasuhiko
    IEEE ACCESS, 2021, 9 (09): : 151373 - 151385
  • [33] Approx-IMC: A general-purpose approximate digital in-memory computing framework based on STT-MRAM
    Hajisadeghi, Amir M.
    Momtazpour, Mahmoud
    Zarandi, Hamid R.
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2024, 160 : 40 - 53
  • [34] IMFlexCom: Energy Efficient In-Memory Flexible Computing Using Dual-Mode SOT-MRAM
    Parveen, Farhana
    Angizi, Shaahin
    Fan, Deliang
    ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2018, 14 (03)
  • [35] Towards Efficient SOT-Assisted STT-MRAM Cell Switching Using Reinforcement Learning
    Ender, Johannes
    de Orio, Roberto Lacerda
    Gos, Wolfgang
    Sverdlov, Viktor
    LARGE-SCALE SCIENTIFIC COMPUTATIONS, LSSC 2023, 2024, 13952 : 83 - 90
  • [36] Compute SNDR-Boosted 22-nm MRAM-Based In-Memory Computing Macro Using Statistical Error Compensation
    Roy, Saion K.
    Ou, Han-Mo
    Ahmed, Mostafa G.
    Deaville, Peter
    Zhang, Bonan
    Verma, Naveen
    Hanumolu, Pavan K.
    Shanbhag, Naresh R.
    IEEE JOURNAL OF SOLID-STATE CIRCUITS, 2025, 60 (03) : 1092 - 1102
  • [37] Energy Efficient Computing With High-Density, Field-Free STT-Assisted SOT-MRAM (SAS-MRAM)
    Hwang, William
    Xue, Fen
    Zhang, Fan
    Song, Ming-Yuan
    Lee, Chien-Min
    Turgut, Emrah
    Chen, T. C.
    Bao, Xinyu
    Tsai, Wilman
    Fan, Deliang
    Wang, Shan X.
    IEEE TRANSACTIONS ON MAGNETICS, 2023, 59 (03)
  • [38] Proposal of Analog In-Memory Computing With Magnified Tunnel Magnetoresistance Ratio and Universal STT-MRAM Cell
    Cai, Hao
    Guo, Yanan
    Liu, Bo
    Zhou, Mingyang
    Chen, Juntong
    Liu, Xinning
    Yang, Jun
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2022, 69 (04) : 1519 - 1531
  • [39] An In-Memory Computing Architecture Utilizing Energy-Efficient VGSOT MRAM Device
    Sarkar, Md Rubel
    Yi, Cindy Yang
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2024, 71 (07) : 3258 - 3262
  • [40] Design of High-RA STT-MRAM for Future Energy-Efficient In-Memory Computing
    Hong, Ming-Chun
    Su, Yi-Hui
    Chen, Guan-Long
    Hsin, Yu-Chen
    Chang, Yao-Jen
    Chen, Kuan-Ming
    Yang, Shan-Yi
    Wang, I-Jung
    Rahaman, S. K. Ziaur
    Lee, Hsin-Han
    Wei, Jeng-Hua
    Sheu, Shyh-Shyuan
    Lo, Wei-Chung
    Chang, Shih-Chieh
    Hou, Tuo-Hung
    2023 INTERNATIONAL VLSI SYMPOSIUM ON TECHNOLOGY, SYSTEMS AND APPLICATIONS, VLSI-TSA/VLSI-DAT, 2023,