On-Device Continual Learning With STT-Assisted-SOT MRAM-Based In-Memory Computing

被引:1
|
作者
Zhang, Fan [1 ]
Sridharan, Amitesh [1 ]
Hwang, William [2 ]
Xue, Fen [2 ]
Tsai, Wilman [3 ]
Wang, Shan Xiang [4 ,5 ]
Fan, Deliang [1 ]
机构
[1] Johns Hopkins Univ, Dept Elect & Comp Engn, Baltimore, MD 21218 USA
[2] Stanford Univ, Dept Elect Engn, Stanford, CA 94305 USA
[3] Stanford Univ, Dept Mat Sci & Engn, Stanford, CA 94305 USA
[4] Stanford Univ, Dept Elect Engn, Stanford, CA 94305 USA
[5] Stanford Univ, Dept Mat Sci & Engn, Stanford, CA 94305 USA
基金
美国国家科学基金会;
关键词
Magnetic tunneling; Training; In-memory computing; Task analysis; Quantization (signal); Nonvolatile memory; Resistance; Continual learning; in-memory computing (IMC); MRAM; neural network;
D O I
10.1109/TCAD.2024.3371268
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Due to the separate memory and computation units in traditional von Neumann architecture, massive data transfer dominates the overall computing system's power and latency, known as the "Memory-Wall" issue. Especially with ever-increasing deep learning-based AI model size and computing complexity, it becomes the bottleneck for state-of-the-art AI computing systems. To address this challenge, in-memory computing (IMC)-based Neural Network accelerators have been widely investigated to support AI computing within memory. However, most of those works focus only on inference. The on-device training and continual learning have not been well explored yet. In this work, for the first time, we introduce on-device continual learning with STT-assisted-SOT (SAS) magnetoresistive random-access memory (MRAM)-based IMC system. On the hardware side, we have fabricated a STT-assisted-SOT MRAM (SAS-MRAM) device prototype with 4 magnetic tunnel junctions (MTJs, each at 100 nm x50 nm) sharing a common heavy metal layer, achieving significantly improved memory writing and area efficiency compared to traditional SOT-MRAM. Next, we designed fully digital IMC circuits with our SAS-MRAM to support both neural network inference and on-device learning. To enable efficient on-device continual learning for new task data, we present an 8-bit integer (INT8)-based continual learning algorithm that utilizes our SAS-MRAM IMC-supported bit-serial digital in-memory convolution operations to train a small parallel reprogramming network (Rep-Net) while freezing the major backbone model. Extensive studies have been presented based on our fabricated SAS-MRAM device prototype, cross-layer device-circuit benchmarking and simulation, as well as the on-device continual learning system evaluation.
引用
收藏
页码:2393 / 2404
页数:12
相关论文
共 50 条
  • [1] A STT-Assisted SOT MRAM-Based In-Memory Booth Multiplier for Neural Network Applications
    Wu, Jiayao
    Wang, Yijiao
    Wang, Pengxu
    Wang, Yiming
    Zhao, Weisheng
    IEEE TRANSACTIONS ON NANOTECHNOLOGY, 2024, 23 : 29 - 34
  • [2] A STT-Assisted SOT MRAM-Based In-Memory Booth Multiplier for Neural Network Applications
    Wu, Jiayao
    Wang, Yijiao
    Wang, Pengxu
    Wang, Yiming
    Zhao, Weisheng
    IEEE Transactions on Nanotechnology, 2024, 23 : 29 - 34
  • [3] MRAM-based Analog Sigmoid Function for In-memory Computing
    Amin, Md Hasibul
    Elbtity, Mohammed
    Mohammadi, Mohammadreza
    Zand, Ramtin
    PROCEEDINGS OF THE 32ND GREAT LAKES SYMPOSIUM ON VLSI 2022, GLSVLSI 2022, 2022, : 319 - 323
  • [4] MRIMA: An MRAM-Based In-Memory Accelerator
    Angizi, Shaahin
    He, Zhezhi
    Awad, Amro
    Fan, Deliang
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2020, 39 (05) : 1123 - 1136
  • [5] MRAM-based In-Memory Computing for Efficient Acceleration of Generative Adversarial Networks
    Kaushik, Partha
    Gupta, Avi
    Nehete, Hemkant
    Kaushik, Brajesh Kumar
    2023 IEEE 23RD INTERNATIONAL CONFERENCE ON NANOTECHNOLOGY, NANO, 2023, : 798 - 802
  • [6] Exploring a SOT-MRAM Based In-Memory Computing for Data Processing
    He, Zhezhi
    Zhang, Yang
    Angizi, Shaahin
    Gong, Boqing
    Fan, Deliang
    IEEE TRANSACTIONS ON MULTI-SCALE COMPUTING SYSTEMS, 2018, 4 (04): : 676 - 685
  • [7] SOT-MRAM based Analog in-Memory Computing for DNN inference
    Doevenspeck, J.
    Garello, K.
    Verhoef, B.
    Degraeve, R.
    Van Beek, S.
    Crotti, D.
    Yasin, F.
    Couet, S.
    Jayakumar, G.
    Papistas, I. A.
    Debacker, P.
    Lauwereins, R.
    Dehaene, W.
    Kar, G. S.
    Cosemans, S.
    Mallik, A.
    Verkest, D.
    2020 IEEE SYMPOSIUM ON VLSI TECHNOLOGY, 2020,
  • [8] A 129.83 TOPS/W Area Efficient Digital SOT/STT MRAM-Based Computing-In-Memory for Advanced Edge AI Chips
    Lu, Lu
    Mani, Aarthy
    Anh Tuan Do
    2023 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, ISCAS, 2023,
  • [9] HieIM: Highly Flexible In-Memory Computing using STT MRAM
    Parveen, Farhana
    He, Zhezhi
    Angizi, Shaahin
    Fan, Deliang
    2018 23RD ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2018, : 361 - 366
  • [10] Domain-Specific STT-MRAM-Based In-Memory Computing: A Survey
    Yusuf, Alaba
    Adegbija, Tosiron
    Gajaria, Dhruv
    IEEE ACCESS, 2024, 12 : 28036 - 28056