An In-Memory-Computing Binary Neural Network Architecture With In-Memory Batch Normalization

被引:0
|
作者
Rege, Prathamesh Prashant [1 ]
Yin, Ming [2 ]
Parihar, Sanjay [3 ]
Versaggi, Joseph [2 ]
Nemawarkar, Shashank [3 ]
机构
[1] Northeastern Univ, Boston, MA 80305 USA
[2] GLOBALFOUNDRIES, Malta, NY 12020 USA
[3] GLOBALFOUNDRIES, Austin, TX 78735 USA
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Accuracy; Neural networks; Batch normalization; Convolutional neural networks; Training; Data models; Voltage control; In-memory computing; SRAM chips; binary neural network; edge device; in-memory computing; process variation; SRAM;
D O I
10.1109/ACCESS.2024.3444481
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper describes an in-memory computing architecture that combines full-precision computation for the first and last layers of a neural network while employing binary weights and input activations for the intermediate layers. This unique approach presents an efficient and effective solution for optimizing neural-network computations, reducing complexity, and enhancing energy efficiency. Notably, multiple architecture-level optimization methods are developed to ensure the binary operations thereby eliminating the need for intricate "digital logic" components external to the memory units. One of the key contributions of this study is in-memory batch normalization, which is implemented to provide good accuracy for CIFAR10 classification applications. Despite the inherent challenges posed by the process variations, the proposed design demonstrated an accuracy of 78%. Furthermore, the SRAM layer in the architecture showed an energy efficiency of 1086 TOPS/W and throughput of 23 TOPS, all packed efficiently within an area of 60 TOPS/mm2. This novel in-memory computing architecture offers a promising solution for next-generation efficient and high-performance deep learning applications.
引用
收藏
页码:190889 / 190896
页数:8
相关论文
共 50 条
  • [31] STT-MRAM Architecture with Parallel Accumulator for In-Memory Binary Neural Networks
    Thi-Nhan Pham
    Quang-Kien Trinh
    Chang, Ik-Joon
    Alioto, Massimo
    2021 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2021,
  • [32] Ternary Output Binary Neural Network With Zero-Skipping for MRAM-Based Digital In-Memory Computing
    Na, Taehui
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2023, 70 (07) : 2655 - 2659
  • [33] An In-Memory VLSI Architecture for Convolutional Neural Networks
    Kang, Mingu
    Lim, Sungmin
    Gonugondla, Sujan
    Shanbhag, Naresh R.
    IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2018, 8 (03) : 494 - 505
  • [34] Efficient and lightweight in-memory computing architecture for hardware security
    Ajmi, Hala
    Zayer, Fakhreddine
    Fredj, Amira Hadj
    Belgacem, Hamdi
    Mohammad, Baker
    Werghi, Naoufel
    Dias, Jorge
    JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING, 2024, 190
  • [35] A Flexible In-Memory Computing Architecture for Heterogeneously Quantized CNNs
    Ponzina, Flavio
    Rios, Marco
    Ansaloni, Giovanni
    Levisse, Alexandre
    Atienza, David
    2021 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI 2021), 2021, : 164 - 169
  • [36] An In-Memory-Computing Structure with Quantum-Dot Transistor Toward Neural Network Applications: From Analog Circuits to Memory Arrays
    Zhao, Yang
    Jain, Faquir
    Wang, Lei
    International Journal of High Speed Electronics and Systems, 2024, 33 (2-3)
  • [37] In-memory computing to break the memory wall*
    Huang, Xiaohe
    Liu, Chunsen
    Jiang, Yu-Gang
    Zhou, Peng
    CHINESE PHYSICS B, 2020, 29 (07)
  • [38] In-memory computing to break the memory wall
    黄晓合
    刘春森
    姜育刚
    周鹏
    Chinese Physics B, 2020, 29 (07) : 42 - 62
  • [39] TAICHI: A Tiled Architecture for In-Memory Computing and Heterogeneous Integration
    Wang, Xinxin
    Pinkham, Reid
    Zidan, Mohammed A.
    Meng, Fan-Hsuan
    Flynn, Michael P.
    Zhang, Zhengya
    Lu, Wei D.
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2022, 69 (02) : 559 - 563
  • [40] Data movement elimination with dual in-memory computing architecture
    Kang, Wang
    Kou, Jing
    Zhang, Liang
    DEVICE, 2024, 2 (12):