Parallelizing SRAM Arrays with Customized Bit-Cell for Binary Neural Networks

被引:0
|
作者
Liu, Rui [1 ]
Peng, Xiaochen [1 ]
Sun, Xiaoyu [1 ]
Khwa, Win-San [2 ]
Si, Xin [2 ]
Chen, Jia-Jing [2 ]
Li, Jia-Fang [2 ]
Chang, Meng-Fan [2 ]
Yu, Shimeng [1 ]
机构
[1] Arizona State Univ, Tempe, AZ 85287 USA
[2] Natl Tsing Hua Univ, Hsinchu, Taiwan
关键词
D O I
10.1145/3195970.3196089
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent advances in deep neural networks (DNNs) have shown Binary Neural Networks (BNNs) are able to provide a reasonable accuracy on various image datasets with a significant reduction in computation and memory cost. In this paper, we explore two BNNs: hybrid BNN (HBNN) and XNORBNN, where the weights are binarized to +1/-1 while the neuron activations are binarized to 1/0 and +1/-1, respectively. Two SRAM bit cell designs are proposed, namely, 6T SRAM for HBNN and customized 8T SRAM for XNOR-BNN. In our design, the high-precision multiply-and-accumulate (MAC) is replaced by bitwise multiplication for HBNN or XNOR for XNOR-BNN plus bit-counting operations. To parallelize the weighted sum operation, we activate multiple word lines in the SRAM array simultaneously and digitize the analog voltage developed along the bit line by a multi-level sense amplifier (MLSA). In order to partition the large matrices in DNNs, we investigate the impact of sensing bit-levels of MLSA on the accuracy degradation for different sub-array sizes and propose using the nonlinear quantization technique to mitigate the accuracy degradation. With 64x64 sub-array size and 3-bit MLSA, HBNN and XNORBNN architectures can minimize the accuracy degradation to 2.37% and 0.88%, respectively, for an inspired VGG-16 network on the CIFAR-10 dataset. Design space exploration of SRAM based synaptic architectures with the conventional row-by-row access scheme and our proposed parallel access scheme are also performed, showing significant benefits in the area, latency and energy-efficiency. Finally, we have successfully taped-out and validated the proposed HBNN and XNOR-BNN designs in TSMC 65 nm process with measured silicon data, achieving energy efficiency >100 TOPS/W for HBNN and >50 TOPS/W for XNOR-BNN.
引用
收藏
页数:6
相关论文
共 50 条
  • [31] Tiled Bit Networks: Sub-Bit Neural Network Compression Through Reuse of Learnable Binary Vectors
    Gorbett, Matt
    Shirazi, Hossein
    Ray, Indrakshi
    PROCEEDINGS OF THE 33RD ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2024, 2024, : 674 - 684
  • [32] Predicting Failure Distributions of SRAM Arrays by Using Extreme Value Statistic, Bit Cell Simulation, and Machine Learning
    Pompl, T.
    Bashir, T. Khan
    Voelker, M.
    Last, F.
    Ostermayr, M.
    IEEE TRANSACTIONS ON DEVICE AND MATERIALS RELIABILITY, 2023, 23 (03) : 363 - 369
  • [33] Quadruple Cross-Coupled Latch-Based 10T and 12T SRAM Bit-Cell Designs for Highly Reliable Terrestrial Applications
    Jiang, Jianwei
    Xu, Yiran
    Zhu, Wenyi
    Xiao, Jun
    Zou, Shichang
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2019, 66 (03) : 967 - 977
  • [34] A ReRAM-Based Computing-in-Memory Convolutional-Macro With Customized 2T2R Bit-Cell for AIoT Chip IP Applications
    Tan, Fei
    Wang, Yiming
    Yang, Yiming
    Li, Liran
    Wang, Tian
    Zhang, Feng
    Wang, Xinghua
    Gao, Jianfeng
    Liu, Yongpan
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2020, 67 (09) : 1534 - 1538
  • [35] Implementation of Convolutional Neural Networks in Memristor Crossbar Arrays with Binary Activation and Weight Quantization
    Park, Jinwoo
    Kim, Sungjoon
    Song, Min Suk
    Youn, Sangwook
    Kim, Kyuree
    Kim, Tae-Hyeon
    Kim, Hyungjin
    ACS APPLIED MATERIALS & INTERFACES, 2024, 16 (01) : 1054 - 1065
  • [36] Binary Ensemble Neural Network: More Bits per Network or More Networks per Bit?
    Zhu, Shilin
    Dong, Xin
    Su, Hao
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 4918 - 4927
  • [37] TD-SRAM: Time-Domain-Based In-Memory Computing Macro for Binary Neural Networks
    Song, Jiahao
    Wang, Yuan
    Guo, Minguang
    Ji, Xiang
    Cheng, Kaili
    Hu, Yixuan
    Tang, Xiyuan
    Wang, Runsheng
    Huang, Ru
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2021, 68 (08) : 3377 - 3387
  • [38] Colonnade: A Reconfigurable SRAM-Based Digital Bit-Serial Compute-In-Memory Macro for Processing Neural Networks
    Kim, Hyunjoon
    Yoo, Taegeun
    Kim, Tony Tae-Hyoung
    Kim, Bongjin
    IEEE JOURNAL OF SOLID-STATE CIRCUITS, 2021, 56 (07) : 2221 - 2233
  • [39] Wrap-Gate CNT-MOSFET Based SRAM Bit-Cell with Asymmetrical Ground Gating and Built-In Read-Assist Schemes for Application in Limited-Energy Environments
    Darabi, Abdolreza
    Salehi, Mohammad Reza
    Abiri, Ebrahim
    ECS JOURNAL OF SOLID STATE SCIENCE AND TECHNOLOGY, 2022, 11 (03)
  • [40] DAM SRAM CORE: An Efficient High-Speed and Low-Power CIM SRAM CORE Design for Feature Extraction Convolutional Layers in Binary Neural Networks
    Zhao, Ruiyong
    Gong, Zhenghui
    Liu, Yulan
    Chen, Jing
    MICROMACHINES, 2024, 15 (05)