Parallelizing SRAM Arrays with Customized Bit-Cell for Binary Neural Networks

被引:0
|
作者
Liu, Rui [1 ]
Peng, Xiaochen [1 ]
Sun, Xiaoyu [1 ]
Khwa, Win-San [2 ]
Si, Xin [2 ]
Chen, Jia-Jing [2 ]
Li, Jia-Fang [2 ]
Chang, Meng-Fan [2 ]
Yu, Shimeng [1 ]
机构
[1] Arizona State Univ, Tempe, AZ 85287 USA
[2] Natl Tsing Hua Univ, Hsinchu, Taiwan
关键词
D O I
10.1145/3195970.3196089
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent advances in deep neural networks (DNNs) have shown Binary Neural Networks (BNNs) are able to provide a reasonable accuracy on various image datasets with a significant reduction in computation and memory cost. In this paper, we explore two BNNs: hybrid BNN (HBNN) and XNORBNN, where the weights are binarized to +1/-1 while the neuron activations are binarized to 1/0 and +1/-1, respectively. Two SRAM bit cell designs are proposed, namely, 6T SRAM for HBNN and customized 8T SRAM for XNOR-BNN. In our design, the high-precision multiply-and-accumulate (MAC) is replaced by bitwise multiplication for HBNN or XNOR for XNOR-BNN plus bit-counting operations. To parallelize the weighted sum operation, we activate multiple word lines in the SRAM array simultaneously and digitize the analog voltage developed along the bit line by a multi-level sense amplifier (MLSA). In order to partition the large matrices in DNNs, we investigate the impact of sensing bit-levels of MLSA on the accuracy degradation for different sub-array sizes and propose using the nonlinear quantization technique to mitigate the accuracy degradation. With 64x64 sub-array size and 3-bit MLSA, HBNN and XNORBNN architectures can minimize the accuracy degradation to 2.37% and 0.88%, respectively, for an inspired VGG-16 network on the CIFAR-10 dataset. Design space exploration of SRAM based synaptic architectures with the conventional row-by-row access scheme and our proposed parallel access scheme are also performed, showing significant benefits in the area, latency and energy-efficiency. Finally, we have successfully taped-out and validated the proposed HBNN and XNOR-BNN designs in TSMC 65 nm process with measured silicon data, achieving energy efficiency >100 TOPS/W for HBNN and >50 TOPS/W for XNOR-BNN.
引用
收藏
页数:6
相关论文
共 50 条
  • [41] A Novel 8T XNOR-SRAM: Computing-in-Memory Design for Binary/Ternary Deep Neural Networks
    Alnatsheh, Nader
    Kim, Youngbae
    Cho, Jaeik
    Choi, Kyuwon Ken
    ELECTRONICS, 2023, 12 (04)
  • [42] Double MAC on a Cell: A 22-nm 8T-SRAM-Based Analog In-Memory Accelerator for Binary/Ternary Neural Networks Featuring Split Wordline
    Tagata, Hiroto
    Sato, Takashi
    Awano, Hiromitsu
    IEEE OPEN JOURNAL OF CIRCUITS AND SYSTEMS, 2024, 5 : 328 - 340
  • [43] Ternary In-Memory MAC Accelerator With Dual-6T SRAM Cell for Deep Neural Networks
    Wang, Xudong
    Li, Geng
    Sun, Jiacong
    Fan, Huanjie
    Chen, Yong
    Jiao, Hailong
    2022 IEEE ASIA PACIFIC CONFERENCE ON CIRCUITS AND SYSTEMS, APCCAS, 2022, : 246 - 250
  • [44] Memory Requirement Reduction of Deep Neural Networks for Field Programmable Gate Arrays Using Low-Bit Quantization of Parameters
    Nicodemo, Niccolo
    Naithani, Gaurav
    Drossos, Konstantinos
    Virtanen, Tuomas
    Saletti, Roberto
    28TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2020), 2021, : 466 - 470
  • [45] A Novel Ultra-Low Power 8T SRAM-Based Compute-in-Memory Design for Binary Neural Networks
    Kim, Youngbae
    Li, Shuai
    Yadav, Nandakishor
    Choi, Kyuwon Ken
    ELECTRONICS, 2021, 10 (17)
  • [46] An Energy-Efficient Reconfigurable Processor for Binary- and Ternary-Weight Neural Networks With Flexible Data Bit Width
    Yin, Shouyi
    Ouyang, Peng
    Yang, Jianxun
    Lu, Tianyi
    Li, Xiudong
    Liu, Leibo
    Wei, Shaojun
    IEEE JOURNAL OF SOLID-STATE CIRCUITS, 2019, 54 (04) : 1120 - 1136
  • [47] Newly energy-efficient SRAM bit-cell using GAA CNT-GDI method with asymmetrical write and built-in read-assist schemes for QR code-based multimedia applications
    Darabi, Abdolreza
    Salehi, Mohammad Reza
    Abiri, Ebrahim
    MICROELECTRONICS JOURNAL, 2021, 114
  • [48] A 290-mV, 7-nm Ultra-Low-Voltage One-Port SRAM Compiler Design Using a 12T Write Contention and Read Upset Free Bit-Cell
    Sinangil, Mahmut E.
    Lin, Yen-Ting
    Liao, Hung-Jen
    Chang, Jonathan
    IEEE JOURNAL OF SOLID-STATE CIRCUITS, 2019, 54 (04) : 1152 - 1160
  • [49] Single RRAM Cell-based In-Memory Accelerator Architecture for Binary Neural Networks
    Oh, Hyunmyung
    Kim, Hyungjun
    Kang, Nameun
    Kim, Yulhwa
    Park, Jihoon
    Kim, Jae-Joon
    2021 IEEE 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS), 2021,
  • [50] X-Fault: Impact of Faults on Binary Neural Networks in Memristor-Crossbar Arrays with Logic-in-Memory Computation
    Staudigl, Felix
    Sturm, Karl J. X.
    Bartel, Maximilian
    Fetz, Thorben
    Sisejkovic, Dominik
    Joseph, Jan Moritz
    Pohls, Leticia Bolzani
    Leupers, Rainer
    2022 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS 2022): INTELLIGENT TECHNOLOGY IN THE POST-PANDEMIC ERA, 2022, : 174 - 177