RNSiM: Efficient Deep Neural Network Accelerator Using Residue Number Systems

被引:7
|
作者
Roohi, Arman [1 ]
Taheri, MohammadReza
Angizi, Shaahin [2 ]
Fan, Deliang [3 ]
机构
[1] Univ Nebraska, Dept Comp Sci & Engn, Lincoln, NE 68588 USA
[2] New Jersey Inst Technol, Dept Elect & Comp Engn, Newark, NJ 07102 USA
[3] Arizona State Univ, Sch Elect Comp & Energy Engn, Tempe, AZ USA
关键词
residue number system; processing-in-Memory; convolutional neural network; accelerator;
D O I
10.1109/ICCAD51958.2021.9643531
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we propose an efficient convolutional neural network (CNN) accelerator design, entitled RNSiM, based on the Residue Number System (RNS) as an alternative for the conventional binary number representation. Instead of traditional arithmetic implementation that suffers from the inevitable lengthy carry propagation chain, the novelty of RNSiM lies in that all the data, including stored weights and communication/computation, are performed in the RNS domain. Due to the inherent parallelism of the RNS arithmetic, power and latency are significantly reduced. Moreover, an enhanced integrated intermodulo operation core is developed to decrease the overhead imposed by non-modular operations. Further improvement in systems' performance efficiency is achieved by developing efficient Processing-in-Memory (PIM) designs using various volatile CMOS and non-volatile Post-CMOS technologies to accelerate RNS-based multiplication-and-accumulations (MACs). The RNSiM accelerator's performance on different datasets, including MNIST, SVHN, and CIFAR-10, is evaluated. With almost the same accuracy to the baseline CNN, the RNSiM accelerator can significantly increase both energy-efficiency and speedup compared with the state-of-the-art FPGA, GPU, and PIM designs. RNSiM and other RNS-PIMs, based on our method, reduce the energy consumption by orders of 28-77x and 331-897x compared with the FPGA and the GPU platforms, respectively.
引用
收藏
页数:9
相关论文
共 50 条
  • [21] Neural Synaptic Plasticity-Inspired Computing: A High Computing Efficient Deep Convolutional Neural Network Accelerator
    Xia, Zihan
    Chen, Jienan
    Huang, Qiu
    Luo, Jinting
    Hu, Jianhao
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2021, 68 (02) : 728 - 740
  • [22] EFFICIENT CONVERTERS FOR RESIDUE AND QUADRATIC-RESIDUE NUMBER-SYSTEMS
    STOURAITIS, T
    IEE PROCEEDINGS-G CIRCUITS DEVICES AND SYSTEMS, 1992, 139 (06): : 626 - 634
  • [23] A Visual Tracking Deep Convolutional Neural Network Accelerator
    Qin, Zhiyong
    Yu, Lixin
    PROCEEDINGS OF THE 2017 2ND INTERNATIONAL CONFERENCE ON AUTOMATION, MECHANICAL CONTROL AND COMPUTATIONAL ENGINEERING (AMCCE 2017), 2017, 118 : 493 - 499
  • [24] Systems Modeling Using Deep Elman Neural Network
    Salah, Latifa Belhaj
    Fourati, Fathi
    ENGINEERING TECHNOLOGY & APPLIED SCIENCE RESEARCH, 2019, 9 (02) : 3881 - 3886
  • [25] FPGA-Based Deep Convolutional Neural Network Accelerator Design Techniques for the Handwritten Number Recognizer
    Yoo, Yechan
    Park, Yoonjin
    Kim, Injung
    Yi, Kang
    ADVANCED SCIENCE LETTERS, 2018, 24 (03) : 2152 - 2155
  • [26] A High Efficient Architecture for Convolution Neural Network Accelerator
    Kong Anmin
    Zhao Bin
    2019 2ND INTERNATIONAL CONFERENCE ON INTELLIGENT AUTONOMOUS SYSTEMS (ICOIAS 2019), 2019, : 131 - 134
  • [27] An Efficient Accelerator Unit for Sparse Convolutional Neural Network
    Zhao, Yulin
    Wang, Donghui
    Wang, Leiou
    TENTH INTERNATIONAL CONFERENCE ON DIGITAL IMAGE PROCESSING (ICDIP 2018), 2018, 10806
  • [28] An efficient classification of malware behavior using deep neural network
    Hai, Quan Tran
    Hwang, Seong Oun
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2018, 35 (06) : 5801 - 5814
  • [29] CondenseNeXt: An Ultra-Efficient Deep Neural Network for Embedded Systems
    Kalgaonkar, Priyank
    El-Sharkawy, Mohamed
    2021 IEEE 11TH ANNUAL COMPUTING AND COMMUNICATION WORKSHOP AND CONFERENCE (CCWC), 2021, : 524 - 528
  • [30] Activation in Network for NoC-based Deep Neural Network Accelerator
    Zhu, Wenyao
    Chen, Yizhi
    Lu, Zhonghai
    2024 INTERNATIONAL VLSI SYMPOSIUM ON TECHNOLOGY, SYSTEMS AND APPLICATIONS, VLSI TSA, 2024,