RNSiM: Efficient Deep Neural Network Accelerator Using Residue Number Systems

被引:7
|
作者
Roohi, Arman [1 ]
Taheri, MohammadReza
Angizi, Shaahin [2 ]
Fan, Deliang [3 ]
机构
[1] Univ Nebraska, Dept Comp Sci & Engn, Lincoln, NE 68588 USA
[2] New Jersey Inst Technol, Dept Elect & Comp Engn, Newark, NJ 07102 USA
[3] Arizona State Univ, Sch Elect Comp & Energy Engn, Tempe, AZ USA
关键词
residue number system; processing-in-Memory; convolutional neural network; accelerator;
D O I
10.1109/ICCAD51958.2021.9643531
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we propose an efficient convolutional neural network (CNN) accelerator design, entitled RNSiM, based on the Residue Number System (RNS) as an alternative for the conventional binary number representation. Instead of traditional arithmetic implementation that suffers from the inevitable lengthy carry propagation chain, the novelty of RNSiM lies in that all the data, including stored weights and communication/computation, are performed in the RNS domain. Due to the inherent parallelism of the RNS arithmetic, power and latency are significantly reduced. Moreover, an enhanced integrated intermodulo operation core is developed to decrease the overhead imposed by non-modular operations. Further improvement in systems' performance efficiency is achieved by developing efficient Processing-in-Memory (PIM) designs using various volatile CMOS and non-volatile Post-CMOS technologies to accelerate RNS-based multiplication-and-accumulations (MACs). The RNSiM accelerator's performance on different datasets, including MNIST, SVHN, and CIFAR-10, is evaluated. With almost the same accuracy to the baseline CNN, the RNSiM accelerator can significantly increase both energy-efficiency and speedup compared with the state-of-the-art FPGA, GPU, and PIM designs. RNSiM and other RNS-PIMs, based on our method, reduce the energy consumption by orders of 28-77x and 331-897x compared with the FPGA and the GPU platforms, respectively.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] DNNARA: A Deep Neural Network Accelerator using Residue Arithmetic and Integrated Photonics
    Peng, Jiaxin
    Alkabani, Yousra
    Sun, Shuai
    Sorger, Volker J.
    El-Ghazawi, Tarek
    PROCEEDINGS OF THE 49TH INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING, ICPP 2020, 2020,
  • [2] A Deep Neural Network Accelerator using Residue Arithmetic in a Hybrid Optoelectronic System
    Peng, Jiaxin
    Alkabani, Yousra
    Puri, Krunal
    Ma, Xiaoxuan
    Sorger, Volker
    El-Ghazawi, Tarek
    ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2022, 18 (04)
  • [3] Design of an Efficient Deep Neural Network Accelerator Based on Block Posit Number Representation
    Hsiao, Shen-Fu
    Lin, Sin-Chen
    Chen, Guan-Lin
    Yang, Shih-Hua
    Yuan, Yen-Che
    Chen, Kun-Chih
    2024 INTERNATIONAL VLSI SYMPOSIUM ON TECHNOLOGY, SYSTEMS AND APPLICATIONS, VLSI TSA, 2024,
  • [4] An efficient deep neural network accelerator using controlled ferroelectric domain dynamics
    Majumdar, Sayani
    NEUROMORPHIC COMPUTING AND ENGINEERING, 2022, 2 (04):
  • [5] An Energy-Efficient Deep Neural Network Accelerator Design
    Jung, Jueun
    Lee, Kyuho Jason
    2020 54TH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS, 2020, : 272 - 276
  • [6] Efficient Hardware Accelerator for Compressed Sparse Deep Neural Network
    Xiao, Hao
    Zhao, Kaikai
    Liu, Guangzhu
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2021, E104D (05) : 772 - 775
  • [7] A Deep Convolutional Neural Network Based on Nested Residue Number System
    Nakahara, Hiroki
    Sasao, Tsutomu
    2015 25TH INTERNATIONAL CONFERENCE ON FIELD PROGRAMMABLE LOGIC AND APPLICATIONS, 2015,
  • [8] Area-Efficient FPGA Implementation of Minimalistic Convolutional Neural Network Using Residue Number System
    Chervyakov, Nikolay I.
    Lyakhov, Pavel A.
    Valueva, Maria V.
    Valuev, Georgii V.
    Kaplun, Dmitrii I.
    Efimenko, George A.
    Gnezdilov, Denis V.
    PROCEEDINGS OF THE 2018 23RD CONFERENCE OF OPEN INNOVATIONS ASSOCIATION (FRUCT), 2018, : 112 - 118
  • [9] Increasing of Convolutional Neural Network Performance Using Residue Number System
    Chervyakov, N. I.
    Lyakhov, P. A.
    Valueva, M. V.
    2017 INTERNATIONAL MULTI-CONFERENCE ON ENGINEERING, COMPUTER AND INFORMATION SCIENCES (SIBIRCON), 2017, : 135 - 140
  • [10] An Efficient Accelerator for Deep Convolutional Neural Networks
    Kuo, Yi-Xian
    Lai, Yeong-Kang
    2020 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS - TAIWAN (ICCE-TAIWAN), 2020,