Optimize FPGA-based Neural Network Accelerator with Bit-shift Quantization

被引:0
|
作者
Liu, Yu [1 ]
Liu, XueJiao [1 ]
Liang, Luhong [1 ]
机构
[1] Hong Kong Appl Sci & Technol Res Inst ASTRI, Hong Kong, Peoples R China
关键词
D O I
10.1109/iscas45731.2020.9180919
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Well-programmed Field Programmable Gate Arrays (FPGAs) can accelerate Deep Neural Network (DNN) with high power efficiency. The dominant workloads of DNNs are Multiply Accumulates (MACs), which can be directly mapped to Digital Signal Processors (DSPs) in the FPGA. A DNN accelerator pursuing high performance can consume almost all the DSPs, but with a considerable amount of Look-up Tables (LUTs) in the FPGA unused or performing MACs inefficiently. To solve this problem, we present a Bit-Shift method for FPGA-based DNN accelerator to fully utilize the resources in the FPGA. The MAC is converted to a limited number of shift-and-add operations, which can be implemented by LUTs with significant improvement of efficiency. A quantization method based on Minimum Mean Absolute Error (MMAE) is proposed to preserve the accuracy of the DNN inference in the conversion of DNN parameters without re-training. The quantized parameters can be compressed to a fixed and fewer number of bits to reduce the memory bandwidth. Accordingly, a Bit-Shift architecture is designed to load the compressed parameters and perform the converted MAC calculations without extra decompression module. A large scale DNN accelerator with the proposed Bit-Shift architecture is implemented in a Xilinx VU095 FPGA. Experimental results show that the proposed method can boost the processing speed by 32% and reach 331 GOPS at 190MHz clock frequency for ResNet-34.
引用
收藏
页数:5
相关论文
共 50 条
  • [1] A reconfigurable FPGA-based spiking neural network accelerator
    Yin, Mingqi
    Cui, Xiaole
    Wei, Feng
    Liu, Hanqing
    Jiang, Yuanyuan
    Cui, Xiaoxin
    MICROELECTRONICS JOURNAL, 2024, 152
  • [2] An FPGA-based Accelerator Platform Implements for Convolutional Neural Network
    Meng, Xiao
    Yu, Lixin
    Qin, Zhiyong
    2019 THE 3RD INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE COMPILATION, COMPUTING AND COMMUNICATIONS (HP3C 2019), 2019, : 25 - 28
  • [3] FPGA-based Accelerator for Convolutional Neural Network Application in Mobile Robotics
    Mazzetto, Lucas F. R.
    Castanho, Jose E. C.
    2023 LATIN AMERICAN ROBOTICS SYMPOSIUM, LARS, 2023 BRAZILIAN SYMPOSIUM ON ROBOTICS, SBR, AND 2023 WORKSHOP ON ROBOTICS IN EDUCATION, WRE, 2023, : 433 - 438
  • [4] A FPGA-based Accelerator of Convolutional Neural Network for Face Feature Extraction
    Ding, Ru
    Su, Guangda
    Bai, Guoqiang
    Xu, Wei
    Su, Nan
    Wu, Xingjun
    2019 IEEE INTERNATIONAL CONFERENCE ON ELECTRON DEVICES AND SOLID-STATE CIRCUITS (EDSSC), 2019,
  • [5] FPGA-Based Unified Accelerator for Convolutional Neural Network and Vision Transformer
    Li T.
    Zhang F.
    Wang S.
    Cao W.
    Chen L.
    Dianzi Yu Xinxi Xuebao/Journal of Electronics and Information Technology, 2024, 46 (06): : 2663 - 2672
  • [6] An FPGA-based accelerator for deep neural network with novel reconfigurable architecture
    Jia, Han
    Ren, Daming
    Zou, Xuecheng
    IEICE ELECTRONICS EXPRESS, 2021, 18 (04):
  • [7] FPGA-based Training Accelerator Utilizing Sparseness of Convolutional Neural Network
    Nakahara, Hiroki
    Sada, Youki
    Shimoda, Masayuki
    Sayama, Kouki
    Jinguji, Akira
    Sato, Shimpei
    2019 29TH INTERNATIONAL CONFERENCE ON FIELD-PROGRAMMABLE LOGIC AND APPLICATIONS (FPL), 2019, : 180 - 186
  • [8] An FPGA-Based Computation-Efficient Convolutional Neural Network Accelerator
    Archana, V. S.
    2022 IEEE INTERNATIONAL POWER AND RENEWABLE ENERGY CONFERENCE, IPRECON, 2022,
  • [9] A High-Efficiency FPGA-Based Accelerator for Binarized Neural Network
    Guo, Peng
    Ma, Hong
    Chen, Ruizhi
    Wang, Donglin
    JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2019, 28
  • [10] An Efficient FPGA-Based Dilated and Transposed Convolutional Neural Network Accelerator
    Wu, Tsung-Hsi
    Shu, Chang
    Liu, Tsung-Te
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2024, 71 (11) : 5178 - 5186