Optimize FPGA-based Neural Network Accelerator with Bit-shift Quantization

被引:0
|
作者
Liu, Yu [1 ]
Liu, XueJiao [1 ]
Liang, Luhong [1 ]
机构
[1] Hong Kong Appl Sci & Technol Res Inst ASTRI, Hong Kong, Peoples R China
关键词
D O I
10.1109/iscas45731.2020.9180919
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Well-programmed Field Programmable Gate Arrays (FPGAs) can accelerate Deep Neural Network (DNN) with high power efficiency. The dominant workloads of DNNs are Multiply Accumulates (MACs), which can be directly mapped to Digital Signal Processors (DSPs) in the FPGA. A DNN accelerator pursuing high performance can consume almost all the DSPs, but with a considerable amount of Look-up Tables (LUTs) in the FPGA unused or performing MACs inefficiently. To solve this problem, we present a Bit-Shift method for FPGA-based DNN accelerator to fully utilize the resources in the FPGA. The MAC is converted to a limited number of shift-and-add operations, which can be implemented by LUTs with significant improvement of efficiency. A quantization method based on Minimum Mean Absolute Error (MMAE) is proposed to preserve the accuracy of the DNN inference in the conversion of DNN parameters without re-training. The quantized parameters can be compressed to a fixed and fewer number of bits to reduce the memory bandwidth. Accordingly, a Bit-Shift architecture is designed to load the compressed parameters and perform the converted MAC calculations without extra decompression module. A large scale DNN accelerator with the proposed Bit-Shift architecture is implemented in a Xilinx VU095 FPGA. Experimental results show that the proposed method can boost the processing speed by 32% and reach 331 GOPS at 190MHz clock frequency for ResNet-34.
引用
收藏
页数:5
相关论文
共 50 条
  • [41] A hardware-efficient computing engine for FPGA-based deep convolutional neural network accelerator
    Li, Xueming
    Huang, Hongmin
    Chen, Taosheng
    Gao, Huaien
    Hu, Xianghong
    Xiong, Xiaoming
    MICROELECTRONICS JOURNAL, 2022, 128
  • [42] An FPGA-based MobileNet Accelerator Considering Network Structure Characteristics
    Yan, Shun
    Liu, Zhengyan
    Wang, Yun
    Zeng, Chenglong
    Liu, Qiang
    Cheng, Bowen
    Cheung, Ray C. C.
    2021 31ST INTERNATIONAL CONFERENCE ON FIELD-PROGRAMMABLE LOGIC AND APPLICATIONS (FPL 2021), 2021, : 17 - 23
  • [43] An FPGA-Based Reconfigurable Convolutional Neural Network Accelerator for Tiny YOLO-V3
    Tsai, Tsung-Han
    Tung, Nai-Chieh
    Chen, Chun-Yu
    CIRCUITS SYSTEMS AND SIGNAL PROCESSING, 2025, : 3388 - 3409
  • [44] Deep Neural Network Accelerator based on FPGA
    Thang Viet Huynh
    2017 4TH NAFOSTED CONFERENCE ON INFORMATION AND COMPUTER SCIENCE (NICS), 2017, : 254 - 257
  • [45] FPGA-Based Deep Convolutional Neural Network Accelerator Design Techniques for the Handwritten Number Recognizer
    Yoo, Yechan
    Park, Yoonjin
    Kim, Injung
    Yi, Kang
    ADVANCED SCIENCE LETTERS, 2018, 24 (03) : 2152 - 2155
  • [46] Design of FPGA-Based Accelerator for Convolutional Neural Network under Heterogeneous Computing Framework with OpenCL
    Luo, Li
    Wu, Yakun
    Qiao, Fei
    Yang, Yi
    Wei, Qi
    Zhou, Xiaobo
    Fan, Yongkai
    Xu, Shuzheng
    Liu, Xinjun
    Yang, Huazhong
    INTERNATIONAL JOURNAL OF RECONFIGURABLE COMPUTING, 2018, 2018
  • [47] An energy-efficient convolutional neural network accelerator for speech classification based on FPGA and quantization
    Dong Wen
    Jingfei Jiang
    Yong Dou
    Jinwei Xu
    Tao Xiao
    CCF Transactions on High Performance Computing, 2021, 3 : 4 - 16
  • [48] An energy-efficient convolutional neural network accelerator for speech classification based on FPGA and quantization
    Wen, Dong
    Jiang, Jingfei
    Dou, Yong
    Xu, Jinwei
    Xiao, Tao
    CCF TRANSACTIONS ON HIGH PERFORMANCE COMPUTING, 2021, 3 (01) : 4 - 16
  • [49] Low-complexity JPEG quantization table requiring only bit-shift operations
    Araar, Chaouki
    Chabbi, Samir
    IMAGING SCIENCE JOURNAL, 2022, 70 (08): : 556 - 563
  • [50] High-Performance FPGA-based Accelerator for Bayesian Neural Networks
    Fan, Hongxiang
    Ferianc, Martin
    Rodrigues, Miguel
    Zhou, Hongyu
    Niu, Xinyu
    Luk, Wayne
    2021 58TH ACM/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2021, : 1063 - 1068