FxP-QNet: A Post-Training Quantizer for the Design of Mixed Low-Precision DNNs With Dynamic Fixed-Point Representation

被引:4
|
作者
Shawahna, Ahmad [1 ]
Sait, Sadiq M. [1 ,2 ]
El-Maleh, Aiman [1 ,2 ]
Ahmad, Irfan [2 ,3 ]
机构
[1] King Fand Univ Petr & Minerals, Dept Comp Engn, Dhahran 31261, Saudi Arabia
[2] King Fand Univ Petr & Minerals, Interdisciplinary Res Ctr Intelligent Secure Syst, Dhahran 31261, Saudi Arabia
[3] King Fand Univ Petr & Minerals, Informat & Comp Sci Dept, Dhahran 31261, Saudi Arabia
关键词
Neural networks; model compression; deep learning; quantization; fixed-point arithmetic; mixed-precision; acceleration; accuracy; efficient inference; resource-constrained devices; NEURAL-NETWORKS; CNN;
D O I
10.1109/ACCESS.2022.3157893
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks (DNNs) have demonstrated their effectiveness in a wide range of computer vision tasks, with the state-of-the-art results obtained through complex and deep structures that require intensive computation and memory. In the past, graphic processing units enabled these breakthroughs because of their greater computational speed. Now-a-days, efficient model inference is crucial for consumer applications on resource-constrained platforms. As a result, there is much interest in the research and development of dedicated deep learning (DL) hardware to improve the throughput and energy efficiency of DNNs. Low-precision representation of DNN data-structures through quantization would bring great benefits to specialized DL hardware especially when expensive floating-point operations can be avoided and replaced by more efficient fixed-point operations. However, the rigorous quantization leads to a severe accuracy drop. As such, quantization opens a large hyper-parameter space at bit-precision levels, the exploration of which is a major challenge. In this paper, we propose a novel framework referred to as the Fixed-Point Quantizer of deep neural Networks (FxP-QNet) that flexibly designs a mixed low-precision DNN for integer-arithmetic-only deployment. Specifically, the FxP-QNet gradually adapts the quantization level for each data-structure of each layer based on the trade-off between the network accuracy and the low-precision requirements. Additionally, it employs post-training self-distillation and network prediction error statistics to optimize the quantization of floating-point values into fixed-point numbers. Examining FxP-QNet on state-of-the-art architectures and the benchmark ImageNet dataset, we empirically demonstrate the effectiveness of FxP-QNet in achieving the accuracy-compression trade-off without the need for training. The results show that FxP-QNet-quantized AlexNet, VGG-16, and ResNet-18 reduce the overall memory requirements of their full-precision counterparts by 7.16x, 10.36x, and 6.44x with less than 0.95%, 0.95%, and 1.99% accuracy drop, respectively.
引用
收藏
页码:30202 / 30231
页数:30
相关论文
共 1 条
  • [1] Training Neural Networks with Low Precision Dynamic Fixed-Point
    Jo, Sujeong
    Park, Hanmin
    Lee, Gunhee
    Choi, Kiyoung
    2018 IEEE 36TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD), 2018, : 405 - 408