Optimizing the Deep Neural Networks by Layer-Wise Refined Pruning and the Acceleration on FPGA

被引:16
|
作者
Li, Hengyi [1 ]
Yue, Xuebin [1 ]
Wang, Zhichen [1 ]
Chai, Zhilei [2 ]
Wang, Wenwen [3 ]
Tomiyama, Hiroyuki [1 ]
Meng, Lin [1 ]
机构
[1] Ritsumeikan Univ, Dept Elect & Comp Engn, Kusatsu, Shiga, Japan
[2] Jiangnan Univ, Sch AI & Comp Sci, Wuxi, Peoples R China
[3] Univ Georgia, Dept Comp Sci, Athens, GA USA
关键词
MODEL;
D O I
10.1155/2022/8039281
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
To accelerate the practical applications of artificial intelligence, this paper proposes a high efficient layer-wise refined pruning method for deep neural networks at the software level and accelerates the inference process at the hardware level on a field-programmable gate array (FPGA). The refined pruning operation is based on the channel-wise importance indexes of each layer and the layer-wise input sparsity of convolutional layers. The method utilizes the characteristics of the native networks without introducing any extra workloads to the training phase. In addition, the operation is easy to be extended to various state-of-the-art deep neural networks. The effectiveness of the method is verified on ResNet architecture and VGG networks in terms of dataset CIFAR10, CIFAR100, and ImageNet100. Experimental results show that in terms of ResNet50 on CIFAR10 and ResNet101 on CIFAR100, more than 85% of parameters and Floating-Point Operations are pruned with only 0.35% and 0.40% accuracy loss, respectively. As for the VGG network, 87.05% of parameters and 75.78% of Floating-Point Operations are pruned with only 0.74% accuracy loss for VGG13BN on CIFAR10. Furthermore, we accelerate the networks at the hardware level on the FPGA platform by utilizing the tool Vitis AI. For two threads mode in FPGA, the throughput/fps of the pruned VGG13BN and ResNet101 achieves 151.99 fps and 124.31 fps, respectively, and the pruned networks achieve about 4.3x and 1.8x speed up for VGG13BN and ResNet101, respectively, compared with the original networks on FPGA.
引用
收藏
页数:22
相关论文
共 50 条
  • [1] Differential evolution based layer-wise weight pruning for compressing deep neural networks
    Wu, Tao
    Li, Xiaoyang
    Zhou, Deyun
    Li, Na
    Shi, Jiao
    Sensors (Switzerland), 2021, 21 (03): : 1 - 20
  • [2] Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks
    Wu, Tao
    Li, Xiaoyang
    Zhou, Deyun
    Li, Na
    Shi, Jiao
    SENSORS, 2021, 21 (03) : 1 - 20
  • [3] Layer-Wise Weight Decay for Deep Neural Networks
    Ishii, Masato
    Sato, Atsushi
    IMAGE AND VIDEO TECHNOLOGY (PSIVT 2017), 2018, 10749 : 276 - 289
  • [4] Stochastic Layer-Wise Precision in Deep Neural Networks
    Lacey, Griffin
    Taylor, Graham W.
    Areibi, Shawki
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2018, : 663 - 672
  • [5] Shallowing Deep Networks: Layer-wise Pruning based on Feature Representations
    Chen, Shi
    Zhao, Qi
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2019, 41 (12) : 3048 - 3056
  • [6] Pruning Ratio Optimization with Layer-Wise Pruning Method for Accelerating Convolutional Neural Networks
    Kamma, Koji
    Inoue, Sarimu
    Wada, Toshikazu
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2022, E105D (01) : 161 - 169
  • [7] LAYER-WISE DEEP NEURAL NETWORK PRUNING VIA ITERATIVELY REWEIGHTED OPTIMIZATION
    Jiang, Tao
    Yang, Xiangyu
    Shi, Yuanming
    Wang, Hao
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 5606 - 5610
  • [8] Unsupervised Layer-Wise Model Selection in Deep Neural Networks
    Ludovic, Arnold
    Helene, Paugam-Moisy
    Michele, Sebag
    ECAI 2010 - 19TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2010, 215 : 915 - 920
  • [9] Collaborative Layer-Wise Discriminative Learning in Deep Neural Networks
    Jin, Xiaojie
    Chen, Yunpeng
    Dong, Jian
    Feng, Jiashi
    Yan, Shuicheng
    COMPUTER VISION - ECCV 2016, PT VII, 2016, 9911 : 733 - 749
  • [10] Post-training deep neural network pruning via layer-wise calibration
    Lazarevich, Ivan
    Kozlov, Alexander
    Malinin, Nikita
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2021), 2021, : 798 - 805