Optimizing the Deep Neural Networks by Layer-Wise Refined Pruning and the Acceleration on FPGA

被引:16
|
作者
Li, Hengyi [1 ]
Yue, Xuebin [1 ]
Wang, Zhichen [1 ]
Chai, Zhilei [2 ]
Wang, Wenwen [3 ]
Tomiyama, Hiroyuki [1 ]
Meng, Lin [1 ]
机构
[1] Ritsumeikan Univ, Dept Elect & Comp Engn, Kusatsu, Shiga, Japan
[2] Jiangnan Univ, Sch AI & Comp Sci, Wuxi, Peoples R China
[3] Univ Georgia, Dept Comp Sci, Athens, GA USA
关键词
MODEL;
D O I
10.1155/2022/8039281
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
To accelerate the practical applications of artificial intelligence, this paper proposes a high efficient layer-wise refined pruning method for deep neural networks at the software level and accelerates the inference process at the hardware level on a field-programmable gate array (FPGA). The refined pruning operation is based on the channel-wise importance indexes of each layer and the layer-wise input sparsity of convolutional layers. The method utilizes the characteristics of the native networks without introducing any extra workloads to the training phase. In addition, the operation is easy to be extended to various state-of-the-art deep neural networks. The effectiveness of the method is verified on ResNet architecture and VGG networks in terms of dataset CIFAR10, CIFAR100, and ImageNet100. Experimental results show that in terms of ResNet50 on CIFAR10 and ResNet101 on CIFAR100, more than 85% of parameters and Floating-Point Operations are pruned with only 0.35% and 0.40% accuracy loss, respectively. As for the VGG network, 87.05% of parameters and 75.78% of Floating-Point Operations are pruned with only 0.74% accuracy loss for VGG13BN on CIFAR10. Furthermore, we accelerate the networks at the hardware level on the FPGA platform by utilizing the tool Vitis AI. For two threads mode in FPGA, the throughput/fps of the pruned VGG13BN and ResNet101 achieves 151.99 fps and 124.31 fps, respectively, and the pruned networks achieve about 4.3x and 1.8x speed up for VGG13BN and ResNet101, respectively, compared with the original networks on FPGA.
引用
收藏
页数:22
相关论文
共 50 条
  • [21] Exploiting potential of deep neural networks by layer-wise fine-grained parallelism
    Jiang, Wenbin
    Zhang, Yangsong
    Liu, Pai
    Peng, Jing
    Yang, Laurence T.
    Ye, Geyan
    Jin, Hai
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2020, 102 : 210 - 221
  • [22] Learning to Prune Deep Neural Networks via Layer-wise Optimal Brain Surgeon
    Dong, Xin
    Chen, Shangyu
    Pan, Sinno Jialin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [23] REINFORCEMENT LEARNING-BASED LAYER-WISE QUANTIZATION FOR LIGHTWEIGHT DEEP NEURAL NETWORKS
    Jung, Juri
    Kim, Jonghee
    Kim, Youngeun
    Kim, Changick
    2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 3070 - 3074
  • [24] Optimizing Loop Operation and Dataflow in FPGA Acceleration of Deep Convolutional Neural Networks
    Ma, Yufei
    Cao, Yu
    Vrudhula, Sarma
    Seo, Jae-sun
    FPGA'17: PROCEEDINGS OF THE 2017 ACM/SIGDA INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE GATE ARRAYS, 2017, : 45 - 54
  • [25] AutoLR: Layer-wise Pruning and Auto-tuning of Learning Rates in Fine-tuning of Deep Networks
    Ro, Younmgin
    Choi, Jin Young
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 2486 - 2494
  • [26] An FSCV Deep Neural Network: Development, Pruning, and Acceleration on an FPGA
    Zhang, Zhichao
    Oh, Yoonbae
    Adams, Scott D.
    Bennet, Kevin E.
    Kouzani, Abbas Z.
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2021, 25 (06) : 2248 - 2259
  • [27] Stochastic Neural Networks with Layer-Wise Adjustable Sequence Length
    Wang, Ziheng
    Reviriego, Pedro
    Niknia, Farzad
    Liu, Shanshan
    Gao, Zhen
    Lombardi, Fabrizio
    2024 IEEE 24TH INTERNATIONAL CONFERENCE ON NANOTECHNOLOGY, NANO 2024, 2024, : 436 - 441
  • [28] Layer-Wise Training to Create Efficient Convolutional Neural Networks
    Zeng, Linghua
    Tian, Xinmei
    NEURAL INFORMATION PROCESSING (ICONIP 2017), PT II, 2017, 10635 : 631 - 641
  • [29] NeuronMotif: Deciphering cis-regulatory codes by layer-wise demixing of deep neural networks
    Wei, Zheng
    Hua, Kui
    Wei, Lei
    Ma, Shining
    Jiang, Rui
    Zhang, Xuegong
    Li, Yanda
    Wong, Wing H.
    Wang, Xiaowo
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2023, 120 (15)
  • [30] A Dynamic Layer-Wise Gradient Sparsity and Gradient Merging Optimization Method for Deep Neural Networks
    Ju, Tao
    Kang, Heting
    Liu, Shuai
    Huo, Jiuyuan
    Hsi-An Chiao Tung Ta Hsueh/Journal of Xi'an Jiaotong University, 2024, 58 (09): : 105 - 116