Optimizing the Deep Neural Networks by Layer-Wise Refined Pruning and the Acceleration on FPGA

被引:16
|
作者
Li, Hengyi [1 ]
Yue, Xuebin [1 ]
Wang, Zhichen [1 ]
Chai, Zhilei [2 ]
Wang, Wenwen [3 ]
Tomiyama, Hiroyuki [1 ]
Meng, Lin [1 ]
机构
[1] Ritsumeikan Univ, Dept Elect & Comp Engn, Kusatsu, Shiga, Japan
[2] Jiangnan Univ, Sch AI & Comp Sci, Wuxi, Peoples R China
[3] Univ Georgia, Dept Comp Sci, Athens, GA USA
关键词
MODEL;
D O I
10.1155/2022/8039281
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
To accelerate the practical applications of artificial intelligence, this paper proposes a high efficient layer-wise refined pruning method for deep neural networks at the software level and accelerates the inference process at the hardware level on a field-programmable gate array (FPGA). The refined pruning operation is based on the channel-wise importance indexes of each layer and the layer-wise input sparsity of convolutional layers. The method utilizes the characteristics of the native networks without introducing any extra workloads to the training phase. In addition, the operation is easy to be extended to various state-of-the-art deep neural networks. The effectiveness of the method is verified on ResNet architecture and VGG networks in terms of dataset CIFAR10, CIFAR100, and ImageNet100. Experimental results show that in terms of ResNet50 on CIFAR10 and ResNet101 on CIFAR100, more than 85% of parameters and Floating-Point Operations are pruned with only 0.35% and 0.40% accuracy loss, respectively. As for the VGG network, 87.05% of parameters and 75.78% of Floating-Point Operations are pruned with only 0.74% accuracy loss for VGG13BN on CIFAR10. Furthermore, we accelerate the networks at the hardware level on the FPGA platform by utilizing the tool Vitis AI. For two threads mode in FPGA, the throughput/fps of the pruned VGG13BN and ResNet101 achieves 151.99 fps and 124.31 fps, respectively, and the pruned networks achieve about 4.3x and 1.8x speed up for VGG13BN and ResNet101, respectively, compared with the original networks on FPGA.
引用
收藏
页数:22
相关论文
共 50 条
  • [41] Deep Layer-wise Networks Have Closed-Form Weights
    Wu, Chieh
    Masoomi, Aria
    Gretton, Arthur
    Dy, Jennifer
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151 : 188 - 225
  • [42] Compressing Neural Networks: Towards Determining the Optimal Layer-wise Decomposition
    Liebenuein, Lucas
    Maalouf, Alaa
    Gal, Oren
    Feldman, Dan
    Rus, Daniela
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [43] Effects of depth, width, and initialization: A convergence analysis of layer-wise training for deep linear neural networks
    Shin, Yeonjong
    ANALYSIS AND APPLICATIONS, 2022, 20 (01) : 73 - 119
  • [44] Network with Sub-networks: Layer-wise Detachable Neural Network
    Fuengfusin, Ninnart
    Tamukoh, Hakaru
    JOURNAL OF ROBOTICS NETWORKING AND ARTIFICIAL LIFE, 2021, 7 (04): : 240 - 244
  • [45] Layer-Wise Relevance Propagation for Neural Networks with Local Renormalization Layers
    Binder, Alexander
    Montavon, Gregoire
    Lapuschkin, Sebastian
    Mueller, Klaus-Robert
    Samek, Wojciech
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2016, PT II, 2016, 9887 : 63 - 71
  • [46] Straightforward Layer-Wise Pruning for More Efficient Visual Adaptation
    Han, Ruizi
    Tang, Jinglei
    COMPUTER VISION - ECCV 2024, PT LXXII, 2025, 15130 : 236 - 252
  • [47] Acceleration of Deep Recurrent Neural Networks with an FPGA cluster
    Sun, Yuxi
    Ben Ahmed, Akram
    Amano, Hideharu
    PROCEEDINGS OF THE 10TH INTERNATIONAL SYMPOSIUM ON HIGHLY EFFICIENT ACCELERATORS AND RECONFIGURABLE TECHNOLOGIES (HEART), 2019,
  • [48] AutoNet-Generated Deep Layer-Wise Convex Networks for ECG Classification
    Shen, Yanting
    Lu, Lei
    Zhu, Tingting
    Wang, Xinshao
    Clifton, Lei
    Chen, Zhengming
    Clarke, Robert
    Clifton, David A.
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (10) : 6542 - 6558
  • [49] Evaluating Layer-wise Relevance Propagation Explainability Maps for Artificial Neural Networks
    Ranguelova, Elena
    Pauwels, Eric J.
    Berkhout, Joost
    2018 IEEE 14TH INTERNATIONAL CONFERENCE ON E-SCIENCE (E-SCIENCE 2018), 2018, : 377 - 378
  • [50] Explanation of Multi-Label Neural Networks with Layer-Wise Relevance Propagation
    Bello, Marilyn
    Napoles, Gonzalo
    Vanhoof, Koen
    Garcia, Maria M.
    Bello, Rafael
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,