Filter pruning with a feature map entropy importance criterion for convolution neural networks compressing

被引:30
|
作者
Wang, Jielei [1 ]
Jiang, Ting [2 ]
Cui, Zongyong [1 ]
Cao, Zongjie [1 ]
机构
[1] Univ Elect Sci & Technol China, Chengdu 611731, Peoples R China
[2] Megvii Technol Ltd, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Convolutional neural network; Model compression; Model pruning; Model acceleration; Entropy; GRADIENT;
D O I
10.1016/j.neucom.2021.07.034
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep Neural Networks (DNN) has made significant progress in recent years. However, its high computing and storage costs make it challenging to apply on resource-limited platforms or edge computation scenarios. Recent studies have shown that model pruning is an effective method to solve this problem. Typically, the model pruning method is a three-stage pipeline: training, pruning, and fine-tuning. In this work, a novel structured pruning method for Convolutional Neural Networks (CNN) compression is proposed, where filter-level redundant weights are pruned according to entropy importance criteria (termed FPEI). In short, the FPEI criterion, which works in the stage of pruning, defines the importance of the filter according to the entropy of feature maps. If a feature map contains very little information, it should not contribute much to the whole network. By removing these uninformative feature maps, their corresponding filters in the current layer and kernels in the next layer can be removed simultaneously. Consequently, the computing and storage costs are significantly reduced. Moreover, because our method cannot show the advantages of the existing ResNet pruning strategy, we propose a dimensionality reduction (DR) pruning strategy for ResNet structured networks. Experiments on several datasets demonstrate that our method is effective. In the experiment about the VGG-16 model on the SVHN dataset, we removed 91.31% of the parameters, from 14.73M to 1.28M, achieving a 63.77% reduction in the FLOPs, from 313.4M to 113.5M, and 1.73 times speedups of model inference. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页码:41 / 54
页数:14
相关论文
共 50 条
  • [21] Entropy Induced Pruning Framework for Convolutional Neural Networks
    Lu, Yiheng
    Guan, Ziyu
    Yang, Yaming
    Zhao, Wei
    Gong, Maoguo
    Xu, Cai
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 4, 2024, : 3918 - 3926
  • [22] Pruning feature maps for efficient convolutional neural networks
    Guo, Xiao-ting
    Xie, Xin-shu
    Lang, Xun
    OPTIK, 2023, 281
  • [23] On the Information of Feature Maps and Pruning of Deep Neural Networks
    Soltani, Mohammadreza
    Wu, Suya
    Ding, Jie
    Ravier, Robert
    Tarokh, Vahid
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 6988 - 6995
  • [24] Holistic Filter Pruning for Efficient Deep Neural Networks
    Enderich, Lukas
    Timm, Fabian
    Burgard, Wolfram
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WACV 2021, 2021, : 2595 - 2604
  • [25] Pruning Convolution Neural Network (SqueezeNet) using Taylor Expansion-based Criterion
    Gaikwad, Akash Sunil
    E-Sharkawy, Mohamed
    2018 IEEE INTERNATIONAL SYMPOSIUM ON SIGNAL PROCESSING AND INFORMATION TECHNOLOGY (ISSPIT), 2018,
  • [26] Filter pruning-based two-step feature map reconstruction
    Liang, Yongsheng
    Liu, Wei
    Yi, Shuangyan
    Yang, Huoxiang
    He, Zhenyu
    SIGNAL IMAGE AND VIDEO PROCESSING, 2021, 15 (07) : 1555 - 1563
  • [27] Learning compact ConvNets through filter pruning based on the saliency of a feature map
    Liu, Zhoufeng
    Liu, Xiaohui
    Li, Chunlei
    Ding, Shumin
    Liao, Liang
    IET IMAGE PROCESSING, 2022, 16 (01) : 123 - 133
  • [28] Filter pruning-based two-step feature map reconstruction
    Yongsheng Liang
    Wei Liu
    Shuangyan Yi
    Huoxiang Yang
    Zhenyu He
    Signal, Image and Video Processing, 2021, 15 : 1555 - 1563
  • [29] OPQ: Compressing Deep Neural Networks with One-shot Pruning-Quantization
    Hu, Peng
    Peng, Xi
    Zhu, Hongyuan
    Aly, Mohamed M. Sabry
    Lin, Jie
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 7780 - 7788
  • [30] Entropy-based pruning method for convolutional neural networks
    Hur, Cheonghwan
    Kang, Sanggil
    JOURNAL OF SUPERCOMPUTING, 2019, 75 (06): : 2950 - 2963