Filter pruning with a feature map entropy importance criterion for convolution neural networks compressing

被引:30
|
作者
Wang, Jielei [1 ]
Jiang, Ting [2 ]
Cui, Zongyong [1 ]
Cao, Zongjie [1 ]
机构
[1] Univ Elect Sci & Technol China, Chengdu 611731, Peoples R China
[2] Megvii Technol Ltd, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Convolutional neural network; Model compression; Model pruning; Model acceleration; Entropy; GRADIENT;
D O I
10.1016/j.neucom.2021.07.034
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep Neural Networks (DNN) has made significant progress in recent years. However, its high computing and storage costs make it challenging to apply on resource-limited platforms or edge computation scenarios. Recent studies have shown that model pruning is an effective method to solve this problem. Typically, the model pruning method is a three-stage pipeline: training, pruning, and fine-tuning. In this work, a novel structured pruning method for Convolutional Neural Networks (CNN) compression is proposed, where filter-level redundant weights are pruned according to entropy importance criteria (termed FPEI). In short, the FPEI criterion, which works in the stage of pruning, defines the importance of the filter according to the entropy of feature maps. If a feature map contains very little information, it should not contribute much to the whole network. By removing these uninformative feature maps, their corresponding filters in the current layer and kernels in the next layer can be removed simultaneously. Consequently, the computing and storage costs are significantly reduced. Moreover, because our method cannot show the advantages of the existing ResNet pruning strategy, we propose a dimensionality reduction (DR) pruning strategy for ResNet structured networks. Experiments on several datasets demonstrate that our method is effective. In the experiment about the VGG-16 model on the SVHN dataset, we removed 91.31% of the parameters, from 14.73M to 1.28M, achieving a 63.77% reduction in the FLOPs, from 313.4M to 113.5M, and 1.73 times speedups of model inference. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页码:41 / 54
页数:14
相关论文
共 50 条
  • [41] Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks
    Wu, Tao
    Li, Xiaoyang
    Zhou, Deyun
    Li, Na
    Shi, Jiao
    SENSORS, 2021, 21 (03) : 1 - 20
  • [42] Differential evolution based layer-wise weight pruning for compressing deep neural networks
    Wu, Tao
    Li, Xiaoyang
    Zhou, Deyun
    Li, Na
    Shi, Jiao
    Sensors (Switzerland), 2021, 21 (03): : 1 - 20
  • [43] Spectral Pruning: Compressing Deep Neural Networks via Spectral Analysis and its Generalization Error
    Suzuki, Taiji
    Abe, Hiroshi
    Murata, Tomoya
    Horiuchi, Shingo
    Ito, Kotaro
    Wachi, Tokuma
    Hirai, So
    Yukishima, Masatoshi
    Nishimura, Tomoaki
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 2839 - 2846
  • [44] The entropy of stationary solutions' map of cellular neural networks
    Wang, Y
    Chen, FY
    INTERNATIONAL JOURNAL OF BIFURCATION AND CHAOS, 2004, 14 (12): : 4317 - 4323
  • [45] Activation-Based Weight Significance Criterion for Pruning Deep Neural Networks
    Dong, Jiayu
    Zheng, Huicheng
    Lian, Lina
    IMAGE AND GRAPHICS (ICIG 2017), PT II, 2017, 10667 : 62 - 73
  • [46] Compressing Deep Convolutional Neural Networks by Stacking Low-dimensional Binary Convolution Filters
    Lan, Weichao
    Lan, Liang
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 8235 - 8242
  • [47] Orthophoto Map Feature Extraction Based on Neural Networks
    Horak, Zdenek
    Kudelka, Milos
    Snasel, Vaclay
    Vozenilek, Vit
    DATESO 2011: DATABASES, TEXTS, SPECIFICATIONS, OBJECTS, 2011, 706 : 216 - 225
  • [48] Auto-Balanced Filter Pruning for Efficient Convolutional Neural Networks
    Ding, Xiaohan
    Ding, Guiguang
    Han, Jungong
    Tang, Sheng
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 6797 - 6804
  • [49] Acceleration of Deep Convolutional Neural Networks Using Adaptive Filter Pruning
    Singh, Pravendra
    Verma, Vinay Kumar
    Rai, Piyush
    Namboodiri, Vinay P.
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2020, 14 (04) : 838 - 847
  • [50] Filter Pruning for Efficient Transfer Learning in Deep Convolutional Neural Networks
    Reinhold, Caique
    Roisenberg, Mauro
    ARTIFICIAL INTELLIGENCEAND SOFT COMPUTING, PT I, 2019, 11508 : 191 - 202