A "Network Pruning Network" Approach to Deep Model Compression

被引:0
|
作者
Verma, Vinay Kumar [1 ]
Singh, Pravendra [1 ]
Namboodiri, Vinay P. [1 ]
Rai, Piyush [1 ]
机构
[1] IIT Kanpur, Dept Comp Sci & Engn, Kanpur, Uttar Pradesh, India
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a filter pruning approach for deep model compression, using a multitask network. Our approach is based on learning a a pruner network to prune a pre-trained target network. The pruner is essentially a multitask deep neural network with binary outputs that help identify the filters from each layer of the original network that do not have any significant contribution to the model and can therefore be pruned. The pruner network has the same architecture as the original network except that it has a multitask/multi-output last layer containing binary-valued outputs (one per filter), which indicate which filters have to be pruned. The pruner's goal is to minimize the number of filters from the original network by assigning zero weights to the corresponding output feature-maps. In contrast to most of the existing methods, instead of relying on iterative pruning, our approach can prune the network (original network) in one go and, moreover, does not require specifying the degree of pruning for each layer (and can learn it instead). The compressed model produced by our approach is generic and does not need any special hardware/software support. Moreover, augmenting with other methods such as knowledge distillation, quantization, and connection pruning can increase the degree of compression for the proposed approach. We show the efficacy of our proposed approach for classification and object detection tasks.
引用
收藏
页码:2998 / 3007
页数:10
相关论文
共 50 条
  • [41] Hybrid network compression through tensor decompositions and pruning
    Van Tien Pham
    Zniyed, Yassine
    Thanh Phuong Nguyen
    32ND EUROPEAN SIGNAL PROCESSING CONFERENCE, EUSIPCO 2024, 2024, : 1052 - 1056
  • [42] CLIP-Q: Deep Network Compression Learning by In-Parallel Pruning-Quantization
    Tung, Frederick
    Mori, Greg
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 7873 - 7882
  • [43] An efficient deep learning model using network pruning for fake banknote recognition
    Pachon, Cesar G.
    Ballesteros, Dora M.
    Renza, Diego
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 233
  • [44] Network Implosion: Effective Model Compression for ResNets via Static Layer Pruning and Retraining
    Ida, Yasutoshi
    Fujiwara, Yasuhiro
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [45] Pruning Convolutional Neural Network with Distinctiveness Approach
    Li, Wenrui
    Plested, Jo
    NEURAL INFORMATION PROCESSING, ICONIP 2019, PT V, 2019, 1143 : 448 - 455
  • [46] Pruning and quantization for deep neural network acceleration: A survey
    Liang, Tailin
    Glossner, John
    Wang, Lei
    Shi, Shaobo
    Zhang, Xiaotong
    NEUROCOMPUTING, 2021, 461 : 370 - 403
  • [47] Linearly Replaceable Filters for Deep Network Channel Pruning
    Joo, Donggyu
    Yi, Eojindl
    Baek, Sunghyun
    Kim, Junmo
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 8021 - 8029
  • [48] Deep Neural Network Pruning Using Persistent Homology
    Watanabe, Satoru
    Yamana, Hayato
    2020 IEEE THIRD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND KNOWLEDGE ENGINEERING (AIKE 2020), 2020, : 153 - 156
  • [49] Literature Review of Deep Network Compression
    Alqahtani, Ali
    Xie, Xianghua
    Jones, Mark W.
    INFORMATICS-BASEL, 2021, 8 (04):
  • [50] Fuzzy qualitative deep compression network
    Tan, Wei Ren
    Chan, Chee Seng
    Aguirre, Hernan E.
    Tanaka, Kiyoshi
    NEUROCOMPUTING, 2017, 251 : 1 - 15