A "Network Pruning Network" Approach to Deep Model Compression

被引:0
|
作者
Verma, Vinay Kumar [1 ]
Singh, Pravendra [1 ]
Namboodiri, Vinay P. [1 ]
Rai, Piyush [1 ]
机构
[1] IIT Kanpur, Dept Comp Sci & Engn, Kanpur, Uttar Pradesh, India
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a filter pruning approach for deep model compression, using a multitask network. Our approach is based on learning a a pruner network to prune a pre-trained target network. The pruner is essentially a multitask deep neural network with binary outputs that help identify the filters from each layer of the original network that do not have any significant contribution to the model and can therefore be pruned. The pruner network has the same architecture as the original network except that it has a multitask/multi-output last layer containing binary-valued outputs (one per filter), which indicate which filters have to be pruned. The pruner's goal is to minimize the number of filters from the original network by assigning zero weights to the corresponding output feature-maps. In contrast to most of the existing methods, instead of relying on iterative pruning, our approach can prune the network (original network) in one go and, moreover, does not require specifying the degree of pruning for each layer (and can learn it instead). The compressed model produced by our approach is generic and does not need any special hardware/software support. Moreover, augmenting with other methods such as knowledge distillation, quantization, and connection pruning can increase the degree of compression for the proposed approach. We show the efficacy of our proposed approach for classification and object detection tasks.
引用
收藏
页码:2998 / 3007
页数:10
相关论文
共 50 条
  • [21] CCAP: Cooperative Context Aware Pruning for Neural Network Model Compression
    Wang, Li-Yun
    Akhtar, Zahid
    23RD IEEE INTERNATIONAL SYMPOSIUM ON MULTIMEDIA (ISM 2021), 2021, : 257 - 260
  • [22] A Neural Network Model Compression Approach Based on Deep Feature Map Transfer
    Guo, Zhibo
    Yao, Xin
    Xu, Yixuan
    Zhang, Ying
    Wang, Linghao
    IEEE ACCESS, 2020, 8 : 158026 - 158035
  • [23] DEEP NETWORK PRUNING FOR OBJECT DETECTION
    Ghosh, Sanjukta
    Srinivasa, Shashi K. K.
    Amon, Peter
    Hutter, Andreas
    Kaup, Andre
    2019 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2019, : 3915 - 3919
  • [24] Quantisation and Pruning for Neural Network Compression and Regularisation
    Paupamah, Kimessha
    James, Steven
    Klein, Richard
    2020 INTERNATIONAL SAUPEC/ROBMECH/PRASA CONFERENCE, 2020, : 295 - 300
  • [25] ON THE ROLE OF STRUCTURED PRUNING FOR NEURAL NETWORK COMPRESSION
    Bragagnolo, Andrea
    Tartaglione, Enzo
    Fiandrotti, Attilio
    Grangetto, Marco
    2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 3527 - 3531
  • [26] Group Fisher Pruning for Practical Network Compression
    Liu, Liyang
    Zhang, Shilong
    Kuang, Zhanghui
    Zhou, Aojun
    Xue, Jing-Hao
    Wang, Xinjiang
    Chen, Yimin
    Yang, Wenming
    Liao, Qingmin
    Zhang, Wayne
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [27] Neural Network Compression and Acceleration by Federated Pruning
    Pei, Songwen
    Wu, Yusheng
    Qiu, Meikang
    ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2020, PT II, 2020, 12453 : 173 - 183
  • [28] NFP: A No Fine-tuning Pruning Approach for Convolutional Neural Network Compression
    Liu, Rui
    Cao, Jian
    Li, Pu
    Sun, Wenyu
    Zhang, Ying
    Wang, Yuan
    2020 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND BIG DATA (ICAIBD 2020), 2020, : 74 - 77
  • [29] HFPQ: deep neural network compression by hardware-friendly pruning-quantization
    YingBo Fan
    Wei Pang
    ShengLi Lu
    Applied Intelligence, 2021, 51 : 7016 - 7028
  • [30] HFPQ: deep neural network compression by hardware-friendly pruning-quantization
    Fan, YingBo
    Pang, Wei
    Lu, ShengLi
    APPLIED INTELLIGENCE, 2021, 51 (10) : 7016 - 7028