A "Network Pruning Network" Approach to Deep Model Compression

被引:0
|
作者
Verma, Vinay Kumar [1 ]
Singh, Pravendra [1 ]
Namboodiri, Vinay P. [1 ]
Rai, Piyush [1 ]
机构
[1] IIT Kanpur, Dept Comp Sci & Engn, Kanpur, Uttar Pradesh, India
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a filter pruning approach for deep model compression, using a multitask network. Our approach is based on learning a a pruner network to prune a pre-trained target network. The pruner is essentially a multitask deep neural network with binary outputs that help identify the filters from each layer of the original network that do not have any significant contribution to the model and can therefore be pruned. The pruner network has the same architecture as the original network except that it has a multitask/multi-output last layer containing binary-valued outputs (one per filter), which indicate which filters have to be pruned. The pruner's goal is to minimize the number of filters from the original network by assigning zero weights to the corresponding output feature-maps. In contrast to most of the existing methods, instead of relying on iterative pruning, our approach can prune the network (original network) in one go and, moreover, does not require specifying the degree of pruning for each layer (and can learn it instead). The compressed model produced by our approach is generic and does not need any special hardware/software support. Moreover, augmenting with other methods such as knowledge distillation, quantization, and connection pruning can increase the degree of compression for the proposed approach. We show the efficacy of our proposed approach for classification and object detection tasks.
引用
收藏
页码:2998 / 3007
页数:10
相关论文
共 50 条
  • [31] Pruning by explaining: A novel criterion for deep neural network pruning
    Yeom, Seul-Ki
    Seegerer, Philipp
    Lapuschkin, Sebastian
    Binder, Alexander
    Wiedemann, Simon
    Mueller, Klaus-Robert
    Samek, Wojciech
    PATTERN RECOGNITION, 2021, 115
  • [32] DEEP NEURAL NETWORK MODEL COMPRESSION AND SIGNAL PROCESSING
    Ukil, Arijit
    Majumdar, Angshul
    Jara, Antonio J.
    Gama, Joao
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING WORKSHOPS, ICASSPW 2024, 2024, : 179 - 183
  • [33] A Probabilistic Approach to Neural Network Pruning
    Qian, Xin
    Klabjan, Diego
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [34] An Efficient Approach to Iterative Network Pruning
    Huang, Chuan-Shun
    Tang, Wuqian
    Chen, Yung-Chih
    Li, Yi-Ting
    Chang, Shih-Chieh
    Wang, Chun-Yao
    2024 INTERNATIONAL VLSI SYMPOSIUM ON TECHNOLOGY, SYSTEMS AND APPLICATIONS, VLSI TSA, 2024,
  • [35] Pruning the deep neural network by similar function
    Liu, Hanqing
    Xin, Bo
    Mu, Senlin
    Zhu, Zhangqing
    2018 INTERNATIONAL SYMPOSIUM ON POWER ELECTRONICS AND CONTROL ENGINEERING (ISPECE 2018), 2019, 1187
  • [36] Overview of Deep Convolutional Neural Network Pruning
    Li, Guang
    Liu, Fang
    Xia, Yuping
    2020 INTERNATIONAL CONFERENCE ON IMAGE, VIDEO PROCESSING AND ARTIFICIAL INTELLIGENCE, 2020, 11584
  • [37] Deep Capsule Network Based on Pruning Optimization
    Zheng X.-P.
    Liang X.
    Jisuanji Xuebao/Chinese Journal of Computers, 2022, 45 (07): : 1557 - 1570
  • [38] Revisiting Random Channel Pruning for Neural Network Compression
    Li, Yawei
    Adamczewski, Kamil
    Li, Wen
    Gu, Shuhang
    Timofte, Radu
    Van Gool, Luc
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 191 - 201
  • [39] Enhanced Network Compression Through Tensor Decompositions and Pruning
    Pham, Van Tien
    Zniyed, Yassine
    Nguyen, Thanh Phuong
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, : 1 - 13
  • [40] Enhanced Network Compression Through Tensor Decompositions and Pruning
    Pham, Van Tien
    Zniyed, Yassine
    Thanh Phuong Nguyen
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (03) : 4358 - 4370