Stability Based Filter Pruning for Accelerating Deep CNNs

被引:22
|
作者
Singh, Pravendra [1 ]
Kadi, Vinay Sameer Raja [2 ]
Verma, Nikhil [2 ]
Namboodiri, Vinay P. [1 ]
机构
[1] IIT Kanpur, Kanpur, Uttar Pradesh, India
[2] Samsung R&D Inst, Delhi, India
来源
2019 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV) | 2019年
关键词
D O I
10.1109/WACV.2019.00129
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Convolutional neural networks (CNN) have achieved impressive performance on the wide variety of tasks (classification, detection, etc.) across multiple domains at the cost of high computational and memory requirements. Thus, leveraging CNNs for real-time applications necessitates model compression approaches that not only reduce the total number of parameters but reduce the overall computation as well. In this work, we present a stability-based approach for filter-level pruning of CNNs. We evaluate our proposed approach on different architectures (LeNet, VGG-16, ResNet, and Faster RCNN) and datasets and demonstrate its generalizability through extensive experiments. Moreover, our compressed models can be used at run-time without requiring any special libraries or hardware. Our model compression method reduces the number of FLOPS by an impressive factor of 6.03X and GPU memory footprint by more than 17X, significantly outperforming other state-of-the-art filter pruning methods.
引用
收藏
页码:1166 / 1174
页数:9
相关论文
共 50 条
  • [21] Accelerating Convolutional Networks via Global & Dynamic Filter Pruning
    Lin, Shaohui
    Ji, Rongrong
    Li, Yuchao
    Wu, Yongjian
    Huang, Feiyue
    Zhang, Baochang
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 2425 - 2432
  • [22] Channel Pruning for Accelerating Very Deep Neural Networks
    He, Yihui
    Zhang, Xiangyu
    Sun, Jian
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 1398 - 1406
  • [23] SFP: Similarity-based filter pruning for deep neural networks
    Li, Guoqing
    Li, Rengang
    Li, Tuo
    Shen, Chaoyao
    Zou, Xiaofeng
    Wang, Jiuyang
    Wang, Changhong
    Li, Nanjun
    INFORMATION SCIENCES, 2025, 689
  • [24] Robust pruning for efficient CNNs A
    Ide, Hidenori
    Kobayashi, Takumi
    Watanabe, Kenji
    Kurita, Takio
    PATTERN RECOGNITION LETTERS, 2020, 135 : 90 - 98
  • [25] Compressing CNNs Using Multilevel Filter Pruning for the Edge Nodes of Multimedia Internet of Things
    Liu, Xingang
    Wu, Lishuai
    Dai, Cheng
    Chao, Han-Chieh
    IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (14) : 11041 - 11051
  • [26] Soft Taylor Pruning for Accelerating Deep Convolutional Neural Networks
    Rong, Jintao
    Yu, Xiyi
    Zhang, Mingyang
    Ou, Linlin
    IECON 2020: THE 46TH ANNUAL CONFERENCE OF THE IEEE INDUSTRIAL ELECTRONICS SOCIETY, 2020, : 5343 - 5349
  • [27] Accelerating Deep Unsupervised Domain Adaptation with Transfer Channel Pruning
    Yu, Chaohui
    Wang, Jindong
    Chen, Yidiang
    Wu, Zijing
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [28] CURATING: A multi-objective based pruning technique for CNNs
    Pattanayak, Santanu
    Nag, Subhrajit
    Mittal, Sparsh
    JOURNAL OF SYSTEMS ARCHITECTURE, 2021, 116
  • [29] Thermodynamics modeling of deep learning systems for a temperature based filter pruning technique
    Lapenna, M.
    Faglioni, F.
    Fioresi, R.
    FRONTIERS IN PHYSICS, 2023, 11
  • [30] Deep neural network compression through interpretability-based filter pruning
    Yao, Kaixuan
    Cao, Feilong
    Leung, Yee
    Liang, Jiye
    PATTERN RECOGNITION, 2021, 119