Stability Based Filter Pruning for Accelerating Deep CNNs

被引:22
|
作者
Singh, Pravendra [1 ]
Kadi, Vinay Sameer Raja [2 ]
Verma, Nikhil [2 ]
Namboodiri, Vinay P. [1 ]
机构
[1] IIT Kanpur, Kanpur, Uttar Pradesh, India
[2] Samsung R&D Inst, Delhi, India
来源
2019 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV) | 2019年
关键词
D O I
10.1109/WACV.2019.00129
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Convolutional neural networks (CNN) have achieved impressive performance on the wide variety of tasks (classification, detection, etc.) across multiple domains at the cost of high computational and memory requirements. Thus, leveraging CNNs for real-time applications necessitates model compression approaches that not only reduce the total number of parameters but reduce the overall computation as well. In this work, we present a stability-based approach for filter-level pruning of CNNs. We evaluate our proposed approach on different architectures (LeNet, VGG-16, ResNet, and Faster RCNN) and datasets and demonstrate its generalizability through extensive experiments. Moreover, our compressed models can be used at run-time without requiring any special libraries or hardware. Our model compression method reduces the number of FLOPS by an impressive factor of 6.03X and GPU memory footprint by more than 17X, significantly outperforming other state-of-the-art filter pruning methods.
引用
收藏
页码:1166 / 1174
页数:9
相关论文
共 50 条
  • [41] Pruning Filter in Filter
    Meng, Fanxu
    Cheng, Hao
    Li, Ke
    Luo, Huixiang
    Guo, Xiaowei
    Lu, Guangming
    Sun, Xing
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [42] A Novel Filter-Level Deep Convolutional Neural Network Pruning Method Based on Deep Reinforcement Learning
    Feng, Yihao
    Huang, Chao
    Wang, Long
    Luo, Xiong
    Li, Qingwen
    APPLIED SCIENCES-BASEL, 2022, 12 (22):
  • [43] Preserving the Essential Features in CNNs: Pruning and Analysis
    Lopez-Gonzalez, Clara, I
    Gomez-Silva, Maria J.
    Besada-Portas, Eva
    Pajares, Gonzalo
    ADVANCES IN ARTIFICIAL INTELLIGENCE, CAEPIA 2024, 2024, : 131 - 141
  • [44] DWMAcc: Accelerating Shift-based CNNs with Domain Wall Memories
    Chen, Zhengguo
    Deng, Quan
    Xiao, Nong
    Pruhs, Kirk
    Zhang, Youtao
    ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2019, 18 (05)
  • [45] Noise Floor Estimation Based on Deep CNNs
    Huang, Hao
    Li, Jianqing
    Wang, Jiao
    Wang, Hong
    2020 13TH INTERNATIONAL CONGRESS ON IMAGE AND SIGNAL PROCESSING, BIOMEDICAL ENGINEERING AND INFORMATICS (CISP-BMEI 2020), 2020, : 574 - 579
  • [46] Filter Pruning via Feature Discrimination in Deep Neural Networks
    He, Zhiqiang
    Qian, Yaguan
    Wang, Yuqi
    Wang, Bin
    Guan, Xiaohui
    Gu, Zhaoquan
    Ling, Xiang
    Zeng, Shaoning
    Wang, Haijiang
    Zhou, Wujie
    COMPUTER VISION, ECCV 2022, PT XXI, 2022, 13681 : 245 - 261
  • [47] Asymptotic Soft Filter Pruning for Deep Convolutional Neural Networks
    He, Yang
    Dong, Xuanyi
    Kang, Guoliang
    Fu, Yanwei
    Yan, Chenggang
    Yang, Yi
    IEEE TRANSACTIONS ON CYBERNETICS, 2020, 50 (08) : 3594 - 3604
  • [48] Play and Prune: Adaptive Filter Pruning for Deep Model Compression
    Singh, Pravendra
    Verma, Vinay Kumar
    Rai, Piyush
    Namboodiri, Vinay P.
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 3460 - 3466
  • [49] D-Pruner: Filter-Based Pruning Method for Deep Convolutional Neural Network
    Huynh, Loc N.
    Lee, Youngki
    Balan, Rajesh Krishna
    PROCEEDINGS OF THE 2018 INTERNATIONAL WORKSHOP ON EMBEDDED AND MOBILE DEEP LEARNING (EMDL '18), 2018, : 7 - 12
  • [50] COP: customized correlation-based Filter level pruning method for deep CNN compression
    Wang, Wenxiao
    Yu, Zhengxu
    Fu, Cong
    Cai, Deng
    He, Xiaofei
    NEUROCOMPUTING, 2021, 464 : 533 - 545