Stability Based Filter Pruning for Accelerating Deep CNNs

被引:22
|
作者
Singh, Pravendra [1 ]
Kadi, Vinay Sameer Raja [2 ]
Verma, Nikhil [2 ]
Namboodiri, Vinay P. [1 ]
机构
[1] IIT Kanpur, Kanpur, Uttar Pradesh, India
[2] Samsung R&D Inst, Delhi, India
关键词
D O I
10.1109/WACV.2019.00129
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Convolutional neural networks (CNN) have achieved impressive performance on the wide variety of tasks (classification, detection, etc.) across multiple domains at the cost of high computational and memory requirements. Thus, leveraging CNNs for real-time applications necessitates model compression approaches that not only reduce the total number of parameters but reduce the overall computation as well. In this work, we present a stability-based approach for filter-level pruning of CNNs. We evaluate our proposed approach on different architectures (LeNet, VGG-16, ResNet, and Faster RCNN) and datasets and demonstrate its generalizability through extensive experiments. Moreover, our compressed models can be used at run-time without requiring any special libraries or hardware. Our model compression method reduces the number of FLOPS by an impressive factor of 6.03X and GPU memory footprint by more than 17X, significantly outperforming other state-of-the-art filter pruning methods.
引用
收藏
页码:1166 / 1174
页数:9
相关论文
共 50 条
  • [1] Accelerating and Pruning CNNs for Semantic Segmentation on FPGA
    Mori, Pierpaolo
    Vemparala, Manoj-Rohit
    Fasfous, Nael
    Mitra, Saptarshi
    Sarkar, Sreetama
    Frickenstein, Alexander
    Frickenstein, Lukas
    Helms, Domenik
    Nagaraja, Naveen Shankar
    Stechele, Walter
    Passerone, Claudio
    PROCEEDINGS OF THE 59TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC 2022, 2022, : 145 - 150
  • [2] A Study of Filter Duplication for CNNs Filter Pruning
    Ikuta, Ryosuke
    Yata, Noriko
    Manabe, Yoshitsugu
    INTERNATIONAL WORKSHOP ON ADVANCED IMAGING TECHNOLOGY, IWAIT 2024, 2024, 13164
  • [3] FALF ConvNets: Fatuous auxiliary loss based filter-pruning for efficient deep CNNs
    Singh, Pravendra
    Kadi, Vinay Sameer Raja
    Namboodiri, Vinay P.
    IMAGE AND VISION COMPUTING, 2020, 93
  • [4] Soft Filter Pruning for Accelerating Deep Convolutional Neural Networks
    He, Yang
    Kang, Guoliang
    Dong, Xuanyi
    Fu, Yanwei
    Yang, Yi
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 2234 - 2240
  • [5] COMPRESSING AUDIO CNNS WITH GRAPH CENTRALITY BASED FILTER PRUNING
    King, James A.
    Singh, Arshdeep
    Plumbley, Mark D.
    2023 IEEE WORKSHOP ON APPLICATIONS OF SIGNAL PROCESSING TO AUDIO AND ACOUSTICS, WASPAA, 2023,
  • [6] A Hybrid Statistics-based Channel Pruning Method for Deep CNNs
    Zhou, Yan
    Liu, Guangyi
    Wang, Dongli
    2019 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC), 2019, : 780 - 785
  • [7] Norm-Correlation based filter pruning to accelerating networks
    Hong, Minsoo
    Kim, Sungjei
    Jeong, Jinwoo
    12TH INTERNATIONAL CONFERENCE ON ICT CONVERGENCE (ICTC 2021): BEYOND THE PANDEMIC ERA WITH ICT CONVERGENCE INNOVATION, 2021, : 1393 - 1396
  • [8] Filter Pruning via Probabilistic Model-based Optimization for Accelerating Deep Convolutional Neural Networks
    Li, Qinghua
    Li, Cuiping
    Chen, Hong
    WSDM '21: PROCEEDINGS OF THE 14TH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2021, : 653 - 661
  • [9] Filter Pruning by Switching to Neighboring CNNs With Good Attributes
    He, Yang
    Liu, Ping
    Zhu, Linchao
    Yang, Yi
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (10) : 8044 - 8056
  • [10] FPWT: Filter pruning via wavelet transform for CNNs
    Liu, Yajun
    Fan, Kefeng
    Zhou, Wenju
    NEURAL NETWORKS, 2024, 179