Toward Compact ConvNets via Structure-Sparsity Regularized Filter Pruning

被引:113
|
作者
Lin, Shaohui [1 ]
Ji, Rongrong [1 ,2 ]
Li, Yuchao [1 ]
Deng, Cheng [3 ]
Li, Xuelong [4 ,5 ]
机构
[1] Xiamen Univ, Sch Informat Sci & Engn, Fujian Key Lab Sensing & Comp Smart City, Xiamen 361005, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518055, Peoples R China
[3] Xidian Univ, Sch Elect Engn, Xian 710071, Peoples R China
[4] Northwestern Polytech Univ, Sch Comp Sci, Xian 710072, Peoples R China
[5] Northwestern Polytech Univ, Ctr OPT IMagery Anal & Learning OPTIMAL, Xian 710072, Peoples R China
基金
国家重点研发计划; 中国博士后科学基金;
关键词
Convolutional neural networks (CNNs); CNN acceleration; CNN compression; structured sparsity; NEURAL-NETWORKS;
D O I
10.1109/TNNLS.2019.2906563
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The success of convolutional neural networks (CNNs) in computer vision applications has been accompanied by a significant increase of computation and memory costs, which prohibits their usage on resource-limited environments, such as mobile systems or embedded devices. To this end, the research of CNN compression has recently become emerging. In this paper, we propose a novel filter pruning scheme, termed structured sparsity regularization (SSR), to simultaneously speed up the computation and reduce the memory overhead of CNNs, which can be well supported by various off-the-shelf deep learning libraries. Concretely, the proposed scheme incorporates two different regularizers of structured sparsity into the original objective function of filter pruning, which fully coordinates the global output and local pruning operations to adaptively prune filters. We further propose an alternative updating with Lagrange multipliers (AULM) scheme to efficiently solve its optimization. AULM follows the principle of alternating direction method of multipliers (ADMM) and alternates between promoting the structured sparsity of CNNs and optimizing the recognition loss, which leads to a very efficient solver (2.5x to the most recent work that directly solves the group sparsity-based regularization). Moreover, by imposing the structured sparsity, the online inference is extremely memory-light since the number of filters and the output feature maps are simultaneously reduced. The proposed scheme has been deployed to a variety of state-of-the-art CNN structures, including LeNet, AlexNet, VGGNet, ResNet, and GoogLeNet, over different data sets. Quantitative results demonstrate that the proposed scheme achieves superior performance over the state-of-the-art methods. We further demonstrate the proposed compression scheme for the task of transfer learning, including domain adaptation and object detection, which also show exciting performance gains over the state-of-the-art filter pruning methods.
引用
收藏
页码:574 / 588
页数:15
相关论文
共 10 条
  • [1] Global probability distribution structure-sparsity filter pruning for edge fault diagnosis in resource constrained wireless sensor networks
    Zhao, Chunhua
    Tang, Baoping
    Deng, Lei
    Huang, Yi
    Tan, Hao
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 133
  • [2] Learning compact ConvNets through filter pruning based on the saliency of a feature map
    Liu, Zhoufeng
    Liu, Xiaohui
    Li, Chunlei
    Ding, Shumin
    Liao, Liang
    IET IMAGE PROCESSING, 2022, 16 (01) : 123 - 133
  • [3] Filter pruning via separation of sparsity search and model training
    Lian, Youzao
    Peng, Peng
    Xu, Weisheng
    NEUROCOMPUTING, 2021, 462 (462) : 185 - 194
  • [4] Learning Compact Compositional Fmbeddings via Regularized Pruning for Recommendation
    Liang, Xurong
    Chen, Tong
    Quoc Viet Hung Nguyen
    Li, Jianxin
    Yin, Hongzhi
    23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING, ICDM 2023, 2023, : 378 - 387
  • [5] Towards efficient filter pruning via adaptive automatic structure search
    Xu, Xiaozhou
    Chen, Jun
    Li, Zhishan
    Su, Hongye
    Xie, Lei
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 133
  • [6] COP: Customized Deep Model Compression via Regularized Correlation-Based Filter-Level Pruning
    Wang, Wenxiao
    Fu, Cong
    Guo, Jishun
    Cai, Deng
    He, Xiaofei
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 3785 - 3791
  • [7] A Via-less Compact Bandpass Filter with Improved Selectivity using Metamaterial Structure
    Choudhary, Dilip Kumar
    Mishra, Naveen
    Kumar, Rajkishor
    Chaudhary, Raghvendra Kumar
    2017 IEEE ASIA PACIFIC MICROWAVE CONFERENCE (APMC), 2017, : 1321 - 1324
  • [8] Ultrahigh-Rate Supercapacitor Based on Carbon Nano-Onion/Graphene Hybrid Structure toward Compact Alternating Current Filter
    Zhang, Chenguang
    Du, Haozhe
    Ma, Ke
    Yuan, Zhihao
    ADVANCED ENERGY MATERIALS, 2020, 10 (43)
  • [9] Toward high-performance fibrillated cellulose-based air filter via constructing spider-web-like structure with the aid of TBA during freeze-drying process
    Zhaoqing Lu
    Zhiping Su
    Shunxi Song
    Yongsheng Zhao
    Shanshan Ma
    Meiyun Zhang
    Cellulose, 2018, 25 : 619 - 629
  • [10] Toward high-performance fibrillated cellulose-based air filter via constructing spider-web-like structure with the aid of TBA during freeze-drying process
    Lu, Zhaoqing
    Su, Zhiping
    Song, Shunxi
    Zhao, Yongsheng
    Ma, Shanshan
    Zhang, Meiyun
    CELLULOSE, 2018, 25 (01) : 619 - 629