Automatic Compression Ratio Allocation for Pruning Convolutional Neural Networks

被引:0
|
作者
Liu, Yunfeng [1 ]
Kong, Huihui [1 ]
Yu, Peihua [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Neural Networks; Network Pruning; Model Compression; Computer Vision;
D O I
10.1145/3387168.3387184
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Convolutional neural networks (CNNs) have demonstrated significant performance improvement in many application scenarios. However, the high computational complexity and model size have limited its application on the mobile and embedded devices. Various approaches have been proposed to compress CNNs. Filter pruning is widely considered as a promising solution, which can significantly speed up the inference and reduce memory consumption. To this end, most approaches tend to prune filters by manually allocating compression ratio, which highly relies on individual expertise and not friendly to non-professional users. In this paper, we propose an Automatic Compression Ratio Allocation (ACRA) scheme based on binary search algorithm to prune convolutional neural networks. Specifically, ACRA provides two strategies for allocating compression ratio automatically. First, uniform pruning strategy allocates the same compression ratio to each layer, which is obtained by binary search based on target FLOPs reduction of the whole networks. Second, sensitivity-based pruning strategy allocates appropriate compression ratio to each layer based on the sensitivity to accuracy. Experimental results from VGG11 and VGG-16, demonstrate that our scheme can reduce FLOPs significantly while maintaining a high accuracy level. Specifically, for the VGG16 on CIFAR-10 dataset, we reduce 29.18% FLOPs with only 1.24% accuracy decrease.
引用
收藏
页数:6
相关论文
共 50 条
  • [41] Soft Filter Pruning for Accelerating Deep Convolutional Neural Networks
    He, Yang
    Kang, Guoliang
    Dong, Xuanyi
    Fu, Yanwei
    Yang, Yi
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 2234 - 2240
  • [42] PRUNING OF CONVOLUTIONAL NEURAL NETWORKS USING ISING ENERGY MODEL
    Salehinejad, Hojjat
    Valaee, Shahrokh
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 3935 - 3939
  • [43] A Filter Rank Based Pruning Method for Convolutional Neural Networks
    Liu, Hao
    Guan, Zhenyu
    Lei, Peng
    2021 IEEE 20TH INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS (TRUSTCOM 2021), 2021, : 1318 - 1322
  • [44] Soft Taylor Pruning for Accelerating Deep Convolutional Neural Networks
    Rong, Jintao
    Yu, Xiyi
    Zhang, Mingyang
    Ou, Linlin
    IECON 2020: THE 46TH ANNUAL CONFERENCE OF THE IEEE INDUSTRIAL ELECTRONICS SOCIETY, 2020, : 5343 - 5349
  • [45] Reborn Filters: Pruning Convolutional Neural Networks with Limited Data
    Tang, Yehui
    You, Shan
    Xu, Chang
    Han, Jin
    Qian, Chen
    Shi, Boxin
    Xu, Chao
    Zhang, Changshui
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 5972 - 5980
  • [46] FRACTIONAL STEP DISCRIMINANT PRUNING: A FILTER PRUNING FRAMEWORK FOR DEEP CONVOLUTIONAL NEURAL NETWORKS
    Gkalelis, Nikolaos
    Mezaris, Vasileios
    2020 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO WORKSHOPS (ICMEW), 2020,
  • [47] 1xN Pattern for Pruning Convolutional Neural Networks
    Lin, Mingbao
    Zhang, Yuxin
    Li, Yuchao
    Chen, Bohong
    Chao, Fei
    Wang, Mengdi
    Li, Shen
    Tian, Yonghong
    Ji, Rongrong
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (04) : 3999 - 4008
  • [48] Filter pruning for convolutional neural networks in semantic image segmentation
    Lopez-Gonzalez, Clara I.
    Gasco, Esther
    Barrientos-Espillco, Fredy
    Besada-Portas, Eva
    Pajares, Gonzalo
    NEURAL NETWORKS, 2024, 169 : 713 - 732
  • [49] Compressing Convolutional Neural Networks by Pruning Density Peak Filters
    Jang, Yunseok
    Lee, Sangyoun
    Kim, Jaeseok
    IEEE ACCESS, 2021, 9 : 8278 - 8285
  • [50] Convolutional Neural Networks for Enhanced Compression Techniques
    Gnacek, Matthew
    Heatwole, Cory
    Fan, David
    Hoffman, Marc
    PROCEEDINGS OF THE 2021 IEEE NATIONAL AEROSPACE AND ELECTRONICS CONFERENCE (NAECON), 2021, : 392 - 399