Automatic Compression Ratio Allocation for Pruning Convolutional Neural Networks

被引:0
|
作者
Liu, Yunfeng [1 ]
Kong, Huihui [1 ]
Yu, Peihua [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Neural Networks; Network Pruning; Model Compression; Computer Vision;
D O I
10.1145/3387168.3387184
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Convolutional neural networks (CNNs) have demonstrated significant performance improvement in many application scenarios. However, the high computational complexity and model size have limited its application on the mobile and embedded devices. Various approaches have been proposed to compress CNNs. Filter pruning is widely considered as a promising solution, which can significantly speed up the inference and reduce memory consumption. To this end, most approaches tend to prune filters by manually allocating compression ratio, which highly relies on individual expertise and not friendly to non-professional users. In this paper, we propose an Automatic Compression Ratio Allocation (ACRA) scheme based on binary search algorithm to prune convolutional neural networks. Specifically, ACRA provides two strategies for allocating compression ratio automatically. First, uniform pruning strategy allocates the same compression ratio to each layer, which is obtained by binary search based on target FLOPs reduction of the whole networks. Second, sensitivity-based pruning strategy allocates appropriate compression ratio to each layer based on the sensitivity to accuracy. Experimental results from VGG11 and VGG-16, demonstrate that our scheme can reduce FLOPs significantly while maintaining a high accuracy level. Specifically, for the VGG16 on CIFAR-10 dataset, we reduce 29.18% FLOPs with only 1.24% accuracy decrease.
引用
收藏
页数:6
相关论文
共 50 条
  • [21] Redundancy-Aware Pruning of Convolutional Neural Networks
    Xie, Guotian
    NEURAL COMPUTATION, 2020, 32 (12) : 2482 - 2506
  • [22] Automatic Pruning Rate Derivation for Structured Pruning of Deep Neural Networks
    Sakai, Yasufumi
    Iwakawa, Akinori
    Tabaru, Tsuguchika
    Inoue, Atsuki
    Kawaguchi, Hiroshi
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2561 - 2567
  • [23] EDP: An Efficient Decomposition and Pruning Scheme for Convolutional Neural Network Compression
    Ruan, Xiaofeng
    Liu, Yufan
    Yuan, Chunfeng
    Li, Bing
    Hu, Weiming
    Li, Yangxi
    Maybank, Stephen
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (10) : 4499 - 4513
  • [24] EACP: An effective automatic channel pruning for neural networks
    Liu, Yajun
    Wu, Dakui
    Zhou, Wenju
    Fan, Kefeng
    Zhou, Zhiheng
    NEUROCOMPUTING, 2023, 526 : 131 - 142
  • [25] Convolutional Neural Network Compression via Dynamic Parameter Rank Pruning
    Sharma, Manish
    Heard, Jamison
    Saber, Eli
    Markopoulos, Panagiotis
    IEEE ACCESS, 2025, 13 : 18441 - 18456
  • [26] Frequency-Domain Dynamic Pruning for Convolutional Neural Networks
    Liu, Zhenhua
    Xu, Jizheng
    Peng, Xiulian
    Xiong, Ruiqin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [27] Dynamic connection pruning for densely connected convolutional neural networks
    Xinyi Hu
    Hangxiang Fang
    Ling Zhang
    Xue Zhang
    Howard H. Yang
    Dongxiao Yang
    Bo Peng
    Zheyang Li
    Haoji Hu
    Applied Intelligence, 2023, 53 : 19505 - 19521
  • [28] Entropy-based pruning method for convolutional neural networks
    Hur, Cheonghwan
    Kang, Sanggil
    JOURNAL OF SUPERCOMPUTING, 2019, 75 (06): : 2950 - 2963
  • [29] Cross-Entropy Pruning for Compressing Convolutional Neural Networks
    Bao, Rongxin
    Yuan, Xu
    Chen, Zhikui
    Ma, Ruixin
    NEURAL COMPUTATION, 2018, 30 (11) : 3128 - 3149
  • [30] Pruning Deep Convolutional Neural Networks Architectures with Evolution Strategy
    Fernandes, Francisco E., Jr.
    Yen, Gary G.
    INFORMATION SCIENCES, 2021, 552 : 29 - 47