Efficient Joint Optimization of Layer-Adaptive Weight Pruning in Deep Neural Networks

被引:7
|
作者
Xu, Kaixin [1 ,2 ]
Wang, Zhe [1 ,2 ]
Geng, Xue [1 ]
Wu, Min [1 ]
Li, Xiaoli [1 ,2 ]
Lin, Weisi [2 ]
机构
[1] ASTAR, Inst Infocomm Res I2R, 1 Fusionopolis Way, Singapore 138632, Singapore
[2] Nanyang Technol Univ, Singapore, Singapore
关键词
D O I
10.1109/ICCV51070.2023.01600
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose a novel layer-adaptive weight-pruning approach for Deep Neural Networks (DNNs) that addresses the challenge of optimizing the output distortion minimization while adhering to a target pruning ratio constraint. Our approach takes into account the collective influence of all layers to design a layer-adaptive pruning scheme. We discover and utilize a very important additivity property of output distortion caused by pruning weights on multiple layers. This property enables us to formulate the pruning as a combinatorial optimization problem and efficiently solve it through dynamic programming. By decomposing the problem into sub-problems, we achieve linear time complexity, making our optimization algorithm fast and feasible to run on CPUs. Our extensive experiments demonstrate the superiority of our approach over existing methods on the ImageNet and CIFAR-10 datasets. On CIFAR-10, our method achieves remarkable improvements, outperforming others by up to 1.0% for ResNet-32, 0.5% for VGG-16, and 0.7% for DenseNet-121 in terms of top-1 accuracy. On ImageNet, we achieve up to 4.7% and 4.6% higher top-1 accuracy compared to other methods for VGG16 and ResNet-50, respectively. These results highlight the effectiveness and practicality of our approach for enhancing DNN performance through layer-adaptive weight pruning. Code will be available on https://github.com/Akimoto-Cris/RD_VIT_PRUNE.
引用
收藏
页码:17401 / 17411
页数:11
相关论文
共 50 条
  • [1] LAYER-ADAPTIVE SPARSITY FOR THE MAGNITUDE-BASED PRUNING
    Lee, Jaeho
    Park, Sejun
    Mo, Sangwoo
    Ahn, Sungsoo
    Shin, Jinwoo
    ICLR 2021 - 9th International Conference on Learning Representations, 2021,
  • [2] Towards Deeper Graph Neural Networks via Layer-Adaptive
    Xu, Bingbing
    Xie, Bin
    Shen, Huawei
    COMPANION OF THE WORLD WIDE WEB CONFERENCE, WWW 2023, 2023, : 103 - 106
  • [3] ACP: ADAPTIVE CHANNEL PRUNING FOR EFFICIENT NEURAL NETWORKS
    Zhang, Yuan
    Yuan, Yuan
    Wang, Qi
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 4488 - 4492
  • [4] Trained Rank Pruning for Efficient Deep Neural Networks
    Xu, Yuhui
    Li, Yuxi
    Zhang, Shuai
    Wen, Wei
    Wang, Botao
    Dai, Wenrui
    Qi, Yingyong
    Chen, Yiran
    Lin, Weiyao
    Xiong, Hongkai
    FIFTH WORKSHOP ON ENERGY EFFICIENT MACHINE LEARNING AND COGNITIVE COMPUTING - NEURIPS EDITION (EMC2-NIPS 2019), 2019, : 14 - 17
  • [5] Holistic Filter Pruning for Efficient Deep Neural Networks
    Enderich, Lukas
    Timm, Fabian
    Burgard, Wolfram
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WACV 2021, 2021, : 2595 - 2604
  • [6] An Adaptive Layer Expansion Algorithm for Efficient Training of Deep Neural Networks
    Chen, Yi-Long
    Liu, Pangfeng
    Wu, Jan-Jan
    2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2020, : 420 - 425
  • [7] Differential evolution based layer-wise weight pruning for compressing deep neural networks
    Wu, Tao
    Li, Xiaoyang
    Zhou, Deyun
    Li, Na
    Shi, Jiao
    Sensors (Switzerland), 2021, 21 (03): : 1 - 20
  • [8] Differential Evolution Based Layer-Wise Weight Pruning for Compressing Deep Neural Networks
    Wu, Tao
    Li, Xiaoyang
    Zhou, Deyun
    Li, Na
    Shi, Jiao
    SENSORS, 2021, 21 (03) : 1 - 20
  • [9] TRP: Trained Rank Pruning for Efficient Deep Neural Networks
    Xu, Yuhui
    Li, Yuxi
    Zhang, Shuai
    Wen, Wei
    Wang, Botao
    Qi, Yingyong
    Chen, Yiran
    Lin, Weiyao
    Xiong, Hongkai
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 977 - 983
  • [10] Adaptive Weight Decay for Deep Neural Networks
    Nakamura, Kensuke
    Hong, Byung-Woo
    IEEE ACCESS, 2019, 7 : 118857 - 118865