Dynamical Channel Pruning by Conditional Accuracy Change for Deep Neural Networks

被引:53
|
作者
Chen, Zhiqiang [1 ,2 ]
Xu, Ting-Bing [2 ,3 ]
Du, Changde [1 ,2 ]
Liu, Cheng-Lin [2 ,3 ,4 ]
He, Huiguang [2 ,4 ,5 ]
机构
[1] Chinese Acad Sci CASIA, Inst Automat, Res Ctr Brain Inspired Intelligence, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci UCAS, Sch Artificial Intelligence, Beijing 100049, Peoples R China
[3] Chinese Acad Sci CASIA, Inst Automat, Natl Lab Pattern Recognit, Beijing 100190, Peoples R China
[4] Chinese Acad Sci, Ctr Excellence Brain Sci & Intelligence Technol, Beijing 100190, Peoples R China
[5] Chinese Acad Sci CASIA, Res Ctr Brain Inspired Intelligence, Natl Lab Pattern Recognit, Inst Automat, Beijing 100190, Peoples R China
基金
中国国家自然科学基金;
关键词
Training; Channel estimation; Logic gates; Computer architecture; Convolution; Biological neural networks; Automation; Conditional accuracy change (CAC); direct criterion; dynamical channel pruning; neural network compression; structure shaping;
D O I
10.1109/TNNLS.2020.2979517
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Channel pruning is an effective technique that has been widely applied to deep neural network compression. However, many existing methods prune from a pretrained model, thus resulting in repetitious pruning and fine-tuning processes. In this article, we propose a dynamical channel pruning method, which prunes unimportant channels at the early stage of training. Rather than utilizing some indirect criteria (e.g., weight norm, absolute weight sum, and reconstruction error) to guide connection or channel pruning, we design criteria directly related to the final accuracy of a network to evaluate the importance of each channel. Specifically, a channelwise gate is designed to randomly enable or disable each channel so that the conditional accuracy changes (CACs) can be estimated under the condition of each channel disabled. Practically, we construct two effective and efficient criteria to dynamically estimate CAC at each iteration of training; thus, unimportant channels can be gradually pruned during the training process. Finally, extensive experiments on multiple data sets (i.e., ImageNet, CIFAR, and MNIST) with various networks (i.e., ResNet, VGG, and MLP) demonstrate that the proposed method effectively reduces the parameters and computations of baseline network while yielding the higher or competitive accuracy. Interestingly, if we Double the initial Channels and then Prune Half (DCPH) of them to baseline's counterpart, it can enjoy a remarkable performance improvement by shaping a more desirable structure.
引用
收藏
页码:799 / 813
页数:15
相关论文
共 50 条
  • [31] Structured Pruning for Deep Convolutional Neural Networks: A Survey
    He, Yang
    Xiao, Lingao
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (05) : 2900 - 2919
  • [32] Class-dependent Pruning of Deep Neural Networks
    Entezari, Rahim
    Saukh, Olga
    2020 IEEE SECOND WORKSHOP ON MACHINE LEARNING ON EDGE IN SENSOR SYSTEMS (SENSYS-ML 2020), 2020, : 13 - 18
  • [33] Holistic Filter Pruning for Efficient Deep Neural Networks
    Enderich, Lukas
    Timm, Fabian
    Burgard, Wolfram
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WACV 2021, 2021, : 2595 - 2604
  • [34] On the Information of Feature Maps and Pruning of Deep Neural Networks
    Soltani, Mohammadreza
    Wu, Suya
    Ding, Jie
    Ravier, Robert
    Tarokh, Vahid
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 6988 - 6995
  • [35] Self-distilled Pruning of Deep Neural Networks
    Neill, James O'
    Dutta, Sourav
    Assem, Haytham
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2022, PT II, 2023, 13714 : 655 - 670
  • [36] QLP: Deep Q-Learning for Pruning Deep Neural Networks
    Camci, Efe
    Gupta, Manas
    Wu, Min
    Lin, Jie
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (10) : 6488 - 6501
  • [37] Exploiting similarity-induced redundancies in correlation topology for channel pruning in deep convolutional neural networks
    Liu J.
    Shao H.J.
    Deng X.
    Jiang Y.T.
    International Journal of Computers and Applications, 2023, 45 (05) : 379 - 390
  • [38] Joint Pruning and Channel-Wise Mixed-Precision Quantization for Efficient Deep Neural Networks
    Motetti, Beatrice Alessandra
    Risso, Matteo
    Burrello, Alessio
    Macii, Enrico
    Poncino, Massimo
    Pagliari, Daniele Jahier
    IEEE TRANSACTIONS ON COMPUTERS, 2024, 73 (11) : 2619 - 2633
  • [39] MULTI-LOSS-AWARE CHANNEL PRUNING OF DEEP NETWORKS
    Hu, Yiming
    Sun, Siyang
    Li, Jianquan
    Zhu, Jiagang
    Wang, Xingang
    Gu, Qingyi
    2019 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2019, : 889 - 893
  • [40] Exploiting Gaussian distribution in channel pruning for convolutional neural networks
    Liu, Yuzhou
    Liu, Bo
    Lin, Weiwei
    Yan, Yuanchao
    Zhang, Li
    JOURNAL OF SUPERCOMPUTING, 2025, 81 (05):