Dynamic sparsity and model feature learning enhanced training for convolutional neural network-pruning

被引:0
|
作者
Ruan X. [1 ,2 ]
Hu W. [1 ,2 ,3 ]
Liu Y. [1 ,2 ]
Li B. [1 ]
机构
[1] National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of Sciences, Beijing
[2] School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing
[3] CAS Center for Excellence in Brain Science and Intelligence Technology, Shanghai
关键词
Deep convolutional neural network; Feature learning; Model compression; Pruning; Structured sparsity;
D O I
10.1360/SST-2021-0088
中图分类号
学科分类号
摘要
Recently, model-pruning approaches have become popular in reducing the high burden of deep neural networks in real-world applications. However, several existing pruning methods simply use a well-trained model to initialize parameters without considering its feature representation. Thus, we propose a label-free and dynamic pruning method based on model feature learning enhanced training. Furthermore, we use the category-level information and features of intermediate layers (well-trained model) to guide the task learning of the compression models, which enhances their ability to learn the features of the well-trained model. Additionally, we use different submodels (compression models) output information to learn from one another, promoting the feature learning ability between different submodels. Moreover, we use a structured sparsity-inducing regularization in a dynamic sparsity manner. The expected pruning parameters are identified using Taylor series-based channel sensitivity criteria. The proposed method solves the optimization problem using an iterative shrinkage-thresholding algorithm with dynamic sparsity. After the training is complete, the proposed method only eliminates redundant parameters without fine-tuning. Extensive experimental results show that the proposed method achieves good compression performance on multiple datasets and networks. © 2022, Science China Press. All right reserved.
引用
收藏
页码:667 / 681
页数:14
相关论文
共 40 条
  • [1] He K, Zhang X, Ren S, Et al., Deep residual learning for image recognition, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770-778, (2016)
  • [2] Ji R R, Lin S H, Chao F, Et al., Deep neural network compression and acceleration: A review, J Comput Res Develop, 55, pp. 1871-1888, (2018)
  • [3] He Y, Zhang X, Sun J., Channel pruning for accelerating very deep neural networks, Proceedings of the IEEE International Conference on Computer Vision, pp. 1389-1397, (2017)
  • [4] Luo J H, Zhang H, Zhou H Y, Et al., Thinet: Pruning CNN filters for a thinner net, IEEE Trans Pattern Anal Mach Intell, 41, pp. 2525-2538, (2019)
  • [5] Lin S, Ji R, Li Y, Et al., Accelerating convolutional networks via global & dynamic filter pruning, Proceedings of International Joint Conference on Artificial Intelligence, pp. 2425-2432, (2018)
  • [6] Luo J H, Wu J, Lin W., Thinet: A filter level pruning method for deep neural network compression, Proceedings of the IEEE International Conference on Computer Vision, pp. 5058-5066, (2017)
  • [7] Li H, Kadav A, Durdanovic I, Et al., Pruning filters for efficient ConvNets, Proceedings of International Conference on Learning Representations (ICLR), (2017)
  • [8] Liu Z, Li J, Shen Z, Et al., Learning efficient convolutional networks through network slimming, Proceedings of the IEEE International Conference on Computer Vision, pp. 2736-2744, (2017)
  • [9] Molchanov P, Mallya A, Tyree S, Et al., Importance estimation for neural network pruning, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 11264-11272, (2019)
  • [10] Li B, Wu B, Su J, Et al., EagleEye: Fast sub-net evaluation for efficient neural network pruning, Proceedings of the European Conference on Computer Vision, pp. 639-654, (2020)