On the compression of neural networks using e0-norm regularization and weight pruning

被引:6
|
作者
Oliveira, Felipe Dennis de Resende [1 ]
Batista, Eduardo Luiz Ortiz [1 ]
Seara, Rui [1 ]
机构
[1] Univ Fed Santa Catarina, Dept Elect Engn, LINSE Circuits & Signal Proc Lab, BR-88040900 Florianopolis, SC, Brazil
关键词
Machine learning; Neural networks; Network compression; Norm regularization; Weight pruning;
D O I
10.1016/j.neunet.2023.12.019
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Despite the growing availability of high-capacity computational platforms, implementation complexity still has been a great concern for the real-world deployment of neural networks. This concern is not exclusively due to the huge costs of state-of-the-art network architectures, but also due to the recent push towards edge intelligence and the use of neural networks in embedded applications. In this context, network compression techniques have been gaining interest due to their ability for reducing deployment costs while keeping inference accuracy at satisfactory levels. The present paper is dedicated to the development of a novel compression scheme for neural networks. To this end, a new form of e0-norm-based regularization is firstly developed, which is capable of inducing strong sparseness in the network during training. Then, targeting the smaller weights of the trained network with pruning techniques, smaller yet highly effective networks can be obtained. The proposed compression scheme also involves the use of e2-norm regularization to avoid overfitting as well as fine tuning to improve the performance of the pruned network. Experimental results are presented aiming to show the effectiveness of the proposed scheme as well as to make comparisons with competing approaches.
引用
收藏
页码:343 / 352
页数:10
相关论文
共 50 条
  • [31] Weight Pruning Techniques Towards Photonic Implementation of Nonlinear Impairment Compensation Using Neural Networks
    Fujisawa, Shinsuke
    Yaman, Fatih
    Batshon, Hussam G.
    Tanio, Masaaki
    Ishii, Naoto
    Huang, Chaoran
    de Lima, Thomas Ferreira
    Inada, Yoshihisa
    Prucnal, Paul R.
    Kamiya, Norifumi
    Wang, Ting
    JOURNAL OF LIGHTWAVE TECHNOLOGY, 2022, 40 (05) : 1273 - 1282
  • [32] PRUNING ARTIFICIAL NEURAL NETWORKS USING NEURAL COMPLEXITY MEASURES
    Jorgensen, Thomas D.
    Haynes, Barry P.
    Norlund, Charlotte C. F.
    INTERNATIONAL JOURNAL OF NEURAL SYSTEMS, 2008, 18 (05) : 389 - 403
  • [33] L0 Regularization based Fine-grained Neural Network Pruning Method
    Xie, Qixin
    Li, Chao
    Diao, Boyu
    An, Zhulin
    Xu, Yongjun
    PROCEEDINGS OF THE 11TH INTERNATIONAL CONFERENCE ON ELECTRONICS, COMPUTERS AND ARTIFICIAL INTELLIGENCE (ECAI-2019), 2019,
  • [34] Smooth Group L1/2 Regularization for Pruning Convolutional Neural Networks
    Bao, Yuan
    Liu, Zhaobin
    Luo, Zhongxuan
    Yang, Sibo
    SYMMETRY-BASEL, 2022, 14 (01):
  • [35] STRUCTURED PRUNING FOR GROUP REGULARIZED CONVOLUTIONAL NEURAL NETWORKS VIA DYNAMIC REGULARIZATION FACTOR
    Li, Feng
    Li, Bo
    Zhu, Meijiao
    Ma, Junchi
    Yuan, Jinlong
    JOURNAL OF INDUSTRIAL AND MANAGEMENT OPTIMIZATION, 2025, 21 (02) : 1440 - 1455
  • [36] Deeper Weight Pruning without Accuracy Loss in Deep Neural Networks
    Ahn, Byungmin
    Kim, Taewhan
    PROCEEDINGS OF THE 2020 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2020), 2020, : 73 - 78
  • [37] SPLITTABLE PATTERN-SPECIFIC WEIGHT PRUNING FOR DEEP NEURAL NETWORKS
    Liu, Yiding
    Teng, Yinglei
    Niu, Tao
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 1439 - 1444
  • [38] REGULARIZATION OF CONVOLUTIONAL NEURAL NETWORKS USING SHUFFLENODE
    Chen, Yihao
    Wang, Hanli
    Long, Yu
    2017 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2017, : 355 - 360
  • [39] Latent Weight-based Pruning for Small Binary Neural Networks
    Chen, Tianen
    Anderson, Noah
    Kim, Younghyun
    2023 28TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE, ASP-DAC, 2023, : 751 - 756
  • [40] Harmonious Coexistence of Structured Weight Pruning and Ternarization for Deep Neural Networks
    Yang, Li
    He, Zhezhi
    Fan, Deliang
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 6623 - 6630