SGD method for entropy error function with smoothing l0 regularization for neural networks

被引:0
|
作者
Nguyen, Trong-Tuan [1 ]
Thang, Van-Dat [2 ]
Nguyen, Van Thin [3 ]
Nguyen, Phuong T. [4 ]
机构
[1] VNPT AI, Hanoi, Vietnam
[2] Viettel High Technol Ind Corp, Hanoi, Vietnam
[3] Thai Nguyen Univ Educ, 20 Luong Ngoc Quyen St, Thai Nguyen City, Vietnam
[4] Univ Aquila, Dept Informat Engn Comp Sci & Math, Via Vetoio Snc, Laquila, Italy
关键词
Neural networks; l0; regularization; Entropy function; L-1/2; REGULARIZATION; GRADIENT DESCENT; APPROXIMATION; LAYER;
D O I
10.1007/s10489-024-05564-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The entropy error function has been widely used in neural networks. Nevertheless, the network training based on this error function generally leads to a slow convergence rate, and can easily be trapped in a local minimum or even with the incorrect saturation problem in practice. In fact, there are many results based on entropy error function in neural network and its applications. However, the theory of such an algorithm and its convergence have not been fully studied so far. To tackle the issue, this works proposes a novel entropy function with smoothing l(0) regularization for feed-forward neural networks. An empirical evaluation has been conducted on real-world datasets to demonstrate that the newly conceived algorithm allows us to substantially improve the prediction performance of the considered neural networks. More importantly, the experimental results also show that the proposed function brings in more precise classifications, compared to well-founded baselines. The work is novel as it enables neural networks to learn effectively, producing more accurate predictions compared to state-of-the-art algorithms. In this respect, it is expected that the algorithm will contribute to existing studies in the field, advancing research in Machine Learning and Deep Learning.
引用
收藏
页码:7213 / 7228
页数:16
相关论文
共 50 条
  • [31] Batch gradient method with smoothing L1/2 regularization for training of feedforward neural networks
    Wu, Wei
    Fan, Qinwei
    Zurada, Jacek M.
    Wang, Jian
    Yang, Dakun
    Liu, Yan
    NEURAL NETWORKS, 2014, 50 : 72 - 78
  • [32] L0 Optimization Using Laplacian Operator for Image Smoothing
    Li M.
    Gao S.
    Han H.
    Zhang C.
    Jisuanji Fuzhu Sheji Yu Tuxingxue Xuebao/Journal of Computer-Aided Design and Computer Graphics, 2021, 33 (07): : 1000 - 1014
  • [33] Change Detection Using L0 Smoothing and Superpixel Techniques
    Shi, Xiaoliang
    Xu, Yingying
    Zhang, Guixu
    Shen, Chaomin
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, KSEM 2015, 2015, 9403 : 600 - 611
  • [34] Image smoothing via truncated l0 gradient regularisation
    He, Liangtian
    Wang, Yilun
    IET IMAGE PROCESSING, 2018, 12 (02) : 226 - 234
  • [35] EXPLORING THE EFFECT OF l0/l2 REGULARIZATION IN NEURAL NETWORK PRUNING USING THE LC TOOLKIT
    Idelbayev, Yerlan
    Carreira-Perpinan, Miguel A.
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 3373 - 3377
  • [36] Image decomposition model OSV with L0 sparse regularization
    Wang, Guodong
    Xu, Jie
    Pan, Zhenkuan
    Zhang, Weizhong
    Diao, Zhaojing
    Journal of Information and Computational Science, 2015, 12 (02): : 743 - 750
  • [37] CAPPED lp APPROXIMATIONS FOR THE COMPOSITE l0 REGULARIZATION PROBLEM
    Li, Qia
    Zhang, Na
    INVERSE PROBLEMS AND IMAGING, 2018, 12 (05) : 1219 - 1243
  • [38] Convergence of Batch Gradient Method Based on the Entropy Error Function for Feedforward Neural Networks
    Xiong, Yan
    Tong, Xin
    NEURAL PROCESSING LETTERS, 2020, 52 (03) : 2687 - 2695
  • [39] Adaptive L0 Regularization for Sparse Support Vector Regression
    Christou, Antonis
    Artemiou, Andreas
    MATHEMATICS, 2023, 11 (13)
  • [40] Convergence of Batch Gradient Method Based on the Entropy Error Function for Feedforward Neural Networks
    Yan Xiong
    Xin Tong
    Neural Processing Letters, 2020, 52 : 2687 - 2695