SGD method for entropy error function with smoothing l0 regularization for neural networks

被引:0
|
作者
Nguyen, Trong-Tuan [1 ]
Thang, Van-Dat [2 ]
Nguyen, Van Thin [3 ]
Nguyen, Phuong T. [4 ]
机构
[1] VNPT AI, Hanoi, Vietnam
[2] Viettel High Technol Ind Corp, Hanoi, Vietnam
[3] Thai Nguyen Univ Educ, 20 Luong Ngoc Quyen St, Thai Nguyen City, Vietnam
[4] Univ Aquila, Dept Informat Engn Comp Sci & Math, Via Vetoio Snc, Laquila, Italy
关键词
Neural networks; l0; regularization; Entropy function; L-1/2; REGULARIZATION; GRADIENT DESCENT; APPROXIMATION; LAYER;
D O I
10.1007/s10489-024-05564-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The entropy error function has been widely used in neural networks. Nevertheless, the network training based on this error function generally leads to a slow convergence rate, and can easily be trapped in a local minimum or even with the incorrect saturation problem in practice. In fact, there are many results based on entropy error function in neural network and its applications. However, the theory of such an algorithm and its convergence have not been fully studied so far. To tackle the issue, this works proposes a novel entropy function with smoothing l(0) regularization for feed-forward neural networks. An empirical evaluation has been conducted on real-world datasets to demonstrate that the newly conceived algorithm allows us to substantially improve the prediction performance of the considered neural networks. More importantly, the experimental results also show that the proposed function brings in more precise classifications, compared to well-founded baselines. The work is novel as it enables neural networks to learn effectively, producing more accurate predictions compared to state-of-the-art algorithms. In this respect, it is expected that the algorithm will contribute to existing studies in the field, advancing research in Machine Learning and Deep Learning.
引用
收藏
页码:7213 / 7228
页数:16
相关论文
共 50 条
  • [21] Motion Estimation with L0 Norm Regularization
    Chen, Jun
    Cai, Zemin
    Xie, Xiaohua
    Lai, Jianhuang
    2021 IEEE 7TH INTERNATIONAL CONFERENCE ON VIRTUAL REALITY (ICVR 2021), 2021, : 127 - 134
  • [22] An Adaptive Ridge Procedure for L0 Regularization
    Frommlet, Florian
    Nuel, Gregory
    PLOS ONE, 2016, 11 (02):
  • [23] A communication-efficient method for l0 regularization linear regression models
    Kang, Lican
    Li, Xuerui
    Liu, Yanyan
    Luo, Yuan
    Zou, Zhikang
    JOURNAL OF STATISTICAL COMPUTATION AND SIMULATION, 2023, 93 (04) : 533 - 555
  • [24] A communication-efficient method for generalized linear regression with l0 regularization
    Wang, Kunpeng
    Li, Xuerui
    Liu, Yanyan
    Kang, Lican
    COMMUNICATIONS IN STATISTICS-SIMULATION AND COMPUTATION, 2024, 53 (08) : 3850 - 3867
  • [25] Image Smoothing via L0 Gradient Minimization
    Xu, Li
    Lu, Cewu
    Xu, Yi
    Jia, Jiaya
    ACM TRANSACTIONS ON GRAPHICS, 2011, 30 (06):
  • [26] Smoothing neural network for L0 regularized optimization problem with general convex constraints
    Li, Wenjing
    Bian, Wei
    NEURAL NETWORKS, 2021, 143 : 678 - 689
  • [27] Parameterized L0 Image Smoothing With Unsupervised Learning
    Yang, Yang
    Tang, Ling
    Yan, Tao
    Zeng, Lanling
    Shen, Xiangjun
    Zhan, Yongzhao
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024, 8 (02): : 1938 - 1951
  • [28] Robust Linear Regression via l0 Regularization
    Liu, Jing
    Cosman, Pamela C.
    Rao, Bhaskar D.
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2018, 66 (03) : 698 - 713
  • [29] Sparse Unmixing using an approximate L0 Regularization
    Guo, Yang
    Gao, Tai
    Deng, Chengzhi
    Wang, Shengqian
    Xiao, JianPing
    PROCEEDINGS OF THE FIRST INTERNATIONAL CONFERENCE ON INFORMATION SCIENCES, MACHINERY, MATERIALS AND ENERGY (ICISMME 2015), 2015, 126 : 900 - 904
  • [30] CT image restoration method via total variation and L0 smoothing filter
    Yin, Hai
    Li, Xianyun
    Liu, Zhi
    Peng, Wei
    Wang, Chengxiang
    Yu, Wei
    JOURNAL OF INVERSE AND ILL-POSED PROBLEMS, 2024, 32 (05): : 927 - 938