A Line Search Based Proximal Stochastic Gradient Algorithm with Dynamical Variance Reduction

被引:8
|
作者
Franchini, Giorgia [1 ]
Porta, Federica [1 ]
Ruggiero, Valeria [2 ]
Trombini, Ilaria [2 ,3 ]
机构
[1] Univ Modena & Reggio Emilia, Dept Phys Informat & Math, Via Campi 213-B, I-41125 Modena, Italy
[2] Univ Ferrara, Dept Math & Comp Sci, Via Machiavelli 30, I-44121 Ferrara, Italy
[3] Univ Parma, Dept Math Phys & Comp Sci, Parco Area Sci 7-A, I-43124 Parma, Italy
关键词
First order stochastic methods; Stochastic proximal methods; Machine learning; Green artificial intelligence; CONVERGENCE;
D O I
10.1007/s10915-022-02084-3
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
Many optimization problems arising from machine learning applications can be cast as the minimization of the sum of two functions: the first one typically represents the expected risk, and in practice it is replaced by the empirical risk, and the other one imposes a priori information on the solution. Since in general the first term is differentiable and the second one is convex, proximal gradient methods are very well suited to face such optimization problems. However, when dealing with large-scale machine learning issues, the computation of the full gradient of the differentiable term can be prohibitively expensive by making these algorithms unsuitable. For this reason, proximal stochastic gradient methods have been extensively studied in the optimization area in the last decades. In this paper we develop a proximal stochastic gradient algorithm which is based on two main ingredients. We indeed combine a proper technique to dynamically reduce the variance of the stochastic gradients along the iterative process with a descent condition in expectation for the objective function, aimed to fix the value for the steplength parameter at each iteration. For general objective functionals, the a.s. convergence of the limit points of the sequence generated by the proposed scheme to stationary points can be proved. For convex objective functionals, both the a.s. convergence of the whole sequence of the iterates to a minimum point and an O(1/k) convergence rate for the objective function values have been shown. The practical implementation of the proposed method does not need neither the computation of the exact gradient of the empirical risk during the iterations nor the tuning of an optimal value for the step length. An extensive numerical experimentation highlights that the proposed approach appears robust with respect to the setting of the hyper parameters and competitive compared to state-of-the-art methods.
引用
收藏
页数:35
相关论文
共 50 条
  • [1] A Line Search Based Proximal Stochastic Gradient Algorithm with Dynamical Variance Reduction
    Giorgia Franchini
    Federica Porta
    Valeria Ruggiero
    Ilaria Trombini
    Journal of Scientific Computing, 2023, 94
  • [2] Correction to: A Line Search Based Proximal Stochastic Gradient Algorithm with Dynamical Variance Reduction
    Giorgia Franchini
    Federica Porta
    Valeria Ruggiero
    Ilaria Trombini
    Journal of Scientific Computing, 2023, 96
  • [3] A Line Search Based Proximal Stochastic Gradient Algorithm with Dynamical Variance Reduction (vol 94, 23, 2023)
    Franchini, Giorgia
    Porta, Federica
    Ruggiero, Valeria
    Trombini, Ilaria
    JOURNAL OF SCIENTIFIC COMPUTING, 2023, 96 (02)
  • [4] A Proximal Stochastic Quasi-Newton Algorithm with Dynamical Sampling and Stochastic Line Search
    Zhang, Mengxiang
    Li, Shengjie
    JOURNAL OF SCIENTIFIC COMPUTING, 2025, 102 (01)
  • [5] PROXIMAL STOCHASTIC GRADIENT METHOD WITH PROGRESSIVE VARIANCE REDUCTION
    Xiao, Lin
    Zhang, Tong
    SIAM JOURNAL ON OPTIMIZATION, 2014, 24 (04) : 2057 - 2075
  • [6] Stochastic Conjugate Gradient Algorithm With Variance Reduction
    Jin, Xiao-Bo
    Zhang, Xu-Yao
    Huang, Kaizhu
    Geng, Guang-Gang
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (05) : 1360 - 1369
  • [7] Nonconvex optimization with inertial proximal stochastic variance reduction gradient
    He, Lulu
    Ye, Jimin
    Jianwei, E.
    INFORMATION SCIENCES, 2023, 648
  • [8] A VARIANCE-BASED PROXIMAL BACKWARD-FORWARD ALGORITHM WITH LINE SEARCH FOR STOCHASTIC MIXED VARIATIONAL INEQUALITIES
    He, Yue-Hong
    Long, Xian-Jun
    PACIFIC JOURNAL OF OPTIMIZATION, 2022, 18 (04): : 713 - 735
  • [9] A SEMISMOOTH NEWTON STOCHASTIC PROXIMAL POINT ALGORITHM WITH VARIANCE REDUCTION
    Milzarek, Andre
    Schaipp, Fabian
    Ulbrich, Michael
    SIAM JOURNAL ON OPTIMIZATION, 2024, 34 (01) : 1157 - 1185
  • [10] General inertial proximal stochastic variance reduction gradient for nonconvex nonsmooth optimization
    Sun, Shuya
    He, Lulu
    JOURNAL OF INEQUALITIES AND APPLICATIONS, 2023, 2023 (01)