Adaptive Powerball Stochastic Conjugate Gradient for Large-Scale Learning

被引:3
|
作者
Yang, Zhuang [1 ]
机构
[1] Soochow Univ, Sch Comp Sci & Technol, Suzhou 215006, Peoples R China
关键词
Machine learning algorithms; Sensitivity; Machine learning; Ordinary differential equations; Information retrieval; Robustness; Computational complexity; Adaptive learning rate; conjugate gradient; large-scale learning; powerball function; stochastic optimization; QUASI-NEWTON METHOD;
D O I
10.1109/TBDATA.2023.3300546
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The extreme success of stochastic optimization (SO) in large-scale machine learning problems, information retrieval, bioinformatics, etc., has been widely reported, especially in recent years. As an effective tactic, conjugate gradient (CG) has been gaining its popularity in accelerating SO algorithms. This paper develops a novel type of stochastic conjugate gradient descent (SCG) algorithms from the perspective of the Powerball strategy and the hypergradient descent (HD) technique. The crucial idea behind the resulting methods is inspired by pursuing the equilibrium of ordinary differential equations (ODEs). We elucidate the effect of the Powerball strategy in SCG algorithms. The introduction of HD, on the other side, makes the resulting methods work with an online learning rate. Meanwhile, we provide a comprehension of the theoretical results for the resulting algorithms under non-convex assumptions. As a byproduct, we bridge the gap between the learning rate and powered stochastic optimization (PSO) algorithms, which is still an open problem. Resorting to numerical experiments on numerous benchmark datasets, we test the parameter sensitivity of the proposed methods and demonstrate the superior performance of our new algorithms over state-of-the-art algorithms.
引用
收藏
页码:1598 / 1606
页数:9
相关论文
共 50 条
  • [41] Testing different conjugate gradient methods for large-scale unconstrained optimization
    Dai, YH
    Ni, Q
    JOURNAL OF COMPUTATIONAL MATHEMATICS, 2003, 21 (03) : 311 - 320
  • [42] Parallel stochastic gradient algorithms for large-scale matrix completion
    Recht B.
    Ré C.
    Mathematical Programming Computation, 2013, 5 (2) : 201 - 226
  • [43] A large-scale stochastic gradient descent algorithm over a graphon
    Chen, Yan
    Li, Tao
    2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 4806 - 4811
  • [44] Stochastic Gradient Descent for Large-scale Linear Nonparallel SVM
    Tang, Jingjing
    Tian, Yingjie
    Wu, Guoqiang
    Li, Dewei
    2017 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON WEB INTELLIGENCE (WI 2017), 2017, : 980 - 983
  • [45] A stochastic recursive gradient algorithm integrating momentum and the powerball function with adaptive step sizes
    Qin, Chuandong
    Cai, Zilin
    Guo, Yuhang
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2025,
  • [46] A Modified Hestenes-Stiefel Conjugate Gradient Algorithm for Large-Scale Optimization
    Yuan, Gonglin
    Zhang, Maojun
    NUMERICAL FUNCTIONAL ANALYSIS AND OPTIMIZATION, 2013, 34 (08) : 914 - 937
  • [47] A Modified Nonlinear Conjugate Gradient Algorithm for Large-Scale Nonsmooth Convex Optimization
    Woldu, Tsegay Giday
    Zhang, Haibin
    Zhang, Xin
    Fissuh, Yemane Hailu
    JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 2020, 185 (01) : 223 - 238
  • [48] Conjugate Gradient Methods with Sufficient Descent Condition for Large-scale Unconstrained Optimization
    Ling, Mei Mei
    Leong, Wah June
    INTERNATIONAL CONFERENCE ON QUANTITATIVE SCIENCES AND ITS APPLICATIONS (ICOQSIA 2014), 2014, 1635 : 629 - 633
  • [49] A Modified Nonlinear Conjugate Gradient Algorithm for Large-Scale Nonsmooth Convex Optimization
    Tsegay Giday Woldu
    Haibin Zhang
    Xin Zhang
    Yemane Hailu Fissuh
    Journal of Optimization Theory and Applications, 2020, 185 : 223 - 238
  • [50] Solving Large-Scale Unconstrained Optimization Problems with an Efficient Conjugate Gradient Class
    Bojari, Sanaz
    Paripour, Mahmoud
    JOURNAL OF MATHEMATICS, 2024, 2024