Adaptive Powerball Stochastic Conjugate Gradient for Large-Scale Learning

被引:3
|
作者
Yang, Zhuang [1 ]
机构
[1] Soochow Univ, Sch Comp Sci & Technol, Suzhou 215006, Peoples R China
关键词
Machine learning algorithms; Sensitivity; Machine learning; Ordinary differential equations; Information retrieval; Robustness; Computational complexity; Adaptive learning rate; conjugate gradient; large-scale learning; powerball function; stochastic optimization; QUASI-NEWTON METHOD;
D O I
10.1109/TBDATA.2023.3300546
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The extreme success of stochastic optimization (SO) in large-scale machine learning problems, information retrieval, bioinformatics, etc., has been widely reported, especially in recent years. As an effective tactic, conjugate gradient (CG) has been gaining its popularity in accelerating SO algorithms. This paper develops a novel type of stochastic conjugate gradient descent (SCG) algorithms from the perspective of the Powerball strategy and the hypergradient descent (HD) technique. The crucial idea behind the resulting methods is inspired by pursuing the equilibrium of ordinary differential equations (ODEs). We elucidate the effect of the Powerball strategy in SCG algorithms. The introduction of HD, on the other side, makes the resulting methods work with an online learning rate. Meanwhile, we provide a comprehension of the theoretical results for the resulting algorithms under non-convex assumptions. As a byproduct, we bridge the gap between the learning rate and powered stochastic optimization (PSO) algorithms, which is still an open problem. Resorting to numerical experiments on numerous benchmark datasets, we test the parameter sensitivity of the proposed methods and demonstrate the superior performance of our new algorithms over state-of-the-art algorithms.
引用
收藏
页码:1598 / 1606
页数:9
相关论文
共 50 条
  • [1] The Powerball Method With Biased Stochastic Gradient Estimation for Large-Scale Learning Systems
    Yang, Zhuang
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024,
  • [2] Painless Stochastic Conjugate Gradient for Large-Scale Machine Learning
    Yang, Zhuang
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (10) : 14645 - 14658
  • [3] Large-scale machine learning with fast and stable stochastic conjugate gradient
    Yang, Zhuang
    COMPUTERS & INDUSTRIAL ENGINEERING, 2022, 173
  • [4] An adaptive conjugate gradient algorithm for large-scale unconstrained optimization
    Andrei, Neculai
    JOURNAL OF COMPUTATIONAL AND APPLIED MATHEMATICS, 2016, 292 : 83 - 91
  • [5] Large-Scale Machine Learning with Stochastic Gradient Descent
    Bottou, Leon
    COMPSTAT'2010: 19TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL STATISTICS, 2010, : 177 - 186
  • [6] Adaptive stochastic conjugate gradient for machine learning
    Yang, Zhuang
    EXPERT SYSTEMS WITH APPLICATIONS, 2022, 206
  • [7] Adaptive stochastic conjugate gradient for machine learning
    Yang, Zhuang
    Expert Systems with Applications, 2022, 206
  • [8] MEAN-NORMALIZED STOCHASTIC GRADIENT FOR LARGE-SCALE DEEP LEARNING
    Wiesler, Simon
    Richard, Alexander
    Schlueter, Ralf
    Ney, Hermann
    2014 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2014,
  • [9] An online conjugate gradient algorithm for large-scale data analysis in machine learning
    Xue, Wei
    Wan, Pengcheng
    Li, Qiao
    Zhong, Ping
    Yu, Gaohang
    Tao, Tao
    AIMS MATHEMATICS, 2021, 6 (02): : 1515 - 1537
  • [10] Adaptive step size rules for stochastic optimization in large-scale learning
    Zhuang Yang
    Li Ma
    Statistics and Computing, 2023, 33