Accelerating Stochastic Variance Reduced Gradient Using Mini-Batch Samples on Estimation of Average Gradient

被引:1
|
作者
Huang, Junchu [1 ]
Zhou, Zhiheng [1 ]
Xu, Bingyuan [1 ]
Huang, Yu [1 ]
机构
[1] South China Univ Technol, Sch Elect & Informat Engn, Guangzhou, Guangdong, Peoples R China
来源
基金
中国国家自然科学基金;
关键词
Optimization algorithms; Stochastic gradient descent; Machine learning;
D O I
10.1007/978-3-319-59072-1_41
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Stochastic gradient descent (SGD) is popular for large scale optimization but has slow convergence. To remedy this problem, stochastic variance reduced gradient (SVRG) is proposed, which adopts average gradient to reduce the effect of variance. Since its expensive computational cost, average gradient is maintained between m iterations, where m is set to the same order of data size. For large scale problems, the efficiency will be decreased due to the prediction on average gradient maybe not accurate enough. We propose a method of using a mini-batch of samples to estimate average gradient, called stochastic mini-batch variance reduced gradient (SMVRG). SMVRG greatly reduces the computational cost of prediction on average gradient, therefore it is possible to estimate average gradient frequently thus more accurate. Numerical experiments show the effectiveness of our method in terms of convergence rate and computation cost.
引用
收藏
页码:346 / 353
页数:8
相关论文
共 50 条
  • [11] Accelerating variance-reduced stochastic gradient methods
    Derek Driggs
    Matthias J. Ehrhardt
    Carola-Bibiane Schönlieb
    Mathematical Programming, 2022, 191 : 671 - 715
  • [12] Comparing Stochastic Gradient Descent and Mini-batch Gradient Descent Algorithms in Loan Risk Assessment
    Adigun, Abodunrin AbdulGafar
    Yinka-Banjo, Chika
    INFORMATICS AND INTELLIGENT APPLICATIONS, 2022, 1547 : 283 - 296
  • [13] Gaussian Process Parameter Estimation Using Mini-batch Stochastic Gradient Descent: Convergence Guarantees and Empirical Benefits
    Chen, Hao
    Zheng, Lili
    Kontar, Raed Al
    Raskutti, Garvesh
    Journal of Machine Learning Research, 2022, 23
  • [14] Gaussian Process Parameter Estimation Using Mini-batch Stochastic Gradient Descent: Convergence Guarantees and Empirical Benefits
    Chen, Hao
    Zheng, Lili
    Al Kontar, Raed
    Raskutti, Garvesh
    JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [15] Stochastic Variance-Reduced Algorithms for PCA with Arbitrary Mini-Batch Sizes
    Kim, Cheolmin
    Klabjan, Diego
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 4302 - 4311
  • [16] Asynchronous Mini-Batch Gradient Descent with Variance Reduction for Non-Convex Optimization
    Huo, Zhouyuan
    Huang, Heng
    THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 2043 - 2049
  • [17] Gradient preconditioned mini-batch SGD for ridge regression
    Zhang, Zhuan
    Zhou, Shuisheng
    Li, Dong
    Yang, Ting
    NEUROCOMPUTING, 2020, 413 : 284 - 293
  • [18] Scalable Hardware Accelerator for Mini-Batch Gradient Descent
    Rasoori, Sandeep
    Akella, Venkatesh
    PROCEEDINGS OF THE 2018 GREAT LAKES SYMPOSIUM ON VLSI (GLSVLSI'18), 2018, : 159 - 164
  • [19] Adaptive Natural Gradient Method for Learning of Stochastic Neural Networks in Mini-Batch Mode
    Park, Hyeyoung
    Lee, Kwanyong
    APPLIED SCIENCES-BASEL, 2019, 9 (21):
  • [20] A Mini-Batch Proximal Stochastic Recursive Gradient Algorithm with Diagonal Barzilai–Borwein Stepsize
    Teng-Teng Yu
    Xin-Wei Liu
    Yu-Hong Dai
    Jie Sun
    Journal of the Operations Research Society of China, 2023, 11 : 277 - 307