Accelerating Stochastic Variance Reduced Gradient Using Mini-Batch Samples on Estimation of Average Gradient

被引:1
|
作者
Huang, Junchu [1 ]
Zhou, Zhiheng [1 ]
Xu, Bingyuan [1 ]
Huang, Yu [1 ]
机构
[1] South China Univ Technol, Sch Elect & Informat Engn, Guangzhou, Guangdong, Peoples R China
来源
基金
中国国家自然科学基金;
关键词
Optimization algorithms; Stochastic gradient descent; Machine learning;
D O I
10.1007/978-3-319-59072-1_41
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Stochastic gradient descent (SGD) is popular for large scale optimization but has slow convergence. To remedy this problem, stochastic variance reduced gradient (SVRG) is proposed, which adopts average gradient to reduce the effect of variance. Since its expensive computational cost, average gradient is maintained between m iterations, where m is set to the same order of data size. For large scale problems, the efficiency will be decreased due to the prediction on average gradient maybe not accurate enough. We propose a method of using a mini-batch of samples to estimate average gradient, called stochastic mini-batch variance reduced gradient (SMVRG). SMVRG greatly reduces the computational cost of prediction on average gradient, therefore it is possible to estimate average gradient frequently thus more accurate. Numerical experiments show the effectiveness of our method in terms of convergence rate and computation cost.
引用
收藏
页码:346 / 353
页数:8
相关论文
共 50 条
  • [21] Statistical Analysis of Fixed Mini-Batch Gradient Descent Estimator
    Qi, Haobo
    Wang, Feifei
    Wang, Hansheng
    JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2023, 32 (04) : 1348 - 1360
  • [22] AN OPTIMAL DESIGN OF RANDOM SURFACES IN SOLAR CELLS VIA MINI-BATCH STOCHASTIC GRADIENT APPROACH
    Wang, Dan
    Li, Qiang
    Shen, Jihong
    COMMUNICATIONS IN MATHEMATICAL SCIENCES, 2022, 20 (03) : 747 - 762
  • [23] Automatic Setting of Learning Rate and Mini-batch Size in Momentum and AdaM Stochastic Gradient Methods
    Franchini, Giorgia
    Porta, Federica
    INTERNATIONAL CONFERENCE ON NUMERICAL ANALYSIS AND APPLIED MATHEMATICS 2022, ICNAAM-2022, 2024, 3094
  • [24] HYPERSPECTRAL UNMIXING VIA PROJECTED MINI-BATCH GRADIENT DESCENT
    Li, Jing
    Li, Xiaorun
    Zhao, Liaoying
    2017 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM (IGARSS), 2017, : 1133 - 1136
  • [25] A Mini-Batch Proximal Stochastic Recursive Gradient Algorithm with Diagonal Barzilai-Borwein Stepsize
    Yu, Teng-Teng
    Liu, Xin-Wei
    Dai, Yu-Hong
    Sun, Jie
    JOURNAL OF THE OPERATIONS RESEARCH SOCIETY OF CHINA, 2023, 11 (02) : 277 - 307
  • [26] Breast cancer detection using Histopathology Image with Mini-Batch Stochastic Gradient Descent and Convolutional Neural Network
    Sasirekha, N.
    Karuppaiah, Jayakumar
    Shekhar, Himanshu
    Saranya, N. Naga
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2023, 45 (03) : 4651 - 4667
  • [27] Efficient distance metric learning by adaptive sampling and mini-batch stochastic gradient descent (SGD)
    Qian, Qi
    Jin, Rong
    Yi, Jinfeng
    Zhang, Lijun
    Zhu, Shenghuo
    MACHINE LEARNING, 2015, 99 (03) : 353 - 372
  • [28] Efficient distance metric learning by adaptive sampling and mini-batch stochastic gradient descent (SGD)
    Qi Qian
    Rong Jin
    Jinfeng Yi
    Lijun Zhang
    Shenghuo Zhu
    Machine Learning, 2015, 99 : 353 - 372
  • [29] Tuple-oriented Compression for Large-scale Mini-batch Stochastic Gradient Descent
    Li, Fengan
    Chen, Lingjiao
    Zeng, Yijing
    Kumar, Arun
    Wu, Xi
    Naughton, Jeffrey F.
    Patel, Jignesh M.
    SIGMOD '19: PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON MANAGEMENT OF DATA, 2019, : 1517 - 1534
  • [30] Multi-Pass Sequential Mini-Batch Stochastic Gradient Descent Algorithms for Noise Covariance Estimation in Adaptive Kalman Filtering
    Kim, Hee-Seung
    Zhang, Lingyi
    Bienkowski, Adam
    Pattipati, Krishna R.
    IEEE ACCESS, 2021, 9 : 99220 - 99234