Accelerating Stochastic Variance Reduced Gradient Using Mini-Batch Samples on Estimation of Average Gradient

被引:1
|
作者
Huang, Junchu [1 ]
Zhou, Zhiheng [1 ]
Xu, Bingyuan [1 ]
Huang, Yu [1 ]
机构
[1] South China Univ Technol, Sch Elect & Informat Engn, Guangzhou, Guangdong, Peoples R China
来源
基金
中国国家自然科学基金;
关键词
Optimization algorithms; Stochastic gradient descent; Machine learning;
D O I
10.1007/978-3-319-59072-1_41
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Stochastic gradient descent (SGD) is popular for large scale optimization but has slow convergence. To remedy this problem, stochastic variance reduced gradient (SVRG) is proposed, which adopts average gradient to reduce the effect of variance. Since its expensive computational cost, average gradient is maintained between m iterations, where m is set to the same order of data size. For large scale problems, the efficiency will be decreased due to the prediction on average gradient maybe not accurate enough. We propose a method of using a mini-batch of samples to estimate average gradient, called stochastic mini-batch variance reduced gradient (SMVRG). SMVRG greatly reduces the computational cost of prediction on average gradient, therefore it is possible to estimate average gradient frequently thus more accurate. Numerical experiments show the effectiveness of our method in terms of convergence rate and computation cost.
引用
收藏
页码:346 / 353
页数:8
相关论文
共 50 条
  • [31] Boundedness and Convergence of Mini-batch Gradient Method with Cyclic Dropconnect and Penalty
    Junling Jing
    Cai Jinhang
    Huisheng Zhang
    Wenxia Zhang
    Neural Processing Letters, 56
  • [32] Carbon Emission Forecasting Study Based on Influence Factor Mining and Mini-Batch Stochastic Gradient Optimization
    Yang, Wei
    Yuan, Qiheng
    Wang, Yongli
    Zheng, Fei
    Shi, Xin
    Li, Yi
    ENERGIES, 2024, 17 (01)
  • [33] Boundedness and Convergence of Mini-batch Gradient Method with Cyclic Dropconnect and Penalty
    Jing, Junling
    Jinhang, Cai
    Zhang, Huisheng
    Zhang, Wenxia
    NEURAL PROCESSING LETTERS, 2024, 56 (02)
  • [34] Mini-batch gradient descent: faster convergence under data sparsity
    Khirirat, Sarit
    Feyzmahdavian, Hamid Reza
    Johansson, Mikael
    2017 IEEE 56TH ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2017,
  • [35] Stochastic Variance-Reduced Policy Gradient
    Papini, Matteo
    Binaghi, Damiano
    Canonaco, Giuseppe
    Pirotta, Matteo
    Restelli, Marcello
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [36] Variance Reduced Stochastic Gradient Descent with Neighbors
    Hofmann, Thomas
    Lucchi, Aurelien
    Lacoste-Julien, Simon
    McWilliams, Brian
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [37] Mini-batch optimized full waveform inversion with geological constrained gradient filtering
    Yang, Hui
    Jia, Junxiong
    Wu, Bangyu
    Gao, Jinghuai
    JOURNAL OF APPLIED GEOPHYSICS, 2018, 152 : 9 - 16
  • [38] Budgeted Mini-Batch Parallel Gradient Descent for Support Vector Machines on Spark
    Tao, Hang
    Wu, Bin
    Lin, Xiuqin
    2014 20TH IEEE INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS (ICPADS), 2014, : 945 - 950
  • [39] Mini-Batch Gradient-Based MCMC for Decentralized Massive MIMO Detection
    Zhou, Xingyu
    Liang, Le
    Zhang, Jing
    Wen, Chao-Kai
    Jin, Shi
    IEEE TRANSACTIONS ON COMMUNICATIONS, 2025, 73 (01) : 677 - 692
  • [40] Efficient mini-batch stochastic gradient descent with Centroidal Voronoi Tessellation for PDE-constrained optimization under uncertainty
    Chen, Liuhong
    Xiong, Meixin
    Ming, Ju
    He, Xiaoming
    PHYSICA D-NONLINEAR PHENOMENA, 2024, 467