Accelerating Stochastic Variance Reduced Gradient Using Mini-Batch Samples on Estimation of Average Gradient

被引:1
|
作者
Huang, Junchu [1 ]
Zhou, Zhiheng [1 ]
Xu, Bingyuan [1 ]
Huang, Yu [1 ]
机构
[1] South China Univ Technol, Sch Elect & Informat Engn, Guangzhou, Guangdong, Peoples R China
来源
基金
中国国家自然科学基金;
关键词
Optimization algorithms; Stochastic gradient descent; Machine learning;
D O I
10.1007/978-3-319-59072-1_41
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Stochastic gradient descent (SGD) is popular for large scale optimization but has slow convergence. To remedy this problem, stochastic variance reduced gradient (SVRG) is proposed, which adopts average gradient to reduce the effect of variance. Since its expensive computational cost, average gradient is maintained between m iterations, where m is set to the same order of data size. For large scale problems, the efficiency will be decreased due to the prediction on average gradient maybe not accurate enough. We propose a method of using a mini-batch of samples to estimate average gradient, called stochastic mini-batch variance reduced gradient (SMVRG). SMVRG greatly reduces the computational cost of prediction on average gradient, therefore it is possible to estimate average gradient frequently thus more accurate. Numerical experiments show the effectiveness of our method in terms of convergence rate and computation cost.
引用
收藏
页码:346 / 353
页数:8
相关论文
共 50 条
  • [41] A novel short-term load forecasting method based on mini-batch stochastic gradient descent regression model
    Lizhen, Wu
    Yifan, Zhao
    Gang, Wang
    Xiaohong, Hao
    ELECTRIC POWER SYSTEMS RESEARCH, 2022, 211
  • [42] Differentially Private Variance Reduced Stochastic Gradient Descent
    Lee, Jaewoo
    2017 INTERNATIONAL CONFERENCE ON NEW TRENDS IN COMPUTING SCIENCES (ICTCS), 2017, : 161 - 166
  • [43] A Learning Algorithm with a Gradient Normalization and a Learning Rate Adaptation for the Mini-batch Type Learning
    Ito, Daiki
    Okamoto, Takashi
    Koakutsu, Seiichi
    2017 56TH ANNUAL CONFERENCE OF THE SOCIETY OF INSTRUMENT AND CONTROL ENGINEERS OF JAPAN (SICE), 2017, : 811 - 816
  • [44] Grounding grid corrosion detection based on mini-batch gradient descent and greedy method
    Xie, Hongpeng
    Yang, Fan
    Hua, Mingsheng
    Liu, Sen
    Hu, Jiayuan
    He, Yifan
    AIP ADVANCES, 2021, 11 (06)
  • [45] A stochastic variance reduced gradient method with adaptive step for stochastic optimization
    Li, Jing
    Xue, Dan
    Liu, Lei
    Qi, Rulei
    OPTIMAL CONTROL APPLICATIONS & METHODS, 2024, 45 (03): : 1327 - 1342
  • [46] Federated Learning Using Variance Reduced Stochastic Gradient for Probabilistically Activated Agents
    Rostami, Mohammadreza
    Kia, Solmaz S.
    2023 AMERICAN CONTROL CONFERENCE, ACC, 2023, : 861 - 866
  • [47] Unmixing of large-scale hyperspectral data based on projected mini-batch gradient descent
    Li, Jing
    Li, Xiaorun
    Zhao, Liaoying
    INTERNATIONAL JOURNAL OF WAVELETS MULTIRESOLUTION AND INFORMATION PROCESSING, 2017, 15 (06)
  • [48] N-SVRG: Stochastic Variance Reduction Gradient with Noise Reduction Ability for Small Batch Samples
    Pan, Haijie
    Zheng, Lirong
    CMES-COMPUTER MODELING IN ENGINEERING & SCIENCES, 2022, 131 (01): : 493 - 512
  • [49] Efficient Mini-Batch Training on Memristor Neural Network Integrating Gradient Calculation and Weight Update
    Yamamori, Satoshi
    Hiromoto, Masayuki
    Sato, Takashi
    IEICE TRANSACTIONS ON FUNDAMENTALS OF ELECTRONICS COMMUNICATIONS AND COMPUTER SCIENCES, 2018, E101A (07) : 1092 - 1100
  • [50] Traffic Scheduling Optimization in Cognitive Radio based Smart Grid Network using Mini-batch Gradient Descent Method
    Khan, Muhammad Waqas
    Zeeshan, Muhammad
    Usman, Muhammad
    2019 14TH IBERIAN CONFERENCE ON INFORMATION SYSTEMS AND TECHNOLOGIES (CISTI), 2019,