Parallel Asynchronous Stochastic Variance Reduction for Nonconvex Optimization

被引:0
|
作者
Fang, Cong
Lin, Zhouchen [1 ]
机构
[1] Peking Univ, Sch EECS, Key Lab Machine Percept MOE, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Nowadays, asynchronous parallel algorithms have received much attention in the optimization field due to the crucial demands for modern large-scale optimization problems. However, most asynchronous algorithms focus on convex problems. Analysis on nonconvex problems is lacking. For the Asynchronous Stochastic Descent (ASGD) algorithm, the best result from (Lian et al. 2015) can only achieve an asymptotic O(1/epsilon(2)) rate (convergence to the stationary points, namely, parallel to del f(x)parallel to(2) <= c) on nonconvex problems. In this paper, we study Stochastic Variance Reduced Gradient (SVRG) in the asynchronous setting. We propose the Asynchronous Stochastic Variance Reduced Gradient (ASVRG) algorithm for nonconvex finite-sum problems. We develop two schemes for ASVRG, depending on whether the parameters are updated as an atom or not. We prove that both of the two schemes can achieve linear speed up(1)(a non-asymptotic O(n(2/3)/epsilon) rate to the stationary points) for nonconvex problems when the delay parameter tau < n(1/3), where n is the number of training samples. We also establish a non-asymptotic O(n(2/3) tau(1/3)/epsilon) rate (convergence to the stationary points) for our algorithm without assumptions on t. This further demonstrates that even with asynchronous updating, SVRG has less number of Incremental First-order Oracles (IFOs) compared with Stochastic Gradient Descent and Gradient Descent. We also conduct experiments on a shared memory multi-core system to demonstrate the efficiency of our algorithm.
引用
收藏
页码:794 / 800
页数:7
相关论文
共 50 条
  • [21] Distributed and asynchronous Stochastic Gradient Descent with variance reduction
    Ming, Yuewei
    Zhao, Yawei
    Wu, Chengkun
    Li, Kuan
    Yin, Jianping
    NEUROCOMPUTING, 2018, 281 : 27 - 36
  • [22] Stochastic Variance-Reduced Cubic Regularization for Nonconvex Optimization
    Wang, Zhe
    Zhou, Yi
    Liang, Yingbin
    Lan, Guanghui
    22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [23] Asynchronous Parallel Nonconvex Optimization Under the Polyak-Lojasiewicz Condition
    Yazdani, Kasra
    Hale, Matthew
    IEEE CONTROL SYSTEMS LETTERS, 2022, 6 : 524 - 529
  • [24] A Variance Reduced Nonconvex Stochastic Optimization framework for Online Kernel Learning
    Pradhan, Hrusikesha
    Rajawat, Ketan
    2022 56TH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS, 2022, : 1281 - 1285
  • [25] On Variance Reduction in Stochastic Gradient Descent and its Asynchronous Variants
    Reddi, Sashank J.
    Hefny, Ahmed
    Sra, Suvrit
    Poczos, Barnabas
    Smola, Alex
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [26] STOCHASTIC ALTERNATING STRUCTURE-ADAPTED PROXIMAL GRADIENT DESCENT METHOD WITH VARIANCE REDUCTION FOR NONCONVEX NONSMOOTH OPTIMIZATION
    Jia, Zehui
    Zhang, Wenxing
    Cai, Xingju
    Han, Deren
    MATHEMATICS OF COMPUTATION, 2024, 93 (348) : 1677 - 1714
  • [27] Cyclic Block Coordinate Descent With Variance Reduction for Composite Nonconvex Optimization
    Cai, Xufeng
    Song, Chaobing
    Wright, Stephen J.
    Diakonikolas, Jelena
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 202, 2023, 202
  • [28] Cyclic Block Coordinate Descent With Variance Reduction for Composite Nonconvex Optimization
    Cai, Xufeng
    Song, Chaobing
    Wright, Stephen J.
    Diakonikolas, Jelena
    Proceedings of Machine Learning Research, 2023, 202 : 3469 - 3494
  • [29] The Sound of APALM Clapping: Faster Nonsmooth Nonconvex Optimization with Stochastic Asynchronous PALM
    Davis, Damek
    Udell, Madeleine
    Edmunds, Brent
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [30] Improved Asynchronous Parallel Optimization Analysis for Stochastic Incremental Methods
    Leblond, Remi
    Pedregosa, Fabian
    Lacoste-Julien, Simon
    JOURNAL OF MACHINE LEARNING RESEARCH, 2018, 19