Faster First-Order Methods for Stochastic Non-Convex Optimization on Riemannian Manifolds

被引:15
|
作者
Zhou, Pan [1 ]
Yuan, Xiao-Tong [2 ]
Yan, Shuicheng [1 ]
Feng, Jiashi [1 ]
机构
[1] Natl Univ Singapore, Dept Elect & Comp Engn, Singapore, Singapore
[2] Nanjing Univ Informat Sci & Technol, Sch Automat, Nanjing 210044, Peoples R China
关键词
Optimization; Complexity theory; Manifolds; Convergence; Signal processing algorithms; Stochastic processes; Minimization; Riemannian optimization; stochastic variance-reduced algorithm; non-convex optimization; online learning; ILLUMINATION; COMPLETION;
D O I
10.1109/TPAMI.2019.2933841
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
First-order non-convex Riemannian optimization algorithms have gained recent popularity in structured machine learning problems including principal component analysis and low-rank matrix completion. The current paper presents an efficient Riemannian Stochastic Path Integrated Differential EstimatoR (R-SPIDER) algorithm to solve the finite-sum and online Riemannian non-convex minimization problems. At the core of R-SPIDER is a recursive semi-stochastic gradient estimator that can accurately estimate Riemannian gradient under not only exponential mapping and parallel transport, but also general retraction and vector transport operations. Compared with prior Riemannian algorithms, such a recursive gradient estimation mechanism endows R-SPIDER with lower computational cost in first-order oracle complexity. Specifically, for finite-sum problems with n components, R-SPIDER is proved to converge to an epsilon-approximate stationary point within O(min(n + root n/epsilon(2),1/epsilon(3))) stochastic gradient evaluations, beating the best-known complexity O(n+1/epsilon(4)); for online optimization, R-SPIDER is shown to converge with O(1/epsilon(3)) complexity which is, to the best of our knowledge, the first non-asymptotic result for online Riemannian optimization. For the special case of gradient dominated functions, we further develop a variant of R-SPIDER with improved linear rate of convergence. Extensive experimental results demonstrate the advantage of the proposed algorithms over the state-of-the-art Riemannian non-convex optimization methods.
引用
收藏
页码:459 / 472
页数:14
相关论文
共 50 条
  • [31] Gradient Methods for Non-convex Optimization
    Prateek Jain
    Journal of the Indian Institute of Science, 2019, 99 : 247 - 256
  • [32] On Graduated Optimization for Stochastic Non-Convex Problems
    Hazan, Elad
    Levy, Kfir Y.
    Shalev-Shwartz, Shai
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [33] Lower bounds for non-convex stochastic optimization
    Yossi Arjevani
    Yair Carmon
    John C. Duchi
    Dylan J. Foster
    Nathan Srebro
    Blake Woodworth
    Mathematical Programming, 2023, 199 : 165 - 214
  • [34] Lower bounds for non-convex stochastic optimization
    Arjevani, Yossi
    Carmon, Yair
    Duchi, John C.
    Foster, Dylan J.
    Srebro, Nathan
    Woodworth, Blake
    MATHEMATICAL PROGRAMMING, 2023, 199 (1-2) : 165 - 214
  • [35] Distributed Non-Convex First-Order Optimization and Information Processing: Lower Complexity Bounds and Rate Optimal Algorithms
    Sun, Haoran
    Hong, Mingyi
    2018 CONFERENCE RECORD OF 52ND ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS, 2018, : 38 - 42
  • [36] Distributed Non-Convex First-Order Optimization and Information Processing: Lower Complexity Bounds and Rate Optimal Algorithms
    Sun, Haoran
    Hong, Mingyi
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2019, 67 (22) : 5912 - 5928
  • [37] Accelerated Zeroth-Order Algorithm for Stochastic Distributed Non-Convex Optimization
    Zhang, Shengjun
    Bailey, Colleen P.
    2022 AMERICAN CONTROL CONFERENCE, ACC, 2022, : 4274 - 4279
  • [38] Second-Order Information in Non-Convex Stochastic Optimization: Power and Limitations
    Arjevani, Yossi
    Carmon, Yair
    Duchi, John C.
    Foster, Dylan J.
    Sekhari, Ayush
    Sridharan, Karthik
    CONFERENCE ON LEARNING THEORY, VOL 125, 2020, 125
  • [39] Stochastic Proximal Methods for Non-Smooth Non-Convex Constrained Sparse Optimization
    Metel, Michael R.
    Takeda, Akiko
    JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 22
  • [40] Simple Stochastic Gradient Methods for Non-Smooth Non-Convex Regularized Optimization
    Metel, Michael R.
    Takeda, Akiko
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97