Faster First-Order Methods for Stochastic Non-Convex Optimization on Riemannian Manifolds

被引:15
|
作者
Zhou, Pan [1 ]
Yuan, Xiao-Tong [2 ]
Yan, Shuicheng [1 ]
Feng, Jiashi [1 ]
机构
[1] Natl Univ Singapore, Dept Elect & Comp Engn, Singapore, Singapore
[2] Nanjing Univ Informat Sci & Technol, Sch Automat, Nanjing 210044, Peoples R China
关键词
Optimization; Complexity theory; Manifolds; Convergence; Signal processing algorithms; Stochastic processes; Minimization; Riemannian optimization; stochastic variance-reduced algorithm; non-convex optimization; online learning; ILLUMINATION; COMPLETION;
D O I
10.1109/TPAMI.2019.2933841
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
First-order non-convex Riemannian optimization algorithms have gained recent popularity in structured machine learning problems including principal component analysis and low-rank matrix completion. The current paper presents an efficient Riemannian Stochastic Path Integrated Differential EstimatoR (R-SPIDER) algorithm to solve the finite-sum and online Riemannian non-convex minimization problems. At the core of R-SPIDER is a recursive semi-stochastic gradient estimator that can accurately estimate Riemannian gradient under not only exponential mapping and parallel transport, but also general retraction and vector transport operations. Compared with prior Riemannian algorithms, such a recursive gradient estimation mechanism endows R-SPIDER with lower computational cost in first-order oracle complexity. Specifically, for finite-sum problems with n components, R-SPIDER is proved to converge to an epsilon-approximate stationary point within O(min(n + root n/epsilon(2),1/epsilon(3))) stochastic gradient evaluations, beating the best-known complexity O(n+1/epsilon(4)); for online optimization, R-SPIDER is shown to converge with O(1/epsilon(3)) complexity which is, to the best of our knowledge, the first non-asymptotic result for online Riemannian optimization. For the special case of gradient dominated functions, we further develop a variant of R-SPIDER with improved linear rate of convergence. Extensive experimental results demonstrate the advantage of the proposed algorithms over the state-of-the-art Riemannian non-convex optimization methods.
引用
收藏
页码:459 / 472
页数:14
相关论文
共 50 条
  • [21] A Random Walk Approach to First-Order Stochastic Convex Optimization
    Vakili, Sattar
    Zhao, Qing
    2019 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2019, : 395 - 399
  • [22] SYNCHRONIZATION OF FIRST-ORDER AUTONOMOUS OSCILLATORS ON RIEMANNIAN MANIFOLDS
    Fiori, Simone
    DISCRETE AND CONTINUOUS DYNAMICAL SYSTEMS-SERIES B, 2019, 24 (04): : 1725 - 1741
  • [23] Avoiding Synchronization in First-Order Methods for Sparse Convex Optimization
    Devarakonda, Aditya
    Demmel, James
    Fountoulakis, Kimon
    Mahoney, Michael W.
    2018 32ND IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS), 2018, : 409 - 418
  • [24] Fast First-Order Methods for Composite Convex Optimization with Backtracking
    Scheinberg, Katya
    Goldfarb, Donald
    Bai, Xi
    FOUNDATIONS OF COMPUTATIONAL MATHEMATICS, 2014, 14 (03) : 389 - 417
  • [25] RELATIVELY SMOOTH CONVEX OPTIMIZATION BY FIRST-ORDER METHODS, AND APPLICATIONS
    Lu, Haihao
    Freund, Robert M.
    Nesterov, Yurii
    SIAM JOURNAL ON OPTIMIZATION, 2018, 28 (01) : 333 - 354
  • [26] Fast First-Order Methods for Composite Convex Optimization with Backtracking
    Katya Scheinberg
    Donald Goldfarb
    Xi Bai
    Foundations of Computational Mathematics, 2014, 14 : 389 - 417
  • [27] First-order methods of smooth convex optimization with inexact oracle
    Olivier Devolder
    François Glineur
    Yurii Nesterov
    Mathematical Programming, 2014, 146 : 37 - 75
  • [28] First-order methods of smooth convex optimization with inexact oracle
    Devolder, Olivier
    Glineur, Francois
    Nesterov, Yurii
    MATHEMATICAL PROGRAMMING, 2014, 146 (1-2) : 37 - 75
  • [29] Stochastic Successive Convex Approximation for Non-Convex Constrained Stochastic Optimization
    Liu, An
    Lau, Vincent K. N.
    Kananian, Borna
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2019, 67 (16) : 4189 - 4203
  • [30] Gradient Methods for Non-convex Optimization
    Jain, Prateek
    JOURNAL OF THE INDIAN INSTITUTE OF SCIENCE, 2019, 99 (02) : 247 - 256