CONVERGENCE AND DYNAMICAL BEHAVIOR OF THE ADAM ALGORITHM FOR NONCONVEX STOCHASTIC OPTIMIZATION

被引:52
|
作者
Barakat, Anas [1 ]
Bianchi, Pascal [1 ]
机构
[1] Inst Polytech Paris, Telecom Paris, LTCI, F-91120 Palaiseau, France
关键词
stochastic approximation; dynamical systems; adaptive gradient methods;
D O I
10.1137/19M1263443
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
Adam is a popular variant of stochastic gradient descent for finding a local minimizer of a function. In the constant stepsize regime, assuming that the objective function is differentiable and nonconvex, we establish the convergence in the long run of the iterates to a stationary point under a stability condition. The key ingredient is the introduction of a continuous-time version of Adam, under the form of a nonautonomous ordinary differential equation. This continuous-time system is a relevant approximation of the Adam iterates, in the sense that the interpolated Adam process converges weakly toward the solution to the ODE. The existence and the uniqueness of the solution are established. We further show the convergence of the solution toward the critical points of the objective function and quantify its convergence rate under a Lojasiewicz assumption. Then, we introduce a novel decreasing stepsize version of Adam. Under mild assumptions, it is shown that the iterates are almost surely bounded and converge almost surely to critical points of the objective function. Finally, we analyze the fluctuations of the algorithm by means of a conditional central limit theorem.
引用
收藏
页码:244 / 274
页数:31
相关论文
共 50 条
  • [41] Stochastic consensus dynamics for nonconvex optimization on the Stiefel manifold: Mean-field limit and convergence
    Ha, Seung-Yeal
    Kang, Myeongju
    Kim, Dohyun
    Kim, Jeongho
    Yang, Insoon
    MATHEMATICAL MODELS & METHODS IN APPLIED SCIENCES, 2022, 32 (03): : 533 - 617
  • [42] Stochastic Bigger Subspace Algorithms for Nonconvex Stochastic Optimization
    Yuan, Gonglin
    Zhou, Yingjie
    Wang, Liping
    Yang, Qingyuan
    IEEE ACCESS, 2021, 9 : 119818 - 119829
  • [43] A Stochastic Proximal Point Algorithm: Convergence and Application to Convex Optimization
    Bianchi, Pascal
    2015 IEEE 6TH INTERNATIONAL WORKSHOP ON COMPUTATIONAL ADVANCES IN MULTI-SENSOR ADAPTIVE PROCESSING (CAMSAP), 2015,
  • [44] A Global Convergence Algorithm with Stochastic Search for Constrained Optimization Problems
    Zhou Changyin
    He Guoping
    SECOND INTERNATIONAL CONFERENCE ON GENETIC AND EVOLUTIONARY COMPUTING: WGEC 2008, PROCEEDINGS, 2008, : 75 - 78
  • [45] A Decentralized Stochastic Algorithm for Coupled Composite Optimization With Linear Convergence
    Lu, Qingguo
    Liao, Xiaofeng
    Deng, Shaojiang
    Li, Huaqing
    IEEE TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING OVER NETWORKS, 2022, 8 : 627 - 640
  • [46] Graphical Convergence of Subgradients in Nonconvex Optimization and Learning
    Davis, Damek
    Drusvyatskiy, Dmitriy
    MATHEMATICS OF OPERATIONS RESEARCH, 2022, 47 (01) : 209 - 231
  • [47] Quantized Distributed Nonconvex Optimization with Linear Convergence
    Xu, Lei
    Yi, Xinlei
    Sun, Jiayue
    Shi, Yang
    Johansson, Karl Henrik
    Yang, Tao
    Proceedings of the IEEE Conference on Decision and Control, 2022, 2022-December : 5837 - 5842
  • [48] Global Convergence of ADMM in Nonconvex Nonsmooth Optimization
    Yu Wang
    Wotao Yin
    Jinshan Zeng
    Journal of Scientific Computing, 2019, 78 : 29 - 63
  • [49] Quantized Distributed Nonconvex Optimization with Linear Convergence
    Xu, Lei
    Yi, Xinlei
    Sun, Jiayue
    Shi, Yang
    Johansson, Karl Henrik
    Yang, Tao
    2022 IEEE 61ST CONFERENCE ON DECISION AND CONTROL (CDC), 2022, : 5837 - 5842
  • [50] A hybrid stochastic optimization framework for composite nonconvex optimization
    Quoc Tran-Dinh
    Pham, Nhan H.
    Phan, Dzung T.
    Nguyen, Lam M.
    MATHEMATICAL PROGRAMMING, 2022, 191 (02) : 1005 - 1071