ON THE DIVERGENCE OF DECENTRALIZED NONCONVEX OPTIMIZATION

被引:4
|
作者
Hong, M. I. N. G. Y. I. [1 ]
Zeng, S. I. L. I. A. N. G. [1 ]
Zhang, J. U. N. Y. U. [2 ]
Sun, H. A. O. R. A. N. [1 ]
机构
[1] Univ Minnesota Twin Cities, Dept ECE, Minneapolis, MN 55455 USA
[2] Natl Univ Singapore, Dept ISEM, Singapore 119007, Singapore
关键词
decentralized optimization; nonconvex problems; Lipschitz continuous gradient; DISTRIBUTED OPTIMIZATION; CONVERGENCE; ALGORITHM; STRATEGIES;
D O I
10.1137/20M1353149
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
agents jointly optimize the nonconvex objective function f(u) := 1/N\sumN Abstract. In this work, we study a generic class of decentralized algorithms in which N i=1 fi(u), while only communicating with their neighbors. This class of problems has become popular in modeling many signal processing and decentralized machine learning applications, and efficient algorithms have been proposed for such a type of problem. However, most of the existing decentralized algorithms require that the local function gradients V fi's as well as the average function gradient Vf are Lipschitz, that is, the local Lipschitz conditions (LLC) and global Lipschitz condition (GLC) are satisfied. In this work, we first demonstrate the importance of the above Lipschitzness assumptions on the state-of-the-art decentralized algorithms. First, by constructing a series of examples, we show that when the LLC on the local function gradient V fi's are not satisfied, a number of state-of-the-art decentralized algorithms diverge, even if the global Lipschitz condition (GLC) still holds. This observation brings out a fundamental theoretical issue of the existing decentralized algorithms---their convergence conditions are strictly stronger than centralized algorithms such as the gradient descent, which only requires the GLC. Our observation raises an important open question: How to design decentralized algorithms when the LLC, or even the GLC, is not satisfied? To address this question, we design two first-order algorithms, which are capable of computing stationary solutions of the original problem with neither the LLC nor the GLC condition. In particular, we show that the proposed algorithms converge sublinearly to a certain \epsilon -stationary solution, where the precise rate depends on various algorithmic and problem parameters. In particular, if the local function fi's are lower bounded Qth order polynomials, then the rate becomes 0(1/\epsilonQ-1) for Q \geq 2 (where the 0 notation hides some constants such as dependency on the network topology). Such a rate is tight for the special case of Q = 2 where each fi satisfies LLC. To our knowledge, this is the first attempt that studies decentralized nonconvex optimization problems with neither the LLC nor the GLC.
引用
收藏
页码:2879 / 2908
页数:30
相关论文
共 50 条
  • [11] GNSD: A GRADIENT-TRACKING BASED NONCONVEX STOCHASTIC ALGORITHM FOR DECENTRALIZED OPTIMIZATION
    Lu, Songtao
    Zhang, Xinwei
    Sun, Haoran
    Hong, Mingyi
    2019 IEEE DATA SCIENCE WORKSHOP (DSW), 2019, : 315 - 321
  • [12] An Efficient Stochastic Algorithm for Decentralized Nonconvex-Strongly-Concave Minimax Optimization
    Chen, Lesi
    Ye, Haishan
    Luo, Luo
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [13] FAST DECENTRALIZED NONCONVEX FINITE-SUM OPTIMIZATION WITH RECURSIVE VARIANCE REDUCTION
    Xin, Ran
    Khan, Usman A.
    Kar, Soummya
    SIAM JOURNAL ON OPTIMIZATION, 2022, 32 (01) : 1 - 28
  • [14] BEER: Fast O(1/T) Rate for Decentralized Nonconvex Optimization with Communication Compression
    Zhao, Haoyu
    Li, Boyue
    Li, Zhize
    Richtarik, Peter
    Chi, Yuejie
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [15] Efficient Decentralized Stochastic Gradient Descent Method for Nonconvex Finite-Sum Optimization Problems
    Zhan, Wenkang
    Wu, Gang
    Gao, Hongchang
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 9006 - 9013
  • [16] D-SPIDER-SFO: A Decentralized Optimization Algorithm with Faster Convergence Rate for Nonconvex Problems
    Pan, Taoxing
    Liu, Jun
    Wang, Jie
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 1619 - 1626
  • [17] A Faster Decentralized Algorithm for Nonconvex Minimax Problems
    Xian, Wenhan
    Huang, Feihu
    Zhang, Yanfu
    Huang, Heng
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [18] Decentralized Riemannian Algorithm for Nonconvex Minimax Problems
    Wu, Xidong
    Hu, Zhengmian
    Huang, Heng
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 9, 2023, : 10370 - 10378
  • [19] Generalized-Smooth Nonconvex Optimization is As Efficient As Smooth Nonconvex Optimization
    Chen, Ziyi
    Zhou, Yi
    Liang, Yingbin
    Lu, Zhaosong
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 202, 2023, 202
  • [20] DESTRESS: Computation-Optimal and Communication-Efficient Decentralized Nonconvex Finite-Sum Optimization
    Li, Boyue
    Li, Zhize
    Chi, Yuejie
    SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2022, 4 (03): : 1031 - 1051