Revisiting Optimal Convergence Rate for Smooth and Non-convex Stochastic Decentralized Optimization

被引:0
|
作者
Yuan, Kun [1 ,3 ]
Huang, Xinmeng [2 ]
Chen, Yiming [1 ,4 ]
Zhang, Xiaohan [2 ]
Zhang, Yingya [1 ]
Pan, Pan [1 ]
机构
[1] DAMO Acad, Alibaba Grp, Beijing, Peoples R China
[2] Univ Penn, Philadelphia, PA 19104 USA
[3] Peking Univ, Beijing, Peoples R China
[4] MetaCarbon, Beijing, Peoples R China
关键词
DISTRIBUTED OPTIMIZATION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Decentralized optimization is effective to save communication in large-scale machine learning. Although numerous algorithms have been proposed with theoretical guarantees and empirical successes, the performance limits in decentralized optimization, especially the influence of network topology and its associated weight matrix on the optimal convergence rate, have not been fully understood. While Lu and Sa [44] have recently provided an optimal rate for non-convex stochastic decentralized optimization with weight matrices defined over linear graphs, the optimal rate with general weight matrices remains unclear. This paper revisits non-convex stochastic decentralized optimization and establishes an optimal convergence rate with general weight matrices. In addition, we also establish the optimal rate when non-convex loss functions further satisfy the Polyak-Lojasiewicz (PL) condition. Following existing lines of analysis in literature cannot achieve these results. Instead, we leverage the Ring-Lattice graph to admit general weight matrices while maintaining the optimal relation between the graph diameter and weight matrix connectivity. Lastly, we develop a new decentralized algorithm to nearly attain the above two optimal rates under additional mild conditions.
引用
收藏
页数:14
相关论文
共 50 条
  • [41] Conditions for linear convergence of the gradient method for non-convex optimization
    Abbaszadehpeivasti, Hadi
    de Klerk, Etienne
    Zamani, Moslem
    OPTIMIZATION LETTERS, 2023, 17 (05) : 1105 - 1125
  • [42] Conditions for linear convergence of the gradient method for non-convex optimization
    Hadi Abbaszadehpeivasti
    Etienne de Klerk
    Moslem Zamani
    Optimization Letters, 2023, 17 : 1105 - 1125
  • [43] KKT OPTIMALITY CONDITIONS IN NON-SMOOTH, NON-CONVEX OPTIMIZATION
    Sisarat, Nithirat
    Wangkeeree, Rabian
    Lee, Gue Myung
    JOURNAL OF NONLINEAR AND CONVEX ANALYSIS, 2018, 19 (08) : 1319 - 1329
  • [44] ON A NEW SMOOTHING TECHNIQUE FOR NON-SMOOTH, NON-CONVEX OPTIMIZATION
    Yilmaz, Nurullah
    Sahiner, Ahmet
    NUMERICAL ALGEBRA CONTROL AND OPTIMIZATION, 2020, 10 (03): : 317 - 330
  • [45] On the Almost Sure Convergence of Stochastic Gradient Descent in Non-Convex Problems
    Mertikopoulos, Panayotis
    Hallak, Nadav
    Kavis, Ali
    Cevher, Volkan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [46] Riemannian Stochastic Recursive Momentum Method for non-Convex Optimization
    Han, Andi
    Gao, Junbin
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2505 - 2511
  • [47] Differentially Private Stochastic Optimization: New Results in Convex and Non-Convex Settings
    Bassily, Raef
    Guzman, Cristobal
    Menart, Michael
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [48] Private Stochastic Non-convex Optimization with Improved Utility Rates
    Zhang, Qiuchen
    Ma, Jing
    Lou, Jian
    Xiong, Li
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 3370 - 3376
  • [49] A method to construct a quasi-normal cone for non-convex and non-smooth set and its applications to non-convex and non-smooth optimization
    Li, Hongwei
    Zhou, Dequn
    Liu, Qinghuai
    WCICA 2006: SIXTH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-12, CONFERENCE PROCEEDINGS, 2006, : 1585 - +
  • [50] Relaxed Majorization-Minimization for Non-Smooth and Non-Convex Optimization
    Xu, Chen
    Lin, Zhouchen
    Zhao, Zhenyu
    Zha, Hongbin
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 812 - 818