Accelerated Bregman proximal gradient methods for relatively smooth convex optimization

被引:19
|
作者
Hanzely, Filip [1 ,2 ]
Richtarik, Peter [1 ,3 ]
Xiao, Lin [4 ]
机构
[1] King Abdullah Univ Sci & Technol KAUST, Div Comp Elect & Math Sci & Engn CEMSE, Thuwal, Saudi Arabia
[2] Toyota Technol Inst Chicago TTIC, Chicago, IL USA
[3] Moscow Inst Phys & Technol, Dolgoprudnyi, Russia
[4] Microsoft Res, Redmond, WA 98052 USA
关键词
Convex optimization; Relative smoothness; Bregman divergence; Proximal gradient methods; Accelerated gradient methods; 1ST-ORDER METHODS; MINIMIZATION ALGORITHM; DESIGNS;
D O I
10.1007/s10589-021-00273-8
中图分类号
C93 [管理学]; O22 [运筹学];
学科分类号
070105 ; 12 ; 1201 ; 1202 ; 120202 ;
摘要
We consider the problem of minimizing the sum of two convex functions: one is differentiable and relatively smooth with respect to a reference convex function, and the other can be nondifferentiable but simple to optimize. We investigate a triangle scaling property of the Bregman distance generated by the reference convex function and present accelerated Bregman proximal gradient (ABPG) methods that attain an O(k(-gamma)) convergence rate, where gamma is an element of (0, 2] is the triangle scaling exponent (TSE) of the Bregman distance. For the Euclidean distance, we have gamma = 2 and recover the convergence rate of Nesterov's accelerated gradient methods. For non-Euclidean Bregman distances, the TSE can be much smaller (say gamma <= 1), but we show that a relaxed definition of intrinsic TSE is always equal to 2. We exploit the intrinsic TSE to develop adaptive ABPG methods that converge much faster in practice. Although theoretical guarantees on a fast convergence rate seem to be out of reach in general, our methods obtain empirical O(k(-2)) rates in numerical experiments on several applications and provide posterior numerical certificates for the fast rates.
引用
收藏
页码:405 / 440
页数:36
相关论文
共 50 条
  • [31] An accelerated proximal gradient method for multiobjective optimization
    Hiroki Tanabe
    Ellen H. Fukuda
    Nobuo Yamashita
    Computational Optimization and Applications, 2023, 86 : 421 - 455
  • [32] ACCELERATED UZAWA METHODS FOR CONVEX OPTIMIZATION
    Tao, Min
    Yuan, Xiaoming
    MATHEMATICS OF COMPUTATION, 2017, 86 (306) : 1821 - 1845
  • [33] Accelerated primal–dual proximal block coordinate updating methods for constrained convex optimization
    Yangyang Xu
    Shuzhong Zhang
    Computational Optimization and Applications, 2018, 70 : 91 - 128
  • [34] An accelerated proximal gradient method for multiobjective optimization
    Tanabe, Hiroki
    Fukuda, Ellen H.
    Yamashita, Nobuo
    COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2023, 86 (02) : 421 - 455
  • [35] Accelerated Proximal Gradient Methods for Nonconvex Programming
    Li, Huan
    Lin, Zhouchen
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [36] Distributed Accelerated Proximal Coordinate Gradient Methods
    Ren, Yong
    Zhu, Jun
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 2655 - 2661
  • [37] RAPID: RAPIDLY ACCELERATED PROXIMAL GRADIENT ALGORITHMS FOR CONVEX MINIMIZATION
    Zhang, Ziming
    Saligrama, Venkatesh
    2015 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING (ICASSP), 2015, : 3796 - 3800
  • [38] A PROXIMAL GRADIENT METHOD WITH BREGMAN DISTANCE IN MULTI-OBJECTIVE OPTIMIZATION*
    Chen, Kangming
    Fukuda, Ellen H.
    Yamashitat, Nobuo
    PACIFIC JOURNAL OF OPTIMIZATION, 2024, 20 (04): : 809 - 826
  • [39] A Unified Convergence Analysis of Stochastic Bregman Proximal Gradient and Extragradient Methods
    Xiantao Xiao
    Journal of Optimization Theory and Applications, 2021, 188 : 605 - 627
  • [40] A Unified Convergence Analysis of Stochastic Bregman Proximal Gradient and Extragradient Methods
    Xiao, Xiantao
    JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 2021, 188 (03) : 605 - 627