Accelerated Bregman proximal gradient methods for relatively smooth convex optimization

被引:19
|
作者
Hanzely, Filip [1 ,2 ]
Richtarik, Peter [1 ,3 ]
Xiao, Lin [4 ]
机构
[1] King Abdullah Univ Sci & Technol KAUST, Div Comp Elect & Math Sci & Engn CEMSE, Thuwal, Saudi Arabia
[2] Toyota Technol Inst Chicago TTIC, Chicago, IL USA
[3] Moscow Inst Phys & Technol, Dolgoprudnyi, Russia
[4] Microsoft Res, Redmond, WA 98052 USA
关键词
Convex optimization; Relative smoothness; Bregman divergence; Proximal gradient methods; Accelerated gradient methods; 1ST-ORDER METHODS; MINIMIZATION ALGORITHM; DESIGNS;
D O I
10.1007/s10589-021-00273-8
中图分类号
C93 [管理学]; O22 [运筹学];
学科分类号
070105 ; 12 ; 1201 ; 1202 ; 120202 ;
摘要
We consider the problem of minimizing the sum of two convex functions: one is differentiable and relatively smooth with respect to a reference convex function, and the other can be nondifferentiable but simple to optimize. We investigate a triangle scaling property of the Bregman distance generated by the reference convex function and present accelerated Bregman proximal gradient (ABPG) methods that attain an O(k(-gamma)) convergence rate, where gamma is an element of (0, 2] is the triangle scaling exponent (TSE) of the Bregman distance. For the Euclidean distance, we have gamma = 2 and recover the convergence rate of Nesterov's accelerated gradient methods. For non-Euclidean Bregman distances, the TSE can be much smaller (say gamma <= 1), but we show that a relaxed definition of intrinsic TSE is always equal to 2. We exploit the intrinsic TSE to develop adaptive ABPG methods that converge much faster in practice. Although theoretical guarantees on a fast convergence rate seem to be out of reach in general, our methods obtain empirical O(k(-2)) rates in numerical experiments on several applications and provide posterior numerical certificates for the fast rates.
引用
收藏
页码:405 / 440
页数:36
相关论文
共 50 条
  • [21] New inertial proximal gradient methods for unconstrained convex optimization problems
    Peichao Duan
    Yiqun Zhang
    Qinxiong Bu
    Journal of Inequalities and Applications, 2020
  • [22] Accelerated Quasi-Newton Proximal Extragradient: Faster Rate for Smooth Convex Optimization
    Jiang, Ruichen
    Mokhtari, Aryan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36, NEURIPS 2023, 2023,
  • [23] Unified Analysis of Stochastic Gradient Methods for Composite Convex and Smooth Optimization
    Ahmed Khaled
    Othmane Sebbouh
    Nicolas Loizou
    Robert M. Gower
    Peter Richtárik
    Journal of Optimization Theory and Applications, 2023, 199 (2) : 499 - 540
  • [24] Unified Analysis of Stochastic Gradient Methods for Composite Convex and Smooth Optimization
    Khaled, Ahmed
    Sebbouh, Othmane
    Loizou, Nicolas
    Gower, Robert M.
    Richtarik, Peter
    JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 2023, 199 (02) : 499 - 540
  • [25] Fast Proximal Gradient Methods for Nonsmooth Convex Optimization for Tomographic Image Reconstruction
    Elias S. Helou
    Marcelo V. W. Zibetti
    Gabor T. Herman
    Sensing and Imaging, 2020, 21
  • [26] Fast Proximal Gradient Methods for Nonsmooth Convex Optimization for Tomographic Image Reconstruction
    Helou, Elias S.
    Zibetti, Marcelo V. W.
    Herman, Gabor T.
    SENSING AND IMAGING, 2020, 21 (01):
  • [27] Accelerated Gradient Methods for Geodesically Convex Optimization: Tractable Algorithms and Convergence Analysis
    Kim, Jungbin
    Yang, Insoon
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [28] Gradient flows and proximal splitting methods: A unified view on accelerated and stochastic optimization
    Franca, Guilherme
    Robinson, Daniel P.
    Vidal, Rene
    PHYSICAL REVIEW E, 2021, 103 (05)
  • [29] Accelerated gradient sliding for structured convex optimization
    Lan, Guanghui
    Ouyang, Yuyuan
    COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2022, 82 (02) : 361 - 394
  • [30] Accelerated gradient sliding for structured convex optimization
    Guanghui Lan
    Yuyuan Ouyang
    Computational Optimization and Applications, 2022, 82 : 361 - 394