Stochastic Primal-Dual Proximal ExtraGradient descent for compositely regularized optimization

被引:6
|
作者
Lin, Tianyi [1 ]
Qiao, Linbo [2 ]
Zhang, Teng [3 ]
Feng, Jiashi [4 ]
Zhang, Bofeng [2 ]
机构
[1] Univ Calif Berkeley, Dept Ind Engn & Operat Res, Berkeley, CA USA
[2] Natl Univ Def Technol, Coll Comp, Changsha, Hunan, Peoples R China
[3] Stanford Univ, Dept Management Sci & Engn, Stanford, CA 94305 USA
[4] Natl Univ Singapore, Dept ECE, Singapore, Singapore
关键词
Compositely regularized optimization; Stochastic Primal-Dual Proximal; ExtraGradient descent; SADDLE-POINT; COMPLEXITY; INEQUALITIES;
D O I
10.1016/j.neucom.2017.07.066
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider a wide range of regularized stochastic minimization problems with two regularization terms, one of which is composed with a linear function. This optimization model abstracts a number of important applications in artificial intelligence and machine learning, such as fused Lasso, fused logistic regression, and a class of graph-guided regularized minimization. The computational challenges of this model are in two folds. On one hand, the closed-form solution of the proximal mapping associated with the composed regularization term or the expected objective function is not available. On the other hand, the calculation of the full gradient of the expectation in the objective is very expensive when the number of input data samples is considerably large. To address these issues, we propose a stochastic variant of extra-gradient type methods, namely Stochastic Primal-Dual Proximal ExtraGradient descent (SPDPEG), and analyze its convergence property for both convex and strongly convex objectives. For general convex objectives, the uniformly average iterates generated by SPDPEG converge in expectation with O (1/root t) rate. While for strongly convex objectives, the uniformly and non-uniformly average iterates generated by SPDPEG converge with O (log (t)/t) and O (1/t) rates, respectively. The order of the rate of the proposed algorithm is known to match the best convergence rate for first-order stochastic algorithms. Experiments on fused logistic regression and graph-guided regularized logistic regression problems show that the proposed algorithm performs very efficiently and consistently outperforms other competing algorithms. (C) 2017 Elsevier B.V. All rights reserved.
引用
收藏
页码:516 / 525
页数:10
相关论文
共 50 条
  • [21] A Coordinate Descent Primal-Dual Algorithm and Application to Distributed Asynchronous Optimization
    Bianchi, Pascal
    Hachem, Walid
    Iutzeler, Franck
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2016, 61 (10) : 2947 - 2957
  • [22] Adaptive Stochastic Primal-Dual Coordinate Descent for Separable Saddle Point Problems
    Zhu, Zhanxing
    Storkey, Amos J.
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2015, PT I, 2015, 9284 : 645 - 658
  • [23] A stochastic primal-dual method for a class of nonconvex constrained optimization
    Jin, Lingzi
    Wang, Xiao
    COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2022, 83 (01) : 143 - 180
  • [24] Primal-dual stochastic distributed algorithm for constrained convex optimization
    Niu, Youcheng
    Wang, Haijing
    Wang, Zheng
    Xia, Dawen
    Li, Huaqing
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2019, 356 (16): : 9763 - 9787
  • [25] A Primal-Dual Algorithm for Distributed Stochastic Optimization with Equality Constraints
    Du, Kai-Xin
    Chen, Xing-Min
    2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 5586 - 5591
  • [26] Random extrapolation for primal-dual coordinate descent
    Alacaoglu, Ahmet
    Fercoq, Olivier
    Cevher, Volkan
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [27] A stochastic primal-dual method for a class of nonconvex constrained optimization
    Lingzi Jin
    Xiao Wang
    Computational Optimization and Applications, 2022, 83 : 143 - 180
  • [28] A stochastic primal-dual algorithm for composite optimization with a linear operator
    Wen, Meng
    Zhang, Yongqiang
    Tang, Yuchao
    Cui, Angang
    Peng, Jigen
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 267
  • [29] A Stochastic Primal-Dual algorithm for Distributed Asynchronous Composite Optimization
    Bianchi, Pascal
    Hachem, Walid
    Iutzeler, Franck
    2014 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP), 2014, : 732 - 736
  • [30] Primal-Dual Optimization for Fluids
    Inglis, T.
    Eckert, M. -L.
    Gregson, J.
    Thuerey, N.
    COMPUTER GRAPHICS FORUM, 2017, 36 (08) : 354 - 368