Constrained Stochastic Gradient Descent for Large-scale Least Squares Problem

被引:0
|
作者
Mu, Yang [1 ]
Ding, Wei [1 ]
Zhou, Tianyi [2 ]
Tao, Dacheng [2 ]
机构
[1] Univ Massachusetts, 100 Morrissey Blvd, Boston, MA 02125 USA
[2] Univ Technol Sydney, Ultimo, NSW 2007, Australia
关键词
Stochastic optimization; Large-scale least squares; online learning; APPROXIMATION; ALGORITHMS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The least squares problem is one of the most important regression problems in statistics, machine learning and data mining. In this paper, we present the Constrained Stochastic Gradient Descent (CSGD) algorithm to solve the large-scale least squares problem. CSGD improves the Stochastic Gradient Descent (SGD) by imposing a provable constraint that the linear regression line passes through the mean point of all the data points. It results in the best regret bound o(logT), and fastest convergence speed among all first order approaches. Empirical studies justify the effectiveness of CSGD by comparing it with SGD and other state-of-the-art approaches. An example is also given to show how to use CSGD to optimize SGD based least squares problems to achieve a better performance.
引用
收藏
页码:883 / 891
页数:9
相关论文
共 50 条
  • [31] Stochastic Mirror Descent for Large-Scale Sparse Recovery
    Ilandarideva, Sasila
    Bekri, Yannis
    Juditsky, Anatoli
    Perchet, Vianney
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 206, 2023, 206
  • [32] Tuple-oriented Compression for Large-scale Mini-batch Stochastic Gradient Descent
    Li, Fengan
    Chen, Lingjiao
    Zeng, Yijing
    Kumar, Arun
    Wu, Xi
    Naughton, Jeffrey F.
    Patel, Jignesh M.
    SIGMOD '19: PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON MANAGEMENT OF DATA, 2019, : 1517 - 1534
  • [33] Quantum gradient descent for linear systems and least squares
    Kerenidis, Iordanis
    Prakash, Anupam
    PHYSICAL REVIEW A, 2020, 101 (02)
  • [34] The turbo decoder as a least squares cost gradient descent
    Walsh, JM
    Johnson, CR
    Regalia, PA
    2005 IEEE 6TH WORKSHOP ON SIGNAL PROCESSING ADVANCES IN WIRELESS COMMUNICATIONS, 2005, : 675 - 679
  • [35] Stochastic subgradient descent method for large-scale robust chance-constrained support vector machines
    Wang, Ximing
    Fan, Neng
    Pardalos, Panos M.
    OPTIMIZATION LETTERS, 2017, 11 (05) : 1013 - 1024
  • [36] Core-elements for large-scale least squares estimation
    Li, Mengyu
    Yu, Jun
    Li, Tao
    Meng, Cheng
    STATISTICS AND COMPUTING, 2024, 34 (06)
  • [37] Stochastic subgradient descent method for large-scale robust chance-constrained support vector machines
    Ximing Wang
    Neng Fan
    Panos M. Pardalos
    Optimization Letters, 2017, 11 : 1013 - 1024
  • [38] A Universal Analysis of Large-Scale Regularized Least Squares Solutions
    Panahi, Ashkan
    Hassibi, Babak
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [39] Large-Scale Regression: A Partition Analysis of the Least Squares Multisplitting
    Inghelbrecht, Gilles
    Pintelon, Rik
    Barbe, Kurt
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2020, 69 (06) : 2635 - 2647
  • [40] Solution of large-scale weighted least-squares problems
    Baryamureeba, V
    NUMERICAL LINEAR ALGEBRA WITH APPLICATIONS, 2002, 9 (02) : 93 - 106