Constrained Stochastic Gradient Descent for Large-scale Least Squares Problem

被引:0
|
作者
Mu, Yang [1 ]
Ding, Wei [1 ]
Zhou, Tianyi [2 ]
Tao, Dacheng [2 ]
机构
[1] Univ Massachusetts, 100 Morrissey Blvd, Boston, MA 02125 USA
[2] Univ Technol Sydney, Ultimo, NSW 2007, Australia
关键词
Stochastic optimization; Large-scale least squares; online learning; APPROXIMATION; ALGORITHMS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The least squares problem is one of the most important regression problems in statistics, machine learning and data mining. In this paper, we present the Constrained Stochastic Gradient Descent (CSGD) algorithm to solve the large-scale least squares problem. CSGD improves the Stochastic Gradient Descent (SGD) by imposing a provable constraint that the linear regression line passes through the mean point of all the data points. It results in the best regret bound o(logT), and fastest convergence speed among all first order approaches. Empirical studies justify the effectiveness of CSGD by comparing it with SGD and other state-of-the-art approaches. An example is also given to show how to use CSGD to optimize SGD based least squares problems to achieve a better performance.
引用
收藏
页码:883 / 891
页数:9
相关论文
共 50 条
  • [11] Localization with TOA as a Constrained Robust Stochastic Least Squares Problem
    Korkmaz, Sayit
    van der Veen, Alle-Jan
    2008 IEEE INTERNATIONAL CONFERENCE ON ULTRA-WIDEBAND, VOL 1, PROCEEDINGS, 2008, 1 : 197 - 200
  • [12] Large-Scale Least Squares Twin SVMs
    Tanveer, M.
    Sharma, S.
    Muhammad, K.
    ACM TRANSACTIONS ON INTERNET TECHNOLOGY, 2021, 21 (02)
  • [13] Model reduction for large-scale dynamical systems via equality constrained least squares
    An, Yu'e
    Gu, Chuanqing
    JOURNAL OF COMPUTATIONAL AND APPLIED MATHEMATICS, 2010, 234 (08) : 2420 - 2431
  • [14] Algorithmic Stability of Heavy-Tailed Stochastic Gradient Descent on Least Squares
    Raj, Anant
    Barsbey, Melih
    Gurbuzbalaban, Mert
    Zhu, Lingjiong
    Simsekli, Umut
    INTERNATIONAL CONFERENCE ON ALGORITHMIC LEARNING THEORY, VOL 201, 2023, 201 : 1292 - 1342
  • [15] On Projected Stochastic Gradient Descent Algorithm with Weighted Averaging for Least Squares Regression
    Cohen, Kobi
    Nedic, Angelia
    Srikant, R.
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2017, 62 (11) : 5974 - 5981
  • [16] On Projected Stochastic Gradient Descent Algorithm with Weighted Averaging for Least Squares Regression
    Cohen, Kobi
    Nedic, Angelia
    Srikant, R.
    2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 2314 - 2318
  • [17] LARGE SCALE RANKING USING STOCHASTIC GRADIENT DESCENT
    Tas, Engin
    COMPTES RENDUS DE L ACADEMIE BULGARE DES SCIENCES, 2022, 75 (10): : 1419 - 1427
  • [18] VARIANCE REDUCTION IN STOCHASTIC METHODS FOR LARGE-SCALE REGULARIZED LEAST-SQUARES PROBLEMS
    Pilavci, Yusuf Yigit
    Amblard, Pierre-Olivier
    Barthelme, Simon
    Tremblay, Nicolas
    2022 30TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2022), 2022, : 1771 - 1775
  • [19] Adaptive Alternating Stochastic Gradient Descent Algorithms for Large-Scale Latent Factor Analysis
    Qin, Wen
    Luo, Xin
    Zhou, MengChu
    2021 IEEE INTERNATIONAL CONFERENCE ON SERVICES COMPUTING (SCC 2021), 2021, : 285 - 290
  • [20] Breaking the curse of kernelization: Budgeted stochastic gradient descent for large-scale SVM training
    Wang, Z. (zhuang.wang@siemens.com), 1600, Microtome Publishing (13):