On Early Stopping in Gradient Descent Learning

被引:1
|
作者
Yuan Yao
Lorenzo Rosasco
Andrea Caponnetto
机构
[1] Department of Mathematics,
[2] University of California,undefined
[3] C.B.C.L.,undefined
[4] Massachusetts Institute of Technology,undefined
[5] Bldg. E25-201,undefined
[6] 45 Carleton St.,undefined
[7] DISI,undefined
[8] Universita di Genova,undefined
[9] Via Dodecaneso 35,undefined
来源
关键词
Convergence Rate; Gradient Descent; Tikhonov Regularization; Reproduce Kernel Hilbert Space; Gradient Descent Method;
D O I
暂无
中图分类号
学科分类号
摘要
In this paper we study a family of gradient descent algorithms to approximate the regression function from reproducing kernel Hilbert spaces (RKHSs), the family being characterized by a polynomial decreasing rate of step sizes (or learning rate). By solving a bias-variance trade-off we obtain an early stopping rule and some probabilistic upper bounds for the convergence of the algorithms. We also discuss the implication of these results in the context of classification where some fast convergence rates can be achieved for plug-in classifiers. Some connections are addressed with Boosting, Landweber iterations, and the online learning algorithms as stochastic approximations of the gradient descent method.
引用
收藏
页码:289 / 315
页数:26
相关论文
共 50 条
  • [41] Natural gradient descent for on-line learning
    Phys Rev Lett, 24 (5461):
  • [42] Learning with Gradient Descent and Weakly Convex Losses
    Richards, Dominic
    Rabbat, Mike
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130
  • [43] Learning rates of gradient descent algorithm for classification
    Dong, Xue-Mei
    Chen, Di-Rong
    JOURNAL OF COMPUTATIONAL AND APPLIED MATHEMATICS, 2009, 224 (01) : 182 - 192
  • [44] Robust supervised learning with coordinate gradient descent
    Merad, Ibrahim
    Gaiffas, Stephane
    STATISTICS AND COMPUTING, 2023, 33 (05)
  • [45] From Gradient Flow on Population Loss to Learning with Stochastic Gradient Descent
    Sekhari, Ayush
    Kale, Satyen
    Lee, Jason D.
    De Sa, Chris
    Sridharan, Karthik
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [46] Stochastic Gradient Descent for Nonconvex Learning Without Bounded Gradient Assumptions
    Lei, Yunwen
    Hu, Ting
    Li, Guiying
    Tang, Ke
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (10) : 4394 - 4400
  • [47] Towards Learning Stochastic Population Models by Gradient Descent
    Kreikemeyer, Justin N.
    Andelfinger, Philipp
    Uhrmacher, Adelinde M.
    PROCEEDINGS OF THE 38TH ACM SIGSIM INTERNATIONAL CONFERENCE ON PRINCIPLES OF ADVANCED DISCRETE SIMULATION, ACM SIGSIM-PADS 2024, 2024, : 88 - 92
  • [48] Multileave Gradient Descent for Fast Online Learning to Rank
    Schuth, Anne
    Oosterhuis, Harrie
    Whiteson, Shimon
    de Rijke, Maarten
    PROCEEDINGS OF THE NINTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING (WSDM'16), 2016, : 457 - 466
  • [49] Empirical Analysis of Generalization and Learning in XCS with Gradient Descent
    Lanzi, Pier Luca
    Butz, Martin V.
    Goldberg, David E.
    GECCO 2007: GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, VOL 1 AND 2, 2007, : 1814 - +
  • [50] Random Walk Gradient Descent for Decentralized Learning on Graphs
    Ayache, Ghadir
    El Rouayheb, Salim
    2019 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS (IPDPSW), 2019, : 926 - 931