On the Trend-corrected Variant of Adaptive Stochastic Optimization Methods

被引:0
|
作者
Zhou, Bingxin [1 ]
Zheng, Xuebin [1 ]
Gao, Junbin [1 ]
机构
[1] Univ Sydney, Business Sch, Sydney, NSW, Australia
关键词
Stochastic Gradient Descent; ADAM; Deep Learning; Optimization;
D O I
10.1109/ijcnn48605.2020.9207166
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Adam-type optimizers, as a class of adaptive moment estimation methods with the exponential moving average scheme, have been successfully used in many applications of deep learning. Such methods are appealing due to the capability on large-scale sparse datasets with high computational efficiency. In this paper, we present a new framework for Adam-type methods with the trend information when updating the parameters with the adaptive step size and gradients. The additional terms in the algorithm promise an efficient movement on the complex cost surface, and thus the loss would converge more rapidly. We show empirically the importance of adding the trend component, where our framework outperforms the conventional Adam and AMSGrad methods constantly on the classical models with several real-world datasets.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Alternatives to polynomial trend-corrected differences-in-differences models
    Vandenberghe, Vincent
    APPLIED ECONOMICS LETTERS, 2019, 26 (05) : 358 - 361
  • [2] Adaptive Subgradient Methods for Online Learning and Stochastic Optimization
    Duchi, John
    Hazan, Elad
    Singer, Yoram
    JOURNAL OF MACHINE LEARNING RESEARCH, 2011, 12 : 2121 - 2159
  • [3] Adaptive subgradient methods for online learning and stochastic optimization
    Duchi, John
    Hazan, Elad
    Singer, Yoram
    Journal of Machine Learning Research, 2011, 12 : 2121 - 2159
  • [4] Robust adaptive control of robots by stochastic optimization methods
    Marti, K
    IDENTIFICATION, CONTROL AND OPTIMISATION OF ENGINEERING STRUCTURES, 2000, : 15 - 28
  • [5] Corrected Gradient Methods for Distributed Optimization
    Qiu, Zhirong
    Xie, Lihua
    You, Keyou
    PROCEEDINGS OF THE 38TH CHINESE CONTROL CONFERENCE (CCC), 2019, : 6148 - 6153
  • [6] Stochastic optimization methods
    Gazioglu, S.
    JOURNAL OF THE OPERATIONAL RESEARCH SOCIETY, 2007, 58 (06) : 836 - 837
  • [7] Adaptive Stochastic Optimization: A Framework for Analyzing Stochastic Optimization Algorithms
    Curtis, Frank E.
    Scheinberg, Katya
    IEEE SIGNAL PROCESSING MAGAZINE, 2020, 37 (05) : 32 - 42
  • [8] Adaptive Sequential Stochastic Optimization
    Wilson, Craig
    Veeravalli, Venugopal V.
    Nedic, Angelia
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2019, 64 (02) : 496 - 509
  • [9] Stochastic Adaptive Optimization With Dithers
    Xie, Siyu
    Liang, Shu
    Wang, Le Yi
    Yin, George
    Chen, Wen
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2022, 67 (01) : 189 - 202
  • [10] Adaptive Biased Stochastic Optimization
    Yang, Zhuang
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2025, 47 (04) : 3067 - 3078