An adaptive mechanism to achieve learning rate dynamically

被引:1
|
作者
Jinjing Zhang
Fei Hu
Li Li
Xiaofei Xu
Zhanbo Yang
Yanbin Chen
机构
[1] Southwest University,School of Computer and Information Science
[2] Chongqing University of Education,Network Centre
[3] Chongqing University,School of Computer Science
来源
关键词
Adaptive mechanism; Learning rate; Adaptive exponential decay rates; Gradient;
D O I
暂无
中图分类号
学科分类号
摘要
Gradient descent is prevalent for large-scale optimization problems in machine learning; especially it nowadays plays a major role in computing and correcting the connection strength of neural networks in deep learning. However, many gradient-based optimization methods contain more sensitive hyper-parameters which require endless ways of configuring. In this paper, we present a novel adaptive mechanism called adaptive exponential decay rate (AEDR). AEDR uses an adaptive exponential decay rate rather than a fixed and preconfigured one, and it can allow us to eliminate one otherwise tuning sensitive hyper-parameters. AEDR also can be used to calculate exponential decay rate adaptively by employing the moving average of both gradients and squared gradients over time. The mechanism is then applied to Adadelta and Adam; it reduces the number of hyper-parameters of Adadelta and Adam to only a single one to be turned. We use neural network of long short-term memory and LeNet to demonstrate how learning rate adapts dynamically. We show promising results compared with other state-of-the-art methods on four data sets, the IMDB (movie reviews), SemEval-2016 (sentiment analysis in twitter) (IMDB), CIFAR-10 and Pascal VOC-2012.
引用
收藏
页码:6685 / 6698
页数:13
相关论文
共 50 条
  • [1] An adaptive mechanism to achieve learning rate dynamically
    Zhang, Jinjing
    Hu, Fei
    Li, Li
    Xu, Xiaofei
    Yang, Zhanbo
    Chen, Yanbin
    NEURAL COMPUTING & APPLICATIONS, 2019, 31 (10): : 6685 - 6698
  • [2] AEDR: An Adaptive Mechanism to Achieve Online Learning Rate Dynamically
    Zhang, Jinjing
    Hu, Fei
    Li, Li
    Xu, Xiaofei
    Yang, Zhanbo
    2017 IEEE 29TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2017), 2017, : 30 - 36
  • [3] Adaptive Activity Learning with Dynamically Available Context
    Wen, Jiahui
    Indulska, Jadwiga
    Zhong, Mingyang
    2016 IEEE INTERNATIONAL CONFERENCE ON PERVASIVE COMPUTING AND COMMUNICATIONS (PERCOM), 2016,
  • [4] Dynamically Adaptive Policies for Dynamically Adaptive Telecommunications Networks
    van der Meer, Sven
    Keeney, John
    Fallon, Liam
    2015 11TH INTERNATIONAL CONFERENCE ON NETWORK AND SERVICE MANAGEMENT (CNSM), 2015, : 182 - 186
  • [5] Segmentation of Tracking Sequences Using Dynamically Updated Adaptive Learning
    Michailovich, Oleg
    Tannenbaum, Allen
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2008, 17 (12) : 2403 - 2412
  • [6] TEACHERS' REQUIREMENTS IN DYNAMICALLY ADAPTIVE E-LEARNING SYSTEMS
    Ognjanovic, Ivana
    Sendelj, Ramo
    EDULEARN12: 4TH INTERNATIONAL CONFERENCE ON EDUCATION AND NEW LEARNING TECHNOLOGIES, 2012, : 2252 - 2261
  • [7] On the learning mechanism of adaptive filters
    Nascimento, VH
    Sayed, AH
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2000, 48 (06) : 1609 - 1625
  • [8] Dynamically Adaptive Scrubbing Mechanism for Improved Reliability in Reconfigurable Embedded Systems
    Santos, Rui
    Venkataraman, Shyamsundar
    Kumar, Akash
    2015 52ND ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2015,
  • [9] Adaptive learning rate clipping stabilizes learning
    Ede, Jeffrey M.
    Beanland, Richard
    MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2020, 1 (01):
  • [10] Functions of Learning Rate in Adaptive Reward Learning
    Wu, Xi
    Wang, Ting
    Liu, Chang
    Wu, Tao
    Jiang, Jiefeng
    Zhou, Dong
    Zhou, Jiliu
    FRONTIERS IN HUMAN NEUROSCIENCE, 2017, 11