Evolution and Role of Optimizers in Training Deep Learning Models

被引:2
|
作者
Wen, XiaoHao [1 ]
Zhou, MengChu [2 ,3 ]
机构
[1] Guangxi Normal Univ, Guilin 541004, Peoples R China
[2] Zhejiang Gongshang Univ, Sch Informat & Elect Engn, Hangzhou 310018, Peoples R China
[3] New Jersey Inst Technol, Helen & John C Hartmann Dept Elect & Comp Engn, Newark, NJ 07102 USA
关键词
D O I
10.1109/JAS.2024.124806
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
To perform well, deep learning (DL) models have to be trained well. Which optimizer should be adopted? We answer this question by discussing how optimizers have evolved from traditional methods like gradient descent to more advanced techniques to address challenges posed by high-dimensional and non-convex problem space. Ongoing challenges include their hyperparameter sensitivity, balancing between convergence and generalization performance, and improving interpretability of optimization processes. Researchers continue to seek robust, efficient, and universally applicable optimizers to advance the field of DL across various domains.
引用
收藏
页码:2039 / 2042
页数:4
相关论文
共 50 条
  • [1] Evolution and Role of Optimizers in Training Deep Learning Models
    XiaoHao Wen
    MengChu Zhou
    IEEE/CAA Journal of Automatica Sinica, 2024, 11 (10) : 2039 - 2042
  • [2] Early prediction of chronic kidney disease based on ensemble of deep learning models and optimizers
    Dina Saif
    Amany M. Sarhan
    Nada M. Elshennawy
    Journal of Electrical Systems and Information Technology, 11 (1)
  • [3] Experimental Comparison of Stochastic Optimizers in Deep Learning
    Okewu, Emmanuel
    Adewole, Philip
    Sennaike, Oladipupo
    COMPUTATIONAL SCIENCE AND ITS APPLICATIONS - ICCSA 2019, PT V: 19TH INTERNATIONAL CONFERENCE, SAINT PETERSBURG, RUSSIA, JULY 14, 2019, PROCEEDINGS, PART V, 2019, 11623 : 704 - 715
  • [4] A Comparative Study of Recently Deep Learning Optimizers
    Liu, Yan
    Zhang, Maojun
    Zhong, Zhiwei
    Zeng, Xiangrong
    Long, Xin
    INTERNATIONAL CONFERENCE ON ALGORITHMS, HIGH PERFORMANCE COMPUTING, AND ARTIFICIAL INTELLIGENCE (AHPCAI 2021), 2021, 12156
  • [5] Continuous Training and Deployment of Deep Learning Models
    Prapas, Ioannis
    Derakhshan, Behrouz
    Mahdiraji, Alireza Rezaei
    Markl, Volker
    Datenbank-Spektrum, 2021, 21 (03) : 203 - 212
  • [6] Towards Training Reproducible Deep Learning Models
    Chen, Boyuan
    Wen, Mingzhi
    Shi, Yong
    Lin, Dayi
    Rajbahadur, Gopi Krishnan
    Jiang, Zhen Ming
    2022 ACM/IEEE 44TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING (ICSE 2022), 2022, : 2202 - 2214
  • [7] Tensor Normal Training for Deep Learning Models
    Ren, Yi
    Goldfarb, Donald
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [8] NeuroEvoBench: Benchmarking Evolutionary Optimizers for Deep Learning Applications
    Lange, Robert Tjarko
    Tang, Yujin
    Tian, Yingtao
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [9] Anderson Acceleration for Distributed Training of Deep Learning Models
    Pasini, Massimiliano Lupo
    Yin, Junqi
    Reshniak, Viktor
    Stoyanov, Miroslav K.
    SOUTHEASTCON 2022, 2022, : 289 - 295
  • [10] Pulsed Thermography Dataset for Training Deep Learning Models
    Wei, Ziang
    Osman, Ahmad
    Valeske, Bernd
    Maldague, Xavier
    APPLIED SCIENCES-BASEL, 2023, 13 (05):