Experimental Comparison of Stochastic Optimizers in Deep Learning

被引:18
|
作者
Okewu, Emmanuel [1 ]
Adewole, Philip [2 ]
Sennaike, Oladipupo [2 ]
机构
[1] Univ Lagos, Ctr Informat Technol & Syst, Lagos, Nigeria
[2] Univ Lagos, Dept Comp Sci, Lagos, Nigeria
关键词
Deep learning; Deep neural networks; Error function; Neural network parameters; Stochastic optimization; NEURAL-NETWORKS;
D O I
10.1007/978-3-030-24308-1_55
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
The stochastic optimization problem in deep learning involves finding optimal values of loss function and neural network parameters using a meta-heuristic search algorithm. The fact that these values cannot be reasonably obtained by using a deterministic optimization technique underscores the need for an iterative method that randomly picks data segments, arbitrarily determines initial values of optimization (network) parameters and steadily computes series of error functions until a tolerable error is attained. The typical stochastic optimization algorithm for training deep neural networks as a non-convex optimization problem is gradient descent. It has existing extensions like Stochastic Gradient Descent, Adagrad, Adadelta, RMSProp and Adam. In terms of accuracy, convergence rate and training time, each of these stochastic optimizers represents an improvement. However, there is room for further improvement. This paper presents outcomes of series of experiments conducted with a view to providing empirical evidences of successes made so far. We used Python deep learning libaries (Tensorflow and Keras API) for our experiments. Each algorithm is executed, results collated, and a case made for further research in deep learning to improve training time and convergence rate of deep neural network, as well as accuracy of outcomes. This is in response to the growing demands for deep learning in mission-critical and highly sophisticated decision making processes across industry verticals.
引用
收藏
页码:704 / 715
页数:12
相关论文
共 50 条
  • [41] A comprehensive comparison of large scale global optimizers
    LaTorre, Antonio
    Muelas, Santiago
    Pena, Jose-Maria
    INFORMATION SCIENCES, 2015, 316 : 517 - 549
  • [42] Automated pneumonia detection on chest X-ray images: A deep learning approach with different optimizers and transfer learning architectures
    Manickam, Adhiyaman
    Jiang, Jianmin
    Zhou, Yu
    Sagar, Abhinav
    Soundrapandiyan, Rajkumar
    Samuel, R. Dinesh Jackson
    MEASUREMENT, 2021, 184
  • [43] An experimental comparison of Deep Learning strategies for AUV navigation in DVL-denied environments
    Topini, Edoardo
    Fanelli, Francesco
    Topini, Alberto
    Pebody, Miles
    Ridolfi, Alessandro
    Phillips, Alexander B.
    Allotta, Benedetto
    OCEAN ENGINEERING, 2023, 274
  • [44] Online measurement and segmentation algorithm of coarse aggregate based on deep learning and experimental comparison
    Hu, Xiang
    Fang, Huaiying
    Yang, Jianhong
    Fan, Lulu
    Lin, Wenhua
    Li, Jiantao
    CONSTRUCTION AND BUILDING MATERIALS, 2022, 327
  • [45] SIMOP efficient reactive distillation optimization using stochastic optimizers
    Lima, RM
    Salcedo, RL
    Barbosa, D
    CHEMICAL ENGINEERING SCIENCE, 2006, 61 (05) : 1718 - 1739
  • [46] Generalization Bounds using Lower Tail Exponents in Stochastic Optimizers
    Hodgkinson, Liam
    Simsekli, Umut
    Khanna, Rajiv
    Mahoney, Michael W.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [47] Modeling and optimization of chemical processes: Ascend IV and stochastic optimizers
    Silva, Heider G.
    Salcedo, Romualdo L. R.
    PROCEEDINGS OF THE 17TH IASTED INTERNATIONAL CONFERENCE ON MODELLING AND SIMULATION, 2006, : 493 - +
  • [48] Experimental comparison of stochastic MIMO channel models
    Pardonche, JF
    Berbineau, M
    Seguinot, C
    ANNALS OF TELECOMMUNICATIONS, 2005, 60 (5-6) : 649 - 680
  • [49] Towards Learning Universal Hyperparameter Optimizers with Transformers
    Chen, Yutian
    Song, Xingyou
    Lee, Chansoo
    Wang, Zi
    Zhang, Qiuyi
    Dohan, David
    Kawakami, Kazuya
    Kochanski, Greg
    Doucet, Arnaud
    Ranzato, Marc'aurelio
    Perel, Sagi
    de Freitas, Nando
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [50] Learning effective stochastic differential equations from microscopic simulations: Linking stochastic numerics to deep learning
    Dietrich, Felix
    Makeev, Alexei
    Kevrekidis, George
    Evangelou, Nikolaos
    Bertalan, Tom
    Reich, Sebastian
    Kevrekidis, Ioannis G.
    CHAOS, 2023, 33 (02)