First-Order Optimization (Training) Algorithms in Deep Learning

被引:0
|
作者
Rudenko, Oleg [1 ]
Bezsonov, Oleksandr [1 ]
Oliinyk, Kyrylo [1 ]
机构
[1] Kharkiv Natl Univ Radio Elect, Nauky Ave 14, UA-61166 Kharkiv, Ukraine
关键词
Convolution; Optimization; Neural Network; Algorithm; Gradient; Training; Image Recognition;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The use of artificial neural networks (ANN) requires solving structural and parametric identification problems corresponding to the choice of the optimal network topology and its training (parameter settings). In contrast to the problem of determining the structure, which is a discrete optimization (combinatorial), the search for optimal parameters is carried out in continuous space using some optimization methods. The most widely used optimization method in deep learning is the first-order algorithm that based on gradient descent (GD). In the given paper a comparative analysis of convolutional neural networks training algorithms that are used in tasks of image recognition is provided. Comparison of training algorithms was carried out on the Oxford17 category flower dataset with TensorFlow framework usage. Studies show that for this task a simple gradient descent algorithm is quite effective. At the same time, however, the problem of selecting the optimal values of the algorithms parameters that provide top speed of learning still remains open.
引用
收藏
页数:15
相关论文
共 50 条
  • [31] Variance amplification of accelerated first-order algorithms for strongly convex quadratic optimization problems
    Mohammadi, Hesameddin
    Razaviyayn, Meisam
    Jovanovic, Mihailo R.
    2018 IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2018, : 5753 - 5758
  • [32] FlexPD: A Flexible Framework of First-Order Primal-Dual Algorithms for Distributed Optimization
    Mansoori, Fatemeh
    Wei, Ermin
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2021, 69 : 3500 - 3512
  • [33] Robustness and Convergence Analysis of First-Order Distributed Optimization Algorithms over Subspace Constraints
    Marquis, Dennis J.
    Abou Jaoude, Dany
    Farhood, Mazen
    Woolsey, Craig A.
    2023 AMERICAN CONTROL CONFERENCE, ACC, 2023, : 4753 - 4758
  • [34] A General Framework of Exact Primal-Dual First-Order Algorithms for Distributed Optimization
    Mansoori, Fatemeh
    Wei, Ermin
    2019 IEEE 58TH CONFERENCE ON DECISION AND CONTROL (CDC), 2019, : 6386 - 6391
  • [35] Learning first-order definitions of functions
    Quinlan, JR
    JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 1996, 5 : 139 - 161
  • [36] Learning first-order Bayesian networks
    Chatpatanasiri, R
    Kijsirikul, B
    ADVANCES IN ARTIFICIAL INTELLIGENCE, PROCEEDINGS, 2003, 2671 : 313 - 328
  • [37] Inexact first-order primal-dual algorithms
    Rasch, Julian
    Chambolle, Antonin
    COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2020, 76 (02) : 381 - 430
  • [38] Comparing Unification Algorithms in First-Order Theorem Proving
    Hoder, Krystof
    Voronkov, Andrei
    KI 2009: ADVANCES IN ARTIFICIAL INTELLIGENCE, PROCEEDINGS, 2009, 5803 : 435 - 443
  • [39] First-Order Logic and First-Order Functions
    Freire, Rodrigo A.
    LOGICA UNIVERSALIS, 2015, 9 (03) : 281 - 329
  • [40] Efficient First-Order Algorithms for Adaptive Signal Denoising
    Ostrovskii, Dmitrii
    Harchaoui, Zaid
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80