A Population-Based Hybrid Approach for Hyperparameter Optimization of Neural Networks

被引:10
|
作者
Japa, Luis [1 ]
Serqueira, Marcello [2 ]
Mendonca, Israel [1 ]
Aritsugi, Masayoshi [3 ]
Bezerra, Eduardo [2 ]
Gonzalez, Pedro Henrique [4 ]
机构
[1] Kumamoto Univ, Grad Sch Sci & Technol, Kumamoto 8608555, Japan
[2] Fed Ctr Technol Educ Rio De Janeiro CEFET RJ, BR-20271110 Rio De Janeiro, Brazil
[3] Kumamoto Univ, Fac Adv Sci & Technol, Kumamoto 8608555, Japan
[4] Univ Fed Rio de Janeiro, Syst Engn & Comp Sci Postgrad Program, BR-21941914 Rio De Janeiro, Brazil
关键词
Genetic algorithms; hyperparameter optimization; machine learning; KEY GENETIC ALGORITHM;
D O I
10.1109/ACCESS.2023.3277310
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Hyperparameter optimization is a fundamental part of Auto Machine Learning (AutoML) and it has been widely researched in recent years; however, it still remains as one of the main challenges in this area. Motivated by the need of faster and more accurate hyperparameter optimization algorithms we developed HyperBRKGA, a new population-based approach for hyperparameter optimization. HyperBRKGA combines the Biased Random Key Genetic Algorithm with an Exploitation Method in order to search the hyperparameter space more efficiently than other commonly used hyperparameter optimization algorithms, such as Grid Search, Random Search, CMA-ES or Bayesian Optimization. We develop and test two different alternatives for this Exploitation Method: Random Walk and Bayesian Walk. We also discuss and implement other schemes, such as a Training Data Reduction Strategy and a Diversity Control strategy, in order to further improve the efficacy of our method. We performed several computational experiments on 8 different datasets to assess the effectiveness of the proposed approach. Results showed that HyperBRKGA could find hyperparameter configurations that outperformed in terms of predictive quality the baseline methods in 6 out of 8 datasets while showing a reasonable execution time. Lastly, we conducted an ablation study and showed that the addition of every component was relevant to achieving high quality results.
引用
收藏
页码:50752 / 50768
页数:17
相关论文
共 50 条
  • [31] Gradient based hyperparameter optimization in Echo State Networks
    Thiede, Luca Anthony
    Parlitz, Ulrich
    NEURAL NETWORKS, 2019, 115 : 23 - 29
  • [32] Neural network hyperparameter optimization based on improved particle swarm optimization
    谢晓燕
    HE Wanqi
    ZHU Yun
    YU Jinhao
    High Technology Letters, 2023, 29 (04) : 427 - 433
  • [33] Hyperparameter importance and optimization of quantum neural networks across small datasets
    Moussa, Charles
    Patel, Yash J.
    Dunjko, Vedran
    Baeck, Thomas
    van Rijn, Jan N.
    MACHINE LEARNING, 2023, 113 (4) : 1941 - 1966
  • [34] Tensor Network Methods for Hyperparameter Optimization and Compression of Convolutional Neural Networks
    Naumov, A.
    Melnikov, A.
    Perelshtein, M.
    Melnikov, Ar.
    Abronin, V.
    Oksanichenko, F.
    APPLIED SCIENCES-BASEL, 2025, 15 (04):
  • [35] Automatic Hyperparameter Optimization for Arbitrary Neural Networks in Serverless AWS Cloud
    Kaplunovich, Alex
    Yesha, Yelena
    2021 12TH INTERNATIONAL CONFERENCE ON INFORMATION AND COMMUNICATION SYSTEMS (ICICS), 2021, : 69 - 76
  • [36] Hyperparameter importance and optimization of quantum neural networks across small datasets
    Charles Moussa
    Yash J. Patel
    Vedran Dunjko
    Thomas Bäck
    Jan N. van Rijn
    Machine Learning, 2024, 113 : 1941 - 1966
  • [37] A Population-based Ant Colony Optimization Approach for DNA Sequence Optimization
    Kurniawan, Tri Basuki
    Ibrahim, Zuwairie
    Khalid, Noor Khafifah
    Khalid, Marzuki
    2009 THIRD ASIA INTERNATIONAL CONFERENCE ON MODELLING & SIMULATION, VOLS 1 AND 2, 2009, : 246 - 251
  • [38] A general framework for population-based distributed optimization over networks
    Ai, Wu
    Chen, Weisheng
    Xie, Jin
    INFORMATION SCIENCES, 2017, 418 : 136 - 152
  • [39] Nonsmooth Optimization-Based Hyperparameter-Free Neural Networks for Large-Scale Regression
    Karmitsa, Napsu
    Taheri, Sona
    Joki, Kaisa
    Paasivirta, Pauliina
    Bagirov, Adil M.
    Makela, Marko M.
    ALGORITHMS, 2023, 16 (09)
  • [40] Static facial expression recognition using convolutional neural networks based on transfer learning and hyperparameter optimization
    Ozcan, Tayyip
    Basturk, Alper
    MULTIMEDIA TOOLS AND APPLICATIONS, 2020, 79 (35-36) : 26587 - 26604