GEFWA: Gradient-Enhanced Fireworks Algorithm for Optimizing Convolutional Neural Networks

被引:0
|
作者
Chen, Maiyue [1 ,2 ]
Tan, Ying [1 ,2 ,3 ]
机构
[1] Peking Univ, Sch Intelligence Sci & Technol, Beijing, Peoples R China
[2] Peking Univ, Key Lab Machine Perceptron MOE, Beijing, Peoples R China
[3] Peking Univ, Inst Artificial Intelligence, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Fireworks algorithm; Deep learning; Convolutional neural network; Swarm intelligence;
D O I
10.1007/978-3-031-36622-2_26
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The efficacy of evolutionary and swarm intelligence-based black-box optimization algorithms in machine learning has increased their usage, but concerns have been raised about their low sample efficiency owing to their reliance on sampling. Consequently, improving the sample efficiency of conventional black-box optimization algorithms while retaining their strengths is crucial. To this end, we propose a new algorithm called Gradient Enhanced Fireworks Algorithm (GEFWA) that incorporates first-order gradient information into the population-based fireworks algorithm (FWA). We enhance the explosion operator with the gradient-enhanced explosion (GEE) and take advantage of attraction-based cooperation (ABC) for firework collaboration. Experimental results illustrate that GEFWA outperforms traditional first-order stochastic gradient descent-based optimization methods such as Adm and SGD when it comes to optimizing convolutional neural networks. These results demonstrate the potential of integrating gradient information into the FWA framework for addressing large-scale machine learning problems.
引用
收藏
页码:323 / 333
页数:11
相关论文
共 50 条
  • [31] Noise-enhanced convolutional neural networks
    Audhkhasi, Kartik
    Osoba, Osonde
    Kosko, Bart
    NEURAL NETWORKS, 2016, 78 : 15 - 23
  • [32] Fast Convolution Algorithm for Convolutional Neural Networks
    Kim, Tae Sun
    Bae, Jihoon
    Sunwoo, Myung Hoon
    2019 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS 2019), 2019, : 258 - 261
  • [33] Gradient-enhanced physics-informed neural networks based on transfer learning for inverse problems of the variable coefficient differential equations
    Lin, Shuning
    Chen, Yong
    PHYSICA D-NONLINEAR PHENOMENA, 2024, 459
  • [34] Spatio-temporal physics-informed neural networks to solve boundary value problems for classical and gradient-enhanced continua
    Nguyen, Duc-Vinh
    Jebahi, Mohamed
    Chinesta, Francisco
    MECHANICS OF MATERIALS, 2024, 198
  • [35] Optimizing Convolutional Neural Networks for Tomato Leaf Disease Classification
    Septiarini, Anindita
    Puspitasari, Novianti
    Kamila, Vina Zahrotun
    Hamdani, Hamdani
    Wati, Masna
    Latifa, Alda
    9TH INTERNATIONAL CONFERENCE ON MECHATRONICS ENGINEERING, ICOM 2024, 2024, : 442 - 447
  • [36] PBIL for Optimizing Hyperparameters of Convolutional Neural Networks and STL Decomposition
    Vasco-Carofilis, Roberto A.
    Gutierrez-Naranjo, Miguel A.
    Cardenas-Montes, Miguel
    HYBRID ARTIFICIAL INTELLIGENT SYSTEMS, HAIS 2020, 2020, 12344 : 147 - 159
  • [37] Optimizing Convolutional Neural Networks for low-resource devices
    Rusu, Cosmin-Ionut
    Czibula, Gabriela
    2018 IEEE 14TH INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTER COMMUNICATION AND PROCESSING (ICCP), 2018, : 245 - 252
  • [38] Configuring and Optimizing of Convolutional Neural Networks for Analyzing the Structure of Metageosystems
    Yamashkin, S. A.
    Yamashkin, A. A.
    Kamaeva, A. A.
    Yamashkina, E. O.
    DATA SCIENCE AND ALGORITHMS IN SYSTEMS, 2022, VOL 2, 2023, 597 : 346 - 356
  • [39] Optimizing Convolutional Neural Networks for Embedded Systems by Means of Neuroevolution
    Badan, Filip
    Sekanina, Lukas
    THEORY AND PRACTICE OF NATURAL COMPUTING, TPNC 2019, 2019, 11934 : 109 - 121
  • [40] Pansharpening Techniques: Optimizing the Loss Function for Convolutional Neural Networks
    Restaino, Rocco
    REMOTE SENSING, 2025, 17 (01)