GEFWA: Gradient-Enhanced Fireworks Algorithm for Optimizing Convolutional Neural Networks

被引:0
|
作者
Chen, Maiyue [1 ,2 ]
Tan, Ying [1 ,2 ,3 ]
机构
[1] Peking Univ, Sch Intelligence Sci & Technol, Beijing, Peoples R China
[2] Peking Univ, Key Lab Machine Perceptron MOE, Beijing, Peoples R China
[3] Peking Univ, Inst Artificial Intelligence, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Fireworks algorithm; Deep learning; Convolutional neural network; Swarm intelligence;
D O I
10.1007/978-3-031-36622-2_26
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The efficacy of evolutionary and swarm intelligence-based black-box optimization algorithms in machine learning has increased their usage, but concerns have been raised about their low sample efficiency owing to their reliance on sampling. Consequently, improving the sample efficiency of conventional black-box optimization algorithms while retaining their strengths is crucial. To this end, we propose a new algorithm called Gradient Enhanced Fireworks Algorithm (GEFWA) that incorporates first-order gradient information into the population-based fireworks algorithm (FWA). We enhance the explosion operator with the gradient-enhanced explosion (GEE) and take advantage of attraction-based cooperation (ABC) for firework collaboration. Experimental results illustrate that GEFWA outperforms traditional first-order stochastic gradient descent-based optimization methods such as Adm and SGD when it comes to optimizing convolutional neural networks. These results demonstrate the potential of integrating gradient information into the FWA framework for addressing large-scale machine learning problems.
引用
收藏
页码:323 / 333
页数:11
相关论文
共 50 条
  • [41] Optimizing Memory Efficiency for Deep Convolutional Neural Networks on GPUs
    Li, Chao
    Yang, Yi
    Feng, Min
    Chakradhar, Srimat
    Zhou, Huiyang
    SC '16: PROCEEDINGS OF THE INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS, 2016, : 633 - 644
  • [42] Data Dropout: Optimizing Training Data for Convolutional Neural Networks
    Wang, Tianyang
    Huan, Jun
    Li, Bo
    2018 IEEE 30TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI), 2018, : 39 - 46
  • [43] Performance Optimizing Method for Sparse Convolutional Neural Networks on GPU
    Dong X.
    Liu L.
    Li J.
    Feng X.-B.
    Ruan Jian Xue Bao/Journal of Software, 2020, 31 (09): : 2944 - 2964
  • [44] Enhanced gradient learning for deep neural networks
    Yan, Ming
    Yang, Jianxi
    Chen, Cen
    Zhou, Joey Tianyi
    Pan, Yi
    Zeng, Zeng
    IET IMAGE PROCESSING, 2022, 16 (02) : 365 - 377
  • [45] Interactive gradient algorithm for artificial neural networks
    Li, JY
    Luo, SW
    Qi, YJ
    Liu, JQ
    Huang, YP
    6TH WORLD MULTICONFERENCE ON SYSTEMICS, CYBERNETICS AND INFORMATICS, VOL VI, PROCEEDINGS: INDUSTRIAL SYSTEMS AND ENGINEERING I, 2002, : 87 - 90
  • [46] The natural gradient learning algorithm for neural networks
    Amari, S
    THEORETICAL ASPECTS OF NEURAL COMPUTATION: A MULTIDISCIPLINARY PERSPECTIVE, 1998, : 1 - 15
  • [47] GRADIENT BASED EVOLUTION TO OPTIMIZE THE STRUCTURE OF CONVOLUTIONAL NEURAL NETWORKS
    Mitschke, Norbert
    Heizmann, Michael
    Noffz, Klaus-Henning
    Wittmann, Ralf
    2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2018, : 3438 - 3442
  • [48] A Gradient Boosting Approach for Training Convolutional and Deep Neural Networks
    Emami, Seyedsaman
    Martinez-Munoz, Gonzalo
    IEEE OPEN JOURNAL OF SIGNAL PROCESSING, 2023, 4 : 313 - 321
  • [49] Convolutional neural networks for enhanced classification mechanisms of metamodels
    Nguyen, Phuong T.
    Di Ruscio, Davide
    Pierantonio, Alfonso
    Di Rocco, Juri
    Iovino, Ludovico
    JOURNAL OF SYSTEMS AND SOFTWARE, 2021, 172
  • [50] Enhanced Online Convolutional Neural Networks for Object Tracking
    Zhang, Dengzhuo
    Gao, Yun
    Zhou, Hao
    Li, Tianwen
    TENTH INTERNATIONAL CONFERENCE ON MACHINE VISION (ICMV 2017), 2018, 10696