GEFWA: Gradient-Enhanced Fireworks Algorithm for Optimizing Convolutional Neural Networks

被引:0
|
作者
Chen, Maiyue [1 ,2 ]
Tan, Ying [1 ,2 ,3 ]
机构
[1] Peking Univ, Sch Intelligence Sci & Technol, Beijing, Peoples R China
[2] Peking Univ, Key Lab Machine Perceptron MOE, Beijing, Peoples R China
[3] Peking Univ, Inst Artificial Intelligence, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Fireworks algorithm; Deep learning; Convolutional neural network; Swarm intelligence;
D O I
10.1007/978-3-031-36622-2_26
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The efficacy of evolutionary and swarm intelligence-based black-box optimization algorithms in machine learning has increased their usage, but concerns have been raised about their low sample efficiency owing to their reliance on sampling. Consequently, improving the sample efficiency of conventional black-box optimization algorithms while retaining their strengths is crucial. To this end, we propose a new algorithm called Gradient Enhanced Fireworks Algorithm (GEFWA) that incorporates first-order gradient information into the population-based fireworks algorithm (FWA). We enhance the explosion operator with the gradient-enhanced explosion (GEE) and take advantage of attraction-based cooperation (ABC) for firework collaboration. Experimental results illustrate that GEFWA outperforms traditional first-order stochastic gradient descent-based optimization methods such as Adm and SGD when it comes to optimizing convolutional neural networks. These results demonstrate the potential of integrating gradient information into the FWA framework for addressing large-scale machine learning problems.
引用
收藏
页码:323 / 333
页数:11
相关论文
共 50 条
  • [21] Optimizing nonlinear activation function for convolutional neural networks
    Munender Varshney
    Pravendra Singh
    Signal, Image and Video Processing, 2021, 15 : 1323 - 1330
  • [22] Deploying and Optimizing Convolutional Neural Networks on Heterogeneous Architecture
    Jiang, Junning
    Cai, Liang
    Dong, Feng
    Yu, Kehua
    Chen, Ke
    Qu, Wei
    Jiang, Jianfei
    2019 IEEE 13TH INTERNATIONAL CONFERENCE ON ASIC (ASICON), 2019,
  • [23] Optimizing nonlinear activation function for convolutional neural networks
    Varshney, Munender
    Singh, Pravendra
    SIGNAL IMAGE AND VIDEO PROCESSING, 2021, 15 (06) : 1323 - 1330
  • [24] Convolutional neural networks with fractional order gradient method
    Sheng, Dian
    Wei, Yiheng
    Chen, Yuquan
    Wang, Yong
    NEUROCOMPUTING, 2020, 408 : 42 - 50
  • [25] GRADIENT LOCAL BINARY PATTERN FOR CONVOLUTIONAL NEURAL NETWORKS
    Tang, Jialiang
    Jiang, Ning
    Yu, Wenxin
    2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 744 - 748
  • [26] Gradient-Sensitive Optimization for Convolutional Neural Networks
    Liu, Zhipeng
    Feng, Rui
    Li, Xiuhan
    Wang, Wei
    Wu, Xiaoling
    COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2021, 2021
  • [27] Calibrated Stochastic Gradient Descent for Convolutional Neural Networks
    Zhuo, Li'an
    Zhang, Baochang
    Chen, Chen
    Ye, Qixiang
    Liu, Jianzhuang
    Doermann, David
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 9348 - 9355
  • [28] Understanding the Convolutional Neural Networks with Gradient Descent and Backpropagation
    Zhou, XueFei
    2ND INTERNATIONAL CONFERENCE ON MACHINE VISION AND INFORMATION TECHNOLOGY (CMVIT 2018), 2018, 1004
  • [29] Features gradient-based signals selection algorithm of linear complexity for convolutional neural networks
    Omae, Yuto
    Sakai, Yusuke
    Takahashi, Hirotaka
    AIMS MATHEMATICS, 2024, 9 (01): : 792 - 817
  • [30] Convolutional Neural Networks for Enhanced Compression Techniques
    Gnacek, Matthew
    Heatwole, Cory
    Fan, David
    Hoffman, Marc
    PROCEEDINGS OF THE 2021 IEEE NATIONAL AEROSPACE AND ELECTRONICS CONFERENCE (NAECON), 2021, : 392 - 399