Efficient Hyperparameter Optimization for Convolution Neural Networks in Deep Learning: A Distributed Particle Swarm Optimization Approach

被引:55
|
作者
Guo, Yu [1 ]
Li, Jian-Yu [1 ]
Zhan, Zhi-Hui [1 ]
机构
[1] South China Univ Technol, Sch Comp Sci & Engn, Guangzhou 510006, Peoples R China
关键词
Convolution neural network (CNN); deep learning; distributed particle swarm optimization algorithm (DPSO); hyperparameter; particle swarm optimization (PSO); ALGORITHM;
D O I
10.1080/01969722.2020.1827797
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Convolution neural network (CNN) is a kind of powerful and efficient deep learning approach that has obtained great success in many real-world applications. However, due to its complex network structure, the intertwining of hyperparameters, and the time-consuming procedure for network training, finding an efficient network configuration for CNN is a challenging yet tough work. To efficiently solve the hyperparameters setting problem, this paper proposes a distributed particle swarm optimization (DPSO) approach, which can optimize the hyperparameters to find high-performing CNNs. Compared to tedious, historical-experience-based, and personal-preference-based manual designs, the proposed DPSO approach can evolve the hyperparameters automatically and globally to obtain promising CNNs, which provides a new idea and approach for finding the global optimal hyperparameter combination. Moreover, by cooperating with the distributed computing techniques, the DPSO approach can have a considerable speedup when compared with the traditional particle swarm optimization (PSO) algorithm. Extensive experiments on widely-used image classification benchmarks have verified that the proposed DPSO approach can effectively find the CNN model with promising performance, and at the same time, has greatly reduced the computational time when compared with traditional PSO.
引用
收藏
页码:36 / 57
页数:22
相关论文
共 50 条
  • [41] A Population-Based Hybrid Approach for Hyperparameter Optimization of Neural Networks
    Japa, Luis
    Serqueira, Marcello
    Mendonca, Israel
    Aritsugi, Masayoshi
    Bezerra, Eduardo
    Gonzalez, Pedro Henrique
    IEEE ACCESS, 2023, 11 : 50752 - 50768
  • [42] An improved approximation approach incorporating particle swarm optimization and a priori information into neural networks
    Han, Fei
    Ling, Qing-Hua
    Huang, De-Shuang
    NEURAL COMPUTING & APPLICATIONS, 2010, 19 (02): : 255 - 261
  • [43] An improved approximation approach incorporating particle swarm optimization and a priori information into neural networks
    Fei Han
    Qing-Hua Ling
    De-Shuang Huang
    Neural Computing and Applications, 2010, 19 : 255 - 261
  • [44] Online Hyperparameter Optimization for Streaming Neural Networks
    Gunasekara, Nuwan
    Gomes, Heitor Murilo
    Pfahringer, Bernhard
    Bifet, Albert
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [45] Parallel hyperparameter optimization of spiking neural networks
    Firmin, Thomas
    Boulet, Pierre
    Talbi, El-Ghazali
    NEUROCOMPUTING, 2024, 609
  • [46] An effective algorithm for hyperparameter optimization of neural networks
    Diaz, G. I.
    Fokoue-Nkoutche, A.
    Nannicini, G.
    Samulowitz, H.
    IBM JOURNAL OF RESEARCH AND DEVELOPMENT, 2017, 61 (4-5)
  • [47] A hybrid particle swarm optimization and its application in neural networks
    Leung, S. Y. S.
    Tang, Yang
    Wong, W. K.
    EXPERT SYSTEMS WITH APPLICATIONS, 2012, 39 (01) : 395 - 405
  • [48] Training neural networks using Multiobjective Particle Swarm Optimization
    Yusiong, John Paul T.
    Naval, Prospero C., Jr.
    ADVANCES IN NATURAL COMPUTATION, PT 1, 2006, 4221 : 879 - 888
  • [49] Evolving Product Unit Neural Networks with Particle Swarm Optimization
    Huang, Rong
    Tong, Shurong
    PROCEEDINGS OF THE FIFTH INTERNATIONAL CONFERENCE ON IMAGE AND GRAPHICS (ICIG 2009), 2009, : 624 - 628
  • [50] Designing neural networks using hybrid particle swarm optimization
    Liu, B
    Wang, L
    Jin, YH
    Huang, DX
    ADVANCES IN NEURAL NETWORKS - ISNN 2005, PT 1, PROCEEDINGS, 2005, 3496 : 391 - 397