Weight Evolution: Improving Deep Neural Networks Training through Evolving InferiorWeight Values

被引:1
|
作者
Lin, Zhenquan [1 ]
Guo, Kailing [1 ]
Xing, Xiaofen [2 ]
Xu, Xiangmin [3 ]
机构
[1] South China Univ Technol, Guangzhou, Peoples R China
[2] South China Univ Technol, UBTECH SCUT Union Lab, Guangzhou, Peoples R China
[3] South China Univ Technol, Inst Modern Ind Technol, Zhongshan, Peoples R China
基金
中国国家自然科学基金;
关键词
weight evolution; neural networks; training method;
D O I
10.1145/3474085.3475376
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To obtain good performance, convolutional neural networks are usually over-parameterized. This phenomenon has stimulated two interesting topics: pruning the unimportant weights for compression and reactivating the unimportant weights to make full use of network capability. However, current weight reactivation methods usually reactivate the entire filters, which may not be precise enough. Looking back in history, the prosperity of filter pruning is mainly due to its friendliness to hardware implementation, but pruning at a finer structure level, i.e., weight elements, usually leads to better network performance. We study the problem of weight element reactivation in this paper. Motivated by evolution, we select the unimportant filters and update their unimportant elements by combining them with the important elements of important filters, just like gene crossover to produce better offspring, and the proposed method is called weight evolution (WE). WE is mainly composed of four strategies. We propose a global selection strategy and a local selection strategy and combine them to locate the unimportant filters. A forward matching strategy is proposed to find the matched important filters and a crossover strategy is proposed to utilize the important elements of the important filters for updating unimportant filters. WE is plug-in to existing network architectures. Comprehensive experiments show that WE outperforms the other reactivation methods and plug-in training methods with typical convolutional neural networks, especially lightweight networks. Our code is available at https://github.com/BZQLin/Weight-evolution.
引用
收藏
页码:2176 / 2184
页数:9
相关论文
共 50 条
  • [1] Evolving Convolutional Neural Networks through Grammatical Evolution
    Lima, Ricardo H. R.
    Pozo, Aurora T. R.
    PROCEEDINGS OF THE 2019 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION (GECCCO'19 COMPANION), 2019, : 179 - 180
  • [2] Improving learning in neural networks through weight initializations
    Mittal, Apeksha
    Singh, Amit Prakash
    Chandra, Pravin
    JOURNAL OF INFORMATION & OPTIMIZATION SCIENCES, 2021, 42 (05): : 951 - 971
  • [3] DNNViz: Training Evolution Visualization for Deep Neural Networks
    Clavien, Gil
    Alberti, Michele
    Pondenkandath, Vinaychandran
    Ingold, Rolf
    Liwicki, Marcus
    2019 6TH SWISS CONFERENCE ON DATA SCIENCE (SDS), 2019, : 19 - 24
  • [4] Evolving Hyperparameters for Training Deep Neural Networks against Adversarial Attacks
    Liu, Jia
    Jin, Yaochu
    2019 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2019), 2019, : 1778 - 1785
  • [5] Improving Crossover of Neural Networks in Evolution Through Speciation
    Verbancsics, Phillip
    INFORMATION PROCESSING IN CELLS AND TISSUES (ICPAT), 2015, 9303 : 221 - 232
  • [6] Improving the Robustness of Deep Neural Networks via Stability Training
    Zheng, Stephan
    Song, Yang
    Leung, Thomas
    Goodfellow, Ian
    2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 4480 - 4488
  • [7] Latent Weight Quantization for Integerized Training of Deep Neural Networks
    Fei, Wen
    Dai, Wenrui
    Zhang, Liang
    Zhang, Luoming
    Li, Chenglin
    Zou, Junni
    Xiong, Hongkai
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2025, 47 (04) : 2816 - 2832
  • [8] Centered Weight Normalization in Accelerating Training of Deep Neural Networks
    Huang, Lei
    Liu, Xianglong
    Liu, Yang
    Lang, Bo
    Tao, Dacheng
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 2822 - 2830
  • [9] Improving training of deep neural networks via Singular Value Bounding
    Jia, Kui
    Tao, Dacheng
    Gao, Shenghua
    Xu, Xiangmin
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 3994 - 4002
  • [10] Weight Normalization: A Simple Reparameterization to Accelerate Training of Deep Neural Networks
    Salimans, Tim
    Kingma, Diederik P.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29