Weight Evolution: Improving Deep Neural Networks Training through Evolving InferiorWeight Values

被引:1
|
作者
Lin, Zhenquan [1 ]
Guo, Kailing [1 ]
Xing, Xiaofen [2 ]
Xu, Xiangmin [3 ]
机构
[1] South China Univ Technol, Guangzhou, Peoples R China
[2] South China Univ Technol, UBTECH SCUT Union Lab, Guangzhou, Peoples R China
[3] South China Univ Technol, Inst Modern Ind Technol, Zhongshan, Peoples R China
来源
PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021 | 2021年
基金
中国国家自然科学基金;
关键词
weight evolution; neural networks; training method;
D O I
10.1145/3474085.3475376
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To obtain good performance, convolutional neural networks are usually over-parameterized. This phenomenon has stimulated two interesting topics: pruning the unimportant weights for compression and reactivating the unimportant weights to make full use of network capability. However, current weight reactivation methods usually reactivate the entire filters, which may not be precise enough. Looking back in history, the prosperity of filter pruning is mainly due to its friendliness to hardware implementation, but pruning at a finer structure level, i.e., weight elements, usually leads to better network performance. We study the problem of weight element reactivation in this paper. Motivated by evolution, we select the unimportant filters and update their unimportant elements by combining them with the important elements of important filters, just like gene crossover to produce better offspring, and the proposed method is called weight evolution (WE). WE is mainly composed of four strategies. We propose a global selection strategy and a local selection strategy and combine them to locate the unimportant filters. A forward matching strategy is proposed to find the matched important filters and a crossover strategy is proposed to utilize the important elements of the important filters for updating unimportant filters. WE is plug-in to existing network architectures. Comprehensive experiments show that WE outperforms the other reactivation methods and plug-in training methods with typical convolutional neural networks, especially lightweight networks. Our code is available at https://github.com/BZQLin/Weight-evolution.
引用
收藏
页码:2176 / 2184
页数:9
相关论文
共 50 条
  • [31] Enhancing the MapReduce Training of BP Neural Networks Based on Local Weight Matrix Evolution
    Chen, Wanghu
    Li, Xintian
    Li, Jing
    Wang, Jianwu
    2017 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2017, : 2829 - 2835
  • [32] Training Neural Networks by Rational Weight Functions
    Zhang, Daiyuan
    ARTIFICIAL INTELLIGENCE AND COMPUTATIONAL INTELLIGENCE, PT III, 2011, 7004 : 346 - 353
  • [33] Adaptive Weight Decay for Deep Neural Networks
    Nakamura, Kensuke
    Hong, Byung-Woo
    IEEE ACCESS, 2019, 7 : 118857 - 118865
  • [34] Exploring weight symmetry in deep neural networks
    Hu, Shell Xu
    Zagoruyko, Sergey
    Komodakis, Nikos
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2019, 187
  • [35] Weight and Gradient Centralization in Deep Neural Networks
    Fuhl, Wolfgang
    Kasneci, Enkelejda
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT IV, 2021, 12894 : 227 - 239
  • [36] Hierarchical Weight Averaging for Deep Neural Networks
    Gu, Xiaozhe
    Zhang, Zixun
    Jiang, Yuncheng
    Luo, Tao
    Zhang, Ruimao
    Cui, Shuguang
    Li, Zhen
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (09) : 12276 - 12287
  • [37] Evolving Artificial Neural Networks through Topological Complexification
    Jorgensen, Thomas D.
    Haynes, Barry P.
    Norlund, Charlotte C. F.
    ENGINEERING LETTERS, 2009, 17 (01)
  • [38] Evolving modular neural-networks through exaptation
    Mouret, Jean-Baptiste
    Doncieux, Stephane
    2009 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION, VOLS 1-5, 2009, : 1570 - 1577
  • [39] Evolving Neural Networks through a Reverse Encoding Tree
    Zhang, Haoling
    Yang, Chao-Han Huck
    Zenil, Hector
    Kiani, Narsis A.
    Shen, Yue
    Tegner, Jesper N.
    2020 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2020,
  • [40] Evolution of neural networks using weight mapping
    Pujol, JCF
    Poli, R
    GECCO-99: PROCEEDINGS OF THE GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, 1999, : 1170 - 1177