Weight Evolution: Improving Deep Neural Networks Training through Evolving InferiorWeight Values

被引:1
|
作者
Lin, Zhenquan [1 ]
Guo, Kailing [1 ]
Xing, Xiaofen [2 ]
Xu, Xiangmin [3 ]
机构
[1] South China Univ Technol, Guangzhou, Peoples R China
[2] South China Univ Technol, UBTECH SCUT Union Lab, Guangzhou, Peoples R China
[3] South China Univ Technol, Inst Modern Ind Technol, Zhongshan, Peoples R China
来源
PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021 | 2021年
基金
中国国家自然科学基金;
关键词
weight evolution; neural networks; training method;
D O I
10.1145/3474085.3475376
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To obtain good performance, convolutional neural networks are usually over-parameterized. This phenomenon has stimulated two interesting topics: pruning the unimportant weights for compression and reactivating the unimportant weights to make full use of network capability. However, current weight reactivation methods usually reactivate the entire filters, which may not be precise enough. Looking back in history, the prosperity of filter pruning is mainly due to its friendliness to hardware implementation, but pruning at a finer structure level, i.e., weight elements, usually leads to better network performance. We study the problem of weight element reactivation in this paper. Motivated by evolution, we select the unimportant filters and update their unimportant elements by combining them with the important elements of important filters, just like gene crossover to produce better offspring, and the proposed method is called weight evolution (WE). WE is mainly composed of four strategies. We propose a global selection strategy and a local selection strategy and combine them to locate the unimportant filters. A forward matching strategy is proposed to find the matched important filters and a crossover strategy is proposed to utilize the important elements of the important filters for updating unimportant filters. WE is plug-in to existing network architectures. Comprehensive experiments show that WE outperforms the other reactivation methods and plug-in training methods with typical convolutional neural networks, especially lightweight networks. Our code is available at https://github.com/BZQLin/Weight-evolution.
引用
收藏
页码:2176 / 2184
页数:9
相关论文
共 50 条
  • [21] MULTILINGUAL TRAINING OF DEEP NEURAL NETWORKS
    Ghoshal, Arnab
    Swietojanski, Pawel
    Renals, Steve
    2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 7319 - 7323
  • [22] Training deep quantum neural networks
    Beer, Kerstin
    Bondarenko, Dmytro
    Farrelly, Terry
    Osborne, Tobias J.
    Salzmann, Robert
    Scheiermann, Daniel
    Wolf, Ramona
    NATURE COMMUNICATIONS, 2020, 11 (01)
  • [23] Training deep quantum neural networks
    Kerstin Beer
    Dmytro Bondarenko
    Terry Farrelly
    Tobias J. Osborne
    Robert Salzmann
    Daniel Scheiermann
    Ramona Wolf
    Nature Communications, 11
  • [24] NOISY TRAINING FOR DEEP NEURAL NETWORKS
    Meng, Xiangtao
    Liu, Chao
    Zhang, Zhiyong
    Wang, Dong
    2014 IEEE CHINA SUMMIT & INTERNATIONAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (CHINASIP), 2014, : 16 - 20
  • [25] Backpropagation Through States: Training Neural Networks with Sequentially Semiseparable Weight Matrices
    Kissel, Matthias
    Gottwald, Martin
    Gjeroska, Biljana
    Paukner, Philipp
    Diepold, Klaus
    PROGRESS IN ARTIFICIAL INTELLIGENCE, EPIA 2022, 2022, 13566 : 476 - 487
  • [26] Improving the Accuracy of Deep Neural Networks Through Developing New Activation Functions
    Mercioni, Marina Adriana
    Tat, Angel Marcel
    Holban, Stefan
    2020 IEEE 16TH INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTER COMMUNICATION AND PROCESSING (ICCP 2020), 2020, : 385 - 391
  • [27] Evolving Deep Convolutional Neural Networks for Image Classification
    Sun, Yanan
    Xue, Bing
    Zhang, Mengjie
    Yen, Gary G.
    IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2020, 24 (02) : 394 - 407
  • [28] Evolving the Topology of Large Scale Deep Neural Networks
    Assuncao, Filipe
    Lourenco, Nuno
    Machado, Penousal
    Ribeiro, Bernardete
    GENETIC PROGRAMMING (EUROGP 2018), 2018, 10781 : 19 - 34
  • [29] Improving Deep Neural Networks with Multilayer Maxout Networks
    Sun, Weichen
    Su, Fei
    Wang, Leiquan
    2014 IEEE VISUAL COMMUNICATIONS AND IMAGE PROCESSING CONFERENCE, 2014, : 334 - 337
  • [30] Memetic Evolution of Deep Neural Networks
    Lorenzo, Pablo Ribalta
    Nalepa, Jakub
    GECCO'18: PROCEEDINGS OF THE 2018 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, 2018, : 505 - 512