Internal Node Bagging: a Layer-Wise Ensemble Training Method

被引:0
|
作者
Li, Jinhong [1 ]
Yi, Shun [1 ]
机构
[1] North China Univ Technol, Sch Informat, Beijing, Peoples R China
关键词
neural networks; deep learning; ensemble learning; regularization;
D O I
10.1145/3357254.3357268
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
When training neural networks, regularization methods are needed to avoid model overfitting. Dropout is a widely used regularization method, but its working principle is inconclusive and it does not work well for small models. This paper introduced a novel view to understand how dropout works as a layer-wise ensemble training method, that each feature in hidden layers is learned by multiple nodes, and next layer integrates the outputs of these nodes. Basing on the novel understanding of dropout, we proposed a new neural network training algorithm named internal node bagging, which explicitly forces a group of nodes to learn the same feature during training phase and combines these nodes into one node during testing phase. This means that more parameters can be used during training phase to improve the fitting ability of models while keeping model remains small during testing phase. After experimenting on three datasets, it is found that this algorithm can significantly improve the test performance of small models.
引用
收藏
页码:124 / 128
页数:5
相关论文
共 50 条
  • [31] High-dimensional neural feature design for layer-wise reduction of training cost
    Alireza M. Javid
    Arun Venkitaraman
    Mikael Skoglund
    Saikat Chatterjee
    EURASIP Journal on Advances in Signal Processing, 2020
  • [32] FLEXIBLE NETWORK BINARIZATION WITH LAYER-WISE PRIORITY
    Wang, He
    Xu, Yi
    Ni, Bingbing
    Zhuang, Lixue
    Xu, Hongteng
    2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2018, : 2346 - 2350
  • [33] Layer-Wise Personalized Federated Learning with Hypernetwork
    Suxia Zhu
    Tianyu Liu
    Guanglu Sun
    Neural Processing Letters, 2023, 55 (9) : 12273 - 12287
  • [34] Post-training deep neural network pruning via layer-wise calibration
    Lazarevich, Ivan
    Kozlov, Alexander
    Malinin, Nikita
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2021), 2021, : 798 - 805
  • [35] Layer-Wise Personalized Federated Learning with Hypernetwork
    Zhu, Suxia
    Liu, Tianyu
    Sun, Guanglu
    NEURAL PROCESSING LETTERS, 2023, 55 (09) : 12273 - 12287
  • [36] The layer-wise method and the backpropagation hybrid approach to learning a feedforward neural network
    Rubanov, NS
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 2000, 11 (02): : 295 - 305
  • [37] Quantification and Analysis of Layer-wise and Pixel-wise Information Discarding
    Ma, Haotian
    Zhang, Hao
    Zhou, Fan
    Zhang, Yinqing
    Zhang, Quanshi
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [38] Layer-Wise De-Training and Re-Training for ConvS2S Machine Translation
    Yu, Hongfei
    Zhou, Xiaoqing
    Duan, Xiangyu
    Zhang, Min
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2020, 19 (02)
  • [39] First-Order Sensitivity Analysis for Hidden Neuron Selection in Layer-Wise Training of Networks
    Li, Bo
    Chen, Cheng
    NEURAL PROCESSING LETTERS, 2018, 48 (02) : 1105 - 1121
  • [40] First-Order Sensitivity Analysis for Hidden Neuron Selection in Layer-Wise Training of Networks
    Bo Li
    Cheng Chen
    Neural Processing Letters, 2018, 48 : 1105 - 1121