Internal Node Bagging: a Layer-Wise Ensemble Training Method

被引:0
|
作者
Li, Jinhong [1 ]
Yi, Shun [1 ]
机构
[1] North China Univ Technol, Sch Informat, Beijing, Peoples R China
关键词
neural networks; deep learning; ensemble learning; regularization;
D O I
10.1145/3357254.3357268
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
When training neural networks, regularization methods are needed to avoid model overfitting. Dropout is a widely used regularization method, but its working principle is inconclusive and it does not work well for small models. This paper introduced a novel view to understand how dropout works as a layer-wise ensemble training method, that each feature in hidden layers is learned by multiple nodes, and next layer integrates the outputs of these nodes. Basing on the novel understanding of dropout, we proposed a new neural network training algorithm named internal node bagging, which explicitly forces a group of nodes to learn the same feature during training phase and combines these nodes into one node during testing phase. This means that more parameters can be used during training phase to improve the fitting ability of models while keeping model remains small during testing phase. After experimenting on three datasets, it is found that this algorithm can significantly improve the test performance of small models.
引用
收藏
页码:124 / 128
页数:5
相关论文
共 50 条
  • [21] Supervised Greedy Layer-Wise Training for Deep Convolutional Networks with Small Datasets
    Rueda-Plata, Diego
    Ramos-Pollan, Raul
    Gonzalez, Fabio A.
    COMPUTATIONAL COLLECTIVE INTELLIGENCE (ICCCI 2015), PT I, 2015, 9329 : 275 - 284
  • [22] Automated layer-wise solution for ensemble deep randomized feed-forward neural network
    Hu, Minghui
    Gao, Ruobin
    Suganthan, Ponnuthurai N.
    Tanveer, M.
    NEUROCOMPUTING, 2022, 514 : 137 - 147
  • [23] Layer-wise Pre-training Mechanism Based on Neural Network for Epilepsy Detection
    Lin, Zichao
    Gu, Zhenghui
    Li, Yinghao
    Yu, Zhuliang
    Li, Yuanqing
    2020 12TH INTERNATIONAL CONFERENCE ON ADVANCED COMPUTATIONAL INTELLIGENCE (ICACI), 2020, : 224 - 227
  • [24] Transkimmer: Transformer Learns to Layer-wise Skim
    Guan, Yue
    Li, Zhengyi
    Leng, Jingwen
    Lin, Zhouhan
    Guo, Minyi
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 7275 - 7286
  • [25] Layer-Wise Representation Fusion for Compositional Generalization
    Zheng, Yafang
    Lin, Lei
    Li, Shuangtao
    Yuan, Yuxuan
    Lai, Zhaohong
    Liu, Shan
    Fu, Biao
    Chen, Yidong
    Shi, Xiaodong
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 19706 - 19714
  • [26] FedLF: Layer-Wise Fair Federated Learning
    Pan, Zibin
    Li, Chi
    Yu, Fangchen
    Wang, Shuyi
    Wang, Haijin
    Tang, Xiaoying
    Zhao, Junhua
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 13, 2024, : 14527 - 14535
  • [27] Potential Layer-Wise Supervised Learning for Training Multi-Layered Neural Networks
    Kamimura, Ryotaro
    2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 2568 - 2575
  • [28] High-dimensional neural feature design for layer-wise reduction of training cost
    Javid, Alireza M.
    Venkitaraman, Arun
    Skoglund, Mikael
    Chatterjee, Saikat
    EURASIP JOURNAL ON ADVANCES IN SIGNAL PROCESSING, 2020, 2020 (01)
  • [29] Push-pull separability objective for supervised layer-wise training of neural networks
    Szymanski, Lech
    McCane, Brendan
    2012 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2012,
  • [30] LOT: Layer-wise Orthogonal Training on Improving l2 Certified Robustness
    Xu, Xiaojun
    Li, Linyi
    Li, Bo
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,