Internal Node Bagging: a Layer-Wise Ensemble Training Method

被引:0
|
作者
Li, Jinhong [1 ]
Yi, Shun [1 ]
机构
[1] North China Univ Technol, Sch Informat, Beijing, Peoples R China
关键词
neural networks; deep learning; ensemble learning; regularization;
D O I
10.1145/3357254.3357268
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
When training neural networks, regularization methods are needed to avoid model overfitting. Dropout is a widely used regularization method, but its working principle is inconclusive and it does not work well for small models. This paper introduced a novel view to understand how dropout works as a layer-wise ensemble training method, that each feature in hidden layers is learned by multiple nodes, and next layer integrates the outputs of these nodes. Basing on the novel understanding of dropout, we proposed a new neural network training algorithm named internal node bagging, which explicitly forces a group of nodes to learn the same feature during training phase and combines these nodes into one node during testing phase. This means that more parameters can be used during training phase to improve the fitting ability of models while keeping model remains small during testing phase. After experimenting on three datasets, it is found that this algorithm can significantly improve the test performance of small models.
引用
收藏
页码:124 / 128
页数:5
相关论文
共 50 条
  • [1] A Layer-Wise Ensemble Technique for Binary Neural Network
    Xi, Jiazhen
    Yamauchi, Hiroyuki
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2021, 35 (08)
  • [2] SPSA for Layer-Wise Training of Deep Networks
    Wulff, Benjamin
    Schuecker, Jannis
    Bauckhage, Christian
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2018, PT III, 2018, 11141 : 564 - 573
  • [3] Layer-Wise Compressive Training for Convolutional Neural Networks
    Grimaldi, Matteo
    Tenace, Valerio
    Calimera, Andrea
    FUTURE INTERNET, 2019, 11 (01)
  • [4] ResIST: Layer-Wise Decomposition of ResNets for Distributed Training
    Dun, Chen
    Wolfe, Cameron R.
    Jermaine, Christopher M.
    Kyrillidis, Anastasios
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, VOL 180, 2022, 180 : 610 - 620
  • [5] A Layer-wise Training and Pruning Method for Memory Efficient On-chip Learning Hardware
    Lew, Dongwoo
    Park, Jongsun
    2022 19TH INTERNATIONAL SOC DESIGN CONFERENCE (ISOCC), 2022, : 97 - 98
  • [6] Filtering-based Layer-wise Parameter Update Method for Training a Neural Network
    Ji, Siyu
    Zhai, Kaikai
    Wen, Chenglin
    2018 INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND INFORMATION SCIENCES (ICCAIS), 2018, : 389 - 394
  • [7] A Layer-wise Score Level Ensemble Framework for Acoustic Scene Classification
    Singh, Arshdeep
    Thakur, Anshul
    Rajan, Padmanabhan
    Bhavsar, Arnav
    2018 26TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2018, : 837 - 841
  • [8] DRN: Bringing Greedy Layer-Wise Training into Time Dimension
    Li, Xiaoyi
    Jia, Xiaowei
    Li, Hui
    Xiao, Houping
    Gao, Jing
    Zhang, Aidong
    2015 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2015, : 859 - 864
  • [9] Layer-wise Adversarial Training Approach to Improve Adversarial Robustness
    Chen, Xiaoyi
    Zhang, Ni
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [10] Layer-Wise Training to Create Efficient Convolutional Neural Networks
    Zeng, Linghua
    Tian, Xinmei
    NEURAL INFORMATION PROCESSING (ICONIP 2017), PT II, 2017, 10635 : 631 - 641