Neural Networks with Marginalized Corrupted Hidden Layer

被引:1
|
作者
Li, Yanjun [1 ]
Xin, Xin [1 ]
Guo, Ping [1 ,2 ]
机构
[1] Beijing Inst Technol, Sch Comp Sci & Technol, Beijing 100081, Peoples R China
[2] Beijing Normal Univ, Image Proc & Pattern Recognit Lab, Beijing 100875, Peoples R China
来源
关键词
Neural network; Overfitting; Classification; REPRESENTATIONS;
D O I
10.1007/978-3-319-26555-1_57
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Overfitting is an important problem in neural networks (NNs) training. When the number of samples in the training set is limited, explicitly extending the training set with artificially generated samples is an effective solution. However, this method has the problem of high computational costs. In this paper we propose a new learning scheme to train single-hidden layer feedforward neural networks (SLFNs) with implicitly extended training set. The training set is extended by corrupting the hidden layer outputs of training samples with noise from exponential family distribution. When the number of corruption approaches infinity, in objective function explicitly generated samples can be expressed as the form of expectation. Our method, called marginalized corrupted hidden layer (MCHL), trains SLFNs by minimizing the loss function expected values under the corrupting distribution. In this way MCHL is trained with infinite samples. Experimental results on multiple data sets show that MCHL can be trained efficiently, and generalizes better to test data.
引用
收藏
页码:506 / 514
页数:9
相关论文
共 50 条
  • [41] Evolutionary Algorithm for Training Compact Single Hidden Layer Feedforward Neural Networks
    Huynh, Hieu Trung
    Won, Yonggwan
    2008 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-8, 2008, : 3028 - 3033
  • [42] Robust and Resource-Efficient Identification of Two Hidden Layer Neural Networks
    Fornasier, Massimo
    Klock, Timo
    Rauchensteiner, Michael
    CONSTRUCTIVE APPROXIMATION, 2022, 55 (01) : 475 - 536
  • [43] Construction and initialization of a hidden layer of multilayer neural networks using linear programming
    Kim, LS
    CRITICAL TECHNOLOGY: PROCEEDINGS OF THE THIRD WORLD CONGRESS ON EXPERT SYSTEMS, VOLS I AND II, 1996, : 986 - 992
  • [44] Approximation capability of two hidden layer feedforward neural networks with fixed weights
    Guliyev, Namig J.
    Ismailov, Vugar E.
    NEUROCOMPUTING, 2018, 316 : 262 - 269
  • [45] Design of multi-hidden-layer Bayesian neural networks for model updating
    He, Yuxuan
    Yin, Tao
    Wang, Xi
    ADVANCES IN STRUCTURAL ENGINEERING, 2025, 28 (05) : 952 - 972
  • [46] Simultaneous approximations of multivariate functions and their derivatives by neural networks with one hidden layer
    Li, X
    NEUROCOMPUTING, 1996, 12 (04) : 327 - 343
  • [47] Robust adaptive nonlinear control using single hidden layer neural networks
    Nardi, F
    Calise, AJ
    PROCEEDINGS OF THE 39TH IEEE CONFERENCE ON DECISION AND CONTROL, VOLS 1-5, 2000, : 3825 - 3830
  • [48] Improving Rule Extraction from Neural Networks by Modifying Hidden Layer Representations
    Huynh, Thuan Q.
    Reggia, James A.
    IJCNN: 2009 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1- 6, 2009, : 734 - 739
  • [49] INTEGRATING GAUSSIAN MIXTURES INTO DEEP NEURAL NETWORKS: SOFTMAX LAYER WITH HIDDEN VARIABLES
    Tueske, Zoltan
    Tahir, Muhammad Ali
    Schlueter, Ralf
    Ney, Hermann
    2015 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING (ICASSP), 2015, : 4285 - 4289
  • [50] Training Single Hidden Layer Feedforward Neural Networks by Singular Value Decomposition
    Hieu Trung Huynh
    Won, Yonggwan
    ICCIT: 2009 FOURTH INTERNATIONAL CONFERENCE ON COMPUTER SCIENCES AND CONVERGENCE INFORMATION TECHNOLOGY, VOLS 1 AND 2, 2009, : 1300 - 1304