ERROR SURFACES FOR MULTILAYER PERCEPTRONS

被引:34
|
作者
HUSH, DR
HORNE, B
SALAS, JM
机构
[1] Department of Electrical Engineering and Cornputer Engineering, University of New Mexico, Albuquerque, NM
来源
关键词
D O I
10.1109/21.179853
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The paper explores the characteristics of error surfaces for the multilayer perceptron neural network. These characteristics help explain why learning techniques that use hill climbing methods are so slow in these networks. They also help provide insights into techniques that may help speed learning. Several important characteristics are revealed. First, the surface has a stair-step appearance with many very flat and very steep regions. In fact, when the number of training samples is small there is often a one-to-one correspondence between individual training samples and the steps on the surface. As the number of training samples is increased the surface becomes smoother. In addition the surface has flat regions that extend to infinity in all directions making it dangerous to apply learning algorithms that perform line searches. The magnitude of gradients on the surface is found to span several orders of magnitude, strongly supporting the need for floating point representations during learning. The consequences of various weight initialization techniques are also discussed.
引用
收藏
页码:1152 / 1161
页数:10
相关论文
共 50 条
  • [1] Data classification with multilayer perceptrons using a generalized error function
    Silva, Luis M.
    de Sa, J. Marques
    Alexandre, Luis A.
    NEURAL NETWORKS, 2008, 21 (09) : 1302 - 1310
  • [2] Averaged learning equations of error-function-based multilayer perceptrons
    Guo, Weili
    Wei, Haikun
    Zhao, Junsheng
    Zhang, Kanjian
    NEURAL COMPUTING & APPLICATIONS, 2014, 25 (3-4): : 825 - 832
  • [3] A new error function at hidden layers for fast training of multilayer perceptrons
    Oh, SH
    Lee, SY
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 1999, 10 (04): : 960 - 964
  • [4] Averaged learning equations of error-function-based multilayer perceptrons
    Weili Guo
    Haikun Wei
    Junsheng Zhao
    Kanjian Zhang
    Neural Computing and Applications, 2014, 25 : 825 - 832
  • [5] An equalized error backpropagation algorithm for the on-line training of multilayer perceptrons
    Martens, JP
    Weymaere, N
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 2002, 13 (03): : 532 - 541
  • [6] An adaptive learning rate with limited error signals for training of multilayer perceptrons
    Oh, SH
    Lee, SY
    ETRI JOURNAL, 2000, 22 (03) : 10 - 18
  • [7] PREPROGRAMMING MULTILAYER PERCEPTRONS
    SMYTH, SG
    BT TECHNOLOGY JOURNAL, 1992, 10 (03): : 30 - 37
  • [8] ROBUSTNESS IN MULTILAYER PERCEPTRONS
    KERLIRZIN, P
    VALLET, F
    NEURAL COMPUTATION, 1993, 5 (03) : 473 - 482
  • [9] Evolving multilayer perceptrons
    Castillo, PA
    Carpio, J
    Merelo, JJ
    Prieto, A
    Rivas, V
    Romero, G
    NEURAL PROCESSING LETTERS, 2000, 12 (02) : 115 - 127
  • [10] Multilayer perceptrons and fractals
    Murthy, CA
    Pittman, J
    INFORMATION SCIENCES, 1998, 112 (1-4) : 137 - 150