Accelerating neural network training using weight extrapolations

被引:91
|
作者
Kamarthi, SV [1 ]
Pittner, S [1 ]
机构
[1] Northeastern Univ, Dept Mech Ind & Mfg Engn, Snell Eng Ctr 334, Boston, MA 02115 USA
关键词
multilayer neural networks; backpropagation algorithm; convergence acceleration; extrapolation methods; parameter estimation; linear regression; conjugate gradient method; relative entropy;
D O I
10.1016/S0893-6080(99)00072-6
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The backpropagation (BP) algorithm for training feedforward neural networks has proven robust even for difficult problems. However, its high performance results are attained at the expense of a long training time to adjust the network parameters, which can be discouraging in many real-world applications. Even on relatively simple problems, standard BP often requires a lengthy training process in which the complete set of training examples is processed hundreds or thousands of times. In this paper, a universal acceleration technique for the BP algorithm based on extrapolation of each individual interconnection weight is presented. This extrapolation procedure is easy to implement and is activated only a few times in between iterations of the conventional BP algorithm. This procedure, unlike earlier acceleration procedures, minimally alters the computational structure of the BP algorithm. The viability of this new approach is demonstrated on three examples. The results suggest that it leads to significant savings in computation time of the standard BP algorithm. Moreover, the solution computed by the proposed approach is always located in close proximity to the one obtained by the conventional BP procedure. Hence, the proposed method provides a real acceleration of the BP algorithm without degrading the usefulness of its solutions. The performance of the new method is also compared with that of the conjugate gradient algorithm, which is an improved and faster version of the BP algorithm. (C) 1999 Elsevier Science Ltd. All rights reserved.
引用
收藏
页码:1285 / 1299
页数:15
相关论文
共 50 条
  • [31] Evolutionary Based Weight Decaying Method for Neural Network Training
    Tsoulos, Ioannis G.
    Tzallas, Alexandros
    Tsalikakis, Dimitris
    NEURAL PROCESSING LETTERS, 2018, 47 (02) : 463 - 473
  • [32] An improved weight-constrained neural network training algorithm
    Livieris, Ioannis E.
    Pintelas, Panagiotis
    NEURAL COMPUTING & APPLICATIONS, 2020, 32 (09): : 4177 - 4185
  • [33] Weight Regularisation in Particle Swarm Optimisation Neural Network Training
    Rakitianskaia, Anna
    Engelbrecht, Andries
    2014 IEEE SYMPOSIUM ON SWARM INTELLIGENCE (SIS), 2014, : 34 - 41
  • [34] Performence analysis of neural network with improved weight training process
    Zhao, Yifeng
    Lang, Weimin
    Li, Bin
    PROCEEDINGS OF 2019 IEEE 3RD INFORMATION TECHNOLOGY, NETWORKING, ELECTRONIC AND AUTOMATION CONTROL CONFERENCE (ITNEC 2019), 2019, : 1760 - 1764
  • [35] An improved weight-constrained neural network training algorithm
    Ioannis E. Livieris
    Panagiotis Pintelas
    Neural Computing and Applications, 2020, 32 : 4177 - 4185
  • [36] Evolutionary Based Weight Decaying Method for Neural Network Training
    Ioannis G. Tsoulos
    Alexandros Tzallas
    Dimitris Tsalikakis
    Neural Processing Letters, 2018, 47 : 463 - 473
  • [37] Integer weight higher-order neural network training using distributed differential evolution
    Epitropakis, M. G.
    Plagianakos, V. P.
    Vrahatis, M. N.
    RECENT PROGRESS IN COMPUTATIONAL SCIENCES AND ENGINEERING, VOLS 7A AND 7B, 2006, 7A-B : 141 - 144
  • [38] Accelerating the Construction of Neural Network Potential Energy Surfaces: A Fast Hybrid Training Algorithm
    Zhang, Yao-long
    Zhou, Xue-yao
    Jiang, Bin
    CHINESE JOURNAL OF CHEMICAL PHYSICS, 2017, 30 (06) : 727 - 734
  • [39] Accelerating Data-Parallel Neural Network Training with Weighted-Averaging Reparameterisation
    Ramroach, Sterling
    Joshi, Ajay
    PARALLEL PROCESSING LETTERS, 2021, 31 (02)
  • [40] Neural network training using stochastic PSO
    Chen, Xin
    Li, Yangmin
    NEURAL INFORMATION PROCESSING, PT 2, PROCEEDINGS, 2006, 4233 : 1051 - 1060