Accelerating neural network training using weight extrapolations

被引:91
|
作者
Kamarthi, SV [1 ]
Pittner, S [1 ]
机构
[1] Northeastern Univ, Dept Mech Ind & Mfg Engn, Snell Eng Ctr 334, Boston, MA 02115 USA
关键词
multilayer neural networks; backpropagation algorithm; convergence acceleration; extrapolation methods; parameter estimation; linear regression; conjugate gradient method; relative entropy;
D O I
10.1016/S0893-6080(99)00072-6
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The backpropagation (BP) algorithm for training feedforward neural networks has proven robust even for difficult problems. However, its high performance results are attained at the expense of a long training time to adjust the network parameters, which can be discouraging in many real-world applications. Even on relatively simple problems, standard BP often requires a lengthy training process in which the complete set of training examples is processed hundreds or thousands of times. In this paper, a universal acceleration technique for the BP algorithm based on extrapolation of each individual interconnection weight is presented. This extrapolation procedure is easy to implement and is activated only a few times in between iterations of the conventional BP algorithm. This procedure, unlike earlier acceleration procedures, minimally alters the computational structure of the BP algorithm. The viability of this new approach is demonstrated on three examples. The results suggest that it leads to significant savings in computation time of the standard BP algorithm. Moreover, the solution computed by the proposed approach is always located in close proximity to the one obtained by the conventional BP procedure. Hence, the proposed method provides a real acceleration of the BP algorithm without degrading the usefulness of its solutions. The performance of the new method is also compared with that of the conjugate gradient algorithm, which is an improved and faster version of the BP algorithm. (C) 1999 Elsevier Science Ltd. All rights reserved.
引用
收藏
页码:1285 / 1299
页数:15
相关论文
共 50 条
  • [21] Weight training for performance optimization in fuzzy neural network
    Chang, HC
    Juang, YT
    KNOWLEDGE-BASED INTELLIGENT INFORMATION AND ENGINEERING SYSTEMS, PT 1, PROCEEDINGS, 2005, 3681 : 596 - 603
  • [22] Accelerating Recurrent Neural Network Training using Sequence Bucketing and Multi-GPU Data Parallelization
    Khomenko, Viacheslav
    Shyshkov, Oleg
    Radyvonenko, Olga
    Bokhan, Kostiantyn
    PROCEEDINGS OF THE 2016 IEEE FIRST INTERNATIONAL CONFERENCE ON DATA STREAM MINING & PROCESSING (DSMP), 2016, : 100 - 103
  • [23] Accelerating CEST imaging using a model-based deep neural network with synthetic training data
    Xu, Jianping
    Zu, Tao
    Hsu, Yi-Cheng
    Wang, Xiaoli
    Chan, Kannie W. Y.
    Zhang, Yi
    MAGNETIC RESONANCE IN MEDICINE, 2024, 91 (02) : 583 - 599
  • [24] Accelerating Monte Carlo event generation rejection sampling using neural network event-weight estimates
    Danziger, Katharina
    Janssen, Timo
    Schumann, Steffen
    Siegert, Frank
    SCIPOST PHYSICS, 2022, 12 (05):
  • [25] Accelerating Deep Neural Network training for autonomous landing guidance via homotopy
    Ni, Yang
    Pan, Binfeng
    Perez, Pablo Gomez
    ACTA ASTRONAUTICA, 2023, 212 : 654 - 664
  • [26] Accelerating Large-Scale Graph Neural Network Training on Crossbar Diet
    Ogbogu, Chukwufumnanya
    Arka, Aqeeb Iqbal
    Joardar, Biresh Kumar
    Doppa, Janardhan Rao
    Li, Hai
    Chakrabarty, Krishnendu
    Pande, Partha Pratim
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2022, 41 (11) : 3626 - 3637
  • [27] ACCELERATING RECURRENT NEURAL NETWORK TRAINING VIA TWO STAGE CLASSES AND PARALLELIZATION
    Huang, Zhiheng
    Zweig, Geoffrey
    Levit, Michael
    Dumoulin, Benoit
    Oguz, Barlas
    Chang, Shawn
    2013 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING (ASRU), 2013, : 326 - 331
  • [28] BulletTrain: Accelerating Robust Neural Network Training via Boundary Example Mining
    Hua, Weizhe
    Zhang, Yichi
    Guo, Chuan
    Zhang, Zhiru
    Suh, G. Edward
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [29] Accelerating Large-Scale Distributed Neural Network Training with SPMD Parallelism
    Zhang, Shiwei
    Diao, Lansong
    Wu, Chuan
    Wang, Siyu
    Lin, Wei
    PROCEEDINGS OF THE 13TH SYMPOSIUM ON CLOUD COMPUTING, SOCC 2022, 2022, : 403 - 418
  • [30] Accelerating Forwarding Computation of Artificial Neural Network using CUDA
    Park, Jong Hyun
    Ro, Won Woo
    2016 INTERNATIONAL CONFERENCE ON ELECTRONICS, INFORMATION, AND COMMUNICATIONS (ICEIC), 2016,