Deterministic convergence of conjugate gradient method for feedforward neural networks

被引:33
|
作者
Wang, Jian [1 ,2 ,3 ]
Wu, Wei [2 ]
Zurada, Jacek M. [1 ]
机构
[1] Univ Louisville, Dept Elect & Comp Engn, Louisville, KY 40292 USA
[2] Dalian Univ Technol, Sch Math Sci, Dalian 116024, Peoples R China
[3] China Univ Petr, Sch Math & Computat Sci, Dongying 257061, Peoples R China
基金
中国国家自然科学基金;
关键词
Deterministic convergence; Conjugate gradient; Backpropagation; Feedforward neural networks; EXTREME LEARNING-MACHINE; ONLINE; ALGORITHM;
D O I
10.1016/j.neucom.2011.03.016
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Conjugate gradient methods have many advantages in real numerical experiments, such as fast convergence and low memory requirements. This paper considers a class of conjugate gradient learning methods for backpropagation neural networks with three layers. We propose a new learning algorithm for almost cyclic learning of neural networks based on PRP conjugate gradient method. We then establish the deterministic convergence properties for three different learning modes, i.e., batch mode, cyclic and almost cyclic learning. The two deterministic convergence properties are weak and strong convergence that indicate that the gradient of the error function goes to zero and the weight sequence goes to a fixed point, respectively. It is shown that the deterministic convergence results are based on different learning modes and dependent on different selection strategies of learning rate. Illustrative numerical examples are given to support the theoretical analysis. (C) 2011 Elsevier B.V. All rights reserved.
引用
收藏
页码:2368 / 2376
页数:9
相关论文
共 50 条
  • [31] Deterministic Convergence of an Online Gradient Method with Momentum
    Zhang, Naimin
    INTELLIGENT COMPUTING, PART I: INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTING, ICIC 2006, PART I, 2006, 4113 : 94 - 105
  • [32] An online gradient method with momentum for two-layer feedforward neural networks
    Zhang, Naimin
    APPLIED MATHEMATICS AND COMPUTATION, 2009, 212 (02) : 488 - 498
  • [33] Scaled Conjugate Gradient Method for Nonconvex Optimization in Deep Neural Networks
    Sato, Naoki
    Izumi, Koshiro
    Iiduka, Hideaki
    JOURNAL OF MACHINE LEARNING RESEARCH, 2024, 25 : 1 - 37
  • [34] CONVERGENCE OF THE CONJUGATE GRADIENT METHOD WITH UNBOUNDED OPERATORS
    Caruso, Noe
    Michelangeli, Alessandro
    OPERATORS AND MATRICES, 2022, 16 (01): : 35 - 68
  • [35] The Global Convergence Properties of a Conjugate Gradient Method
    Omer, Osman
    Mamat, Mustafa
    Abashar, Abdelrhaman
    Rivaie, Mohd
    PROCEEDINGS OF THE 3RD INTERNATIONAL CONFERENCE ON MATHEMATICAL SCIENCES, 2014, 1602 : 286 - 295
  • [36] GLOBAL CONVERGENCE OF A MODIFIED CONJUGATE GRADIENT METHOD
    Li, Can
    Fang, Ling
    Lu, Peng
    2012 INTERNATIONAL CONFERENCE ON WAVELET ACTIVE MEDIA TECHNOLOGY AND INFORMATION PROCESSING (LCWAMTIP), 2012, : 78 - 81
  • [37] ON CONVERGENCE OF CONJUGATE GRADIENT METHOD IN HILBERT SPACE
    KAWAMURA, K
    VOLZ, RA
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 1969, AC14 (03) : 296 - +
  • [38] SOME CONVERGENCE PROPERTIES OF CONJUGATE GRADIENT METHOD
    POWELL, MJD
    MATHEMATICAL PROGRAMMING, 1976, 11 (01) : 42 - 49
  • [39] New results on the convergence of the conjugate gradient method
    Bouyouli, R.
    Meurant, G.
    Smoch, L.
    Sadok, H.
    NUMERICAL LINEAR ALGEBRA WITH APPLICATIONS, 2009, 16 (03) : 223 - 236
  • [40] Global Convergence of a Nonlinear Conjugate Gradient Method
    Liu Jin-kui
    Zou Li-min
    Song Xiao-qian
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2011, 2011