Adaptive CL-BFGS Algorithms for Complex-Valued Neural Networks

被引:4
|
作者
Zhang, Yongliang [1 ,2 ]
Huang, He [1 ,2 ]
Shen, Gangxiang [1 ,2 ]
机构
[1] Soochow Univ, Sch Elect & Informat Engn, Suzhou 215006, Peoples R China
[2] Jiangsu Engn Res Ctr Novel Opt Fiber Technol & Co, Suzhou 215006, Peoples R China
关键词
Approximation algorithms; Training; Signal processing algorithms; Optimization; Neural networks; Upper bound; Mathematical models; Adaptive complex-valued limited-memory BFGS (ACL-BFGS) algorithm; complex-valued neural networks (CVNNs); moving average; multistep quasi-Newton method; variable memory size; CLASSIFICATION; CONVERGENCE;
D O I
10.1109/TNNLS.2021.3135553
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Complex-valued limited-memory BFGS (CL-BFGS) algorithm is efficient for the training of complex-valued neural networks (CVNNs). As an important parameter, the memory size represents the number of saved vector pairs and would essentially affect the performance of the algorithm. However, the determination of a suitable memory size for the CL-BFGS algorithm remains challenging. To deal with this issue, an adaptive method is proposed in which the memory size is allowed to vary during the iteration process. Basically, at each iteration, with the help of multistep quasi-Newton method, an appropriate memory size is chosen from a variable set {1,2,..., M} by approximating complex Hessian matrix as close as possible. To reduce the computational complexity and ensure desired performance, the upper bound M is adjustable according to the moving average of memory sizes found in previous iterations. The proposed adaptive CL-BFGS (ACL-BFGS) algorithm can be efficiently applied for the training of CVNNs. Moreover, it is suggested to take multiple memory sizes to construct the search direction, which further improves the performance of the ACL-BFGS algorithm. Experimental results on some benchmark problems including the pattern classification, complex function approximation, and nonlinear channel equalization problems are given to illustrate the advantages of the developed algorithms over some previous ones.
引用
收藏
页码:6313 / 6327
页数:15
相关论文
共 50 条
  • [1] Stochastic adaptive CL-BFGS algorithms for fully complex-valued dendritic neuron model
    Wang, Yuelin
    Wang, Zhidong
    Huang, He
    KNOWLEDGE-BASED SYSTEMS, 2023, 277
  • [2] Adaptive complex-valued stepsize based fast learning of complex-valued neural networks
    Zhang, Yongliang
    Huang, He
    NEURAL NETWORKS, 2020, 124 : 233 - 242
  • [3] Conjugate Gradient Algorithms for Complex-Valued Neural Networks
    Popa, Calin-Adrian
    NEURAL INFORMATION PROCESSING, PT II, 2015, 9490 : 412 - 422
  • [4] Complex-valued neural networks
    Department of Electrical Engineering and Information Systems, University of Tokyo, 7-3-1, Hongo, Bunkyo-ku, Tokyo 113-8656, Japan
    IEEJ Trans. Electron. Inf. Syst., 1 (2-8):
  • [5] Enhanced Gradient Descent Algorithms for Complex-Valued Neural Networks
    Popa, Calin-Adrian
    16TH INTERNATIONAL SYMPOSIUM ON SYMBOLIC AND NUMERIC ALGORITHMS FOR SCIENTIFIC COMPUTING (SYNASC 2014), 2014, : 272 - 279
  • [6] A L-BFGS Based Learning Algorithm for Complex-Valued Feedforward Neural Networks
    Rongrong Wu
    He Huang
    Xusheng Qian
    Tingwen Huang
    Neural Processing Letters, 2018, 47 : 1271 - 1284
  • [7] A L-BFGS Based Learning Algorithm for Complex-Valued Feedforward Neural Networks
    Wu, Rongrong
    Huang, He
    Qian, Xusheng
    Huang, Tingwen
    NEURAL PROCESSING LETTERS, 2018, 47 (03) : 1271 - 1284
  • [8] Adaptive Synchronization of Complex-Valued Neural Networks with Time Delay
    Bao, Haibo
    Park, Ju H.
    2016 EIGHTH INTERNATIONAL CONFERENCE ON ADVANCED COMPUTATIONAL INTELLIGENCE (ICACI), 2016, : 283 - 288
  • [9] Complex-Valued Logic for Neural Networks
    Kagan, Evgeny
    Rybalov, Alexander
    Yager, Ronald
    2018 IEEE INTERNATIONAL CONFERENCE ON THE SCIENCE OF ELECTRICAL ENGINEERING IN ISRAEL (ICSEE), 2018,
  • [10] A class of low complexity and fast converging algorithms for complex-valued neural networks
    Goh, SL
    Mandic, DP
    MACHINE LEARNING FOR SIGNAL PROCESSING XIV, 2004, : 13 - 22