Efficient Online Learning Algorithms Based on LSTM Neural Networks

被引:91
|
作者
Ergen, Tolga [1 ]
Kozat, Suleyman Serdar [1 ]
机构
[1] Bilkent Univ, Dept Elect & Elect Engn, TR-06800 Ankara, Turkey
关键词
Gated recurrent unit (GRU); Kalman filtering; long short term memory (LSTM); online learning; particle filtering (PF); regression; stochastic gradient descent (SGD); REGRESSION; FILTERS;
D O I
10.1109/TNNLS.2017.2741598
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We investigate online nonlinear regression and introduce novel regression structures based on the long short term memory (LSTM) networks. For the introduced structures, we also provide highly efficient and effective online training methods. To train these novel LSTM-based structures, we put the underlying architecture in a state space form and introduce highly efficient and effective particle filtering (PF)-based updates. We also provide stochastic gradient descent and extended Kalman filter-based updates. Our PF-based training method guarantees convergence to the optimal parameter estimation in the mean square error sense provided that we have a sufficient number of particles and satisfy certain technical conditions. More importantly, we achieve this performance with a computational complexity in the order of the first-order gradient-based methods by controlling the number of particles. Since our approach is generic, we also introduce a gated recurrent unit (GRU)-based approach by directly replacing the LSTM architecture with the GRU architecture, where we demonstrate the superiority of our LSTM-based approach in the sequential prediction task via different real life data sets. In addition, the experimental results illustrate significant performance improvements achieved by the introduced algorithms with respect to the conventional methods over several different benchmark real life data sets.
引用
收藏
页码:3772 / 3783
页数:12
相关论文
共 50 条
  • [1] Efficient online learning with improved LSTM neural networks
    Mirza, Ali H.
    Kerpicci, Mine
    Kozat, Suleyman S.
    DIGITAL SIGNAL PROCESSING, 2020, 102
  • [2] Energy-Efficient LSTM Networks for Online Learning
    Ergen, Tolga
    Mirza, Ali H.
    Kozat, Suleyman Serdar
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (08) : 3114 - 3126
  • [3] Recurrent Neural Networks Based Online Learning Algorithms for Distributed Systems
    Ergen, Tolga
    Sahin, S. Onur
    Kozat, S. Serdar
    2018 26TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2018,
  • [4] EFFICIENT LEARNING ALGORITHMS FOR NEURAL NETWORKS (ELEANNE)
    KARAYIANNIS, NB
    VENETSANOPOULOS, AN
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS, 1993, 23 (05): : 1372 - 1383
  • [5] ON A CLASS OF EFFICIENT LEARNING ALGORITHMS FOR NEURAL NETWORKS
    BARMANN, F
    BIEGLERKONIG, F
    NEURAL NETWORKS, 1992, 5 (01) : 139 - 144
  • [6] Efficient learning of neural networks with evolutionary algorithms
    Siebel, Nils T.
    Krause, Jochen
    Sommer, Gerald
    PATTERN RECOGNITION, PROCEEDINGS, 2007, 4713 : 466 - +
  • [7] LSTM Neural Networks for Transfer Learning in Online Moderation of Abuse Context
    Bleiweiss, Avi
    PROCEEDINGS OF THE 11TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE (ICAART), VOL 2, 2019, : 112 - 122
  • [8] Neural Networks Based Online Learning
    Ergen, Tolga
    Kozat, Suleyman S.
    2017 25TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2017,
  • [9] Efficient Online Learning with Spiral Recurrent Neural Networks
    Sollacher, Rudolf
    Gao, Huaien
    2008 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-8, 2008, : 2551 - 2558
  • [10] A unified framework of online learning algorithms for training recurrent neural networks
    Marschall, Owen
    Cho, Kyunghyun
    Savin, Cristina
    Journal of Machine Learning Research, 2020, 21