Prescribed attractivity region selection for recurrent neural networks based on deep reinforcement learning

被引:1
|
作者
Bao, Gang [1 ]
Song, Zhenyan [1 ]
Xu, Rui [1 ]
机构
[1] China Three Gorges Univ, Hubei Key Lab Cascaded Hydropower Stat Operat & C, Yichang 443002, Peoples R China
来源
NEURAL COMPUTING & APPLICATIONS | 2024年 / 36卷 / 05期
基金
中国国家自然科学基金;
关键词
Recurrent neural networks; Attractivity region selection; Deep reinforcement learning; GLOBAL EXPONENTIAL STABILITY; TIME-VARYING DELAYS; DESIGN;
D O I
10.1007/s00521-023-09191-8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recurrent neural networks' (RNNs') outputs are the same when network states converge to the same saturation region. Strong external inputs can cause the neural network to converge to a prescribed saturation region. Different from previous works, this paper employs deep reinforcement learning to obtain external inputs to make network states converge to the desired saturation region. Firstly, for five-dimensional neural networks, the deep Q learning (DQN) algorithm is used to compute the optimal external inputs that make the network state converge to the specified saturation region. When scaling to n-dimensional RNNs, the problem of dimensional disaster is encountered. Then, it proposes a batch computation of the external inputs to cope with the curse of dimensionality. At last, the proposed method is validated by numerical examples, and compared with existing methods, it shows that less conservative external inputs conditions can be obtained.
引用
收藏
页码:2399 / 2409
页数:11
相关论文
共 50 条
  • [31] Cooperative Communications With Relay Selection Based on Deep Reinforcement Learning in Wireless Sensor Networks
    Su, Yuhan
    Lu, Xiaozhen
    Zhao, Yifeng
    Huang, Lianfen
    Du, Xiaojiang
    IEEE SENSORS JOURNAL, 2019, 19 (20) : 9561 - 9569
  • [32] Deep Reinforcement Learning of Region Proposal Networks for Object Detection
    Pirinen, Aleksis
    Sminchisescu, Cristian
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 6945 - 6954
  • [33] Reinforcement Learning and Deep Neural Networks for PI Controller Tuning
    Shipman, William J.
    Coetzee, Loutjie C.
    IFAC PAPERSONLINE, 2019, 52 (14): : 111 - 116
  • [34] Deep Auto-Encoder Neural Networks in Reinforcement Learning
    Lange, Sascha
    Riedmiller, Martin
    2010 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS IJCNN 2010, 2010,
  • [35] Learning deep hierarchical and temporal recurrent neural networks with residual learning
    Zia, Tehseen
    Abbas, Assad
    Habib, Usman
    Khan, Muhammad Sajid
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2020, 11 (04) : 873 - 882
  • [36] Learning deep hierarchical and temporal recurrent neural networks with residual learning
    Tehseen Zia
    Assad Abbas
    Usman Habib
    Muhammad Sajid Khan
    International Journal of Machine Learning and Cybernetics, 2020, 11 : 873 - 882
  • [37] Assuring Efficient Path Selection in an Intent-Based Networking System: A Graph Neural Networks and Deep Reinforcement Learning Approach
    Alam, Sajid
    Rivera, Javier Jose Diaz
    Sarwar, Mir Muhammad Suleman
    Muhammad, Afaq
    Song, Wang-Cheol
    JOURNAL OF NETWORK AND SYSTEMS MANAGEMENT, 2024, 32 (02)
  • [38] Integrating recurrent neural networks and reinforcement learning for dynamic service composition
    Wang, Hongbing
    Li, Jiajie
    Yu, Qi
    Hong, Tianjing
    Yan, Jia
    Zhao, Wei
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2020, 107 (107): : 551 - 563
  • [39] Reinforcement Learning for Central Pattern Generation in Dynamical Recurrent Neural Networks
    Yoder, Jason A.
    Anderson, Cooper B.
    Wang, Cehong
    Izquierdo, Eduardo J.
    FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2022, 16
  • [40] An Unsupervised Learning Algorithm for Deep Recurrent Spiking Neural Networks
    Du, Pangao
    Lin, Xianghong
    Pi, Xiaomei
    Wang, Xiangwen
    2020 11TH IEEE ANNUAL UBIQUITOUS COMPUTING, ELECTRONICS & MOBILE COMMUNICATION CONFERENCE (UEMCON), 2020, : 603 - 607