Online adaptive optimal control algorithm based on synchronous integral reinforcement learning with explorations

被引:7
|
作者
Guo, Lei [1 ]
Zhao, Han [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Sch Artificial Intelligence, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Reinforcement learning; Neural networks; Adaptive control; Actor; -critic; Explorations; TIME LINEAR-SYSTEMS; NONLINEAR-SYSTEMS;
D O I
10.1016/j.neucom.2022.11.055
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this study, we present a novel algorithm, based on synchronous policy iteration, to solve the continuous-time infinite-horizon optimal control problem of input affine system dynamics. The integral reinforcement is measured as an excitation signal to estimate the solution to the Hamilton-Jacobi-Bell man equation. In addition, the proposed method is completely model-free, that is, no a priori knowledge of the system is required. Using the adaptive tuning law, the actor and critic neural networks can simultaneously approximate the optimal value function and policy. The persistence of excitation condition is required to guarantee the convergence of the two networks. Unlike in traditional policy iteration algorithms, the restriction of the initial admissible policy was eliminated using this method. The effectiveness of the proposed algorithm is verified through numerical simulations. (c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:250 / 261
页数:12
相关论文
共 50 条
  • [31] PNLMS-based Algorithm for Online Approximated Solution of HJB Equation in the Context of Discrete MIMO Optimal Control and Reinforcement Learning
    Silva, Marcio Eduardo G.
    da Fonseca Neto, Joao Viana
    Chagas de Souza, Francisco das
    2014 UKSIM-AMSS 16TH INTERNATIONAL CONFERENCE ON COMPUTER MODELLING AND SIMULATION (UKSIM), 2014, : 69 - 76
  • [32] A reinforcement learning-based scheme for adaptive optimal control of linear stochastic systems
    Wong, Wee Chin
    Lee, Jay H.
    2008 AMERICAN CONTROL CONFERENCE, VOLS 1-12, 2008, : 57 - 62
  • [33] Reinforcement learning based computational adaptive optimal control and system identification for linear systems
    Subbarao, Kamesh
    Nuthi, Pavan
    Atmeh, Ghassan
    ANNUAL REVIEWS IN CONTROL, 2016, 42 : 319 - 331
  • [34] Adaptive optimal control for a class of uncertain systems with saturating actuators and external disturbance using integral reinforcement learning
    Zhao, Jingang
    Gan, Minggang
    Chen, Jie
    Hou, Dongyang
    Zhang, Meng
    Bai, Yongqiang
    2017 11TH ASIAN CONTROL CONFERENCE (ASCC), 2017, : 1146 - 1151
  • [35] Reinforcement Learning-Based Adaptive Optimal Control for Nonlinear Systems With Asymmetric Hysteresis
    Zheng, Licheng
    Liu, Zhi
    Wang, Yaonan
    Chen, C. L. Philip
    Zhang, Yun
    Wu, Zongze
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (11) : 15800 - 15809
  • [36] Combined control algorithm based on synchronous reinforcement learning for a self-balancing bicycle robot
    Guo, Lei
    Lin, Hongyu
    Jiang, Jiale
    Song, Yuan
    Gan, Dongming
    ISA TRANSACTIONS, 2024, 145 : 479 - 492
  • [37] Stable Reinforcement Learning for Optimal Frequency Control: A Distributed Averaging-Based Integral Approach
    Jiang, Yan
    Cui, Wenqi
    Zhang, Baosen
    Cortes, Jorge
    IEEE OPEN JOURNAL OF CONTROL SYSTEMS, 2022, 1 : 194 - 209
  • [38] Distributed optimal control of nonlinear multi-agent systems based on integral reinforcement learning
    Xu, Ying
    Li, Kewen
    Li, Yongming
    OPTIMAL CONTROL APPLICATIONS & METHODS, 2024, 45 (06): : 2596 - 2612
  • [39] Robust optimal control of the multi-input systems with unknown disturbance based on adaptive integral reinforcement learning Q-function
    Lv, Yongfeng
    Zhao, Jun
    Li, Rong
    Ren, Xuemei
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2024, 34 (06) : 4234 - 4251
  • [40] The Optimal Path Finding Algorithm Based on Reinforcement Learning
    Khekare, Ganesh
    Verma, Pushpneel
    Dhanre, Urvashi
    Raut, Seema
    Sheikh, Shahrukh
    INTERNATIONAL JOURNAL OF SOFTWARE SCIENCE AND COMPUTATIONAL INTELLIGENCE-IJSSCI, 2020, 12 (04): : 1 - 18