Model-Free Optimal Tracking Design With Evolving Control Strategies via Q-Learning

被引:2
|
作者
Wang, Ding [1 ,2 ]
Huang, Haiming [1 ,2 ]
Zhao, Mingming [1 ,2 ]
机构
[1] Beijing Univ Technol, Fac Informat Technol, Beijing Key Lab Computat Intelligence & Intelligen, Beijing Lab Smart Environm Protect, Beijing 100124, Peoples R China
[2] Beijing Univ Technol, Beijing Inst Artificial Intelligence, Beijing 100124, Peoples R China
基金
北京市自然科学基金; 中国国家自然科学基金;
关键词
Adaptive dynamic programming; intelligent control; optimal tracking control system; stability criterion; value-iteration-based Q-learning; DISCRETE-TIME-SYSTEMS;
D O I
10.1109/TCSII.2024.3359258
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
This brief leverages a value-iteration-based Q-learning (VIQL) scheme to tackle optimal tracking problems for nonlinear nonaffine systems. The optimal policy is learned from measured data instead of a precise mathematical model. Furthermore, a novel criterion is proposed to determine the stability of the iterative policy based on measured data. The evolving control algorithm is developed to verify the proposed criterion by employing these stable policies for system control. The advantage of the early elimination of tracking errors is provided by this approach since various stable policies can be employed before obtaining the optimal strategy. Finally, the effectiveness of the developed algorithm is demonstrated by a simulation experiment.
引用
收藏
页码:3373 / 3377
页数:5
相关论文
共 50 条
  • [21] Stochastic linear quadratic optimal control for model-free discrete-time systems based on Q-learning algorithm
    Wang, Tao
    Zhang, Huaguang
    Luo, Yanhong
    NEUROCOMPUTING, 2018, 312 : 1 - 8
  • [22] Model-free extended Q-learning method for H∞, output tracking control of networked control systems with network delays and packet loss
    Hao, Longyan
    Wang, Chaoli
    Liang, Dong
    Li, Shihua
    NEUROCOMPUTING, 2025, 634
  • [23] Optimal Trajectory Output Tracking Control with a Q-learning Algorithm
    Vamvoudakis, Kyriakos G.
    2016 AMERICAN CONTROL CONFERENCE (ACC), 2016, : 5752 - 5757
  • [24] Model-Free Perception-Based Control via Q-Learning with an Application to Heat-Seeking Missile Guidance
    Kovalik, Wade S.
    Zhai, Lijing
    Vamvoudakis, Kyriakos G.
    5TH IEEE CONFERENCE ON CONTROL TECHNOLOGY AND APPLICATIONS (IEEE CCTA 2021), 2021, : 1142 - 1147
  • [25] Model-Free Algorithms for Containment Control of Saturated Discrete-Time Multiagent Systems via Q-Learning Method
    Long, Mingkang
    Su, Housheng
    Zeng, Zhigang
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2022, 52 (02): : 1308 - 1316
  • [26] Model-free Resilient Controller Design based on Incentive Feedback Stackelberg Game and Q-learning
    Shen, Jiajun
    Li, Fengjun
    Hashemi, Morteza
    Fang, Huazhen
    2024 IEEE 7TH INTERNATIONAL CONFERENCE ON INDUSTRIAL CYBER-PHYSICAL SYSTEMS, ICPS 2024, 2024,
  • [27] Q-learning based model-free input-output feedback linearization control method ?
    Sun, Yipu
    Chen, Xin
    He, Wenpeng
    Zhang, Ziying
    Fukushima, Edwardo F.
    She, Jinhua
    IFAC PAPERSONLINE, 2023, 56 (02): : 9534 - 9539
  • [28] Nonlinear neuro-optimal tracking control via stable iterative Q-learning algorithm
    Wei, Qinglai
    Song, Ruizhuo
    Sun, Qiuye
    NEUROCOMPUTING, 2015, 168 : 520 - 528
  • [29] Model-Free H∞ Optimal Tracking Control of Constrained Nonlinear Systems via an Iterative Adaptive Learning Algorithm
    Hou, Jiaxu
    Wang, Ding
    Liu, Derong
    Zhang, Yun
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2020, 50 (11): : 4097 - 4108
  • [30] Output Feedback Optimal Tracking Control Using Reinforcement Q-Learning
    Rizvi, Syed Ali Asad
    Lin, Zongli
    2018 ANNUAL AMERICAN CONTROL CONFERENCE (ACC), 2018, : 3423 - 3428