Model-Free Optimal Tracking Design With Evolving Control Strategies via Q-Learning

被引:2
|
作者
Wang, Ding [1 ,2 ]
Huang, Haiming [1 ,2 ]
Zhao, Mingming [1 ,2 ]
机构
[1] Beijing Univ Technol, Fac Informat Technol, Beijing Key Lab Computat Intelligence & Intelligen, Beijing Lab Smart Environm Protect, Beijing 100124, Peoples R China
[2] Beijing Univ Technol, Beijing Inst Artificial Intelligence, Beijing 100124, Peoples R China
基金
北京市自然科学基金; 中国国家自然科学基金;
关键词
Adaptive dynamic programming; intelligent control; optimal tracking control system; stability criterion; value-iteration-based Q-learning; DISCRETE-TIME-SYSTEMS;
D O I
10.1109/TCSII.2024.3359258
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
This brief leverages a value-iteration-based Q-learning (VIQL) scheme to tackle optimal tracking problems for nonlinear nonaffine systems. The optimal policy is learned from measured data instead of a precise mathematical model. Furthermore, a novel criterion is proposed to determine the stability of the iterative policy based on measured data. The evolving control algorithm is developed to verify the proposed criterion by employing these stable policies for system control. The advantage of the early elimination of tracking errors is provided by this approach since various stable policies can be employed before obtaining the optimal strategy. Finally, the effectiveness of the developed algorithm is demonstrated by a simulation experiment.
引用
收藏
页码:3373 / 3377
页数:5
相关论文
共 50 条
  • [31] Model-Free control performance improvement using virtual reference feedback tuning and reinforcement Q-learning
    Radac, Mircea-Bogdan
    Precup, Radu-Emil
    Roman, Raul-Cristian
    INTERNATIONAL JOURNAL OF SYSTEMS SCIENCE, 2017, 48 (05) : 1071 - 1083
  • [32] Q-Learning Based Optimal Tracking Control of Free-Flying Space Manipulators with Unknown Dynamics
    Zhu, Hongxu
    Wu, Shufan
    Shen, Qiang
    Sun, Ran
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 2357 - 2362
  • [33] Improved model-free H∞ control for batch processes via off-policy 2D game Q-learning
    Jiang, Xueying
    Huang, Min
    Kuang, Hanbin
    Shi, Huiyuan
    Wang, Xingwei
    Lee, Loo Hay
    INTERNATIONAL JOURNAL OF CONTROL, 2023, 96 (10) : 2447 - 2463
  • [34] Data-driven model reference control of MIMO vertical tank systems with model-free VRFT and Q-Learning
    Radac, Mircea-Bogdan
    Precup, Radu-Emil
    Roman, Raul-Cristian
    ISA TRANSACTIONS, 2018, 73 : 227 - 238
  • [35] Model-free nonlinear robust control design via online critic learning
    Wang, Xiaoyang
    Deng, Hao
    Ye, Xiufen
    ISA TRANSACTIONS, 2022, 129 : 446 - 459
  • [36] Off-policy Q-learning: Optimal tracking control for networked control systems
    Li J.-N.
    Yin Z.-X.
    Kongzhi yu Juece/Control and Decision, 2019, 34 (11): : 2343 - 2349
  • [37] Optimal Tracking Control of a Nonlinear Multiagent System Using Q-Learning via Event-Triggered Reinforcement Learning
    Wang, Ziwei
    Wang, Xin
    Tang, Yijie
    Liu, Ying
    Hu, Jun
    ENTROPY, 2023, 25 (02)
  • [38] Q-learning Based Adaptive Optimal Control for Linear Quadratic Tracking Problem
    Sharma, Shashi Kant
    Jha, Sumit Kumar
    Dhawan, Amit
    Tiwari, Manish
    INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2023, 21 (08) : 2718 - 2725
  • [39] Q-learning Based Adaptive Optimal Control for Linear Quadratic Tracking Problem
    Shashi Kant Sharma
    Sumit Kumar Jha
    Amit Dhawan
    Manish Tiwari
    International Journal of Control, Automation and Systems, 2023, 21 : 2718 - 2725
  • [40] Optimal Learning Output Tracking Control: A Model-Free Policy Optimization Method With Convergence Analysis
    Lin, Mingduo
    Zhao, Bo
    Liu, Derong
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (03) : 5574 - 5585