Model-Free Optimal Tracking Design With Evolving Control Strategies via Q-Learning

被引:2
|
作者
Wang, Ding [1 ,2 ]
Huang, Haiming [1 ,2 ]
Zhao, Mingming [1 ,2 ]
机构
[1] Beijing Univ Technol, Fac Informat Technol, Beijing Key Lab Computat Intelligence & Intelligen, Beijing Lab Smart Environm Protect, Beijing 100124, Peoples R China
[2] Beijing Univ Technol, Beijing Inst Artificial Intelligence, Beijing 100124, Peoples R China
基金
北京市自然科学基金; 中国国家自然科学基金;
关键词
Adaptive dynamic programming; intelligent control; optimal tracking control system; stability criterion; value-iteration-based Q-learning; DISCRETE-TIME-SYSTEMS;
D O I
10.1109/TCSII.2024.3359258
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
This brief leverages a value-iteration-based Q-learning (VIQL) scheme to tackle optimal tracking problems for nonlinear nonaffine systems. The optimal policy is learned from measured data instead of a precise mathematical model. Furthermore, a novel criterion is proposed to determine the stability of the iterative policy based on measured data. The evolving control algorithm is developed to verify the proposed criterion by employing these stable policies for system control. The advantage of the early elimination of tracking errors is provided by this approach since various stable policies can be employed before obtaining the optimal strategy. Finally, the effectiveness of the developed algorithm is demonstrated by a simulation experiment.
引用
收藏
页码:3373 / 3377
页数:5
相关论文
共 50 条
  • [1] Model-Free Optimal Tracking Control via Critic-Only Q-Learning
    Luo, Biao
    Liu, Derong
    Huang, Tingwen
    Wang, Ding
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2016, 27 (10) : 2134 - 2144
  • [2] Adjustable Iterative Q-Learning Schemes for Model-Free Optimal Tracking Control
    Qiao, Junfei
    Zhao, Mingming
    Wang, Ding
    Ha, Mingming
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2024, 54 (02): : 1202 - 1213
  • [3] Quantized measurements in Q-learning based model-free optimal control
    Tiistola, Sini
    Ritala, Risto
    Vilkko, Matti
    IFAC PAPERSONLINE, 2020, 53 (02): : 1640 - 1645
  • [4] Iterative Q-Learning for Model-Free Optimal Control With Adjustable Convergence Rate
    Wang, Ding
    Wang, Yuan
    Zhao, Mingming
    Qiao, Junfei
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2024, 71 (04) : 2224 - 2228
  • [5] Model-free optimal tracking control for discrete-time system with delays using reinforcement Q-learning
    Liu, Yang
    Yu, Rui
    ELECTRONICS LETTERS, 2018, 54 (12) : 750 - 751
  • [6] Optimal tracking control for discrete-time systems by model-free off-policy Q-learning approach
    Li, Jinna
    Yuan, Decheng
    Ding, Zhengtao
    2017 11TH ASIAN CONTROL CONFERENCE (ASCC), 2017, : 7 - 12
  • [7] Gaussian Process Based Model-free Control with Q-Learning
    Hauser, Jan
    Pachner, Daniel
    Havlena, Vladimir
    IFAC PAPERSONLINE, 2019, 52 (11): : 236 - 243
  • [8] Model-free optimal chiller loading method based on Q-learning
    Qiu, Shunian
    Li, Zhenhai
    Li, Zhengwei
    Zhang, Xinfang
    SCIENCE AND TECHNOLOGY FOR THE BUILT ENVIRONMENT, 2020, 26 (08) : 1100 - 1116
  • [9] Model-Free Optimal Tracking Control of Nonlinear Input-Affine Discrete-Time Systems via an Iterative Deterministic Q-Learning Algorithm
    Song, Shijie
    Zhu, Minglei
    Dai, Xiaolin
    Gong, Dawei
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (01) : 999 - 1012
  • [10] Model-free H∞ Stochastic Optimal Design for Unknown Linear Networked Control System Zero-sum Games via Q-Learning
    Xu, Hao
    Jagannathan, S.
    2011 IEEE INTERNATIONAL SYMPOSIUM ON INTELLIGENT CONTROL (ISIC), 2011, : 198 - 203