H∞ Tracking Control for Linear Discrete-Time Systems: Model-Free Q-Learning Designs

被引:39
|
作者
Yang, Yunjie [1 ]
Wan, Yan [2 ]
Zhu, Jihong [1 ]
Lewis, Frank L. [3 ]
机构
[1] Tsinghua Univ, Dept Comp Sci & Technol, Beijing 100084, Peoples R China
[2] Univ Texas Arlington, Dept Elect Engn, Arlington, TX 76019 USA
[3] Univ Texas Arlington, UTA Res Inst, Ft Worth, TX 75052 USA
来源
IEEE CONTROL SYSTEMS LETTERS | 2021年 / 5卷 / 01期
基金
中国国家自然科学基金;
关键词
Linear discrete-time systems; H-infinity tracking control; Q-learning; ZERO-SUM GAMES;
D O I
10.1109/LCSYS.2020.3001241
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this letter, a novel model-free Q-learning based approach is developed to solve the H-infinity tracking problem for linear discrete-time systems. A new exponential discounted value function is introduced that includes the cost of the whole control input and tracking error. The tracking Bellman equation and the game algebraic Riccati equation (GARE) are derived. The solution to the GARE leads to the feedback and feedforward parts of the control input. A Q-learning algorithm is then developed to learn the solution of the GARE online without requiring any knowledge of the system dynamics. Convergence of the algorithm is analyzed, and it is also proved that probing noises in maintaining the persistence of excitation (PE) condition do not result in any bias. An example of the F-16 aircraft short period dynamics is developed to validate the proposed algorithm.
引用
收藏
页码:175 / 180
页数:6
相关论文
共 50 条
  • [31] The Adaptive Optimal Output Feedback Tracking Control of Unknown Discrete-Time Linear Systems Using a Multistep Q-Learning Approach
    Dong, Xunde
    Lin, Yuxin
    Suo, Xudong
    Wang, Xihao
    Sun, Weijie
    MATHEMATICS, 2024, 12 (04)
  • [32] Output Feedback Reinforcement Q-learning for Optimal Quadratic Tracking Control of Unknown Discrete-Time Linear Systems and Its Application
    Zhao, Guangyue
    Sun, Weijie
    Cai, He
    Peng, Yunjian
    2018 15TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION (ICARCV), 2018, : 750 - 755
  • [33] MODEL-FREE MULTI-KERNEL LEARNING CONTROL FOR NONLINEAR DISCRETE-TIME SYSTEMS
    Liu, Jiahang
    Xu, Xin
    Huang, Zhenhua
    Lian, Chuanqiang
    INTERNATIONAL JOURNAL OF ROBOTICS & AUTOMATION, 2017, 32 (05): : 538 - 550
  • [34] The model-free learning adaptive control of a class of MISO nonlinear discrete-time systems
    Hou, ZS
    Han, CW
    Huang, WH
    LOW COST AUTOMATION 1998 (LCA'98), 1999, : 227 - 232
  • [35] H∞ Tracking Control of Unknown Discrete-Time Linear Systems via Output-Data-Driven Off-policy Q-learning Algorithm
    Zhang, Kun
    Liu, Xuantong
    Zhang, Lei
    Chen, Qian
    Peng, Yunjian
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 2350 - 2356
  • [36] Model-free extended Q-learning method for H∞, output tracking control of networked control systems with network delays and packet loss
    Hao, Longyan
    Wang, Chaoli
    Liang, Dong
    Li, Shihua
    NEUROCOMPUTING, 2025, 634
  • [37] Explainable data-driven Q-learning control for a class of discrete-time linear autonomous systems
    Perrusquia, Adolfo
    Zou, Mengbang
    Guo, Weisi
    INFORMATION SCIENCES, 2024, 682
  • [38] Online Model-Free Reinforcement Learning for Output Feedback Tracking Control of a Class of Discrete-Time Systems With Input Saturation
    Al-Mahasneh, Ahmad Jobran
    Anavatti, Sreenatha G.
    Garratt, Matthew A.
    IEEE ACCESS, 2022, 10 : 104966 - 104979
  • [39] Adaptive Q-Learning Based Model-Free H∞ Control of Continuous-Time Nonlinear Systems: Theory and Application
    Zhao, Jun
    Lv, Yongfeng
    Wang, Zhangu
    Zhao, Ziliang
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024,
  • [40] Model-Free Optimal Output Regulation for Linear Discrete-Time Lossy Networked Control Systems
    Fan, Jialu
    Wu, Qian
    Jiang, Yi
    Chai, Tianyou
    Lewis, Frank L.
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2020, 50 (11): : 4033 - 4042