Linear quadratic optimal control method based on output feedback inverse reinforcement Q-learning

被引:0
|
作者
Liu, Wen [1 ]
Fan, Jia-Lu [1 ]
Xue, Wen-Qian [1 ]
机构
[1] State Key Laboratory of Synthetical Automation for Process Industries, Northeastern University, Liaoning, Shenyang,110819, China
关键词
u6536u7A3Fu65E5u671F: 2022u221206u221221; u5F55u7528u65E5u671F: 2023u221206u221204. u2020u901Au4FE1u4F5Cu8005. E-mail: jlfan@mail.neu.edu.cn; Tel.: +86 18204058952. u672Cu6587u8D23u4EFBu7F16u59D4: u8D75u51ACu658C. u56FDu5BB6u81EAu7136u79D1u5B66u57FAu91D1u91CDu5927u9879u76EE(61991400); u8FBDu5B81u7701u201Cu5174u8FBDu82F1u624Du8BA1u201Du9879u76EE(XLYC2007135)u8D44u52A9. Supported by the National Natural Science Foundation of China (61991400) and the Liaoning Revitalization Talents Program (XLYC2007135);
D O I
10.7641/CTA.2023.20551
中图分类号
学科分类号
摘要
引用
收藏
页码:1469 / 1479
相关论文
共 50 条
  • [21] Optimal Control Inspired Q-Learning for Switched Linear Systems
    Chen, Hua
    Zheng, Linfang
    Zhang, Wei
    2020 AMERICAN CONTROL CONFERENCE (ACC), 2020, : 4003 - 4010
  • [22] The Adaptive Optimal Output Feedback Tracking Control of Unknown Discrete-Time Linear Systems Using a Multistep Q-Learning Approach
    Dong, Xunde
    Lin, Yuxin
    Suo, Xudong
    Wang, Xihao
    Sun, Weijie
    MATHEMATICS, 2024, 12 (04)
  • [23] Reinforcement Q-learning for optimal tracking control of linear discrete-time systems with unknown dynamics
    Kiumarsi, Bahare
    Lewis, Frank L.
    Modares, Hamidreza
    Karimpour, Ali
    Naghibi-Sistani, Mohammad-Bagher
    AUTOMATICA, 2014, 50 (04) : 1167 - 1175
  • [24] Inverse Reinforcement Learning in Tracking Control Based on Inverse Optimal Control
    Xue, Wenqian
    Kolaric, Patrik
    Fan, Jialu
    Lian, Bosen
    Chai, Tianyou
    Lewis, Frank L.
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (10) : 10570 - 10581
  • [25] Stochastic linear quadratic optimal control for model-free discrete-time systems based on Q-learning algorithm
    Wang, Tao
    Zhang, Huaguang
    Luo, Yanhong
    NEUROCOMPUTING, 2018, 312 : 1 - 8
  • [26] Data-Driven $H_{∞}$ Optimal Output Feedback Control for Linear Discrete-Time Systems Based on Off-Policy Q-Learning
    Zhang, Li
    Fan, Jialu
    Xue, Wenqian
    Lopez, Victor G.
    Li, Jinna
    Chai, Tianyou
    Lewis, Frank L.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (07) : 3553 - 3567
  • [27] An Optimal Control Method for Expressways Entering Ramps Metering Based on Q-Learning
    Ji, Xiaofeng
    He, Zenghui
    ICICTA: 2009 SECOND INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTATION TECHNOLOGY AND AUTOMATION, VOL I, PROCEEDINGS, 2009, : 739 - 741
  • [28] Output Feedback Speed Control for a Wankel Rotary Engine via Q-Learning
    Chen, Anthony Siming
    Herrmann, Guido
    Burgess, Stuart
    Brace, Chris
    IFAC PAPERSONLINE, 2023, 56 (02): : 8278 - 8283
  • [29] Reinforcement learning-based adaptive optimal output feedback control for nonlinear systems with output quantization
    Jin, Yitong
    Wang, Fang
    Lai, Guanyu
    Zhang, Xueyi
    NONLINEAR DYNAMICS, 2024, : 7029 - 7045
  • [30] An Optimal Tracking Control Method with Q-learning for Discrete-time Linear Switched System
    Zhao, Shangwei
    Wang, Jingcheng
    Wang, Hongyuan
    Xu, Haotian
    PROCEEDINGS OF THE 39TH CHINESE CONTROL CONFERENCE, 2020, : 1414 - 1419