Robust Learning-Based Predictive Control for Discrete-Time Nonlinear Systems With Unknown Dynamics and State Constraints

被引:33
|
作者
Zhang, Xinglong [1 ]
Liu, Jiahang [1 ]
Xu, Xin [1 ]
Yu, Shuyou [2 ,3 ]
Chen, Hong [4 ]
机构
[1] Natl Univ Def Technol, Coll Intelligence Sci & Technol, Changsha 410073, Peoples R China
[2] Jilin Univ, State Key Lab Automot Simulat & Control, Changchun 130025, Peoples R China
[3] Jilin Univ, Dept Control Sci & Engn, Changchun 130025, Peoples R China
[4] Tongji Univ, Coll Elect & Informat Engn, Shanghai 201804, Peoples R China
来源
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS | 2022年 / 52卷 / 12期
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
Robustness; Predictive control; Aerospace electronics; Computational modeling; Predictive models; Optimal control; Heuristic algorithms; Model predictive control (MPC); nonlinear systems; reinforcement learning (RL); robustness; state constraints; TRACKING CONTROL; LINEAR-SYSTEMS; STABILITY;
D O I
10.1109/TSMC.2022.3146284
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Robust model predictive control (MPC) is a well-known control technique for model-based control with constraints and uncertainties. In classic robust tube-based MPC approaches, an open-loop control sequence is computed via periodically solving an online nominal MPC problem, which requires prior model information and frequent access to onboard computational resources. In this article, we propose an efficient robust MPC solution based on receding horizon reinforcement learning, called r-LPC, for unknown nonlinear systems with state constraints and disturbances. The proposed r-LPC utilizes a Koopman operator-based prediction model obtained offline from precollected input-output datasets. Unlike classic tube-based MPC, in each prediction time interval of r-LPC, we use an actor-critic structure to learn a near-optimal feedback control policy rather than a control sequence. The resulting closed-loop control policy can be learned offline and deployed online or learned online in an asynchronous way. In the latter case, online learning can be activated whenever necessary; for instance, the safety constraint is violated with the deployed policy. The closed-loop recursive feasibility, robustness, and asymptotic stability are proven under function approximation errors of the actor-critic networks. Simulation and experimental results on two nonlinear systems with unknown dynamics and disturbances have demonstrated that our approach has better or comparable performance when compared with tube-based MPC and linear quadratic regulator, and outperforms a recently developed actor-critic learning approach.
引用
收藏
页码:7314 / 7327
页数:14
相关论文
共 50 条
  • [21] Reinforcement learning-based online adaptive controller design for a class of unknown nonlinear discrete-time systems with time delays
    Yuling Liang
    Huaguang Zhang
    Geyang Xiao
    He Jiang
    Neural Computing and Applications, 2018, 30 : 1733 - 1745
  • [22] Linear Time-Varying Robust Model Predictive Control for Discrete-Time Nonlinear Systems
    Pereira, Goncalo Collares
    Lima, Pedro F.
    Wahlberg, Bo
    Pettersson, Henrik
    Martensson, Jonas
    2018 IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2018, : 2659 - 2666
  • [23] Learning-based T-sHDP(λ) for optimal control of a class of nonlinear discrete-time systems
    Yu, Luyang
    Liu, Weibo
    Liu, Yurong
    Alsaadi, Fawaz E.
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2022, 32 (05) : 2624 - 2643
  • [24] State Control Design for Discrete-time Systems with State Constraints
    Krokavec, Dusan
    Filasova, Anna
    2018 19TH INTERNATIONAL CARPATHIAN CONTROL CONFERENCE (ICCC), 2018, : 217 - 222
  • [25] H∞ Control of Unknown Discrete-Time Nonlinear Systems with Control Constraints Using Adaptive Dynamic Programming
    Liu, Derong
    Li, Hongliang
    Wang, Ding
    2012 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2012,
  • [26] Optimal Tracking Control of Affine Nonlinear Discrete-time Systems with Unknown Internal Dynamics
    Dierks, Travis
    Jagannathan, S.
    PROCEEDINGS OF THE 48TH IEEE CONFERENCE ON DECISION AND CONTROL, 2009 HELD JOINTLY WITH THE 2009 28TH CHINESE CONTROL CONFERENCE (CDC/CCC 2009), 2009, : 6750 - 6755
  • [27] Composite Learning Fixed-Time Control for Nonlinear Servo Systems With State Constraints and Unknown Dynamics
    Wang, Shubo
    Sun, Chuanbin
    Chen, Qiang
    He, Haoran
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2025, 55 (03): : 2332 - 2342
  • [28] Robust iterative learning control for linear discrete-time systems with initial state learning
    Ayatinia, Mojtaba
    Forouzanfar, Mehdi
    Ramezani, Amin
    JOURNAL OF VIBRATION AND CONTROL, 2023, 29 (9-10) : 2242 - 2254
  • [29] Robust safe reinforcement learning control of unknown continuous-time nonlinear systems with state constraints and disturbances
    Zhang, Haoran
    Zhao, Chunhui
    Ding, Jinliang
    JOURNAL OF PROCESS CONTROL, 2023, 128
  • [30] Robust observer-based output feedback for nonlinear discrete-time systems with constraints
    Imsland, L
    Slupphaug, O
    Foss, BA
    NONLINEAR CONTROL SYSTEMS 2001, VOLS 1-3, 2002, : 1247 - 1252