Undiscounted control policy generation for continuous-valued optimal control by approximate dynamic programming

被引:2
|
作者
Lock, Jonathan [1 ]
McKelvey, Tomas [1 ]
机构
[1] Chalmers Univ Technol, Dept Elect Engn, S-41296 Gothenburg, Sweden
关键词
Approximate dynamic programming; control policy; undiscounted infinite-horizon; optimal control;
D O I
10.1080/00207179.2021.1939892
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We present a numerical method for generating the state-feedback control policy associated with general undiscounted, constant-setpoint, infinite-horizon, nonlinear optimal control problems with continuous state variables. The method is based on approximate dynamic programming, and is closely related to approximate policy iteration. Existing methods typically terminate based on the convergence of the control policy and either require a discounted problem formulation or demand the cost function to lie in a specific subclass of functions. The presented method extends on existing termination criteria by requiring both the control policy and the resulting system state to converge, allowing for use with undiscounted cost functions that are bounded and continuous. This paper defines the numerical method, derives the relevant underlying mathematical properties, and validates the numerical method with representative examples. A MATLAB implementation with the shown examples is freely available.
引用
收藏
页码:2854 / 2864
页数:11
相关论文
共 50 条
  • [1] Approximate Dynamic Programming for Continuous State and Control Problems
    Si, Jennie
    Yang, Lei
    Lu, Chao
    Sun, Jian
    Mei, Shengwei
    MED: 2009 17TH MEDITERRANEAN CONFERENCE ON CONTROL & AUTOMATION, VOLS 1-3, 2009, : 1415 - 1420
  • [2] On-policy Approximate Dynamic Programming for Optimal Control of non-linear systems
    Shalini, K.
    Vrushabh, D.
    Sonam, K.
    2020 7TH INTERNATIONAL CONFERENCE ON CONTROL, DECISION AND INFORMATION TECHNOLOGIES (CODIT'20), VOL 1, 2020, : 1058 - 1062
  • [3] Error Bounds of Adaptive Dynamic Programming Algorithms for Solving Undiscounted Optimal Control Problems
    Liu, Derong
    Li, Hongliang
    Wang, Ding
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2015, 26 (06) : 1323 - 1334
  • [4] Convergence Proof of Approximate Policy Iteration for Undiscounted Optimal Control of Discrete-Time Systems
    Zhu, Yuanheng
    Zhao, Dongbin
    He, Haibo
    Ji, Junhong
    COGNITIVE COMPUTATION, 2015, 7 (06) : 763 - 771
  • [5] Convergence Proof of Approximate Policy Iteration for Undiscounted Optimal Control of Discrete-Time Systems
    Yuanheng Zhu
    Dongbin Zhao
    Haibo He
    Junhong Ji
    Cognitive Computation, 2015, 7 : 763 - 771
  • [6] Approximate Dynamic Programming for Optimal Stationary Control with Control-Dependent Noise
    Jiang, Yu
    Jiang, Zhong-Ping
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 2011, 22 (12): : 2392 - 2398
  • [7] Approximate Dynamic Programming with Gaussian Processes for Optimal Control of Continuous-Time Nonlinear Systems
    Beppu, Hirofumi
    Maruta, Ichiro
    Fujimoto, Kenji
    IFAC PAPERSONLINE, 2020, 53 (02): : 6715 - 6722
  • [8] Policy-Iteration-Based Finite-Horizon Approximate Dynamic Programming for Continuous-Time Nonlinear Optimal Control
    Lin, Ziyu
    Duan, Jingliang
    Li, Shengbo Eben
    Ma, Haitong
    Li, Jie
    Chen, Jianyu
    Cheng, Bo
    Ma, Jun
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (09) : 5255 - 5267
  • [9] A Dynamic Programming Approach for Approximate Optimal Control for Cancer Therapy
    Nowakowski, A.
    Popa, A.
    JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 2013, 156 (02) : 365 - 379
  • [10] A Dynamic Programming Approach for Approximate Optimal Control for Cancer Therapy
    A. Nowakowski
    A. Popa
    Journal of Optimization Theory and Applications, 2013, 156 : 365 - 379