Optimized Leader-Follower Consensus Control Using Reinforcement Learning for a Class of Second-Order Nonlinear Multiagent Systems

被引:46
|
作者
Wen, Guoxing [1 ]
Li, Bin [2 ]
机构
[1] Binzhou Univ, Coll Sci, Binzhou 256600, Shandong, Peoples R China
[2] Qilu Univ Technol, Sch Math & Stat, Shandong Acad Sci, Jinan 250353, Shandong, Peoples R China
基金
中国国家自然科学基金;
关键词
Optimal control; Multi-agent systems; Artificial neural networks; Heuristic algorithms; Reinforcement learning; Consensus control; Topology; Double integrator dynamic; multiagent system; neural network (NN); optimal control; reinforcement learning (RL); unknown nonlinear dynamic; HJB EQUATION; NETWORKS;
D O I
10.1109/TSMC.2021.3130070
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this article, an optimized leader-follower consensus control is proposed for a class of second-order unknown nonlinear dynamical multiagent system. Different with the first-order multiagent consensus, the second-order case needs to achieve the agreement not only on position but also on velocity, therefore this optimized control is more challenging and interesting. To derive the control, reinforcement learning (RL) can be a natural consideration because it can overcome the difficulty of solving the Hamilton-Jacobi-Bellman (HJB) equation. To implement RL, it needs to iterate both adaptive critic and actor networks each other. However, if this optimized control learns RL from most existing optimal methods that derives the critic and actor adaptive laws from the negative gradient of square of the approximating function of the HJB equation, this control algorithm will be very intricate, because the HJB equation correlated to a second-order nonlinear multiagent system will become very complex due to strong state coupling and nonlinearity. In this work, since the two RL adaptive laws are derived via implementing the gradient descent method to a simple positive function, which is obtained on the basis of a partial derivative of the HJB equation, this optimized control is significantly simple. Meanwhile, it not merely can avoid the requirement of known dynamic acknowledge, but also can release the condition of persistent excitation, which is demanded in most RL optimization methods for training the adaptive parameter more sufficiently. Finally, the proposed control is demonstrated by both theory and computer simulation.
引用
收藏
页码:5546 / 5555
页数:10
相关论文
共 50 条
  • [1] Distributed leader-follower consensus for a class of semilinear second-order multiagent systems using time scale theory
    Babenko, Serhii
    Defoort, Michael
    Djemai, Mohamed
    Nicaise, Serge
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2019, 29 (02) : 433 - 450
  • [2] Delay-Distribution-Dependent Consensus for Second-Order Leader-Follower Nonlinear Multiagent Systems via Pinning Control
    Li, Hongjie
    Chen, Ming
    Shen, Shigen
    Li, Lin
    ABSTRACT AND APPLIED ANALYSIS, 2013,
  • [3] Fixed-Time Leader-Follower Output Feedback Consensus for Second-Order Multiagent Systems
    Tian, Bailing
    Lu, Hanchen
    Zuo, Zongyu
    Yang, Wen
    IEEE TRANSACTIONS ON CYBERNETICS, 2019, 49 (04) : 1545 - 1550
  • [4] Leader-Follower Consensus of Second-Order Multiagent Systems with Absent Velocity Measurement and Time Delay
    Yang, Panpan
    Tang, Ye
    Yan, Maode
    Zuo, Lei
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2018, 2018
  • [5] Second-Order Super-Twisting Sliding Mode Control for Finite-Time Leader-Follower Consensus with Uncertain Nonlinear Multiagent Systems
    Liu, Nan
    Ling, Rui
    Huang, Qin
    Zhu, Zheren
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2015, 2015
  • [6] Adaptive practical predefined-time leader-follower consensus for second-order multiagent systems with uncertain disturbances
    Mei, Hong
    Wen, Xin
    Ma, Xiaolu
    Tan, Yibo
    Wang, Jian
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2024, 361 (06):
  • [7] Dynamic Leader-Follower Output Containment Control of Heterogeneous Multiagent Systems Using Reinforcement Learning
    Zhang, Huaipin
    Zhao, Wei
    Xie, Xiangpeng
    Yue, Dong
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2024, 54 (09): : 5307 - 5316
  • [8] Leader-follower Consensus Control of a Class of Nonholonomic Systems
    Khoo, Suiyang
    Xie, Lihua
    Man, Zhihong
    11TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION (ICARCV 2010), 2010, : 1381 - 1386
  • [9] Event-Triggered Finite-Time Consensus of Second-Order Leader-Follower Multiagent Systems With Uncertain Disturbances
    Fan, Huijin
    Zheng, Kanghua
    Liu, Lei
    Wang, Bo
    Wang, Wei
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (05) : 3783 - 3793
  • [10] Reinforcement Learning Control for Consensus of the Leader-Follower Multi-Agent Systems
    Chiang, Ming-Li
    Liu, An-Sheng
    Fu, Li-Chen
    PROCEEDINGS OF 2018 IEEE 7TH DATA DRIVEN CONTROL AND LEARNING SYSTEMS CONFERENCE (DDCLS), 2018, : 1152 - 1157