Constrained predictive control for consensus of nonlinear multi-agent systems by using game Q-learning

被引:0
|
作者
Wang, Yan [1 ]
Xue, Huiwen [1 ]
Wen, Jiwei [1 ]
Liu, Jinfeng [2 ]
Luan, Xiaoli [1 ]
机构
[1] Jiangnan Univ, Sch Internet Things Engn, Key Lab Adv Proc Control Light Ind, Minist Educ, Wuxi 214122, Peoples R China
[2] Univ Alberta, Dept Chem & Mat Engn, Edmonton, AB T6G 2R3, Canada
基金
中国国家自然科学基金;
关键词
Multi-agent systems; Joint constraints; Learning predictive control; Barrier function;
D O I
10.1007/s11071-024-10698-5
中图分类号
TH [机械、仪表工业];
学科分类号
0802 ;
摘要
This paper develops constrained learning predictive control for achieving consensus in nonlinear multi-agent systems. First, a general predictive and learning framework is constructed for the optimization of control policies by employing an Identifier-Actor-Critic network. Specifically, the Identifier neural network is utilized to approximately characterize the dynamics of the nonlinear system and generate predictive data for available datasets. Each time point within the predictive horizon, regarded as a participant in a non-zero-sum game (NZSG), executes distributed policy and is fed into the Actor-Critic network. When the constrained control policies at all time points reach optimality via the policy gradient algorithm (PGA), the NZSG achieves Nash equilibrium. Subsequently, a gradient recentered self-concordant barrier function is employed to address the joint constraints on tracking error and control input. Moreover, by introducing incremental adjustments, the learning rate factors within the PGA are optimized to enhance the learning efficiency of the Actor-Critic network. Finally, simulation results demonstrate the effectiveness and the rapidity of achieving consensus of the learning predictive control approach compared to the general predictive control methodology.
引用
收藏
页码:11683 / 11700
页数:18
相关论文
共 50 条
  • [31] Iterative Learning Control of Consensus Problems in Nonlinear Multi-agent Systems with Communication Disturbance
    Wei Yongdong
    Li Zonggang
    Du Yajiang
    2020 IEEE 16TH INTERNATIONAL CONFERENCE ON CONTROL & AUTOMATION (ICCA), 2020, : 830 - 835
  • [32] Distributed model predictive control for consensus of nonlinear second-order multi-agent systems
    Gao, Yulong
    Dai, Li
    Xia, Yuanqing
    Liu, Yuwei
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2017, 27 (05) : 830 - 842
  • [33] Iterative learning consensus control of nonlinear impulsive distributed parameter multi-agent systems
    Wu, Jing
    Dai, Xisheng
    Tian, Senping
    Huang, Qingnan
    EUROPEAN JOURNAL OF CONTROL, 2023, 71
  • [34] Quantized iterative learning control for consensus of switched nonlinear heterogeneous multi-agent systems
    Yang, Song
    Li, Xiao-Dong
    NONLINEAR DYNAMICS, 2024, : 6695 - 6716
  • [35] Consensus Tracking of Multi-Agent Systems with Constrained Time-delay by Iterative Learning Control
    Ri, Yongil
    Fang, Yong
    Yu, Hongwen
    PROCEEDINGS OF THE 28TH CHINESE CONTROL AND DECISION CONFERENCE (2016 CCDC), 2016, : 6580 - 6585
  • [36] Resource Allocation for Multi-user Cognitive Radio Systems using Multi-agent Q-Learning
    Azzouna, Ahmed
    Guezmil, Amel
    Sakly, Anis
    Mtibaa, Abdellatif
    ANT 2012 AND MOBIWIS 2012, 2012, 10 : 46 - 53
  • [37] An Enterprise Multi-agent Model with Game Q-Learning Based on a Single Decision Factor
    Xu, Siying
    Zhang, Gaoyu
    Yuan, Xianzhi
    COMPUTATIONAL ECONOMICS, 2024, 64 (04) : 2523 - 2562
  • [38] Q-Learning Policies for Multi-Agent Foraging Task
    Yogeswaran, M.
    Ponnambalam, S. C.
    TRENDS IN INTELLIGENT ROBOTICS, 2010, 103 : 194 - 201
  • [39] Regularized Softmax Deep Multi-Agent Q-Learning
    Pan, Ling
    Rashid, Tabish
    Peng, Bei
    Huang, Longbo
    Whiteson, Shimon
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [40] Multi-Agent Q-Learning for Drone Base Stations
    Janji, Salim
    Kliks, Adrian
    2023 19TH INTERNATIONAL CONFERENCE ON WIRELESS AND MOBILE COMPUTING, NETWORKING AND COMMUNICATIONS, WIMOB, 2023, : 261 - 266