Constrained predictive control for consensus of nonlinear multi-agent systems by using game Q-learning

被引:0
|
作者
Wang, Yan [1 ]
Xue, Huiwen [1 ]
Wen, Jiwei [1 ]
Liu, Jinfeng [2 ]
Luan, Xiaoli [1 ]
机构
[1] Jiangnan Univ, Sch Internet Things Engn, Key Lab Adv Proc Control Light Ind, Minist Educ, Wuxi 214122, Peoples R China
[2] Univ Alberta, Dept Chem & Mat Engn, Edmonton, AB T6G 2R3, Canada
基金
中国国家自然科学基金;
关键词
Multi-agent systems; Joint constraints; Learning predictive control; Barrier function;
D O I
10.1007/s11071-024-10698-5
中图分类号
TH [机械、仪表工业];
学科分类号
0802 ;
摘要
This paper develops constrained learning predictive control for achieving consensus in nonlinear multi-agent systems. First, a general predictive and learning framework is constructed for the optimization of control policies by employing an Identifier-Actor-Critic network. Specifically, the Identifier neural network is utilized to approximately characterize the dynamics of the nonlinear system and generate predictive data for available datasets. Each time point within the predictive horizon, regarded as a participant in a non-zero-sum game (NZSG), executes distributed policy and is fed into the Actor-Critic network. When the constrained control policies at all time points reach optimality via the policy gradient algorithm (PGA), the NZSG achieves Nash equilibrium. Subsequently, a gradient recentered self-concordant barrier function is employed to address the joint constraints on tracking error and control input. Moreover, by introducing incremental adjustments, the learning rate factors within the PGA are optimized to enhance the learning efficiency of the Actor-Critic network. Finally, simulation results demonstrate the effectiveness and the rapidity of achieving consensus of the learning predictive control approach compared to the general predictive control methodology.
引用
收藏
页码:11683 / 11700
页数:18
相关论文
共 50 条
  • [1] Multi-Agent Cooperation Q-Learning Algorithm Based on Constrained Markov Game
    Ge, Yangyang
    Zhu, Fei
    Huang, Wei
    Zhao, Peiyao
    Liu, Quan
    COMPUTER SCIENCE AND INFORMATION SYSTEMS, 2020, 17 (02) : 647 - 664
  • [2] Neighbor Q-learning based consensus control for discrete-time multi-agent systems
    Zhu, Xiaoxia
    Yuan, Xin
    Dong, Lu
    Wang, Yuanda
    Sun, Changyin
    OPTIMAL CONTROL APPLICATIONS & METHODS, 2023, 44 (03): : 1475 - 1490
  • [3] Pricing in agent economies using multi-agent Q-learning
    Tesauro, G
    Kephart, JO
    AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2002, 5 (03) : 289 - 304
  • [4] Pricing in Agent Economies Using Multi-Agent Q-Learning
    Gerald Tesauro
    Jeffrey O. Kephart
    Autonomous Agents and Multi-Agent Systems, 2002, 5 : 289 - 304
  • [5] Q-learning in Multi-Agent Cooperation
    Hwang, Kao-Shing
    Chen, Yu-Jen
    Lin, Tzung-Feng
    2008 IEEE WORKSHOP ON ADVANCED ROBOTICS AND ITS SOCIAL IMPACTS, 2008, : 239 - 244
  • [6] Multi-Agent Advisor Q-Learning
    Subramanian S.G.
    Taylor M.E.
    Larson K.
    Crowley M.
    Journal of Artificial Intelligence Research, 2022, 74 : 1 - 74
  • [7] Multi-Agent Reinforcement Learning - An Exploration Using Q-Learning
    Graham, Caoimhin
    Bell, David
    Luo, Zhihui
    RESEARCH AND DEVELOPMENT IN INTELLIGENT SYSTEMS XXVI: INCORPORATING APPLICATIONS AND INNOVATIONS IN INTELLIGENT SYSTEMS XVII, 2010, : 293 - 298
  • [8] Multi-Agent Advisor Q-Learning
    Subramanian, Sriram Ganapathi
    Taylor, Matthew E.
    Larson, Kate
    Crowley, Mark
    JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2022, 74 : 1 - 74
  • [9] Multi-Agent Advisor Q-Learning
    Subramanian, Sriram Ganapathi
    Taylor, Matthew E.
    Larson, Kate
    Crowley, Mark
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 6884 - 6889
  • [10] Using Fuzzy Logic and Q-Learning for Trust Modeling in Multi-agent Systems
    Aref, Abdullah
    Tran, Thomas
    FEDERATED CONFERENCE ON COMPUTER SCIENCE AND INFORMATION SYSTEMS, 2014, 2014, 2 : 59 - 66