Distributed Minmax Strategy for Consensus Tracking in Differential Graphical Games: A Model-Free Approach

被引:3
|
作者
Zhou, Yan [1 ]
Zhou, Jialing [2 ]
Wen, Guanghui [3 ]
Gan, Minggang [4 ]
Yang, Tao [5 ]
机构
[1] Southeast Univ, Sch Cyber Sci & Engn, Nanjing 211189, Peoples R China
[2] Beijing Inst Technol, Adv Res Inst Multidisciplinary Sci, Beijing 100081, Peoples R China
[3] Southeast Univ, Dept Syst Sci, Nanjing 211189, Peoples R China
[4] Beijing Inst Technol, State Key Lab Intelligent Control & Decis Complex, Sch Automat, Beijing 100081, Peoples R China
[5] Northeastern Univ, State Key Lab Synthet Automat Proc Ind, Shenyang 110819, Peoples R China
来源
基金
中国国家自然科学基金;
关键词
Asymptotic stability; Sufficient conditions; Heuristic algorithms; Riccati equations; Games; Reinforcement learning; Mathematical models; ADAPTIVE OPTIMAL-CONTROL; SYSTEMS; ITERATION;
D O I
10.1109/MSMC.2023.3282774
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
This article focuses on the design of distributed minmax strategies for multiagent consensus tracking control problems with completely unknown dynamics in the presence of external disturbances or attacks. Each agent obtains its distributed minmax strategy by solving a multiagent zero-sum differential graphical game, which involves both nonadversarial and adversarial behaviors. Solving such a game is equivalent to solving a game algebraic Riccati equation (GARE). By making slight assumptions concerning performance matrices, L-2 stability and asymptotic stability of the closed-loop consensus error systems are strictly proven. Furthermore, inspired by data-driven off-policy reinforcement learning (RL), a model-free policy iteration (PI) algorithm is presented for each follower to generate the minmax strategy. Finally, simulations are performed to demonstrate the effectiveness of the proposed theoretical results.
引用
收藏
页码:53 / 68
页数:16
相关论文
共 50 条
  • [31] PHAST: MODEL-FREE PHASELESS SUBSPACE TRACKING
    Nayer, Seyedehsara
    Vaswani, Namrata
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 3712 - 3716
  • [32] Model-Free Reinforcement Learning for Mean Field Games
    Mishra, Rajesh
    Vasal, Deepanshu
    Vishwanath, Sriram
    IEEE TRANSACTIONS ON CONTROL OF NETWORK SYSTEMS, 2023, 10 (04): : 2141 - 2151
  • [33] A model-free approach for testing association
    Chatterjee, Saptarshi
    Chowdhury, Shrabanti
    Basu, Sanjib
    JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES C-APPLIED STATISTICS, 2021, 70 (03) : 511 - 531
  • [34] A model-free approach for posture classification
    Castiello, C
    D'Orazio, T
    Fanelli, AM
    Spagnolo, P
    Torsello, MA
    AVSS 2005: ADVANCED VIDEO AND SIGNAL BASED SURVEILLANCE, PROCEEDINGS, 2005, : 276 - 281
  • [35] Three-level hierarchical model-free learning approach to trajectory tracking control
    Radac, Mircea-Bogdan
    Precup, Radu-Emil
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2016, 55 : 103 - 118
  • [36] Distributed Model-free Control in Low Voltage Distribution Networks: A Mean Field Approach
    Wei, Boyuan
    Deconinck, Geert
    2020 INTERNATIONAL CONFERENCE ON PROBABILISTIC METHODS APPLIED TO POWER SYSTEMS (PMAPS), 2020,
  • [37] Robust tracking control of hypersonic flight vehicles: A continuous model-free control approach
    Bu, Xiangwei
    Lei, Humin
    Gao, Yupeng
    ACTA ASTRONAUTICA, 2019, 161 : 234 - 240
  • [38] Data-Driven Model-Free Iterative Tuning Approach for Smooth and Accurate Tracking
    Li, Xiaocong
    Chen, Si-Lu
    Ma, Jun
    Teo, Chek Sing
    Tan, Kok Kiong
    2018 IEEE/ASME INTERNATIONAL CONFERENCE ON ADVANCED INTELLIGENT MECHATRONICS (AIM), 2018, : 593 - 598
  • [39] An inverse optimal approach for distributed preview consensus tracking
    Jiang, Shuoying
    Ding, Zhengtao
    INTERNATIONAL JOURNAL OF CONTROL, 2024, 97 (04) : 858 - 870
  • [40] Model-Based Reinforcement Learning in Differential Graphical Games
    Kamalapurkar, Rushikesh
    Klotz, Justin R.
    Walters, Patrick
    Dixon, Warren E.
    IEEE TRANSACTIONS ON CONTROL OF NETWORK SYSTEMS, 2018, 5 (01): : 423 - 433