Resilient adaptive optimal control of distributed multi-agent systems using reinforcement learning

被引:18
|
作者
Moghadam, Rohollah [1 ]
Modares, Hamidreza [2 ]
机构
[1] Missouri Univ Sci, Dept Elect & Comp Engn, Technol, Emerson Hall,301 W 16th St, Rolla, MO 65409 USA
[2] Michigan State Univ, Dept Mech Engn, E Lansing, MI 48824 USA
来源
IET CONTROL THEORY AND APPLICATIONS | 2018年 / 12卷 / 16期
关键词
game theory; Riccati equations; adaptive control; learning (artificial intelligence); multi-agent systems; distributed control; H control; resilient adaptive optimal control; distributed multiagent systems; model-free reinforcement learning; control protocol; leader-follower multiagent systems; optimal control protocols; adversarial input; adverse effects; learning outcome; intact agents; unified RL-based; control frameworks; corrupted sensory data; actual sensory information; leader state; distributed observer; compromised agent; off-policy RL algorithm; control problem; H control problem; ZERO-SUM GAMES; H-INFINITY; CONTINUOUS-TIME; TRACKING CONTROL; CONSENSUS CONTROL; FEEDBACK-CONTROL; DESIGN; AGENTS; SYNCHRONIZATION; ALGORITHM;
D O I
10.1049/iet-cta.2018.0029
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This study presents a unified resilient model-free reinforcement learning (RL) based distributed control protocol for leader-follower multi-agent systems. Although RL has been successfully used to learn optimal control protocols for multi-agent systems, the effects of adversarial inputs are ignored. It is shown in this study, however, that their adverse effects can propagate across the network and impact the learning outcome of other intact agents. To alleviate this problem, a unified RL-based distributed control frameworks is developed for both homogeneous and heterogeneous multi-agent systems to prevent corrupted sensory data from propagating across the network. To this end, only the leader communicates its actual sensory information and other agents estimate the leader' state using a distributed observer and communicate this estimation to their neighbours to achieve consensus on the leader state. The observer cannot be physically affected by any adversarial input. To further improve resiliency, distributed H8 control protocols are designed to attenuate the effect of the adversarial inputs on the compromised agent itself. An off-policy RL algorithm is developed to learn the solutions of the game algebraic Riccati equations arising from solving the H8 control problem. No knowledge of the agent's dynamics is required and it is shown that the proposed RL-based H8 control protocol is resilient against adversarial inputs.
引用
收藏
页码:2165 / 2174
页数:10
相关论文
共 50 条
  • [41] Multi-agent reinforcement learning with adaptive mimetism
    Yamaguchi, T
    Miura, M
    Yachida, M
    ETFA '96 - 1996 IEEE CONFERENCE ON EMERGING TECHNOLOGIES AND FACTORY AUTOMATION, PROCEEDINGS, VOLS 1 AND 2, 1996, : 288 - 294
  • [42] Reinforcement Learning Approach for Cooperative Control of Multi-Agent Systems
    Javalera-Rincon, Valeria
    Puig Cayuela, Vicenc
    Morcego Seix, Bernardo
    Orduna-Cabrera, Fernando
    PROCEEDINGS OF THE 11TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE (ICAART), VOL 2, 2019, : 80 - 91
  • [43] Multi-agent Reinforcement Learning for Control Systems: Challenges and Proposals
    Grana, Manuel
    Fernandez-Gauna, Borja
    INTELLIGENT DATA ENGINEERING AND AUTOMATED LEARNING - IDEAL 2015, 2015, 9375 : 18 - 25
  • [44] Distributed Reconfiguration for Resilient Synchronization of Multi-Agent Systems
    Diaz-Garcia, Gilberto
    Guatibonza, Daniel
    Giraldo, Luis Felipe
    IEEE ACCESS, 2021, 9 : 140235 - 140247
  • [45] Globally Optimal Multi-agent Reinforcement Learning Parameters in Distributed Task Assignment
    Dahlem, Dominik
    Harrison, William
    2009 IEEE/WIC/ACM INTERNATIONAL JOINT CONFERENCES ON WEB INTELLIGENCE (WI) AND INTELLIGENT AGENT TECHNOLOGIES (IAT), VOL 2, 2009, : 28 - 35
  • [46] Multi-agent Deep Reinforcement Learning Based Optimal Dispatch of Distributed Generators
    Zhang J.
    Pu T.
    Li Y.
    Wang X.
    Zhou X.
    Dianwang Jishu/Power System Technology, 2022, 46 (09): : 3496 - 3503
  • [47] Optimal synchronization control for heterogeneous multi-agent systems: Online adaptive learning solutions
    Zhou, Yuanqiang
    Li, Dewei
    Gao, Furong
    ASIAN JOURNAL OF CONTROL, 2022, 24 (05) : 2352 - 2362
  • [48] Multi-Agent Deep Reinforcement Learning-Based Distributed Optimal Generation Control of DC Microgrids
    Fan, Zhen
    Zhang, Wei
    Liu, Wenxin
    IEEE TRANSACTIONS ON SMART GRID, 2023, 14 (05) : 3337 - 3351
  • [49] Deep Reinforcement Learning Agent for Negotiation in Multi-Agent Cooperative Distributed Predictive Control
    Aponte-Rengifo, Oscar
    Vega, Pastora
    Francisco, Mario
    APPLIED SCIENCES-BASEL, 2023, 13 (04):
  • [50] Resilient Multi-Agent Reinforcement Learning with Adversarial Value Decomposition
    Phan, Thomy
    Belzner, Lenz
    Gabor, Thomas
    Sedlmeier, Andreas
    Ritz, Fabian
    Linnhoff-Popien, Claudia
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 11308 - 11316