Autonomous Swarm Robot Coordination via Mean-Field Control Embedding Multi-Agent Reinforcement Learning

被引:0
|
作者
Tang, Huaze [1 ]
Zhang, Hengxi [1 ]
Shi, Zhenpeng [1 ]
Chen, Xinlei [1 ]
Ding, Wenbo [1 ,2 ]
Zhang, Xiao-Ping [1 ,2 ,3 ]
机构
[1] Tsinghua Berkeley Shenzhen Inst, Tsinghua Univ, Tsinghua Shenzhen Int Grad Sch, Shenzhen, Peoples R China
[2] RISC Int Open Source Lab, Shenzhen 518055, Peoples R China
[3] Ryerson Univ, Dept Elect Comp & Biomed Engn, Toronto, ON, Canada
关键词
D O I
10.1109/IROS55552.2023.10341749
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The learning approaches of designing a controller to guide the collective behavior of swarm robots have gained significant attention in recent years. However, the scalability of swarm robots and their inherent stochasticity complicate the control problem due to increasing complexity, unpredictability, and non-linearity. Despite considerable progress made in swarm robotics, addressing these challenges remains a significant issue. In this work, we model the stochastic dynamics of a swarm robot system and then propose a novel control framework based on a mean-field control (MFC) embedding multi-agent reinforcement learning (MARL) approach named MF-MARL to deal with these challenges. While MARL is able to deal with stochasticity statistically, we integrate MFC, allowing MF-MARL to cope with large-scale robots. Moreover, we apply statistical moments of robots' state and control action to discretize continuous input and enable MF-MARL to be applied in continuous scenarios. To demonstrate the effectiveness of MF-MARL, we evaluate the performance of the robots on a specific swarm simulation platform. The experimental results show that our algorithm outperforms the traditional algorithms both in navigation and manipulation tasks. Finally, we demonstrate the adaptability of the proposed algorithm through the component failure test.
引用
收藏
页码:8820 / 8826
页数:7
相关论文
共 50 条
  • [31] On the Approximation of Cooperative Heterogeneous Multi-Agent Reinforcement Learning (MARL) using Mean Field Control (MFC)
    Mondal, Washim Uddin
    Agarwal, Mridul
    Aggarwal, Vaneet
    Ukkusuri, Satish V.
    Journal of Machine Learning Research, 2022, 23
  • [32] On the Approximation of Cooperative Heterogeneous Multi-Agent Reinforcement Learning (MARL) using Mean Field Control (MFC)
    Mondal, Washim Uddin
    Aggarwal, Vaneet
    Ukkusuri, Satish, V
    Agarwal, Mridul
    JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [33] A multi-agent reinforcement learning approach to robot soccer
    Yong Duan
    Bao Xia Cui
    Xin He Xu
    Artificial Intelligence Review, 2012, 38 : 193 - 211
  • [34] A multi-agent reinforcement learning approach to robot soccer
    Duan, Yong
    Cui, Bao Xia
    Xu, Xin He
    ARTIFICIAL INTELLIGENCE REVIEW, 2012, 38 (03) : 193 - 211
  • [35] Autonomous learning of reward distribution for each agent in multi-agent reinforcement learning
    Shibata, K
    Ito, K
    INTELLIGENT AUTONOMOUS SYSTEMS 6, 2000, : 495 - 502
  • [36] Scalable Deep Multi-Agent Reinforcement Learning via Observation Embedding and Parameter Noise
    Zhang, Jian
    Pan, Yaozong
    Yang, Haitao
    Fang, Yuqiang
    IEEE ACCESS, 2019, 7 : 54615 - 54622
  • [37] Multi-agent reinforcement learning for redundant robot control in task-space
    Perrusquia, Adolfo
    Yu, Wen
    Li, Xiaoou
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2021, 12 (01) : 231 - 241
  • [38] Multi-agent reinforcement learning for redundant robot control in task-space
    Adolfo Perrusquía
    Wen Yu
    Xiaoou Li
    International Journal of Machine Learning and Cybernetics, 2021, 12 : 231 - 241
  • [39] Decentralized Multi-agent Formation Control via Deep Reinforcement Learning
    Gutpa, Aniket
    Nallanthighal, Raghava
    ICAART: PROCEEDINGS OF THE 13TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE - VOL 1, 2021, : 289 - 295
  • [40] Multi-agent reinforcement learning for character control
    Li, Cheng
    Fussell, Levi
    Komura, Taku
    VISUAL COMPUTER, 2021, 37 (12): : 3115 - 3123