VMAS: A Vectorized Multi-agent Simulator for Collective Robot Learning

被引:0
|
作者
Bettini, Matteo [1 ]
Kortvelesy, Ryan [1 ]
Blumenkamp, Jan [1 ]
Prorok, Amanda [1 ]
机构
[1] Univ Cambridge, Dept Comp Sci & Technol, Cambridge, England
基金
英国工程与自然科学研究理事会; 欧洲研究理事会;
关键词
simulator; multi-robot learning; vectorization;
D O I
10.1007/978-3-031-51497-5_4
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
While many multi-robot coordination problems can be solved optimally by exact algorithms, solutions are often not scalable in the number of robots. Multi-Agent Reinforcement Learning (MARL) is gaining increasing attention in the robotics community as a promising solution to tackle such problems. Nevertheless, we still lack the tools that allow us to quickly and efficiently find solutions to largescale collective learning tasks. In this work, we introduce the Vectorized Multi-Agent Simulator (VMAS). VMAS is an open-source framework designed for efficient MARL benchmarking. It is comprised of a vectorized 2D physics engine written in PyTorch and a set of twelve challenging multi-robot scenarios. Additional scenarios can be implemented through a simple and modular interface. We demonstrate how vectorization enables parallel simulation on accelerated hardware without added complexity. When comparing VMAS to OpenAI MPE, we show how MPE's execution time increases linearly in the number of simulations while VMAS is able to execute 30,000 parallel simulations in under 10 s, proving more than 100x faster. Using VMAS's RLlib interface, we benchmark our multi-robot scenarios using various Proximal Policy Optimization (PPO)-based MARL algorithms. VMAS's scenarios prove challenging in orthogonal ways for state-of-the-art MARL algorithms. The VMAS framework is available at: https://github.com/proroklab/ VectorizedMultiAgentSimulator. A video of VMAS scenarios and experiments is available https://youtu.be/aaDRYfiesAY
引用
收藏
页码:42 / 56
页数:15
相关论文
共 50 条
  • [41] Collective Intrinsic Motivation of a Multi-agent System Based on Reinforcement Learning Algorithms
    Bolshakov, Vladislav
    Sakulin, Sergey
    Alfimtsev, Alexander
    INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 4, INTELLISYS 2023, 2024, 825 : 655 - 670
  • [42] Modeling collective motion for fish schooling via multi-agent reinforcement learning
    Wang, Xin
    Liu, Shuo
    Yu, Yifan
    Yue, Shengzhi
    Liu, Ying
    Zhang, Fumin
    Lin, Yuanshan
    ECOLOGICAL MODELLING, 2023, 477
  • [43] Multi-agent systems and role games: collective learning processes for ecosystem management
    Bousquet, F
    Barreteau, O
    d'Aquino, P
    Etienne, M
    Boissau, S
    Aubert, S
    Le Page, C
    Babin, D
    Castella, JC
    COMPLEXITY AND ECOSYSTEM MANAGEMENT: THE THEORY AND PRACTICE OF MULTI-AGENT SYSTEMS, 2002, : 248 - 285
  • [44] MULTI-AGENT COORDINATION AND COLLECTIVE ARTIFICIAL PAINTINGS
    Urbano, Paulo
    ICAART 2009: PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE, 2009, : IS19 - IS26
  • [45] Collective surrounding control in multi-agent networks
    魏婷婷
    陈小平
    Chinese Physics B, 2014, (05) : 36 - 39
  • [46] Collective trust estimation in multi-agent systems
    Balas, Cristian
    Karlsen, Robert
    Muench, Paul
    Mikulski, Dariusz
    Mohammed, Utayba
    Al-Holou, Nizar
    UNMANNED SYSTEMS TECHNOLOGY XXI, 2019, 11021
  • [47] A Collective Behaviour Framework for Multi-agent Systems
    Guzel, Mehmet Serdar
    Kayakoku, Hakan
    MECHATRONICS AND ROBOTICS ENGINEERING FOR ADVANCED AND INTELLIGENT MANUFACTURING, 2017, : 61 - 71
  • [48] Collective Decision Making in Multi-Agent Systems
    Aziz, Haris
    IEEE INTELLIGENT SYSTEMS, 2016, 31 (01) : 57 - 57
  • [49] Collective surrounding control in multi-agent networks
    Wei Ting-Ting
    Chen Xiao-Ping
    CHINESE PHYSICS B, 2014, 23 (05)
  • [50] Multi-Agent Cognition Difference Reinforcement Learning for Multi-Agent Cooperation
    Wang, Huimu
    Qiu, Tenghai
    Liu, Zhen
    Pu, Zhiqiang
    Yi, Jianqiang
    Yuan, Wanmai
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,