A Heterogeneous Acceleration System for Attention-Based Multi-Agent Reinforcement Learning

被引:0
|
作者
Wiggins, Samuel [1 ]
Meng, Yuan [1 ]
Iyer, Mahesh A. [2 ]
Prasanna, Viktor [1 ]
机构
[1] Univ Southern Calif, Ming Hsieh Dept Elect & Comp Engn, Los Angeles, CA 90007 USA
[2] Intel Corp, Santa Clara, CA USA
基金
美国国家科学基金会;
关键词
Multi-Agent Reinforcement Learning; Hardware Accelerator; Heterogeneous Computing;
D O I
10.1109/FPL64840.2024.00040
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-Agent Reinforcement Learning (MARL) is an emerging technology that has seen success in many AI applications. Multi-Actor-Attention-Critic (MAAC) is a state-of-the-art MARL algorithm that uses a Multi-Head Attention (MHA) mechanism to learn messages communicated among agents during the training process. Current implementations of MAAC using CPU and CPU-GPU platforms lack fine-grained parallelism among agents, sequentially executing each stage of the training loop, and their performance suffers from costly data movement involved in MHA communication learning. In this work, we develop the first high-throughput accelerator for MARL with attention-based communication on a CPU-FPGA heterogeneous system. We alleviate the limitations of existing implementations through a combination of data- and pipeline-parallel modules in our accelerator design and enable fine-grained system scheduling for exploiting concurrency among heterogeneous resources. Our design increases the overall system throughput by 4.6x and 4.1x compared to CPU and CPU-GPU implementations, respectively.
引用
收藏
页码:236 / 242
页数:7
相关论文
共 50 条
  • [41] Deep Reinforcement Learning for Multi-Agent Power Control in Heterogeneous Networks
    Zhang, Lin
    Liang, Ying-Chang
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2021, 20 (04) : 2551 - 2564
  • [42] Distributed, Heterogeneous, Multi-Agent Social Coordination via Reinforcement Learning
    Shi, Dongqing
    Sauter, Michael Z.
    Kralik, Jerald D.
    2009 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (ROBIO 2009), VOLS 1-4, 2009, : 653 - 658
  • [43] Heterogeneous Multi-Agent Reinforcement Learning for Grid-Interactive Communities
    Wu, Allen
    Nweye, Kingsley
    Nagy, Zoltan
    PROCEEDINGS OF THE 10TH ACM INTERNATIONAL CONFERENCE ON SYSTEMS FOR ENERGY-EFFICIENT BUILDINGS, CITIES, AND TRANSPORTATION, BUILDSYS 2023, 2023, : 314 - 315
  • [44] Scalable Autonomous Separation Assurance With Heterogeneous Multi-Agent Reinforcement Learning
    Brittain, Marc
    Wei, Peng
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2022, 19 (04) : 2837 - 2848
  • [45] Multi-agent cooperative learning research based on reinforcement learning
    Liu, Fei
    Zeng, Guangzhou
    2006 10TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, PROCEEDINGS, VOLS 1 AND 2, 2006, : 1408 - 1413
  • [46] Multi-Agent Based Control of a Heterogeneous System
    Li, Howard
    Karray, Fakhreddine
    Basir, Otman
    Song, Insop
    JOURNAL OF ADVANCED COMPUTATIONAL INTELLIGENCE AND INTELLIGENT INFORMATICS, 2006, 10 (02) : 161 - 167
  • [47] Multi-agent reinforcement learning based on local communication
    Zhang, Wenxu
    Ma, Lei
    Li, Xiaonan
    CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2019, 22 (Suppl 6): : 15357 - 15366
  • [48] Multi-agent Cooperative Search based on Reinforcement Learning
    Sun, Yinjiang
    Zhang, Rui
    Liang, Wenbao
    Xu, Cheng
    PROCEEDINGS OF 2020 3RD INTERNATIONAL CONFERENCE ON UNMANNED SYSTEMS (ICUS), 2020, : 891 - 896
  • [49] Hierarchical Multi-Agent Training Based on Reinforcement Learning
    Wang, Guanghua
    Li, Wenjie
    Wu, Zhanghua
    Guo, Xian
    2024 9TH ASIA-PACIFIC CONFERENCE ON INTELLIGENT ROBOT SYSTEMS, ACIRS, 2024, : 11 - 18
  • [50] Function approximation based multi-agent reinforcement learning
    Abul, O
    Polat, F
    Alhajj, R
    12TH IEEE INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, PROCEEDINGS, 2000, : 36 - 39