A Heterogeneous Acceleration System for Attention-Based Multi-Agent Reinforcement Learning

被引:0
|
作者
Wiggins, Samuel [1 ]
Meng, Yuan [1 ]
Iyer, Mahesh A. [2 ]
Prasanna, Viktor [1 ]
机构
[1] Univ Southern Calif, Ming Hsieh Dept Elect & Comp Engn, Los Angeles, CA 90007 USA
[2] Intel Corp, Santa Clara, CA USA
基金
美国国家科学基金会;
关键词
Multi-Agent Reinforcement Learning; Hardware Accelerator; Heterogeneous Computing;
D O I
10.1109/FPL64840.2024.00040
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-Agent Reinforcement Learning (MARL) is an emerging technology that has seen success in many AI applications. Multi-Actor-Attention-Critic (MAAC) is a state-of-the-art MARL algorithm that uses a Multi-Head Attention (MHA) mechanism to learn messages communicated among agents during the training process. Current implementations of MAAC using CPU and CPU-GPU platforms lack fine-grained parallelism among agents, sequentially executing each stage of the training loop, and their performance suffers from costly data movement involved in MHA communication learning. In this work, we develop the first high-throughput accelerator for MARL with attention-based communication on a CPU-FPGA heterogeneous system. We alleviate the limitations of existing implementations through a combination of data- and pipeline-parallel modules in our accelerator design and enable fine-grained system scheduling for exploiting concurrency among heterogeneous resources. Our design increases the overall system throughput by 4.6x and 4.1x compared to CPU and CPU-GPU implementations, respectively.
引用
收藏
页码:236 / 242
页数:7
相关论文
共 50 条
  • [21] A Multi-group Multi-agent System Based on Reinforcement Learning and Flocking
    Gang Wang
    Jian Xiao
    Rui Xue
    Yongting Yuan
    International Journal of Control, Automation and Systems, 2022, 20 : 2364 - 2378
  • [22] A Multi-group Multi-agent System Based on Reinforcement Learning and Flocking
    Wang, Gang
    Xiao, Jian
    Xue, Rui
    Yuan, Yongting
    INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2022, 20 (07) : 2364 - 2378
  • [23] Cooperative Behavior Acquisition in Multi-agent Reinforcement Learning System Using Attention Degree
    Kobayashi, Kunikazu
    Kurano, Tadashi
    Kuremoto, Takashi
    Obayashi, Masanao
    NEURAL INFORMATION PROCESSING, ICONIP 2012, PT III, 2012, 7665 : 537 - 544
  • [24] Robust Optimal Formation Control of Heterogeneous Multi-Agent System via Reinforcement Learning
    Lin, Wei
    Zhao, Wanbing
    Liu, Hao
    IEEE ACCESS, 2020, 8 (08): : 218424 - 218432
  • [25] Regularization-Adapted Anderson Acceleration for multi-agent reinforcement learning
    Wang, Siying
    Chen, Wenyu
    Huang, Liwei
    Zhang, Fan
    Zhao, Zhitong
    Qu, Hong
    KNOWLEDGE-BASED SYSTEMS, 2023, 275
  • [26] Fast conflict resolution based on reinforcement learning in multi-agent system
    Piao, SH
    Hong, BR
    Chu, HT
    CHINESE JOURNAL OF ELECTRONICS, 2004, 13 (01): : 92 - 95
  • [27] Multi-Agent Reinforcement Learning
    Stankovic, Milos
    2016 13TH SYMPOSIUM ON NEURAL NETWORKS AND APPLICATIONS (NEUREL), 2016, : 43 - 43
  • [28] Hierarchical Multi-Agent Deep Reinforcement Learning with an Attention-based Graph Matching Approach for Multi-Domain VNF-FG Embedding
    Slim, Lotfi
    Bannour, Fetia
    IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 2105 - 2110
  • [29] Attention based Reinforcement Learning for Efficient Communication under Constraint in Multi-Agent Systems
    Mei, Jianguo
    Quan, Zhibin
    Yang, Wankou
    Cao, Xianghui
    2023 35TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2023, : 3867 - 3873
  • [30] Multi-agent Deep Reinforcement Learning Based Adaptive User Association in Heterogeneous Networks
    Yi, Weiwen
    Zhang, Xing
    Wang, Wenbo
    Li, Jing
    COMMUNICATIONS AND NETWORKING, CHINACOM 2018, 2019, 262 : 57 - 67