Joint Trajectory and Communication Optimization for Heterogeneous Vehicles in Maritime SAR: Multi-Agent Reinforcement Learning

被引:0
|
作者
Lei, Chengjia [1 ,2 ]
Wu, Shaohua [2 ,3 ]
Yang, Yi [2 ]
Xue, Jiayin [2 ]
Zhang, Qinyu [2 ,3 ]
机构
[1] Harbin Inst Technol, Dept Elect & Informat Engn, Shenzhen 518055, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518055, Peoples R China
[3] Harbin Inst Technol, Guangdong Prov Key Lab Aerosp Commun & Networking, Shenzhen 518055, Peoples R China
基金
中国国家自然科学基金;
关键词
Maritime search and rescue (SAR); multi-agent reinforcement learning (MARL); efficiency; fault-tolerant communication; unmanned aerial vehicle (UAV); automatic surface vehicle (ASV); SEARCH; LEVEL;
D O I
10.1109/TVT.2024.3388499
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Nowadays, multiple types of equipment, including unmanned aerial vehicles (UAVs) and automatic surface vehicles (ASVs), have been deployed in maritime search and rescue (SAR). However, due to the lack of base stations (BSs), how to complete rescue while maintaining communication between vehicles is an unresolved challenge. In this paper, we design an efficient and fault-tolerant communication solution by jointly optimizing vehicles' trajectory, offloading scheduling, and routing topology for a heterogeneous vehicle system. First, we model several essential factors in maritime SAR, including the impact of ocean currents, the observational behavior of UAVs, the fault tolerance of relay networks, resource management of mobile edge computing (MEC), and energy consumption. A multi-objective optimization problem is formulated, aiming at minimizing time and energy consumption while increasing the fault tolerance of relay networks. Then, we transfer the objective into a decentralized partially observable Markov Decision Process (Dec-POMDP) and introduce multi-agent reinforcement learning (MARL) to search for a collaborative strategy. Specifically, two MARL approaches with different training styles are evaluated, and three techniques are added for improving performance, including sharing parameters, normalized generalized-advantage-estimation (GAE), and preserving-outputs-precisely-while-adaptively-rescaling-targets (Pop-Art). Experimental results demonstrate that our proposed approach, named heterogeneous vehicles multi-agent proximal policy optimization (HVMAPPO), outperforms other baselines in efficiency and fault tolerance of communication.
引用
收藏
页码:12328 / 12344
页数:17
相关论文
共 50 条
  • [1] Joint UAV trajectory and communication design with heterogeneous multi-agent reinforcement learning
    Xuanhan ZHOU
    Jun XIONG
    Haitao ZHAO
    Xiaoran LIU
    Baoquan REN
    Xiaochen ZHANG
    Jibo WEI
    Hao YIN
    ScienceChina(InformationSciences), 2024, 67 (03) : 225 - 245
  • [2] Joint UAV trajectory and communication design with heterogeneous multi-agent reinforcement learning
    Zhou, Xuanhan
    Xiong, Jun
    Zhao, Haitao
    Liu, Xiaoran
    Ren, Baoquan
    Zhang, Xiaochen
    Wei, Jibo
    Yin, Hao
    SCIENCE CHINA-INFORMATION SCIENCES, 2024, 67 (03)
  • [3] Joint Trajectory and Radio Resource Optimization for Autonomous Mobile Robots Exploiting Multi-Agent Reinforcement Learning
    Luo, Ruyu
    Ni, Wanli
    Tian, Hui
    Cheng, Julian
    Chen, Kwang-Cheng
    IEEE TRANSACTIONS ON COMMUNICATIONS, 2023, 71 (09) : 5244 - 5258
  • [4] Autonomous Intersection Management with Heterogeneous Vehicles: A Multi-Agent Reinforcement Learning Approach
    Chen, Kaixin
    Li, Bing
    Zhang, Rongqing
    Cheng, Xiang
    2024 35TH IEEE INTELLIGENT VEHICLES SYMPOSIUM, IEEE IV 2024, 2024, : 2255 - 2260
  • [5] Multi-agent reinforcement learning for electric vehicles joint routing and scheduling strategies
    Wang, Yi
    Qiu, Dawei
    Strbac, Goran
    2022 IEEE 25TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2022, : 3044 - 3049
  • [6] Multi-Agent Reinforcement Learning With Distributed Targeted Multi-Agent Communication
    Xu, Chi
    Zhang, Hui
    Zhang, Ya
    2023 35TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2023, : 2915 - 2920
  • [7] Learning structured communication for multi-agent reinforcement learning
    Sheng, Junjie
    Wang, Xiangfeng
    Jin, Bo
    Yan, Junchi
    Li, Wenhao
    Chang, Tsung-Hui
    Wang, Jun
    Zha, Hongyuan
    AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2022, 36 (02)
  • [8] Learning structured communication for multi-agent reinforcement learning
    Junjie Sheng
    Xiangfeng Wang
    Bo Jin
    Junchi Yan
    Wenhao Li
    Tsung-Hui Chang
    Jun Wang
    Hongyuan Zha
    Autonomous Agents and Multi-Agent Systems, 2022, 36
  • [9] Learning to Schedule Joint Radar-Communication With Deep Multi-Agent Reinforcement Learning
    Lee, Joash
    Niyato, Dusit
    Guan, Yong Liang
    Kim, Dong In
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2022, 71 (01) : 406 - 422
  • [10] Multi-Agent Reinforcement Learning for Autonomous On Demand Vehicles
    Boyali, Ali
    Hashimoto, Naohisa
    John, Vijay
    Acarman, Tankut
    2019 30TH IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV19), 2019, : 1461 - 1468