Joint Trajectory and Communication Optimization for Heterogeneous Vehicles in Maritime SAR: Multi-Agent Reinforcement Learning

被引:0
|
作者
Lei, Chengjia [1 ,2 ]
Wu, Shaohua [2 ,3 ]
Yang, Yi [2 ]
Xue, Jiayin [2 ]
Zhang, Qinyu [2 ,3 ]
机构
[1] Harbin Inst Technol, Dept Elect & Informat Engn, Shenzhen 518055, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518055, Peoples R China
[3] Harbin Inst Technol, Guangdong Prov Key Lab Aerosp Commun & Networking, Shenzhen 518055, Peoples R China
基金
中国国家自然科学基金;
关键词
Maritime search and rescue (SAR); multi-agent reinforcement learning (MARL); efficiency; fault-tolerant communication; unmanned aerial vehicle (UAV); automatic surface vehicle (ASV); SEARCH; LEVEL;
D O I
10.1109/TVT.2024.3388499
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Nowadays, multiple types of equipment, including unmanned aerial vehicles (UAVs) and automatic surface vehicles (ASVs), have been deployed in maritime search and rescue (SAR). However, due to the lack of base stations (BSs), how to complete rescue while maintaining communication between vehicles is an unresolved challenge. In this paper, we design an efficient and fault-tolerant communication solution by jointly optimizing vehicles' trajectory, offloading scheduling, and routing topology for a heterogeneous vehicle system. First, we model several essential factors in maritime SAR, including the impact of ocean currents, the observational behavior of UAVs, the fault tolerance of relay networks, resource management of mobile edge computing (MEC), and energy consumption. A multi-objective optimization problem is formulated, aiming at minimizing time and energy consumption while increasing the fault tolerance of relay networks. Then, we transfer the objective into a decentralized partially observable Markov Decision Process (Dec-POMDP) and introduce multi-agent reinforcement learning (MARL) to search for a collaborative strategy. Specifically, two MARL approaches with different training styles are evaluated, and three techniques are added for improving performance, including sharing parameters, normalized generalized-advantage-estimation (GAE), and preserving-outputs-precisely-while-adaptively-rescaling-targets (Pop-Art). Experimental results demonstrate that our proposed approach, named heterogeneous vehicles multi-agent proximal policy optimization (HVMAPPO), outperforms other baselines in efficiency and fault tolerance of communication.
引用
收藏
页码:12328 / 12344
页数:17
相关论文
共 50 条
  • [41] Multi-Agent Reinforcement Learning
    Stankovic, Milos
    2016 13TH SYMPOSIUM ON NEURAL NETWORKS AND APPLICATIONS (NEUREL), 2016, : 43 - 43
  • [42] Cooperative Behavior by Multi-agent Reinforcement Learning with Abstractive Communication
    Tanda, Jin
    Moustafa, Ahmed
    Ito, Takayuki
    2019 IEEE INTERNATIONAL CONFERENCE ON AGENTS (ICA), 2019, : 8 - 13
  • [43] Diffusion-based Multi-agent Reinforcement Learning with Communication
    Qi, Xinyue
    Tang, Jianhang
    Jin, Jiangming
    Zhang, Yang
    2024 IEEE VTS ASIA PACIFIC WIRELESS COMMUNICATIONS SYMPOSIUM, APWCS 2024, 2024,
  • [44] Trajectory planning of space manipulator based on multi-agent reinforcement learning
    Zhao Y.
    Guan G.
    Guo J.
    Yu X.
    Yan P.
    Hangkong Xuebao/Acta Aeronautica et Astronautica Sinica, 2021, 42 (01):
  • [45] Semantic Communication for Partial Observation Multi-agent Reinforcement Learning
    Do, Hoang Khoi
    Dinh, Thi Quynh
    Nguyen, Minh Duong
    Nguyen, Tien Hoa
    2023 IEEE STATISTICAL SIGNAL PROCESSING WORKSHOP, SSP, 2023, : 319 - 323
  • [46] Multi-agent Pathfinding with Communication Reinforcement Learning and Deadlock Detection
    Ye, Zhaohui
    Li, Yanjie
    Guo, Ronghao
    Gao, Jianqi
    Fu, Wen
    INTELLIGENT ROBOTICS AND APPLICATIONS (ICIRA 2022), PT I, 2022, 13455 : 493 - 504
  • [47] Optimistic sequential multi-agent reinforcement learning with motivational communication
    Huang, Anqi
    Wang, Yongli
    Zhou, Xiaoliang
    Zou, Haochen
    Dong, Xu
    Che, Xun
    NEURAL NETWORKS, 2024, 179
  • [48] Cooperative Multi-agent Reinforcement Learning with Hierachical Communication Architecture
    Liu, Shifan
    Yuan, Quan
    Chen, Bo
    Luo, Guiyang
    Li, Jinglin
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2022, PT II, 2022, 13530 : 14 - 25
  • [49] Deep Hierarchical Communication Graph in Multi-Agent Reinforcement Learning
    Liu, Zeyang
    Wan, Lipeng
    Sui, Xue
    Chen, Zhuoran
    Sun, Kewu
    Lan, Xuguang
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 208 - 216
  • [50] Communication-Efficient and Federated Multi-Agent Reinforcement Learning
    Krouka, Mounssif
    Elgabli, Anis
    Ben Issaid, Chaouki
    Bennis, Mehdi
    IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2022, 8 (01) : 311 - 320