Joint Trajectory and Communication Optimization for Heterogeneous Vehicles in Maritime SAR: Multi-Agent Reinforcement Learning

被引:0
|
作者
Lei, Chengjia [1 ,2 ]
Wu, Shaohua [2 ,3 ]
Yang, Yi [2 ]
Xue, Jiayin [2 ]
Zhang, Qinyu [2 ,3 ]
机构
[1] Harbin Inst Technol, Dept Elect & Informat Engn, Shenzhen 518055, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518055, Peoples R China
[3] Harbin Inst Technol, Guangdong Prov Key Lab Aerosp Commun & Networking, Shenzhen 518055, Peoples R China
基金
中国国家自然科学基金;
关键词
Maritime search and rescue (SAR); multi-agent reinforcement learning (MARL); efficiency; fault-tolerant communication; unmanned aerial vehicle (UAV); automatic surface vehicle (ASV); SEARCH; LEVEL;
D O I
10.1109/TVT.2024.3388499
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Nowadays, multiple types of equipment, including unmanned aerial vehicles (UAVs) and automatic surface vehicles (ASVs), have been deployed in maritime search and rescue (SAR). However, due to the lack of base stations (BSs), how to complete rescue while maintaining communication between vehicles is an unresolved challenge. In this paper, we design an efficient and fault-tolerant communication solution by jointly optimizing vehicles' trajectory, offloading scheduling, and routing topology for a heterogeneous vehicle system. First, we model several essential factors in maritime SAR, including the impact of ocean currents, the observational behavior of UAVs, the fault tolerance of relay networks, resource management of mobile edge computing (MEC), and energy consumption. A multi-objective optimization problem is formulated, aiming at minimizing time and energy consumption while increasing the fault tolerance of relay networks. Then, we transfer the objective into a decentralized partially observable Markov Decision Process (Dec-POMDP) and introduce multi-agent reinforcement learning (MARL) to search for a collaborative strategy. Specifically, two MARL approaches with different training styles are evaluated, and three techniques are added for improving performance, including sharing parameters, normalized generalized-advantage-estimation (GAE), and preserving-outputs-precisely-while-adaptively-rescaling-targets (Pop-Art). Experimental results demonstrate that our proposed approach, named heterogeneous vehicles multi-agent proximal policy optimization (HVMAPPO), outperforms other baselines in efficiency and fault tolerance of communication.
引用
收藏
页码:12328 / 12344
页数:17
相关论文
共 50 条
  • [31] Dynamic Multi-Agent Reinforcement Learning for Control Optimization
    Fagan, Derek
    Meier, Rene
    PROCEEDINGS FIFTH INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS, MODELLING AND SIMULATION, 2014, : 99 - 104
  • [32] Learning Distinct Strategies for Heterogeneous Cooperative Multi-agent Reinforcement Learning
    Wan, Kejia
    Xu, Xinhai
    Li, Yuan
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT IV, 2021, 12894 : 544 - 555
  • [33] Reinforcement learning of coordination in heterogeneous cooperative multi-agent systems
    Kapetanakis, S
    Kudenko, D
    ADAPTIVE AGENTS AND MULTI-AGENT SYSTEMS II: ADAPTATION AND MULTI-AGENT LEARNING, 2005, 3394 : 119 - 131
  • [34] Heterogeneous Observation Aggregation Network for Multi-agent Reinforcement Learning
    Hu, Tianyi
    Ai, Xiaolin
    Pu, Zhiqiang
    Qiu, Tenghai
    Yi, Jianqiang
    2024 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN 2024, 2024,
  • [35] Effective Communications: A Joint Learning and Communication Framework for Multi-Agent Reinforcement Learning Over Noisy Channels
    Tung, Tze-Yang
    Kobus, Szymon
    Roig, Joan Pujol
    Gunduz, Deniz
    IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2021, 39 (08) : 2590 - 2603
  • [36] Efficient multi-agent cooperation: Scalable reinforcement learning with heterogeneous graph networks and limited communication
    Li, Z.
    Yang, Y.
    Cheng, H.
    KNOWLEDGE-BASED SYSTEMS, 2024, 300
  • [37] Privacy-Preserving Joint Edge Association and Power Optimization for the Internet of Vehicles via Federated Multi-Agent Reinforcement Learning
    Lin, Yan
    Bao, Jinming
    Zhang, Yijin
    Li, Jun
    Shu, Feng
    Hanzo, Lajos
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (06) : 8256 - 8261
  • [38] Hierarchical Optimization Scheduling Algorithm for Logistics Transport Vehicles Based on Multi-Agent Reinforcement Learning
    Zhang, Min
    Pan, Chaohong
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (03) : 3108 - 3117
  • [39] Learning multi-agent communication with double attentional deep reinforcement learning
    Mao, Hangyu
    Zhang, Zhengchao
    Xiao, Zhen
    Gong, Zhibo
    Ni, Yan
    AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2020, 34 (01)
  • [40] Learning multi-agent communication with double attentional deep reinforcement learning
    Hangyu Mao
    Zhengchao Zhang
    Zhen Xiao
    Zhibo Gong
    Yan Ni
    Autonomous Agents and Multi-Agent Systems, 2020, 34