Energy Constrained Multi-Agent Reinforcement Learning for Coverage Path Planning

被引:0
|
作者
Zhao, Chenyang [1 ]
Liu, Juan [2 ]
Yoon, Suk-Un [3 ]
Li, Xinde [1 ,4 ]
Li, Heqing [1 ]
Zhang, Zhentong [1 ,4 ]
机构
[1] Southeast Univ, Nanjing 210096, Peoples R China
[2] Samsung Elect China R&D Ctr, Nanjing 210012, Peoples R China
[3] Samsung Elect, Suwon 16677, Gyeonggi Do, South Korea
[4] Nanjing Ctr Appl Math, Nanjing 211135, Peoples R China
基金
中国国家自然科学基金;
关键词
NAVIGATION;
D O I
10.1109/IROS55552.2023.10341412
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For multi-agent area coverage path planning problem, existing researches regard it as a combination of Traveling Salesman Problem (TSP) and Coverage Path Planning (CPP). However, these approaches have disadvantages of poor observation ability in online phase and high computational cost in offline phase, making it difficult to be applied to energy-constrained Unmanned Aerial Vehicles (UAVs) and adjust strategy dynamically. In this paper, we decompose the task into two sub-problems: multi-agent path planning and sub-region CPP. We model the multi-agent path planning problem as a Collective Markov Decision Process (C-MDP), and design an Energy Constrained Multi-Agent Reinforcement Learning (ECMARL) algorithm based on the centralized training and distributed execution concept. Taking into account energy constraint of UAVs, the UAV propulsion power model is established to measure the energy consumption of UAVs, and load balancing strategy is applied to dynamically allocate target areas for each UAV. If the UAV is under energy-depleted situation, ECMARL can adjust the mission strategy in real time according to environmental information and energy storage conditions of other UAVs. When UAVs reach each sub-region of interest, Back-an-Forth Paths (BFPs) are adopted to solve CPP problem, which can ensure full coverage, optimality and complexity of the sub-problem. Comprehensive theoretical analysis and experiments demonstrate that ECMARL is superior to the traditional offline TSP-CPP strategy in terms of solution quality and computational time, and can effectively deal with the energy-constrained UAVs.
引用
下载
收藏
页码:5590 / 5597
页数:8
相关论文
共 50 条
  • [41] Multi-agent Deep Reinforcement Learning for Zero Energy Communities
    Prasad, Amit
    Dusparic, Ivana
    PROCEEDINGS OF 2019 IEEE PES INNOVATIVE SMART GRID TECHNOLOGIES EUROPE (ISGT-EUROPE), 2019,
  • [42] Multi-agent reinforcement learning in a new transactive energy mechanism
    Mohsenzadeh-Yazdi, Hossein
    Kebriaei, Hamed
    Aminifar, Farrokh
    IET GENERATION TRANSMISSION & DISTRIBUTION, 2024, : 2943 - 2955
  • [43] Multi-agent Deep Reinforcement Learning for Microgrid Energy Scheduling
    Zuo, Zhiqiang
    Li, Zhi
    Wang, Yijing
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 6184 - 6189
  • [44] Trajectory planning of space manipulator based on multi-agent reinforcement learning
    Zhao Y.
    Guan G.
    Guo J.
    Yu X.
    Yan P.
    Hangkong Xuebao/Acta Aeronautica et Astronautica Sinica, 2021, 42 (01):
  • [45] Safe multi-agent motion planning via filtered reinforcement learning
    Vinod, Abraham P.
    Safaoui, Sleiman
    Chakrabarty, Ankush
    Quirynen, Rien
    Yoshikawa, Nobuyuki
    Di Cairano, Stefano
    2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA 2022, 2022, : 7270 - 7276
  • [46] Multi-Agent Dynamic Area Coverage Based on Reinforcement Learning with Connected Agents
    Aydemir, Fatih
    Cetin, Aydin
    COMPUTER SYSTEMS SCIENCE AND ENGINEERING, 2023, 45 (01): : 215 - 230
  • [47] A Distributed Multi-Agent Dynamic Area Coverage Algorithm Based on Reinforcement Learning
    Xiao, Jian
    Wang, Gang
    Zhang, Ying
    Cheng, Lei
    IEEE ACCESS, 2020, 8 : 33511 - 33521
  • [48] MATE: Benchmarking Multi-Agent Reinforcement Learning in Distributed Target Coverage Control
    Pan, Xuehai
    Liu, Mickel
    Zhong, Fangwei
    Yang, Yaodong
    Zhu, Song-Chun
    Wang, Yizhou
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [49] Multi-Agent Cognition Difference Reinforcement Learning for Multi-Agent Cooperation
    Wang, Huimu
    Qiu, Tenghai
    Liu, Zhen
    Pu, Zhiqiang
    Yi, Jianqiang
    Yuan, Wanmai
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [50] Multi-Agent Uncertainty Sharing for Cooperative Multi-Agent Reinforcement Learning
    Chen, Hao
    Yang, Guangkai
    Zhang, Junge
    Yin, Qiyue
    Huang, Kaiqi
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,