Constrained Coverage of Unknown Environment Using Safe Reinforcement Learning

被引:1
|
作者
Zhang, Yunlin [1 ]
You, Junjie [1 ]
Shi, Lei [2 ]
Shao, Jinliang [1 ,3 ]
Zheng, Wei Xing [4 ]
机构
[1] Univ Elect Sci & Technol China, Sch Automat Engn, Chengdu 611731, Peoples R China
[2] Henan Univ, Sch Artificial Intelligence, Zhengzhou 450046, Peoples R China
[3] Lab Electromagnet Space Cognit & Intelligent Cont, Beijing 100089, Peoples R China
[4] Western Sydney Univ, Sch Comp Data & Math Sci, Sydney, NSW 2751, Australia
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CDC49753.2023.10383702
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Achieving a connected, collision-free and timeefficient coverage in unknown environments is challenging for multi-agent systems. Particularly, agents with second-order dynamics are supposed to efficiently search and reach the optimal deployment positions over targets whose distribution is unknown, while preserving the distributed connectivity and avoiding collision. In this paper, a safe reinforcement learning based shield method is proposed for unknown environment exploration while correcting actions of agents for safety guarantee and avoiding invalid samples into policy updating. The shield is achieved distributively by a control barrier function and its validity is proved in theory. Moreover, policies of the optimal coverage are centrally learned via reward engineering and executed distributively. Numerical results show that the proposed approach not only achieves zero safety violations during training, but also speeds up the convergence of learning.
引用
收藏
页码:3415 / 3420
页数:6
相关论文
共 50 条
  • [31] Safe Reinforcement Learning for Autonomous Vehicles through Parallel Constrained Policy Optimization
    Wen, Lu
    Duan, Jingliang
    Li, Shengbo Eben
    Xu, Shaobing
    Peng, Huei
    2020 IEEE 23RD INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2020,
  • [32] Safe Reinforcement Learning for Continuous Spaces through Lyapunov-Constrained Behavior
    Fjerdingen, Sigurd A.
    Kyrkjebo, Erik
    ELEVENTH SCANDINAVIAN CONFERENCE ON ARTIFICIAL INTELLIGENCE (SCAI 2011), 2011, 227 : 70 - 79
  • [33] Dynamic Path Planning of Unknown Environment Based on Deep Reinforcement Learning
    Lei, Xiaoyun
    Zhang, Zhian
    Dong, Peifang
    JOURNAL OF ROBOTICS, 2018, 2018
  • [34] Crowd navigation in an unknown and complex environment based on deep reinforcement learning
    Sun, Libo
    Qu, Yuke
    Qin, Wenhu
    COMPUTER ANIMATION AND VIRTUAL WORLDS, 2022, 33 (3-4)
  • [35] Crowd Navigation in an Unknown and Dynamic Environment Based on Deep Reinforcement Learning
    Sun, Libo
    Zhai, Jinfeng
    Qin, Wenhu
    IEEE ACCESS, 2019, 7 : 109544 - 109554
  • [36] Multi-objective Reinforcement Learning Algorithm for MOSDMP in Unknown Environment
    Zhao, Yun
    Chen, Qingwei
    Hu, Weili
    2010 8TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION (WCICA), 2010, : 3190 - 3194
  • [37] Exploration of Unknown Environments Using Deep Reinforcement Learning
    McCalmon, Joseph
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 15970 - 15971
  • [38] Disaster Area Coverage Optimisation using Reinforcement Learning
    Gruffeille, Ciaran
    Perrusquia, Adolfo
    Tsourdos, Antonios
    Guo, Weisi
    2024 INTERNATIONAL CONFERENCE ON UNMANNED AIRCRAFT SYSTEMS, ICUAS, 2024, : 61 - 67
  • [39] Generalized Multiagent Reinforcement Learning for Coverage Path Planning in Unknown, Dynamic, and Hazardous Environments
    Henshall, Andrea M.
    Karaman, Sertac
    AIAA SCITECH 2024 FORUM, 2024,
  • [40] Path Planning for a Mobile Robot in Unknown Dynamic Environments Using Integrated Environment Representation and Reinforcement Learning
    Zhang, Jian
    2019 AUSTRALIAN & NEW ZEALAND CONTROL CONFERENCE (ANZCC), 2019, : 258 - 263