Constrained Coverage of Unknown Environment Using Safe Reinforcement Learning

被引:1
|
作者
Zhang, Yunlin [1 ]
You, Junjie [1 ]
Shi, Lei [2 ]
Shao, Jinliang [1 ,3 ]
Zheng, Wei Xing [4 ]
机构
[1] Univ Elect Sci & Technol China, Sch Automat Engn, Chengdu 611731, Peoples R China
[2] Henan Univ, Sch Artificial Intelligence, Zhengzhou 450046, Peoples R China
[3] Lab Electromagnet Space Cognit & Intelligent Cont, Beijing 100089, Peoples R China
[4] Western Sydney Univ, Sch Comp Data & Math Sci, Sydney, NSW 2751, Australia
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CDC49753.2023.10383702
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Achieving a connected, collision-free and timeefficient coverage in unknown environments is challenging for multi-agent systems. Particularly, agents with second-order dynamics are supposed to efficiently search and reach the optimal deployment positions over targets whose distribution is unknown, while preserving the distributed connectivity and avoiding collision. In this paper, a safe reinforcement learning based shield method is proposed for unknown environment exploration while correcting actions of agents for safety guarantee and avoiding invalid samples into policy updating. The shield is achieved distributively by a control barrier function and its validity is proved in theory. Moreover, policies of the optimal coverage are centrally learned via reward engineering and executed distributively. Numerical results show that the proposed approach not only achieves zero safety violations during training, but also speeds up the convergence of learning.
引用
收藏
页码:3415 / 3420
页数:6
相关论文
共 50 条
  • [41] Achieving Safe Deep Reinforcement Learning via Environment Comprehension Mechanism
    Peng Pai
    Zhu Fei
    Liu Quan
    Zhao Peiyao
    Wu Wen
    CHINESE JOURNAL OF ELECTRONICS, 2021, 30 (06) : 1049 - 1058
  • [42] Safe Reinforcement Learning in Simulated Environment of Self-Driving Laboratory
    Chernov, Andrey V.
    Savvas, Ilias K.
    Butakova, Maria A.
    Kartashov, Oleg O.
    ESSE 2021: THE 2ND EUROPEAN SYMPOSIUM ON SOFTWARE ENGINEERING, 2021, : 78 - 84
  • [43] Achieving Safe Deep Reinforcement Learning via Environment Comprehension Mechanism
    PENG Pai
    ZHU Fei
    LIU Quan
    ZHAO Peiyao
    WU Wen
    ChineseJournalofElectronics, 2021, 30 (06) : 1049 - 1058
  • [44] Constrained attractor selection using deep reinforcement learning
    Wang, Xue-She
    Turner, James D.
    Mann, Brian P.
    JOURNAL OF VIBRATION AND CONTROL, 2021, 27 (5-6) : 502 - 514
  • [45] On Normative Reinforcement Learning via Safe Reinforcement Learning
    Neufeld, Emery A.
    Bartocci, Ezio
    Ciabattoni, Agata
    PRIMA 2022: PRINCIPLES AND PRACTICE OF MULTI-AGENT SYSTEMS, 2023, 13753 : 72 - 89
  • [46] Safe reinforcement learning using risk mapping by similarity
    Serrano-Cuevas, Jonathan
    Morales, Eduardo F.
    Hernandez-Leal, Pablo
    ADAPTIVE BEHAVIOR, 2020, 28 (04) : 213 - 224
  • [47] Batch reinforcement learning for network-safe demand response in unknown electric grids
    Lesage-Landry, Antoine
    Callaway, Duncan S.
    ELECTRIC POWER SYSTEMS RESEARCH, 2022, 212
  • [48] Deep Reinforcement Learning for Safe Local Planning of a Ground Vehicle in Unknown Rough Terrain
    Josef, Shirel
    Degani, Amir
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (04) : 6748 - 6755
  • [49] Reinforcement learning-based optimal control of unknown constrained-input nonlinear systems using simulated experience
    Asl, Hamed Jabbari
    Uchibe, Eiji
    NONLINEAR DYNAMICS, 2023, 111 (17) : 16093 - 16110
  • [50] Optimal tracking control of nonlinear partially-unknown constrained-input systems using integral reinforcement learning
    Modares, Hamidreza
    Lewis, Frank L.
    AUTOMATICA, 2014, 50 (07) : 1780 - 1792