Risk-Sensitive Mobile Robot Navigation in Crowded Environment via Offline Reinforcement Learning

被引:0
|
作者
Wu, Jiaxu [1 ]
Wang, Yusheng [1 ]
Asama, Hajime [1 ]
An, Qi [2 ]
Yamashita, Atsushi [2 ]
机构
[1] Univ Tokyo, Grad Sch Engn, Dept Precis Engn, 7-3-1 Hongo,Bunkyo Ku, Tokyo 1138656, Japan
[2] Univ Tokyo, Grad Sch Frontier Sci, Dept Human & Engn Environm Studies, 5-1-5 Kashiwanoha, Kashiwa, Chiba 2778563, Japan
关键词
COLLISION-AVOIDANCE;
D O I
10.1109/IROS55552.2023.10341948
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Mobile robot navigation in a human-populated environment has been of great interest to the research community in recent years, referred to as crowd navigation. Currently, offline reinforcement learning (RL)-based method has been introduced to this domain, for its ability to alleviate the sim2real gap brought by online RL which relies on simulators to execute training, and its scalability to use the same dataset to train for differently customized rewards. However, the performance of the navigation policy suffered from the distributional shift between the training data and the input during deployment, since when it gets an input out of the training data distribution, the learned policy has the risk of choosing an erroneous action that leads to catastrophic failure such as colliding with a human. To realize risk sensitivity and improve the safety of the offline RL agent during deployment, this work proposes a multipolicy control framework that combines offline RL navigation policy with a risk detector and a force-based risk-avoiding policy. In particular, a Lyapunov density model is learned using the latent feature of the offline RL policy and works as a risk detector to switch the control to the risk-avoiding policy when the robot has a tendency to go out of the area supported by the training data. Experimental results showed that the proposed method was able to learn navigation in a crowded scene from the offline trajectory dataset and the risk detector substantially reduces the collision rate of the vanilla offline RL agent while maintaining the navigation efficiency outperforming the state-of-the-art methods.
引用
收藏
页码:7456 / 7462
页数:7
相关论文
共 50 条
  • [21] Socially Compliant Robot Navigation in Crowded Environment by Human Behavior Resemblance Using Deep Reinforcement Learning
    Samsani, Sunil Srivatsav
    Muhammad, Mannan Saeed
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (03) : 5223 - 5230
  • [22] Sensor-based Mobile Robot Navigation via Deep Reinforcement Learning
    Han, Seungho-Ho
    Choi, Ho-Jin
    Benz, Philipp
    Loaiciga, Jorge
    2018 IEEE INTERNATIONAL CONFERENCE ON BIG DATA AND SMART COMPUTING (BIGCOMP), 2018, : 147 - 154
  • [23] Risk-sensitive Distributional Reinforcement Learning for Flight Control
    Seres, Peter
    Liu, Cheng
    van Kampen, Erik-Jan
    IFAC PAPERSONLINE, 2023, 56 (02): : 2013 - 2018
  • [24] Risk-Sensitive Inhibitory Control for Safe Reinforcement Learning
    Lederer, Armin
    Noorani, Erfaun
    Baras, John S.
    Hirche, Sandra
    2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 1040 - 1045
  • [25] Distributional Model Equivalence for Risk-Sensitive Reinforcement Learning
    Kastner, Tyler
    Erdogdu, Murat A.
    Farahmand, Amir-massoud
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [26] Intelligent mobile robot navigation in unknown and complex environment using reinforcement learning technique
    Ravi Raj
    Andrzej Kos
    Scientific Reports, 14 (1)
  • [27] Fuzzy logic and reinforcement learning based approaches for mobile robot navigation in unknown environment
    Cherroun, Lakhmissi
    Boumehraz, Mohamed
    Mediterranean Journal of Measurement and Control, 2013, 9 (03): : 109 - 117
  • [28] Reinforcement Extreme Learning Machine for Mobile Robot Navigation
    Geng, Hongjie
    Liu, Huaping
    Wang, Bowen
    Sun, Fuchun
    PROCEEDINGS OF ELM-2016, 2018, 9 : 61 - 73
  • [29] Reinforcement learning based on backpropagation for mobile robot navigation
    Jaksa, R
    Majerník, P
    Sincák, P
    COMPUTATIONAL INTELLIGENCE FOR MODELLING, CONTROL & AUTOMATION - NEURAL NETWORKS & ADVANCED CONTROL STRATEGIES, 1999, 54 : 46 - 51
  • [30] Mobile Robot Navigation Using Deep Reinforcement Learning
    Lee, Min-Fan Ricky
    Yusuf, Sharfiden Hassen
    PROCESSES, 2022, 10 (12)