RRT-based maximum entropy inverse reinforcement learning for robust and efficient driving behavior prediction

被引:1
|
作者
Hosoma, Shinpei [1 ]
Sugasaki, Masato [1 ]
Arie, Hiroaki [2 ]
Shimosaka, Masamichi [1 ]
机构
[1] Tokyo Inst Technol, Dept Comp Sci, Tokyo, Japan
[2] DENSO Corp, Tokyo, Japan
关键词
D O I
10.1109/IV51971.2022.9827039
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Advanced driver assistance systems have gained popularity as a safe technology that helps people avoid traffic accidents. To improve system reliability, a lot of research on driving behavior prediction has been extensively researched. Inverse reinforcement learning (IRL) is known as a prominent approach because it can directly learn complicated behaviors from expert demonstrations. Because driving data tend to have a couple of optimal behaviors from the drivers' preferences, i.e., sub-optimality issue, maximum entropy IRL has been getting attention with their capability of considering suboptimality. While accurate modeling and prediction can be expected, standard maximum entropy IRL needs to calculate the partition function, which requires large computational costs. Thus, it is not straightforward to apply this model to a high-dimensional space for detailed car modeling. In addition, existing research attempts to reduce these costs by approximating maximum entropy IRL; however, a combination of the efficient path planning and the proper parameter updating is required for an accurate approximation, and existing methods have not achieved them. In this study, we leverage a rapidly-exploring random tree (RRT) motion planner. With the RRT planner, we propose novel importance sampling for an accurate approximation from the generated trees. This ensures a stable and fast IRL model in a large high-dimensional space. Experimental results on artificial environments show that our approach improves stability and is faster than the existing IRL methods.
引用
收藏
页码:1353 / 1359
页数:7
相关论文
共 50 条
  • [21] Robust Bayesian Inverse Reinforcement Learning with Sparse Behavior Noise
    Zheng, Jiangchuan
    Liu, Siyuan
    Ni, Lionel M.
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2014, : 2198 - 2205
  • [22] Research on Trajectory Prediction of Vehicle Lane Change for Autonomous Driving Based on Inverse Reinforcement Learning
    Zhan, Ming
    Fan, Jingjing
    Jin, Linhao
    [J]. SEVENTH INTERNATIONAL CONFERENCE ON TRAFFIC ENGINEERING AND TRANSPORTATION SYSTEM, ICTETS 2023, 2024, 13064
  • [23] Driving Behavior Modeling Using Naturalistic Human Driving Data With Inverse Reinforcement Learning
    Huang, Zhiyu
    Wu, Jingda
    Lv, Chen
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (08) : 10239 - 10251
  • [24] Outlier-robust Inverse Reinforcement Learning and Reward-based Detection of Anomalous Driving Behaviors
    Li, Dan
    Shehab, Mohamad Louai
    Liu, Zexiang
    Arechiga, Nikos
    DeCastro, Jonathan
    Ozay, Necmiye
    [J]. 2022 IEEE 25TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2022, : 4175 - 4182
  • [25] Future Trajectory Prediction via RNN and Maximum Margin Inverse Reinforcement Learning
    Choi, Dooseop
    An, Taeg-Hyun
    Ahn, Kyounghwan
    Choi, Jeongdan
    [J]. 2018 17TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA), 2018, : 125 - 130
  • [26] Inverse Reinforcement Learning Based Stochastic Driver Behavior Learning
    Ozkan, Mehmet F.
    Rocque, Abishek J.
    Ma, Yao
    [J]. IFAC PAPERSONLINE, 2021, 54 (20): : 882 - 888
  • [27] A behavior fusion method based on inverse reinforcement learning
    Shi, Haobin
    Li, Jingchen
    Chen, Shicong
    Hwang, Kao-Shing
    [J]. INFORMATION SCIENCES, 2022, 609 : 429 - 444
  • [28] Analyzing the Suitability of Cost Functions for Explaining and Imitating Human Driving Behavior based on Inverse Reinforcement Learning
    Naumann, Maximilian
    Sun, Liting
    Zhan, Wei
    Tomizuka, Masayoshi
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 5481 - 5487
  • [29] Predicting Pedestrian Crossing Behavior at Unsignalized Mid-Block Crosswalks Using Maximum Entropy Deep Inverse Reinforcement Learning
    Wang, Yongjie
    Niu, Yuchen
    Zhu, Wenying
    Chen, Wenqiang
    Li, Qiong
    Wang, Tao
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (05) : 3685 - 3698
  • [30] Adaptive Noise-based Evolutionary Reinforcement Learning With Maximum Entropy
    Wang, Jun-Yi
    Wang, Zhi
    Li, Hua-Xiong
    Chen, Chun-Lin
    [J]. Zidonghua Xuebao/Acta Automatica Sinica, 2023, 49 (01): : 54 - 66