Analyzing the Suitability of Cost Functions for Explaining and Imitating Human Driving Behavior based on Inverse Reinforcement Learning

被引:0
|
作者
Naumann, Maximilian [1 ,2 ]
Sun, Liting [3 ]
Zhan, Wei [3 ]
Tomizuka, Masayoshi [3 ]
机构
[1] FZI Res Ctr Informat Technol, D-76131 Karlsruhe, Germany
[2] Karlsruhe Inst Technol KIT, D-76131 Karlsruhe, Germany
[3] Univ Calif Berkeley, Dept Mech Engn, Berkeley, CA 94720 USA
关键词
Automated vehicles; cost function; inverse reinforcement learning; imitation learning; cooperative motion planning;
D O I
10.1109/icra40945.2020.9196795
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Autonomous vehicles are sharing the road with human drivers. In order to facilitate interactive driving and cooperative behavior in dense traffic, a thorough understanding and representation of other traffic participants' behavior are necessary. Cost functions (or reward functions) have been widely used to describe the behavior of human drivers since they can not only explicitly incorporate the rationality of human drivers and the theory of mind (TOM), but also share similarity with the motion planning problem of autonomous vehicles. Hence, more human-like driving behavior and comprehensible trajectories can be generated to enable safer interaction and cooperation. However, the selection of cost functions in different driving scenarios is not trivial, and there is no systematic summary and analysis for cost function selection and learning from a variety of driving scenarios. In this work, we aim to investigate to what extent cost functions are suitable for explaining and imitating human driving behavior. Further, we focus on how cost functions differ from each other in different driving scenarios. Towards this goal, we first comprehensively review existing cost function structures in literature. Based on that, we point out required conditions for demonstrations to be suitable for inverse reinforcement learning (IRL). Finally, we use IRL to explore suitable features and learn cost function weights from human driven trajectories in three different scenarios.
引用
收藏
页码:5481 / 5487
页数:7
相关论文
共 50 条
  • [1] Driving Behavior Modeling Using Naturalistic Human Driving Data With Inverse Reinforcement Learning
    Huang, Zhiyu
    Wu, Jingda
    Lv, Chen
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (08) : 10239 - 10251
  • [2] Modeling framework of human driving behavior based on Deep Maximum Entropy Inverse Reinforcement Learning
    Wang, Yongjie
    Niu, Yuchen
    Xiao, Mei
    Zhu, Wenying
    You, Xinshang
    [J]. PHYSICA A-STATISTICAL MECHANICS AND ITS APPLICATIONS, 2024, 652
  • [3] Predicting driving behavior using inverse reinforcement learning with multiple reward functions towards environmental diversity
    Shimosaka, Masamichi
    Nishi, Kentaro
    Sato, Junichi
    Kataoka, Hirokatsu
    [J]. 2015 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2015, : 567 - 572
  • [4] Conditional Predictive Behavior Planning With Inverse Reinforcement Learning for Human-Like Autonomous Driving
    Huang, Zhiyu
    Liu, Haochen
    Wu, Jingda
    Lv, Chen
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (07) : 7244 - 7258
  • [5] Analyzing Sensor-Based Individual and Population Behavior Patterns via Inverse Reinforcement Learning
    Lin, Beiyu
    Cook, Diane J.
    [J]. SENSORS, 2020, 20 (18) : 1 - 21
  • [6] Inverse Reinforcement Learning Based Stochastic Driver Behavior Learning
    Ozkan, Mehmet F.
    Rocque, Abishek J.
    Ma, Yao
    [J]. IFAC PAPERSONLINE, 2021, 54 (20): : 882 - 888
  • [7] Fast Inverse Reinforcement Learning with Interval Consistent Graph for Driving Behavior Prediction
    Shimosaka, Masamichi
    Sato, Junichi
    Takenaka, Kazuhito
    Hitomi, Kentarou
    [J]. THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 1532 - 1538
  • [8] Probabilistic Prediction of Interactive Driving Behavior via Hierarchical Inverse Reinforcement Learning
    Sun, Liting
    Zhan, Wei
    Tomizuka, Masayoshi
    [J]. 2018 21ST INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2018, : 2111 - 2117
  • [9] A behavior fusion method based on inverse reinforcement learning
    Shi, Haobin
    Li, Jingchen
    Chen, Shicong
    Hwang, Kao-Shing
    [J]. INFORMATION SCIENCES, 2022, 609 : 429 - 444
  • [10] RRT-based maximum entropy inverse reinforcement learning for robust and efficient driving behavior prediction
    Hosoma, Shinpei
    Sugasaki, Masato
    Arie, Hiroaki
    Shimosaka, Masamichi
    [J]. 2022 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2022, : 1353 - 1359