Efficient Sampling-Based Maximum Entropy Inverse Reinforcement Learning With Application to Autonomous Driving

被引:65
|
作者
Wu, Zheng [1 ]
Sun, Liting [1 ]
Zhan, Wei [1 ]
Yang, Chenyu [2 ]
Tomizuka, Masayoshi [1 ]
机构
[1] Univ Calif Berkeley, Dept Mech Engn, Berkeley, CA 94709 USA
[2] Shanghai Jiao Tong Univ, Dept Comp Sci, Shanghai 200240, Peoples R China
关键词
Learning from demonstration; intelligent transportation systems; inverse reinforcement learning; autonomous driving; social human-robot interaction; ALGORITHMS;
D O I
10.1109/LRA.2020.3005126
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
In the past decades, we have witnessed significant progress in the domain of autonomous driving. Advanced techniques based on optimization and reinforcement learning become increasingly powerful when solving the forward problem: given designed reward/cost functions, how we should optimize them and obtain driving policies that interact with the environment safely and efficiently. Such progress has raised another equally important question: what should we optimize? Instead of manually specifying the reward functions, it is desired that we can extract what human drivers try to optimize from real traffic data and assign that to autonomous vehicles to enable more naturalistic and transparent interaction between humans and intelligent agents. To address this issue, we present an efficient sampling-based maximum-entropy inverse reinforcement learning (IRL) algorithm in this letter. Different from existing IRL algorithms, by introducing an efficient continuous-domain trajectory sampler, the proposed algorithm can directly learn the reward functions in the continuous domain while considering the uncertainties in demonstrated trajectories from human drivers. We evaluate the proposed algorithm via real-world driving data, including both non-interactive and interactive scenarios. The experimental results show that the proposed algorithm achieves more accurate prediction performance with faster convergence speed and better generalization compared to other baseline IRL algorithms.
引用
收藏
页码:5355 / 5362
页数:8
相关论文
共 50 条
  • [1] Maximum Entropy Inverse Reinforcement Learning Based on Frenet Frame Sampling for Human-like Autonomous Driving
    Zhang, Tangyike
    Sun, Shuo
    Shi, Jiamin
    Chen, Shitao
    Ang, Marcelo H.
    Xin, Jingmin
    Zheng, Nanning
    2023 IEEE 26TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS, ITSC, 2023, : 1820 - 1827
  • [2] RRT-based maximum entropy inverse reinforcement learning for robust and efficient driving behavior prediction
    Hosoma, Shinpei
    Sugasaki, Masato
    Arie, Hiroaki
    Shimosaka, Masamichi
    2022 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2022, : 1353 - 1359
  • [3] Maximum Entropy Inverse Reinforcement Learning Using Monte Carlo Tree Search for Autonomous Driving
    da Silva, Junior Anderson Rodrigues
    Grassi Jr, Valdir
    Wolf, Denis Fernando
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (09) : 1 - 0
  • [4] Sampling-based Inverse Reinforcement Learning Algorithms with Safety Constraints
    Fischer, Johannes
    Eyberg, Christoph
    Werling, Moritz
    Lauer, Martin
    2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 791 - 798
  • [5] Efficient Sampling-Based Motion Planning for On-Road Autonomous Driving
    Ma, Liang
    Xue, Jianru
    Kawabata, Kuniaki
    Zhu, Jihua
    Ma, Chao
    Zheng, Nanning
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2015, 16 (04) : 1961 - 1976
  • [6] Integrating Algorithmic Sampling-Based Motion Planning with Learning in Autonomous Driving
    Zhang, Yifan
    Zhang, Jinghuai
    Zhang, Jindi
    Wang, Jianping
    Lu, Kejie
    Hong, Jeff
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2022, 13 (03)
  • [7] A Novel Learning Framework for Sampling-Based Motion Planning in Autonomous Driving
    Zhang, Yifan
    Zhang, Jinghuai
    Zhang, Jindi
    Wang, Jianping
    Lu, Kejie
    Hong, Jeff
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 1202 - 1209
  • [8] Modeling framework of human driving behavior based on Deep Maximum Entropy Inverse Reinforcement Learning
    Wang, Yongjie
    Niu, Yuchen
    Xiao, Mei
    Zhu, Wenying
    You, Xinshang
    PHYSICA A-STATISTICAL MECHANICS AND ITS APPLICATIONS, 2024, 652
  • [9] ESRL: Efficient Sampling-Based Reinforcement Learning for Sequence Generation
    Wang, Chenglong
    Zhou, Hang
    Hu, Yimin
    Huo, Yifu
    Li, Bei
    Liu, Tongran
    Xiao, Tong
    Zhu, Jingbo
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 19107 - 19115
  • [10] Maximum causal entropy inverse constrained reinforcement learning
    Mattijs Baert
    Pietro Mazzaglia
    Sam Leroux
    Pieter Simoens
    Machine Learning, 2025, 114 (4)