Inverse Reinforcement Learning of Behavioral Models for Online-Adapting Navigation Strategies

被引:0
|
作者
Herman, Michael [1 ]
Fiseher, Volker [1 ]
Gindele, Tobias [1 ]
Burgard, Wolfram [2 ]
机构
[1] Robert Bosch GmbH, Corp Sect Res & Adv Engn, D-70442 Stuttgart, Germany
[2] Univ Freiburg, Dept Comp Sci, D-79110 Freiburg, Germany
关键词
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
To increase the acceptance of autonomous systems in populated environments, it is indispensable to teach them social behavior. We would expect a social robot, which plans its motions among humans, to consider both the social acceptability of its behavior as well as task constraints, such as time limits. These requirements are often contradictory and therefore resulting in a trade-off. For example, a robot has to decide whether it is more important to quickly achieve its goal or to comply with social conventions, such as the proximity to humans, i.e., the robot has to react adaptively to task-specific priorities. In this paper, we present a method for priority-adaptive navigation of mobile autonomous systems, which optimizes the social acceptability of the behavior while meeting task constraints. We learn acceptability-dependent behavioral models from human demonstrations by using maximum entropy (MaxEnt) inverse reinforcement learning (IRL). These models are generative and describe the learned stochastic behavior. We choose the optimum behavioral model by maximizing the social acceptability under constraints on expected time-limits and reliabilities. This approach is evaluated in the context of driving behaviors based on the highway scenario of Levine et al. [1].
引用
收藏
页码:3215 / 3222
页数:8
相关论文
共 50 条
  • [1] Identification of animal behavioral strategies by inverse reinforcement learning
    Yamaguchi, Shoichiro
    Naoki, Honda
    Ikeda, Muneki
    Tsukada, Yuki
    Nakano, Shunji
    Mori, Ikue
    Ishii, Shin
    [J]. PLOS COMPUTATIONAL BIOLOGY, 2018, 14 (05):
  • [2] Reinforcement learning strategies for vessel navigation
    Daranda, Andrius
    Dzemyda, Gintautas
    [J]. INTEGRATED COMPUTER-AIDED ENGINEERING, 2023, 30 (01) : 53 - 66
  • [3] Neural inverse reinforcement learning in autonomous navigation
    Xia, Chen
    El Kamel, Abdelkader
    [J]. ROBOTICS AND AUTONOMOUS SYSTEMS, 2016, 84 : 1 - 14
  • [4] Learning and Adapting Behavior of Autonomous Vehicles through Inverse Reinforcement Learning
    Trauth, Rainer
    Kaufeld, Marc
    Geisslinger, Maximilian
    Betz, Johannes
    [J]. 2023 IEEE INTELLIGENT VEHICLES SYMPOSIUM, IV, 2023,
  • [5] Online inverse reinforcement learning with limited data
    Self, Ryan
    Mahmud, S. M. Nahid
    Hareland, Katrine
    Kamalapurkar, Rushikesh
    [J]. 2020 59TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2020, : 603 - 608
  • [6] Online Inverse Reinforcement Learning Under Occlusion
    Arora, Saurabh
    Doshi, Prashant
    Banerjee, Bikramjit
    [J]. AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 1170 - 1178
  • [7] Online inverse reinforcement learning for nonlinear systems
    Self, Ryan
    Harlan, Michael
    Kamalapurkar, Rushikesh
    [J]. 2019 3RD IEEE CONFERENCE ON CONTROL TECHNOLOGY AND APPLICATIONS (IEEE CCTA 2019), 2019, : 296 - 301
  • [8] Online inverse reinforcement learning for systems with disturbances
    Self, Ryan
    Abudia, Moad
    Kamalapurkar, Rushikesh
    [J]. 2020 AMERICAN CONTROL CONFERENCE (ACC), 2020, : 1118 - 1123
  • [9] Practical Bayesian Inverse Reinforcement Learning for Robot Navigation
    Okal, Billy
    Arras, Kai O.
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2016, PT III, 2016, 9853 : 271 - 274
  • [10] Adapting Hidden Markov Models for Online Learning
    Chis, Tiberiu
    Harrison, Peter G.
    [J]. ELECTRONIC NOTES IN THEORETICAL COMPUTER SCIENCE, 2015, 318 : 109 - 127