Inverse Reinforcement Learning of Behavioral Models for Online-Adapting Navigation Strategies

被引:0
|
作者
Herman, Michael [1 ]
Fiseher, Volker [1 ]
Gindele, Tobias [1 ]
Burgard, Wolfram [2 ]
机构
[1] Robert Bosch GmbH, Corp Sect Res & Adv Engn, D-70442 Stuttgart, Germany
[2] Univ Freiburg, Dept Comp Sci, D-79110 Freiburg, Germany
关键词
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
To increase the acceptance of autonomous systems in populated environments, it is indispensable to teach them social behavior. We would expect a social robot, which plans its motions among humans, to consider both the social acceptability of its behavior as well as task constraints, such as time limits. These requirements are often contradictory and therefore resulting in a trade-off. For example, a robot has to decide whether it is more important to quickly achieve its goal or to comply with social conventions, such as the proximity to humans, i.e., the robot has to react adaptively to task-specific priorities. In this paper, we present a method for priority-adaptive navigation of mobile autonomous systems, which optimizes the social acceptability of the behavior while meeting task constraints. We learn acceptability-dependent behavioral models from human demonstrations by using maximum entropy (MaxEnt) inverse reinforcement learning (IRL). These models are generative and describe the learned stochastic behavior. We choose the optimum behavioral model by maximizing the social acceptability under constraints on expected time-limits and reliabilities. This approach is evaluated in the context of driving behaviors based on the highway scenario of Levine et al. [1].
引用
收藏
页码:3215 / 3222
页数:8
相关论文
共 50 条
  • [21] A comparison of reinforcement learning models of human spatial navigation
    Qiliang He
    Jancy Ling Liu
    Lou Eschapasse
    Elizabeth H. Beveridge
    Thackery I. Brown
    [J]. Scientific Reports, 12
  • [22] A comparison of reinforcement learning models of human spatial navigation
    He, Qiliang
    Liu, Jancy Ling
    Eschapasse, Lou
    Beveridge, Elizabeth H.
    Brown, Thackery, I
    [J]. SCIENTIFIC REPORTS, 2022, 12 (01)
  • [23] STRATEGIES FOR ADAPTING ONLINE EDUCATION TO SECOND LANGUAGE TEACHING & LEARNING
    Farina, Marcella A.
    Selmo, Laura
    [J]. 12TH INTERNATIONAL TECHNOLOGY, EDUCATION AND DEVELOPMENT CONFERENCE (INTED), 2018, : 7356 - 7364
  • [24] Socially compliant mobile robot navigation via inverse reinforcement learning
    Kretzschmar, Henrik
    Spies, Markus
    Sprunk, Christoph
    Burgard, Wolfram
    [J]. INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2016, 35 (11): : 1352 - 1370
  • [25] Convergence analysis of an incremental approach to online inverse reinforcement learning
    Jin, Zhuo-jun
    Qian, Hui
    Chen, Shen-yi
    Zhu, Miao-liang
    [J]. JOURNAL OF ZHEJIANG UNIVERSITY-SCIENCE C-COMPUTERS & ELECTRONICS, 2011, 12 (01): : 17 - 24
  • [27] Online inverse reinforcement learning for nonlinear systems with adversarial attacks
    Lian, Bosen
    Xue, Wenqian
    Lewis, Frank L.
    Chai, Tianyou
    [J]. INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2021, 31 (14) : 6646 - 6667
  • [28] Strategies for simulating pedestrian navigation with multiple reinforcement learning agents
    Martinez-Gil, Francisco
    Lozano, Miguel
    Fernandez, Fernando
    [J]. AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2015, 29 (01) : 98 - 130
  • [29] Convergence analysis of an incremental approach to online inverse reinforcement learning
    Zhuo-jun Jin
    Hui Qian
    Shen-yi Chen
    Miao-liang Zhu
    [J]. Journal of Zhejiang University SCIENCE C, 2011, 12 : 17 - 24
  • [30] Strategies for simulating pedestrian navigation with multiple reinforcement learning agents
    Francisco Martinez-Gil
    Miguel Lozano
    Fernando Fernández
    [J]. Autonomous Agents and Multi-Agent Systems, 2015, 29 : 98 - 130