State Aware Imitation Learning

被引:0
|
作者
Schroecker, Yannick [1 ]
Isbell, Charles [1 ]
机构
[1] Georgia Inst Technol, Coll Comp, Atlanta, GA 30332 USA
关键词
AVERAGE;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Imitation learning is the study of learning how to act given a set of demonstrations provided by a human expert. It is intuitively apparent that learning to take optimal actions is a simpler undertaking in situations that are similar to the ones shown by the teacher. However, imitation learning approaches do not tend to use this insight directly. In this paper, we introduce State Aware Imitation Learning (SAIL), an imitation learning algorithm that allows an agent to learn how to remain in states where it can confidently take the correct action and how to recover if it is lead astray. Key to this algorithm is a gradient learned using a temporal difference update rule which leads the agent to prefer states similar to the demonstrated states. We show that estimating a linear approximation of this gradient yields similar theoretical guarantees to online temporal difference learning approaches and empirically show that SAIL can effectively be used for imitation learning in continuous domains with non-linear function approximators used for both the policy representation and the gradient estimate.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] LEARNING OF IMITATION AND LEARNING THROUGH IMITATION IN WHITE RAT
    HARUKI, Y
    TSUZUKI, T
    ANNUAL OF ANIMAL PSYCHOLOGY, 1967, 17 (02): : 57 - &
  • [22] Adversarial Imitation Learning from State-only Demonstrations
    Torabi, Faraz
    Warnell, Garrett
    Stone, Peter
    AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 2229 - 2231
  • [23] Adversarial Imitation Learning from Video using a State Observer
    Karnan, Haresh
    Torabi, Faraz
    Warnell, Garrett
    Stone, Peter
    2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022), 2022, : 2452 - 2458
  • [24] iCOIL: Scenario Aware Autonomous Parking Via Integrated Constrained Optimization and Imitation Learning
    Huang, Lexiong
    Han, Ruihua
    Li, Guoliang
    Li, He
    Wang, Shuai
    Wang, Yang
    Xu, Chengzhong
    2023 IEEE 43RD INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS WORKSHOPS, ICDCSW, 2023, : 97 - 102
  • [25] Deep Adversarial Imitation Reinforcement Learning for QoS-Aware Cloud Job Scheduling
    Huang, Yifeng
    Cheng, Long
    Xue, Lianting
    Liu, Cong
    Li, Yuancheng
    Li, Jianbin
    Ward, Tomas
    IEEE SYSTEMS JOURNAL, 2022, 16 (03): : 4232 - 4242
  • [26] Learning by imitation
    Basçi, E
    JOURNAL OF ECONOMIC DYNAMICS & CONTROL, 1999, 23 (9-10): : 1569 - 1585
  • [27] Adversarial Imitation Learning between Agents with Different Numbers of State Dimensions
    Yoshida, Taketo
    Kuniyoshi, Yasuo
    2019 IEEE SECOND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND KNOWLEDGE ENGINEERING (AIKE), 2019, : 179 - 186
  • [28] Domain-Adversarial and -Conditional State Space Model for Imitation Learning
    Okumura, Ryo
    Okada, Masashi
    Taniguchi, Tadahiro
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 5179 - 5186
  • [29] SHAIL: Safety-Aware Hierarchical Adversarial Imitation Learning for Autonomous Driving in Urban Environments
    Jamgochian, Arec
    Buehrle, Etienne
    Fischer, Johannes
    Kochenderfer, Mykel J.
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 1530 - 1536
  • [30] Risk-Aware Self-consistent Imitation Learning for Trajectory Planning in Autonomous Driving
    Fan, Yixuan
    Li, Yali
    Wang, Shengjin
    COMPUTER VISION - ECCV 2024, PT XIII, 2025, 15071 : 270 - 287