Option-Aware Adversarial Inverse Reinforcement Learning for Robotic Control

被引:2
|
作者
Chen, Jiayu [1 ]
Lan, Tian [3 ]
Aggarwal, Vaneet [1 ,2 ]
机构
[1] Purdue Univ, Sch Ind Engn, W Lafayette, IN 47907 USA
[2] KAUST, CS Dept, Thuwal, Saudi Arabia
[3] George Washington Univ, Dept Elect & Comp Engn, Washington, DC 20052 USA
来源
2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA | 2023年
关键词
NEURAL-NETWORKS;
D O I
10.1109/ICRA48891.2023.10160374
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Hierarchical Imitation Learning (HIL) has been proposed to recover highly-complex behaviors in long-horizon tasks from expert demonstrations by modeling the task hierarchy with the option framework. Existing methods either overlook the causal relationship between the subtask and its corresponding policy or cannot learn the policy in an end-to-end fashion, which leads to suboptimality. In this work, we develop a novel HIL algorithm based on Adversarial Inverse Reinforcement Learning and adapt it with the Expectation-Maximization algorithm in order to directly recover a hierarchical policy from the unannotated demonstrations. Further, we introduce a directed information term to the objective function to enhance the causality and propose a Variational Autoencoder framework for learning with our objectives in an end-to-end fashion. Theoretical justifications and evaluations on challenging robotic control tasks are provided to show the superiority of our algorithm. The codes are available at https://github.com/LucasCJYSDL/HierAIRL.
引用
收藏
页码:5902 / 5908
页数:7
相关论文
共 50 条
  • [21] Robotic Control in Adversarial and Sparse Reward Environments: A Robust Goal-Conditioned Reinforcement Learning Approach
    He X.
    Lv C.
    IEEE Transactions on Artificial Intelligence, 2024, 5 (01): : 244 - 253
  • [22] A New Robotic Knee Impedance Control Parameter Optimization Method Facilitated by Inverse Reinforcement Learning
    Liu, Wentao
    Wu, Ruofan
    Si, Jennie
    Huang, He
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (04) : 10882 - 10889
  • [23] Inverse-Reinforcement-Learning-Based Robotic Ultrasound Active Compliance Control in Uncertain Environments
    Ning, Guochen
    Liang, Hanying
    Zhang, Xinran
    Liao, Hongen
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2024, 71 (02) : 1686 - 1696
  • [24] Generative Adversarial Inverse Reinforcement Learning With Deep Deterministic Policy Gradient
    Zhan, Ming
    Fan, Jingjing
    Guo, Jianying
    IEEE ACCESS, 2023, 11 : 87732 - 87746
  • [25] Adversarial Inverse Reinforcement Learning to Estimate Policies from Multiple Experts
    Yamashita K.
    Hamagami T.
    Yamashita, Kodai, 2021, Institute of Electrical Engineers of Japan (141) : 1405 - 1410
  • [26] Adversarial Inverse Reinforcement Learning With Self-Attention Dynamics Model
    Sun, Jiankai
    Yu, Lantao
    Dong, Pinqian
    Lu, Bo
    Zhou, Bolei
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (02) : 1880 - 1886
  • [27] Grasping Living Objects With Adversarial Behaviors Using Inverse Reinforcement Learning
    Hu, Zhe
    Zheng, Yu
    Pan, Jia
    IEEE TRANSACTIONS ON ROBOTICS, 2023, 39 (02) : 1151 - 1163
  • [28] Objective Weight Interval Estimation Using Adversarial Inverse Reinforcement Learning
    Takayama, Naoya
    Arai, Sachiyo
    IEEE ACCESS, 2023, 11 : 58532 - 58538
  • [29] Decentralized reinforcement learning control of a robotic manipulator
    Busoniu, Lucian
    De Schutter, Bart
    Babuska, Robert
    2006 9TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION, VOLS 1- 5, 2006, : 1121 - +
  • [30] Adversarial attacks on reinforcement learning agents for command and control
    Dabholkar, Ahaan
    Hare, James Z.
    Mittrick, Mark
    Richardson, John
    Waytowich, Nicholas
    Narayanan, Priya
    Bagchi, Saurabh
    JOURNAL OF DEFENSE MODELING AND SIMULATION-APPLICATIONS METHODOLOGY TECHNOLOGY-JDMS, 2024,