Learning Variable Impedance Control via Inverse Reinforcement Learning for Force-Related Tasks

被引:65
|
作者
Zhang, Xiang [1 ]
Sun, Liting [1 ]
Kuang, Zhian [1 ,2 ]
Tomizuka, Masayoshi [1 ]
机构
[1] Univ Calif Berkeley, Dept Mech Engn, Berkeley, CA 94720 USA
[2] Harbin Inst Technol, Res Inst Intelligent Control & Syst, Harbin 150001, Peoples R China
关键词
Compliance and impedance control; learning from demonstration; machine learning for robot control;
D O I
10.1109/LRA.2021.3061374
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Many manipulation tasks require robots to interact with unknown environments. In such applications, the ability to adapt the impedance according to different task phases and environment constraints is crucial for safety and performance. Although many approaches based on deep reinforcement learning (RI) and learning from demonstration (LfD) have been proposed to obtain variable impedance skills on contact-rich manipulation tasks, these skills are typically task-specific and could be sensitive to changes in task settings. This letter proposes an inverse reinforcement learning (IRL) based approach to recover both the variable impedance policy and reward function from expert demonstrations. We explore different action space of the reward functions to achieve a more general representation of expert variable impedance skills. Experiments on two variable impedance tasks (Peg-in-Hole and Cup-on-Plate) were conducted in both simulations and on a real FANUC LR Mate 200iD/7 L industrial robot. The comparison results with behavior cloning and force-based IRL proved that the learned reward function in the gain action space has better transferability than in the force space. Experiment videos are available at https://msc.berkeley.edu/research/impedance-irl.html.
引用
收藏
页码:2225 / 2232
页数:8
相关论文
共 50 条
  • [21] Competitive reinforcement learning in continuous control tasks
    Abramson, M
    Pachowicz, P
    Wechsler, H
    PROCEEDINGS OF THE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS 2003, VOLS 1-4, 2003, : 1909 - 1914
  • [22] Haptic Assistance via Inverse Reinforcement Learning
    Scobee, Dexter R. R.
    Royo, Vicenc Rubies
    Tomlin, Claire J.
    Sastry, S. Shankar
    2018 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2018, : 1510 - 1517
  • [23] Variable Admittance Interaction Control of UAVs via Deep Reinforcement Learning
    Feng, Yuting
    Shi, Chuanbeibei
    Du, Jianrui
    Yu, Yushu
    Sun, Fuchun
    Song, Yixu
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 1291 - 1297
  • [24] Inverse Reinforcement Learning in Tracking Control Based on Inverse Optimal Control
    Xue, Wenqian
    Kolaric, Patrik
    Fan, Jialu
    Lian, Bosen
    Chai, Tianyou
    Lewis, Frank L.
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (10) : 10570 - 10581
  • [25] Inverse Reinforcement Learning: A Control Lyapunov Approach
    Tesfazgi, Samuel
    Lederer, Armin
    Hirche, Sandra
    2021 60TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2021, : 3627 - 3632
  • [26] Safety reinforcement learning control via transfer learning
    Zhang, Quanqi
    Wu, Chengwei
    Tian, Haoyu
    Gao, Yabin
    Yao, Weiran
    Wu, Ligang
    AUTOMATICA, 2024, 166
  • [27] Learning to Control Camera Exposure via Reinforcement Learning
    Lee, Kyunghyun
    Shin, Ukcheol
    Lee, Byeong-Uk
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2024, 2024, : 2975 - 2983
  • [28] A New Robotic Knee Impedance Control Parameter Optimization Method Facilitated by Inverse Reinforcement Learning
    Liu, Wentao
    Wu, Ruofan
    Si, Jennie
    Huang, He
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (04) : 10882 - 10889
  • [29] Preference-learning based Inverse Reinforcement Learning for Dialog Control
    Sugiyama, Hiroaki
    Meguro, Toyomi
    Minami, Yasuhiro
    13TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2012 (INTERSPEECH 2012), VOLS 1-3, 2012, : 222 - 225
  • [30] Learning state-action correspondence across reinforcement learning control tasks via partially paired trajectories
    Garcia, Javier
    Rano, Inaki
    Bures, J. Miguel
    Fdez-Vidal, Xose R.
    Iglesias, Roberto
    APPLIED INTELLIGENCE, 2025, 55 (03)