Data-Efficient Reinforcement Learning for Variable Impedance Control

被引:0
|
作者
Anand, Akhil S. [1 ]
Kaushik, Rituraj [2 ]
Gravdahl, Jan Tommy [1 ]
Abu-Dakka, Fares J. [3 ]
机构
[1] Norwegian Univ Sci & Technol NTNU, Dept Engn Cybernet, N-7491 Trondheim, Norway
[2] Aalto Univ, Dept Elect Engn & Automat EEA, Intelligent Robot Grp, Espoo 00076, Finland
[3] Mondragon Univ, Fac Engn, Dept Elect & Informat, Arrasate Mondragon 20500, Spain
关键词
Model-based reinforcement learning; variable impedance learning control; Gaussian processes; covariance matrix adaptation; EVOLUTIONARY OPTIMIZATION; FORCE CONTROL; CONTACT; ROBOT; ENVIRONMENT; ADAPTATION; MOTION;
D O I
10.1109/ACCESS.2024.3355311
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
One of the most crucial steps toward achieving human-like manipulation skills in robots is to incorporate compliance into the robot controller. Compliance not only makes the robot's behaviour safe but also makes it more energy efficient. In this direction, the variable impedance control (VIC) approach provides a framework for a robot to adapt its compliance during execution by employing an adaptive impedance law. Nevertheless, autonomously adapting the compliance profile as demanded by the task remains a challenging problem to be solved in practice. In this work, we introduce a reinforcement learning (RL)-based approach called DEVILC (Data-Efficient Variable Impedance Learning Controller) to learn the variable impedance controller through real-world interaction of the robot. More concretely, we use a model-based RL approach in which, after every interaction, the robot iteratively learns a probabilistic model of its dynamics using the Gaussian process regression model. The model is then used to optimize a neural-network policy that modulates the robot's impedance such that the long-term reward for the task is maximized. Thanks to the model-based RL framework, DEVILC allows a robot to learn the VIC policy with only a few interactions, making it practical for real-world applications. In simulations and experiments, we evaluate DEVILC on a Franka Emika Panda robotic manipulator for different manipulation tasks in the Cartesian space. The results show that DEVILC is a promising direction toward autonomously learning compliant manipulation skills directly in the real world through interactions. A video of the experiments is available in the link: https://youtu.be/_uyr0Vye5no.
引用
收藏
页码:15631 / 15641
页数:11
相关论文
共 50 条
  • [21] Masked and Inverse Dynamics Modeling for Data-Efficient Reinforcement Learning
    Lee, Young Jae
    Kim, Jaehoon
    Park, Young Joon
    Kwak, Mingu
    Kim, Seoung Bum
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,
  • [22] A Data-Efficient Method of Deep Reinforcement Learning for Chinese Chess
    Xu, Changming
    Ding, Hengfeng
    Zhang, Xuejian
    Wang, Cong
    Yang, Hongji
    [J]. 2022 IEEE 22ND INTERNATIONAL CONFERENCE ON SOFTWARE QUALITY, RELIABILITY, AND SECURITY COMPANION, QRS-C, 2022, : 687 - 693
  • [23] DeMis: Data-Efficient Misinformation Detection Using Reinforcement Learning
    Kawintiranon, Kornraphop
    Singh, Lisa
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2022, PT II, 2023, 13714 : 224 - 240
  • [24] Ensemble and Auxiliary Tasks for Data-Efficient Deep Reinforcement Learning
    Maulana, Muhammad Rizki
    Lee, Wee Sun
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, 2021, 12975 : 122 - 138
  • [25] Shielded Planning Guided Data-Efficient and Safe Reinforcement Learning
    Wang, Hao
    Qin, Jiahu
    Kan, Zhen
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, : 1 - 12
  • [26] Unsupervised Salient Patch Selection for Data-Efficient Reinforcement Learning
    Jiang, Zhaohui
    Weng, Paul
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT IV, 2023, 14172 : 556 - 572
  • [27] Data-Efficient Deep Reinforcement Learning-Based Optimal Generation Control in DC Microgrids
    Fan, Zhen
    Zhang, Wei
    Liu, Wenxin
    [J]. IEEE SYSTEMS JOURNAL, 2024, 18 (01): : 426 - 437
  • [28] A Data-Efficient Reinforcement Learning Method Based on Local Koopman Operators
    Song, Lixing
    Wang, Junheng
    Xu, Junhong
    [J]. 20TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2021), 2021, : 515 - 520
  • [29] Data-efficient model-based reinforcement learning with trajectory discrimination
    Tuo Qu
    Fuqing Duan
    Junge Zhang
    Bo Zhao
    Wenzhen Huang
    [J]. Complex & Intelligent Systems, 2024, 10 : 1927 - 1936
  • [30] Data-Efficient Off-Policy Policy Evaluation for Reinforcement Learning
    Thomas, Philip S.
    Brunskill, Emma
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48