Data-Efficient Reinforcement Learning for Variable Impedance Control

被引:0
|
作者
Anand, Akhil S. [1 ]
Kaushik, Rituraj [2 ]
Gravdahl, Jan Tommy [1 ]
Abu-Dakka, Fares J. [3 ]
机构
[1] Norwegian Univ Sci & Technol NTNU, Dept Engn Cybernet, N-7491 Trondheim, Norway
[2] Aalto Univ, Dept Elect Engn & Automat EEA, Intelligent Robot Grp, Espoo 00076, Finland
[3] Mondragon Univ, Fac Engn, Dept Elect & Informat, Arrasate Mondragon 20500, Spain
关键词
Model-based reinforcement learning; variable impedance learning control; Gaussian processes; covariance matrix adaptation; EVOLUTIONARY OPTIMIZATION; FORCE CONTROL; CONTACT; ROBOT; ENVIRONMENT; ADAPTATION; MOTION;
D O I
10.1109/ACCESS.2024.3355311
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
One of the most crucial steps toward achieving human-like manipulation skills in robots is to incorporate compliance into the robot controller. Compliance not only makes the robot's behaviour safe but also makes it more energy efficient. In this direction, the variable impedance control (VIC) approach provides a framework for a robot to adapt its compliance during execution by employing an adaptive impedance law. Nevertheless, autonomously adapting the compliance profile as demanded by the task remains a challenging problem to be solved in practice. In this work, we introduce a reinforcement learning (RL)-based approach called DEVILC (Data-Efficient Variable Impedance Learning Controller) to learn the variable impedance controller through real-world interaction of the robot. More concretely, we use a model-based RL approach in which, after every interaction, the robot iteratively learns a probabilistic model of its dynamics using the Gaussian process regression model. The model is then used to optimize a neural-network policy that modulates the robot's impedance such that the long-term reward for the task is maximized. Thanks to the model-based RL framework, DEVILC allows a robot to learn the VIC policy with only a few interactions, making it practical for real-world applications. In simulations and experiments, we evaluate DEVILC on a Franka Emika Panda robotic manipulator for different manipulation tasks in the Cartesian space. The results show that DEVILC is a promising direction toward autonomously learning compliant manipulation skills directly in the real world through interactions. A video of the experiments is available in the link: https://youtu.be/_uyr0Vye5no.
引用
收藏
页码:15631 / 15641
页数:11
相关论文
共 50 条
  • [31] Data-efficient model-based reinforcement learning with trajectory discrimination
    Qu, Tuo
    Duan, Fuqing
    Zhang, Junge
    Zhao, Bo
    Huang, Wenzhen
    [J]. COMPLEX & INTELLIGENT SYSTEMS, 2024, 10 (02) : 1927 - 1936
  • [32] Gaussian Processes for Data-Efficient Learning in Robotics and Control
    Deisenroth, Marc Peter
    Fox, Dieter
    Rasmussen, Carl Edward
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2015, 37 (02) : 408 - 423
  • [33] Data-Efficient Graph Learning
    Ding, Kaize
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 20, 2024, : 22663 - 22663
  • [34] Data-Efficient Deep Reinforcement Learning for Attitude Control of Fixed-Wing UAVs: Field Experiments
    Bohn, Eivind
    Coates, Erlend M.
    Reinhardt, Dirk
    Johansen, Tor Arne
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (03) : 3168 - 3180
  • [35] Data-Efficient Reinforcement Learning from Controller Guidance with Integrated Self-Supervision for Process Control
    Bougie, Nicolas
    Onishi, Takashi
    Tsuruoka, Yoshimasa
    [J]. IFAC PAPERSONLINE, 2022, 55 (07): : 863 - 868
  • [36] Model-Based Reinforcement Learning With Probabilistic Ensemble Terminal Critics for Data-Efficient Control Applications
    Park, Jonghyeok
    Jeon, Soo
    Han, Soohee
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2024, 71 (08) : 9470 - 9479
  • [37] Guest Editorial Special Issue on Reinforcement Learning-Based Control: Data-Efficient and Resilient Methods
    Gao, Weinan
    Li, Na
    Vamvoudakis, Kyriakos G.
    Yu, Fei Richard
    Jiang, Zhong-Ping
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (03) : 3103 - 3106
  • [38] Data-efficient Co-Adaptation of Morphology and Behaviour with Deep Reinforcement Learning
    Luck, Kevin Sebastian
    Ben Amor, Heni
    Calandra, Roberto
    [J]. CONFERENCE ON ROBOT LEARNING, VOL 100, 2019, 100
  • [39] Data-Efficient Reinforcement Learning for Energy Optimization of Power-Assisted Wheelchairs
    Feng, Guoxi
    Busoniu, Lucian
    Guerra, Thierry-Marie
    Mohammad, Sami
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2019, 66 (12) : 9734 - 9744
  • [40] Robust On-Policy Sampling for Data-Efficient Policy Evaluation in Reinforcement Learning
    Zhong, Rujie
    Zhang, Duohan
    Schafer, Lukas
    Albrecht, Stefano V.
    Hanna, Josiah P.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,