Model-Free Reinforcement Learning of Impedance Control in Stochastic Environments

被引:38
|
作者
Stulp, Freek [1 ,2 ,3 ]
Buchli, Jonas [1 ,4 ]
Ellmer, Alice [1 ]
Mistry, Michael [5 ]
Theodorou, Evangelos A. [1 ]
Schaal, Stefan [1 ]
机构
[1] Univ So Calif, Computat Learning & Motor Control Lab, Los Angeles, CA 90089 USA
[2] ParisTech, Ecole Natl Superi Tech Avancees, F-75015 Paris, France
[3] INRIA Bordeaux Sud Ouest, FLOWERS Res Team, F-33405 Talence, France
[4] Ist Italiano Tecnol, Dept Adv Robot, I-16163 Genoa, Italy
[5] Univ Birmingham, Sch Comp Sci, Birmingham B15 2TT, W Midlands, England
基金
美国国家科学基金会; 瑞士国家科学基金会;
关键词
Force field experiments; motion primitives; motor system and development; reinforcement learning; robots with development and learning skills; using robots to study development and learning; variable impedance control;
D O I
10.1109/TAMD.2012.2205924
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For humans and robots, variable impedance control is an essential component for ensuring robust and safe physical interaction with the environment. Humans learn to adapt their impedance to specific tasks and environments; a capability which we continually develop and improve until we are well into our twenties. In this article, we reproduce functionally interesting aspects of learning impedance control in humans on a simulated robot platform. As demonstrated in numerous force field tasks, humans combine two strategies to adapt their impedance to perturbations, thereby minimizing position error and energy consumption: 1) if perturbations are unpredictable, subjects increase their impedance through cocontraction; and 2) if perturbations are predictable, subjects learn a feed-forward command to offset the perturbation. We show how a 7-DOF simulated robot demonstrates similar behavior with our model-free reinforcement learning algorithm PI2, by applying deterministic and stochastic force fields to the robot's end-effector. We show the qualitative similarity between the robot and human movements. Our results provide a biologically plausible approach to learning appropriate impedances purely from experience, without requiring a model of either body or environment dynamics. Not requiring models also facilitates autonomous development for robots, as prespecified models cannot be provided for each environment a robot might encounter.
引用
收藏
页码:330 / 341
页数:12
相关论文
共 50 条
  • [1] Model-Free Quantum Control with Reinforcement Learning
    Sivak, V. V.
    Eickbusch, A.
    Liu, H.
    Royer, B.
    Tsioutsios, I
    Devoret, M. H.
    [J]. PHYSICAL REVIEW X, 2022, 12 (01):
  • [2] Model-free Reinforcement Learning for Stochastic Stackelberg Security Games
    Mishra, Rajesh K.
    Vasal, Deepanshu
    Vishwanath, Sriram
    [J]. 2020 59TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2020, : 348 - 353
  • [3] Model-free Reinforcement Learning with Stochastic Reward Stabilization for Recommender Systems
    Cai, Tianchi
    Bao, Shenliao
    Jiang, Jiyan
    Zhou, Shiji
    Zhang, Wenpeng
    Gu, Lihong
    Gu, Jinjie
    Zhang, Guannan
    [J]. PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 2179 - 2183
  • [4] Model-free learning control of neutralization processes using reinforcement learning
    Syafiie, S.
    Tadeo, F.
    Martinez, E.
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2007, 20 (06) : 767 - 782
  • [5] Linear Quadratic Control Using Model-Free Reinforcement Learning
    Yaghmaie, Farnaz Adib
    Gustafsson, Fredrik
    Ljung, Lennart
    [J]. IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2023, 68 (02) : 737 - 752
  • [6] Model-Free Recurrent Reinforcement Learning for AUV Horizontal Control
    Huo, Yujia
    Li, Yiping
    Feng, Xisheng
    [J]. 3RD INTERNATIONAL CONFERENCE ON AUTOMATION, CONTROL AND ROBOTICS ENGINEERING (CACRE 2018), 2018, 428
  • [7] On Distributed Model-Free Reinforcement Learning Control With Stability Guarantee
    Mukherjee, Sayak
    Vu, Thanh Long
    [J]. IEEE CONTROL SYSTEMS LETTERS, 2021, 5 (05): : 1615 - 1620
  • [8] Model-Free Control for Soft Manipulators based on Reinforcement Learning
    You, Xuanke
    Zhang, Yixiao
    Chen, Xiaotong
    Liu, Xinghua
    Wang, Zhanchi
    Jiang, Hao
    Chen, Xiaoping
    [J]. 2017 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2017, : 2909 - 2915
  • [9] Model-Free Emergency Frequency Control Based on Reinforcement Learning
    Chen, Chunyu
    Cui, Mingjian
    Li, Fangxing
    Yin, Shengfei
    Wang, Xinan
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (04) : 2336 - 2346
  • [10] On Distributed Model-Free Reinforcement Learning Control with Stability Guarantee
    Mukherjee, Sayak
    Thanh Long Vu
    [J]. 2021 AMERICAN CONTROL CONFERENCE (ACC), 2021, : 2175 - 2180