Model-Free Reinforcement Learning of Impedance Control in Stochastic Environments

被引:38
|
作者
Stulp, Freek [1 ,2 ,3 ]
Buchli, Jonas [1 ,4 ]
Ellmer, Alice [1 ]
Mistry, Michael [5 ]
Theodorou, Evangelos A. [1 ]
Schaal, Stefan [1 ]
机构
[1] Univ So Calif, Computat Learning & Motor Control Lab, Los Angeles, CA 90089 USA
[2] ParisTech, Ecole Natl Superi Tech Avancees, F-75015 Paris, France
[3] INRIA Bordeaux Sud Ouest, FLOWERS Res Team, F-33405 Talence, France
[4] Ist Italiano Tecnol, Dept Adv Robot, I-16163 Genoa, Italy
[5] Univ Birmingham, Sch Comp Sci, Birmingham B15 2TT, W Midlands, England
基金
瑞士国家科学基金会; 美国国家科学基金会;
关键词
Force field experiments; motion primitives; motor system and development; reinforcement learning; robots with development and learning skills; using robots to study development and learning; variable impedance control;
D O I
10.1109/TAMD.2012.2205924
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For humans and robots, variable impedance control is an essential component for ensuring robust and safe physical interaction with the environment. Humans learn to adapt their impedance to specific tasks and environments; a capability which we continually develop and improve until we are well into our twenties. In this article, we reproduce functionally interesting aspects of learning impedance control in humans on a simulated robot platform. As demonstrated in numerous force field tasks, humans combine two strategies to adapt their impedance to perturbations, thereby minimizing position error and energy consumption: 1) if perturbations are unpredictable, subjects increase their impedance through cocontraction; and 2) if perturbations are predictable, subjects learn a feed-forward command to offset the perturbation. We show how a 7-DOF simulated robot demonstrates similar behavior with our model-free reinforcement learning algorithm PI2, by applying deterministic and stochastic force fields to the robot's end-effector. We show the qualitative similarity between the robot and human movements. Our results provide a biologically plausible approach to learning appropriate impedances purely from experience, without requiring a model of either body or environment dynamics. Not requiring models also facilitates autonomous development for robots, as prespecified models cannot be provided for each environment a robot might encounter.
引用
收藏
页码:330 / 341
页数:12
相关论文
共 50 条
  • [31] Recovering Robustness in Model-Free Reinforcement Learning
    Venkataraman, Harish K.
    Seiler, Peter J.
    [J]. 2019 AMERICAN CONTROL CONFERENCE (ACC), 2019, : 4210 - 4216
  • [32] Model-Free Reinforcement Learning Algorithms: A Survey
    Calisir, Sinan
    Pehlivanoglu, Meltem Kurt
    [J]. 2019 27TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2019,
  • [33] DATA-DRIVEN MODEL-FREE ITERATIVE LEARNING CONTROL USING REINFORCEMENT LEARNING
    Song, Bing
    Phan, Minh Q.
    Longman, Richard W.
    [J]. ASTRODYNAMICS 2018, PTS I-IV, 2019, 167 : 2579 - 2597
  • [34] Model-free MIMO control tuning of a chiller process using reinforcement learning
    Rosdahl, Christian
    Bernhardsson, B. O.
    Eisenhower, Bryan
    [J]. SCIENCE AND TECHNOLOGY FOR THE BUILT ENVIRONMENT, 2023, 29 (08) : 782 - 794
  • [35] Online model-free reinforcement learning for the automatic control of a flexible wing aircraft
    Abouheaf, Mohammed
    Gueaieb, Wail
    Lewis, Frank
    [J]. IET CONTROL THEORY AND APPLICATIONS, 2020, 14 (01): : 73 - 84
  • [36] Adaptive optics control with multi-agent model-free reinforcement learning
    Pou, B.
    Ferreira, F.
    Quinones, E.
    Gratadour, D.
    Martin, M.
    [J]. OPTICS EXPRESS, 2022, 30 (02) : 2991 - 3015
  • [37] Model-free Based Reinforcement Learning Control Strategy of Aircraft Attitude Systems
    Huang, Dingcui
    Hu, Jiangping
    Peng, Zhinan
    Chen, Bo
    Hao, Mingrui
    Ghosh, Bijoy Kumar
    [J]. 2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 743 - 748
  • [38] A model-free deep reinforcement learning approach for control of exoskeleton gait patterns
    Rose, Lowell
    Bazzocchi, Michael C. F.
    Nejat, Goldie
    [J]. ROBOTICA, 2022, 40 (07) : 2189 - 2214
  • [39] Model-free Data-driven Predictive Control Using Reinforcement Learning
    Sawant, Shambhuraj
    Reinhardt, Dirk
    Kordabad, Arash Bahari
    Gros, Sebastien
    [J]. 2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 4046 - 4052
  • [40] Control of a Wave Energy Converter Using Model-free Deep Reinforcement Learning
    Chen, Kemeng
    Huang, Xuanrui
    Lin, Zechuan
    Xiao, Xi
    [J]. 2024 UKACC 14TH INTERNATIONAL CONFERENCE ON CONTROL, CONTROL, 2024, : 1 - 6