Development of Push-Recovery control system for humanoid robots using deep reinforcement learning

被引:7
|
作者
Aslan, Emrah [1 ]
Arserim, Muhammet Ali [2 ]
Ucar, Aysegul [3 ]
机构
[1] Dicle Univ, Silvan Vocat Sch, Diyarbakir, Turkiye
[2] Dicle Univ, Engn Fac, Diyarbakir, Turkiye
[3] Firat Univ, Engn Fac, Elazig, Turkiye
关键词
Deep reinforcement learning; deep q network(DQN); double deep q network(DDQN); Humanoid robot; Robotis op2; Push-recovery; GENERATION;
D O I
10.1016/j.asej.2023.102167
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
This paper focuses on the push-recovery problem of bipedal humanoid robots affected by external forces and pushes. Since they are structurally unstable, balance is the most important problem in humanoid robots. Our purpose is to design and implement a completely independent push-recovery control system that can imitate the actions of a human. For humanoid robots to be able to stay in balance while standing or walking, and to prevent balance disorders that may be caused by external forces, an active balance control has been presented. Push-recovery controllers consist of three strategies: ankle strategy, hip strategy, and step strategy. These strategies are biomechanical responses that people show in cases of balance disorder. In our application, both simulation and real-world tests have been performed. The simulation tests of the study were carried out with 3D models in the Webots environment. Real-world tests were performed on the Robotis-OP2 humanoid robot. The gyroscope, accelerometer and motor data from the sensors in our robot were recorded and external pushing force was applied to the robot. The balance of the robot was achieved by using the recorded data and the ankle strategy. To make the robot completely autonomous, Deep Q Network (DQN) and Double Deep Q Network (DDQN) methods from Deep Reinforcement Learning (DPL) algorithms have been applied. The results obtained with the DDQN algorithm yielded 21.03% more successful results compared to the DQN algorithm. The results obtained in the real environment tests showed parallelism to the simulation results.(c) 2023 THE AUTHORS. Published by Elsevier BV on behalf of Faculty of Engineering, Ain Shams University. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/ by-nc-nd/4.0/).
引用
收藏
页数:11
相关论文
共 50 条
  • [41] A classifier system for reinforcement learning control of autonomous robots
    Kuroyama, K
    Svinin, MM
    Ueda, K
    [J]. INTELLIGENT AUTONOMOUS SYSTEMS: IAS-5, 1998, : 304 - 311
  • [42] Deep Reinforcement Learning for Humanoid Robot Behaviors
    Muzio, Alexandre F. V.
    Maximo, Marcos R. O. A.
    Yoneyama, Takashi
    [J]. Journal of Intelligent and Robotic Systems: Theory and Applications, 2022, 105 (01):
  • [43] Motivated Reinforcement Learning for Improved Head Actuation of Humanoid Robots
    Fountain, Jake
    Walker, Josiah
    Budden, David
    Mendes, Alexandre
    Chalup, Stephan K.
    [J]. ROBOCUP 2013: ROBOT WORLD CUP XVII, 2014, 8371 : 268 - 279
  • [44] Cross-overlapping Hierarchical Reinforcement Learning in Humanoid Robots
    Chen, Kuihan
    Liang, Zhiwei
    Liang, Wenzhao
    Zhou, Huijie
    Chen, Li
    Qin, Shiyan
    [J]. PROCEEDINGS OF THE 33RD CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2021), 2021, : 3340 - 3345
  • [45] A novel approach for humanoid push recovery using stereopsis
    Mahani, Mohammad-Ali Nikouei
    Jafari, Shahram
    Rahmatkhah, Hadi
    [J]. ROBOTICA, 2014, 32 (03) : 413 - 431
  • [46] Dynamic Fall Recovery Control for Legged Robots via Reinforcement Learning
    Li, Sicen
    Pang, Yiming
    Bai, Panju
    Hu, Shihao
    Wang, Liquan
    Wang, Gang
    [J]. BIOMIMETICS, 2024, 9 (04)
  • [47] Adaptive Power System Emergency Control Using Deep Reinforcement Learning
    Huang, Qiuhua
    Huang, Renke
    Hao, Weituo
    Tan, Jie
    Fan, Rui
    Huang, Zhenyu
    [J]. IEEE TRANSACTIONS ON SMART GRID, 2020, 11 (02) : 1171 - 1182
  • [48] Development of an HVAC system control method using weather forecasting data with deep reinforcement learning algorithms
    Shin, Minjae
    Kim, Sungsoo
    Kim, Youngjin
    Song, Ahhyun
    Kim, Yeeun
    Kim, Ha Young
    [J]. BUILDING AND ENVIRONMENT, 2024, 248
  • [49] Humanoid robot control based on reinforcement learning
    Iida, S
    Kuwayama, K
    Kanoh, M
    Kato, S
    Kunitachi, T
    Itoh, H
    [J]. PROCEEDINGS OF THE 2004 INTERNATIONAL SYMPOSIUM ON MICRO-NANOMECHATRONICS AND HUMAN SCIENCE, 2004, : 353 - 358
  • [50] Humanoid robot control based on reinforcement learning
    [J]. Iida, S. (iida@ics.nitech.ac.jp), IEEE Robotics and Automation Society; Nagoya University, Japan; City of Nagoya, Japan; Nagoya City Science Museum; Chubu Science and Technology Center (Institute of Electrical and Electronics Engineers Inc.):