Robotic Knee Tracking Control to Mimic the Intact Human Knee Profile Based on Actor-Critic Reinforcement Learning

被引:2
|
作者
Ruofan Wu [1 ]
Zhikai Yao [1 ]
Jennie Si [2 ,1 ]
He(Helen) Huang [2 ,3 ,4 ]
机构
[1] the School of Electrical, Computer and Energy Engineering, Arizona State University
[2] IEEE
[3] the Department of Biomedical Engineering, North Carolina State University
[4] the University of North Carolina at Chapel Hill
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP242 [机器人];
学科分类号
1111 ;
摘要
We address a state-of-the-art reinforcement learning(RL) control approach to automatically configure robotic prosthesis impedance parameters to enable end-to-end, continuous locomotion intended for transfemoral amputee subjects.Specifically, our actor-critic based RL provides tracking control of a robotic knee prosthesis to mimic the intact knee profile. This is a significant advance from our previous RL based automatic tuning of prosthesis control parameters which have centered on regulation control with a designer prescribed robotic knee profile as the target. In addition to presenting the tracking control algorithm based on direct heuristic dynamic programming(d HDP), we provide a control performance guarantee including the case of constrained inputs. We show that our proposed tracking control possesses several important properties, such as weight convergence of the learning networks, Bellman(sub)optimality of the cost-to-go value function and control input, and practical stability of the human-robot system. We further provide a systematic simulation of the proposed tracking control using a realistic human-robot system simulator, the Open Sim, to emulate how the d HDP enables level ground walking, walking on different terrains and at different paces. These results show that our proposed d HDP based tracking control is not only theoretically suitable, but also practically useful.
引用
收藏
页码:19 / 30
页数:12
相关论文
共 50 条
  • [1] Robotic Knee Tracking Control to Mimic the Intact Human Knee Profile Based on Actor-Critic Reinforcement Learning
    Wu, Ruofan
    Yao, Zhikai
    Si, Jennie
    Huang, He Helen
    IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2022, 9 (01) : 19 - 30
  • [2] Actor-Critic Reinforcement Learning for Tracking Control in Robotics
    Pane, Yudha P.
    Nageshrao, Subramanya P.
    Babuska, Robert
    2016 IEEE 55TH CONFERENCE ON DECISION AND CONTROL (CDC), 2016, : 5819 - 5826
  • [3] Actor-critic learning based PID control for robotic manipulators
    Nohooji, Hamed Rahimi
    Zaraki, Abolfazl
    Voos, Holger
    APPLIED SOFT COMPUTING, 2024, 151
  • [4] Actor-Critic based Improper Reinforcement Learning
    Zaki, Mohammadi
    Mohan, Avinash
    Gopalan, Aditya
    Mannor, Shie
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [5] Actor-Critic Reinforcement Learning for Control With Stability Guarantee
    Han, Minghao
    Zhang, Lixian
    Wang, Jun
    Pan, Wei
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (04) : 6217 - 6224
  • [6] Actor-Critic reinforcement learning based on prior knowledge
    Yang, Zhenyu, 1600, Transport and Telecommunication Institute, Lomonosova street 1, Riga, LV-1019, Latvia (18):
  • [7] Trajectory Tracking Control for Robotic Manipulator Based on Soft Actor-Critic and Generative Adversarial Imitation Learning
    Hu, Jintao
    Wang, Fujie
    Li, Xing
    Qin, Yi
    Guo, Fang
    Jiang, Ming
    BIOMIMETICS, 2024, 9 (12)
  • [8] Optimized Adaptive Nonlinear Tracking Control Using Actor-Critic Reinforcement Learning Strategy
    Wen, Guoxing
    Chen, C. L. Philip
    Ge, Shuzhi Sam
    Yang, Hongli
    Liu, Xiaoguang
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2019, 15 (09) : 4969 - 4977
  • [9] Adaptive Assist-as-needed Control Based on Actor-Critic Reinforcement Learning
    Zhang, Yufeng
    Li, Shuai
    Nolan, Karen J.
    Zanotto, Damiano
    2019 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2019, : 4066 - 4071
  • [10] Actor-critic reinforcement learning for the feedback control of a swinging chain
    Dengler, C.
    Lohmann, B.
    IFAC PAPERSONLINE, 2018, 51 (13): : 378 - 383