Robotic Knee Tracking Control to Mimic the Intact Human Knee Profile Based on Actor-Critic Reinforcement Learning

被引:2
|
作者
Ruofan Wu [1 ]
Zhikai Yao [1 ]
Jennie Si [2 ,1 ]
He(Helen) Huang [2 ,3 ,4 ]
机构
[1] the School of Electrical, Computer and Energy Engineering, Arizona State University
[2] IEEE
[3] the Department of Biomedical Engineering, North Carolina State University
[4] the University of North Carolina at Chapel Hill
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP242 [机器人];
学科分类号
1111 ;
摘要
We address a state-of-the-art reinforcement learning(RL) control approach to automatically configure robotic prosthesis impedance parameters to enable end-to-end, continuous locomotion intended for transfemoral amputee subjects.Specifically, our actor-critic based RL provides tracking control of a robotic knee prosthesis to mimic the intact knee profile. This is a significant advance from our previous RL based automatic tuning of prosthesis control parameters which have centered on regulation control with a designer prescribed robotic knee profile as the target. In addition to presenting the tracking control algorithm based on direct heuristic dynamic programming(d HDP), we provide a control performance guarantee including the case of constrained inputs. We show that our proposed tracking control possesses several important properties, such as weight convergence of the learning networks, Bellman(sub)optimality of the cost-to-go value function and control input, and practical stability of the human-robot system. We further provide a systematic simulation of the proposed tracking control using a realistic human-robot system simulator, the Open Sim, to emulate how the d HDP enables level ground walking, walking on different terrains and at different paces. These results show that our proposed d HDP based tracking control is not only theoretically suitable, but also practically useful.
引用
收藏
页码:19 / 30
页数:12
相关论文
共 50 条
  • [31] Optimal synchronized control of nonlinear coupled harmonic oscillators based on actor-critic reinforcement learning
    Gu, Zhiyang
    Fan, Chengli
    Yu, Dengxiu
    Wang, Zhen
    NONLINEAR DYNAMICS, 2023, 111 (22) : 21051 - 21064
  • [32] Symmetric actor-critic deep reinforcement learning for cascade quadrotor flight control
    Han, Haoran
    Cheng, Jian
    Xi, Zhilong
    Lv, Maolong
    NEUROCOMPUTING, 2023, 559
  • [33] Data-Driven Model-Free Tracking Reinforcement Learning Control with VRFT-based Adaptive Actor-Critic
    Radac, Mircea-Bogdan
    Precup, Radu-Emil
    APPLIED SCIENCES-BASEL, 2019, 9 (09):
  • [34] Model-Based Actor-Critic Learning for Optimal Tracking Control of Robots With Input Saturation
    Zhao, Xingwei
    Tao, Bo
    Qian, Lu
    Ding, Han
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2021, 68 (06) : 5046 - 5056
  • [35] Dynamic Content Caching Based on Actor-Critic Reinforcement Learning for IoT Systems
    Lai, Lifeng
    Zheng, Fu-Chun
    Wen, Wanli
    Luo, Jingjing
    Li, Ge
    2022 IEEE 96TH VEHICULAR TECHNOLOGY CONFERENCE (VTC2022-FALL), 2022,
  • [36] REINFORCEMENT LEARNING-BASED ADAPTIVE MOTION CONTROL FOR AUTONOMOUS VEHICLES VIA ACTOR-CRITIC STRUCTURE
    Wang, Honghai
    Wei, Liangfen
    Wang, Xianchao
    He, Shuping
    DISCRETE AND CONTINUOUS DYNAMICAL SYSTEMS-SERIES S, 2024, 17 (09): : 2894 - 2911
  • [37] Quantization-Based Adaptive Actor-Critic Tracking Control With Tracking Error Constraints
    Fan, Quan-Yong
    Yang, Guang-Hong
    Ye, Dan
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (04) : 970 - 980
  • [38] Benchmarking Actor-Critic Deep Reinforcement Learning Algorithms for Robotics Control With Action Constraints
    Kasaura, Kazumi
    Miura, Shuwa
    Kozuno, Tadashi
    Yonetani, Ryo
    Hoshino, Kenta
    Hosoe, Yohei
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (08) : 4449 - 4456
  • [39] DASH Live Video Streaming Control Using Actor-Critic Reinforcement Learning Method
    Wei, Bo
    Song, Hang
    Quang Ngoc Nguyen
    Katto, Jiro
    MOBILE NETWORKS AND MANAGEMENT, MONAMI 2021, 2022, 418 : 17 - 24
  • [40] Online Human Training of a Myoelectric Prosthesis Controller via Actor-Critic Reinforcement Learning
    Pilarski, Patrick M.
    Dawson, Michael R.
    Degris, Thomas
    Fahimi, Farbod
    Carey, Jason P.
    Sutton, Richard S.
    2011 IEEE INTERNATIONAL CONFERENCE ON REHABILITATION ROBOTICS (ICORR), 2011,