Adaptive Observation-Based Efficient Reinforcement Learning for Uncertain Systems

被引:9
|
作者
Ran, Maopeng [1 ]
Xie, Lihua [1 ]
机构
[1] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore 639798, Singapore
基金
新加坡国家研究基金会;
关键词
Optimal control; Observers; Adaptive systems; Adaptation models; Uncertain systems; Estimation; Data models; Adaptive observer; concurrent learning (CL); optimal control; reinforcement learning (RL); uncertain systems; CONTINUOUS-TIME; PARAMETER-ESTIMATION; LINEAR-SYSTEMS; ITERATION;
D O I
10.1109/TNNLS.2021.3070852
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This article develops an adaptive observation-based efficient reinforcement learning (RL) approach for systems with uncertain drift dynamics. A novel concurrent learning adaptive extended observer (CL-AEO) is first designed to jointly estimate the system state and parameter. This observer has a two-time-scale structure and does not require any additional numerical techniques to calculate the state derivative information. The idea of concurrent learning (CL) is leveraged to use the recorded data, which leads to a relaxed verifiable excitation condition for the convergence of parameter estimation. Based on the estimated state and parameter provided by the CL-AEO, a simulation of experience-based RL scheme is developed to online approximate the optimal control policy. Rigorous theoretical analysis is given to show that the practical convergence of the system state to the origin and the developed policy to the ideal optimal policy can be achieved without the persistence of excitation (PE) condition. Finally, the effectiveness and superiority of the developed methodology are demonstrated via comparative simulations.
引用
收藏
页码:5492 / 5503
页数:12
相关论文
共 50 条
  • [21] Disturbance observer-based adaptive reinforcement learning for perturbed uncertain surface vessels
    Vu, Van Tu
    Pham, Thanh Loc
    Dao, Phuong Nam
    ISA TRANSACTIONS, 2022, 130 : 277 - 292
  • [22] Adaptive Critic Control With Knowledge Transfer for Uncertain Nonlinear Dynamical Systems: A Reinforcement Learning Approach
    Zhang, Liangju
    Zhang, Kun
    Xie, Xiang Peng
    Chadli, Mohammed
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2025, 22 : 6752 - 6761
  • [23] Reinforcement-Learning-Based Disturbance Rejection Control for Uncertain Nonlinear Systems
    Ran, Maopeng
    Li, Juncheng
    Xie, Lihua
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (09) : 9621 - 9633
  • [24] Reinforcement learning for control design of uncertain polytopic systems
    Oliveira, Pedro M.
    Palma, Jonathan M.
    Nepomuceno, Erivelton G.
    Lacerda, Marcio J.
    INFORMATION SCIENCES, 2023, 625 : 417 - 429
  • [25] Feedback Linearization for Uncertain Systems via Reinforcement Learning
    Westenbroek, Tyler
    Fridovich-Keil, David
    Mazumdar, Eric
    Arora, Shreyas
    Prabhu, Valmik
    Sastry, S. Shankar
    Tomlin, Claire J.
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 1364 - 1371
  • [26] Reinforcement adaptive fuzzy control for a class of nonlinear uncertain systems
    Kim, YH
    Lewis, FL
    Kwan, C
    NEW TECHNOLOGIES FOR COMPUTER CONTROL 2001, 2002, : 211 - 216
  • [27] Adaptive learning control of uncertain robotic systems
    Park, BH
    Kuc, TY
    Lee, JS
    INTERNATIONAL JOURNAL OF CONTROL, 1996, 65 (05) : 725 - 744
  • [28] Use of machine learning to shorten observation-based screening and diagnosis of autism
    Wall, D. P.
    Kosmicki, J.
    DeLuca, T. F.
    Harstad, E.
    Fusaro, V. A.
    TRANSLATIONAL PSYCHIATRY, 2012, 2 : e100 - e100
  • [29] Adaptive Multiagent Model Based on Reinforcement Learning for Distributed Generation Systems
    Divenyi, Daniel
    Dan, Andras
    2012 23RD INTERNATIONAL WORKSHOP ON DATABASE AND EXPERT SYSTEMS APPLICATIONS (DEXA), 2012, : 303 - 307
  • [30] Reinforcement Learning-Based Adaptive Modulation and Coding for Efficient Underwater Communications
    Su, Wei
    Lin, Jiamin
    Chen, Keyu
    Xiao, Liang
    En, Cheng
    IEEE ACCESS, 2019, 7 (67539-67550) : 67539 - 67550