Adaptive Observation-Based Efficient Reinforcement Learning for Uncertain Systems

被引:9
|
作者
Ran, Maopeng [1 ]
Xie, Lihua [1 ]
机构
[1] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore 639798, Singapore
基金
新加坡国家研究基金会;
关键词
Optimal control; Observers; Adaptive systems; Adaptation models; Uncertain systems; Estimation; Data models; Adaptive observer; concurrent learning (CL); optimal control; reinforcement learning (RL); uncertain systems; CONTINUOUS-TIME; PARAMETER-ESTIMATION; LINEAR-SYSTEMS; ITERATION;
D O I
10.1109/TNNLS.2021.3070852
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This article develops an adaptive observation-based efficient reinforcement learning (RL) approach for systems with uncertain drift dynamics. A novel concurrent learning adaptive extended observer (CL-AEO) is first designed to jointly estimate the system state and parameter. This observer has a two-time-scale structure and does not require any additional numerical techniques to calculate the state derivative information. The idea of concurrent learning (CL) is leveraged to use the recorded data, which leads to a relaxed verifiable excitation condition for the convergence of parameter estimation. Based on the estimated state and parameter provided by the CL-AEO, a simulation of experience-based RL scheme is developed to online approximate the optimal control policy. Rigorous theoretical analysis is given to show that the practical convergence of the system state to the origin and the developed policy to the ideal optimal policy can be achieved without the persistence of excitation (PE) condition. Finally, the effectiveness and superiority of the developed methodology are demonstrated via comparative simulations.
引用
收藏
页码:5492 / 5503
页数:12
相关论文
共 50 条
  • [11] Reinforcement learning-based optimal control of uncertain nonlinear systems
    Garcia, Miguel
    Dong, Wenjie
    INTERNATIONAL JOURNAL OF CONTROL, 2024, 97 (12) : 2839 - 2850
  • [12] Robust Control of Uncertain Linear Systems Based on Reinforcement Learning Principles
    Xu, Dengguo
    Wang, Qinglin
    Li, Yuan
    IEEE ACCESS, 2019, 7 : 16431 - 16443
  • [13] Deep Reinforcement Learning for Adaptive Learning Systems
    Li, Xiao
    Xu, Hanchen
    Zhang, Jinming
    Chang, Hua-hua
    JOURNAL OF EDUCATIONAL AND BEHAVIORAL STATISTICS, 2023, 48 (02) : 220 - 243
  • [14] Adaptive Influence Blocking: Minimizing the Negative Spread by Observation-based Policies
    Shi, Qihao
    Wang, Can
    Ye, Deshi
    Chen, Jiawei
    Feng, Yan
    Chen, Chun
    2019 IEEE 35TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE 2019), 2019, : 1502 - 1513
  • [15] Wavelet Reduced Order Observer based Adaptive Tracking Control for a Class of Uncertain Nonlinear Systems using Reinforcement Learning
    Sharma, Manish
    Verma, Ajay
    INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2013, 11 (03) : 496 - 502
  • [16] Wavelet reduced order observer based adaptive tracking control for a class of uncertain nonlinear systems using reinforcement learning
    Manish Sharma
    Ajay Verma
    International Journal of Control, Automation and Systems, 2013, 11 : 496 - 502
  • [17] Adaptive Caching Policies for Chiplet Systems Based on Reinforcement Learning
    Yang, Chongyi
    Zhang, Zhendong
    Wang, Xiaohang
    Liu, Peng
    2023 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, ISCAS, 2023,
  • [18] Observer-Based Adaptive Optimized Control for Uncertain Cyclic Switched Nonlinear Systems: Reinforcement Learning Algorithm Approach
    Yan, Chengyuan
    Xia, Jianwei
    Park, Ju H.
    Sun, Wei
    Xie, Xiangpeng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2024, 71 (05) : 2203 - 2216
  • [19] Deep reinforcement learning based efficient access scheduling algorithm with an adaptive number of devices for federated learning IoT systems
    Guan, Zheng
    Wang, Zengwen
    Cai, Yu
    Wang, Xue
    INTERNET OF THINGS, 2023, 24
  • [20] Disturbance observer-based adaptive reinforcement learning for perturbed uncertain surface vessels
    Tu Vu, Van
    Pham, Thanh Loc
    Dao, Phuong Nam
    ISA Transactions, 2022, 130 : 277 - 292