Robust control under worst-case uncertainty for unknown nonlinear systems using modified reinforcement learning

被引:32
|
作者
Perrusquia, Adolfo [1 ]
Yu, Wen [1 ]
机构
[1] CINVESTAV IPN, Natl Polytech Inst, Dept Automat Control, Ave IPN 2508, Mexico City 07360, DF, Mexico
关键词
k-nearest neighbors; double estimator; overestimation; robust reward; state-action space; worst-case uncertainty; POLICY; STABILIZATION; DESIGN;
D O I
10.1002/rnc.4911
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Reinforcement learning (RL) is an effectivemethod for the design of robust controllers of unknown nonlinear systems. Normal RLs for robust control, such as actor-critic (AC) algorithms, depend on the estimation accuracy. Uncertainty in the worst case requires a large state-action space, this causes overestimation and computational problems. In this article, the RL method is modified with the k-nearest neighbor and the double Q-learning algorithm. The modified RL does not need the neural estimator as AC and can stabilize the unknown nonlinear system under the worst-case uncertainty. The convergence property of the proposed RL method is analyzed. The simulations and the experimental results show that our modified RLs are much more robust compared with the classic controllers, such as the proportional-integral-derivative, the sliding mode, and the optimal linear quadratic regulator controllers.
引用
收藏
页码:2920 / 2936
页数:17
相关论文
共 50 条
  • [21] Monotonic Quantile Network for Worst-Case Offline Reinforcement Learning
    Bai, Chenjia
    Xiao, Ting
    Zhu, Zhoufan
    Wang, Lingxiao
    Zhou, Fan
    Garg, Animesh
    He, Bin
    Liu, Peng
    Wang, Zhaoran
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (07) : 8954 - 8968
  • [22] Simultaneous process synthesis and control design under uncertainty: A worst-case performance approach
    Sanchez-Sanchez, Kelvyn
    Ricardez-Sandoval, Luis
    [J]. AICHE JOURNAL, 2013, 59 (07) : 2497 - 2514
  • [23] Worst-Case Analysis of Complex Nonlinear Flight Control Designs Using Deep Q-Learning
    Braun, David
    Marb, Michael M.
    Angelov, Jorg
    Wechner, Maximilian
    Holzapfel, Florian
    [J]. JOURNAL OF GUIDANCE CONTROL AND DYNAMICS, 2023, 46 (07) : 1365 - 1377
  • [24] Neural H2 Control Using Reinforcement Learning for Unknown Nonlinear Systems
    Perrusquia, Adolfo
    Yu, Wen
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [25] Worst-case robust design optimization under distributional assumptions
    Padulo, Mattia
    Guenov, Marin D.
    [J]. INTERNATIONAL JOURNAL FOR NUMERICAL METHODS IN ENGINEERING, 2011, 88 (08) : 797 - 816
  • [26] A worst-case approach to optimal tracking control with robust performance
    Toivonen, HT
    Pensar, J
    [J]. INTERNATIONAL JOURNAL OF CONTROL, 1996, 65 (01) : 17 - 32
  • [27] Echo State Network-Based Robust Tracking Control for Unknown Constrained Nonlinear Systems by Using Integral Reinforcement Learning
    Liu, Chong
    Li, Yalun
    Duan, Zhongxing
    Chu, Zhousheng
    Ma, Zongfang
    [J]. IEEE ACCESS, 2024, 12 : 15133 - 15144
  • [28] Robust Adaptive Model Predictive Control with Worst-Case Cost
    Parsi, Anilkumar
    Iannelli, Andrea
    Yin, Mingzhou
    Khosravi, Mohammad
    Smith, Roy S.
    [J]. IFAC PAPERSONLINE, 2020, 53 (02): : 4222 - 4227
  • [29] Optimistic posterior sampling for reinforcement learning: worst-case regret bounds
    Agrawal, Shipra
    Jia, Randy
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [30] Robust assortment optimization using worst-case CVaR under the multinomial logit model
    Li, Xiaolong
    Ke, Jiannan
    [J]. OPERATIONS RESEARCH LETTERS, 2019, 47 (05) : 452 - 457