Adaptive Control for Linearizable Systems Using On-Policy Reinforcement Learning

被引:0
|
作者
Westenbroek, Tyler [1 ,2 ]
Mazumdar, Eric [1 ,2 ]
Fridovich-Keil, David [1 ,2 ]
Prabhu, Valmik [1 ,2 ]
Tomlin, Claire J. [1 ,2 ]
Sastry, S. Shankar [1 ,2 ]
机构
[1] Dept Elect Engn, Berkeley, CA 94720 USA
[2] Univ Calif Berkeley, Berkeley, CA 94720 USA
关键词
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper proposes a framework for adaptively learning a feedback linearization-based tracking controller for an unknown system using discrete-time model-free policy-gradient parameter update rules. The primary advantage of the scheme over standard model-reference adaptive control techniques is that it does not require the learned inverse model to be invertible at all instances of time. This enables the use of general function approximators to approximate the linearizing controller for the system without having to worry about singularities. The overall learning system is stochastic, due to the random nature of the policy gradient updates, thus we combine analysis techniques commonly employed in the machine learning literature alongside stability arguments from adaptive control to demonstrate that with high probability the tracking and parameter errors concentrate near zero, under a standard persistency of excitation condition. A simulated example of a double pendulum demonstrates the utility of the proposed theory.
引用
收藏
页码:118 / 125
页数:8
相关论文
共 50 条
  • [21] Model Predictive Control via On-Policy Imitation Learning
    Ahn, Kwangjun
    Mhammedi, Zakaria
    Mania, Horia
    Hong, Zhang-Wei
    Jadbabaie, Ali
    [J]. LEARNING FOR DYNAMICS AND CONTROL CONFERENCE, VOL 211, 2023, 211
  • [22] Robust On-Policy Sampling for Data-Efficient Policy Evaluation in Reinforcement Learning
    Zhong, Rujie
    Zhang, Duohan
    Schafer, Lukas
    Albrecht, Stefano V.
    Hanna, Josiah P.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [23] Reinforcement learning to adaptive control of nonlinear systems
    Hwang, KS
    Tan, SW
    Tsai, MC
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2003, 33 (03): : 514 - 521
  • [24] Robust adaptive sliding control of linearizable systems
    Xu, HJ
    Mirmirani, M
    Ioannou, PA
    Boussalis, HR
    [J]. PROCEEDINGS OF THE 2001 AMERICAN CONTROL CONFERENCE, VOLS 1-6, 2001, : 4351 - 4356
  • [25] Fuzzy direct adaptive control of linearizable systems
    Xie, WF
    Rad, AB
    [J]. ADVANCES IN DYNAMICS, INSTRUMENTATION AND CONTROL, 2004, : 168 - 177
  • [26] Adaptive composite frequency control of power systems using reinforcement learning
    Mu, Chaoxu
    Wang, Ke
    Ma, Shiqian
    Chong, Zhiqiang
    Ni, Zhen
    [J]. CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY, 2022, 7 (04) : 671 - 684
  • [27] ADAPTIVE-CONTROL OF LINEARIZABLE SYSTEMS - COMMENTS
    SCHWARTZ, CA
    MAREELS, IMY
    [J]. IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 1992, 37 (05) : 698 - 701
  • [28] An adaptive feedback control of linearizable chaotic systems
    Feki, M
    [J]. CHAOS SOLITONS & FRACTALS, 2003, 15 (05) : 883 - 890
  • [29] Two novel on-policy reinforcement learning algorithms based on TD(λ)-methods
    Wiering, Marco A.
    van Hasselt, Hado
    [J]. 2007 IEEE INTERNATIONAL SYMPOSIUM ON APPROXIMATE DYNAMIC PROGRAMMING AND REINFORCEMENT LEARNING, 2007, : 280 - +
  • [30] Practical Critic Gradient based Actor Critic for On-Policy Reinforcement Learning
    Gurumurthy, Swaminathan
    Manchester, Zachary
    Kolter, J. Zico
    [J]. LEARNING FOR DYNAMICS AND CONTROL CONFERENCE, VOL 211, 2023, 211