Data-Driven Model-Free Tracking Reinforcement Learning Control with VRFT-based Adaptive Actor-Critic

被引:35
|
作者
Radac, Mircea-Bogdan [1 ]
Precup, Radu-Emil [1 ]
机构
[1] Politehn Univ Timisoara, Dept Automat & Appl Informat, Timisoara 300006, Romania
来源
APPLIED SCIENCES-BASEL | 2019年 / 9卷 / 09期
关键词
adaptive actor-critic; model-free control; data-driven control; reinforcement learning; approximate dynamic programming; output reference model tracking; multi-input multi-output systems; vertical tank systems; Virtual Reference Feedback Tuning; CONJUGATE-GRADIENT ALGORITHM; TRAJECTORY TRACKING; NEURAL-NETWORKS; CONTROL DESIGN; SYSTEMS; FEEDBACK; ITERATION;
D O I
10.3390/app9091807
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
This paper proposes a neural network (NN)-based control scheme in an Adaptive Actor-Critic (AAC) learning framework designed for output reference model tracking, as a representative deep-learning application. The control learning scheme is model-free with respect to the process model. AAC designs usually require an initial controller to start the learning process; however, systematic guidelines for choosing the initial controller are not offered in the literature, especially in a model-free manner. Virtual Reference Feedback Tuning (VRFT) is proposed for obtaining an initially stabilizing NN nonlinear state-feedback controller, designed from input-state-output data collected from the process in open-loop setting. The solution offers systematic design guidelines for initial controller design. The resulting suboptimal state-feedback controller is next improved under the AAC learning framework by online adaptation of a critic NN and a controller NN. The mixed VRFT-AAC approach is validated on a multi-input multi-output nonlinear constrained coupled vertical two-tank system. Discussions on the control system behavior are offered together with comparisons with similar approaches.
引用
收藏
页数:24
相关论文
共 50 条
  • [1] Actor-Critic Reinforcement Learning for Tracking Control in Robotics
    Pane, Yudha P.
    Nageshrao, Subramanya P.
    Babuska, Robert
    [J]. 2016 IEEE 55TH CONFERENCE ON DECISION AND CONTROL (CDC), 2016, : 5819 - 5826
  • [2] Optimized Adaptive Nonlinear Tracking Control Using Actor-Critic Reinforcement Learning Strategy
    Wen, Guoxing
    Chen, C. L. Philip
    Ge, Shuzhi Sam
    Yang, Hongli
    Liu, Xiaoguang
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2019, 15 (09) : 4969 - 4977
  • [3] Adaptive Assist-as-needed Control Based on Actor-Critic Reinforcement Learning
    Zhang, Yufeng
    Li, Shuai
    Nolan, Karen J.
    Zanotto, Damiano
    [J]. 2019 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2019, : 4066 - 4071
  • [4] Model-free Data-driven Predictive Control Using Reinforcement Learning
    Sawant, Shambhuraj
    Reinhardt, Dirk
    Kordabad, Arash Bahari
    Gros, Sebastien
    [J]. 2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 4046 - 4052
  • [5] DATA-DRIVEN MODEL-FREE ITERATIVE LEARNING CONTROL USING REINFORCEMENT LEARNING
    Song, Bing
    Phan, Minh Q.
    Longman, Richard W.
    [J]. ASTRODYNAMICS 2018, PTS I-IV, 2019, 167 : 2579 - 2597
  • [6] Adaptive Optimal Tracking Control of an Underactuated Surface Vessel Using Actor-Critic Reinforcement Learning
    Chen, Lin
    Dai, Shi-Lu
    Dong, Chao
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (06) : 7520 - 7533
  • [7] A World Model for Actor-Critic in Reinforcement Learning
    Panov, A. I.
    Ugadiarov, L. A.
    [J]. PATTERN RECOGNITION AND IMAGE ANALYSIS, 2023, 33 (03) : 467 - 477
  • [8] Model-Free VRFT-Based Tuning Method for PID Controllers
    Vrancic, Damir
    Oliveira, Paulo Moura
    Bistak, Pavol
    Huba, Mikulas
    [J]. MATHEMATICS, 2023, 11 (03)
  • [9] Network Congestion Control Algorithm Based on Actor-Critic Reinforcement Learning Model
    Xu, Tao
    Gong, Lina
    Zhang, Wei
    Li, Xuhong
    Wang, Xia
    Pan, Wenwen
    [J]. ADVANCES IN MATERIALS, MACHINERY, ELECTRONICS II, 2018, 1955
  • [10] Actor-Critic based Improper Reinforcement Learning
    Zaki, Mohammadi
    Mohan, Avinash
    Gopalan, Aditya
    Mannor, Shie
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,