Data-Driven Model-Free Tracking Reinforcement Learning Control with VRFT-based Adaptive Actor-Critic

被引:35
|
作者
Radac, Mircea-Bogdan [1 ]
Precup, Radu-Emil [1 ]
机构
[1] Politehn Univ Timisoara, Dept Automat & Appl Informat, Timisoara 300006, Romania
来源
APPLIED SCIENCES-BASEL | 2019年 / 9卷 / 09期
关键词
adaptive actor-critic; model-free control; data-driven control; reinforcement learning; approximate dynamic programming; output reference model tracking; multi-input multi-output systems; vertical tank systems; Virtual Reference Feedback Tuning; CONJUGATE-GRADIENT ALGORITHM; TRAJECTORY TRACKING; NEURAL-NETWORKS; CONTROL DESIGN; SYSTEMS; FEEDBACK; ITERATION;
D O I
10.3390/app9091807
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
This paper proposes a neural network (NN)-based control scheme in an Adaptive Actor-Critic (AAC) learning framework designed for output reference model tracking, as a representative deep-learning application. The control learning scheme is model-free with respect to the process model. AAC designs usually require an initial controller to start the learning process; however, systematic guidelines for choosing the initial controller are not offered in the literature, especially in a model-free manner. Virtual Reference Feedback Tuning (VRFT) is proposed for obtaining an initially stabilizing NN nonlinear state-feedback controller, designed from input-state-output data collected from the process in open-loop setting. The solution offers systematic design guidelines for initial controller design. The resulting suboptimal state-feedback controller is next improved under the AAC learning framework by online adaptation of a critic NN and a controller NN. The mixed VRFT-AAC approach is validated on a multi-input multi-output nonlinear constrained coupled vertical two-tank system. Discussions on the control system behavior are offered together with comparisons with similar approaches.
引用
收藏
页数:24
相关论文
共 50 条
  • [41] A Data-Driven Model-Reference Adaptive Control Approach Based on Reinforcement Learning
    Abouheaf, Mohammed
    Gueaieb, Wail
    Spinello, Davide
    Al-Sharhan, Salah
    [J]. 2021 IEEE INTERNATIONAL SYMPOSIUM ON ROBOTIC AND SENSORS ENVIRONMENTS (ROSE 2021), 2021,
  • [42] Speed Tracking Control via Online Continuous Actor-Critic learning
    Huang, Zhenhua
    Xu, Xin
    Sun, Zhenping
    Tan, Jun
    Qian, Lilin
    [J]. PROCEEDINGS OF THE 35TH CHINESE CONTROL CONFERENCE 2016, 2016, : 3172 - 3177
  • [43] Deep reinforcement learning-based model-free path planning and collision avoidance for UAVs: A soft actor-critic with hindsight experience replay approach
    Lee, Myoung Hoon
    Moon, Jun
    [J]. ICT EXPRESS, 2023, 9 (03): : 403 - 408
  • [44] Data-driven set-point tuning of model-free adaptive control
    Lin, Na
    Chi, Ronghu
    Liu, Yang
    Hou, Zhongsheng
    Huang, Biao
    [J]. INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2023, 33 (13) : 7667 - 7686
  • [45] Data-Driven Model-Free Adaptive Control of Twin Rotor Aerodynamic Systems
    Roman, Raul-Cristian
    Radac, Mircea-Bogdan
    Precup, Radu-Emil
    [J]. 2014 IEEE 9TH INTERNATIONAL SYMPOSIUM ON APPLIED COMPUTATIONAL INTELLIGENCE AND INFORMATICS (SACI), 2014, : 25 - 30
  • [46] Data-Driven Based Model-Free Adaptive Optimal Control Method for Hypersonic Morphing Vehicle
    Bao, Cunyu
    Wang, Peng
    Tang, Guojian
    [J]. IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS, 2023, 59 (04) : 3713 - 3725
  • [47] Data-Driven Model-Free Adaptive Control of Z-Source Inverters
    Asadi, Yasin
    Ahmadi, Amirhossein
    Mohammadi, Sasan
    Amani, Ali Moradi
    Marzband, Mousa
    Mohammadi-Ivatloo, Behnam
    [J]. SENSORS, 2021, 21 (22)
  • [48] Data-Driven Model-Free Adaptive Control based on a Novel Double Successive Projection Algorithm
    Liu, Shida
    Hou, Zhongsheng
    Li, Zhenxuan
    [J]. 2016 14TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION (ICARCV), 2016,
  • [49] Data-driven urban traffic model-free adaptive iterative learning control with traffic data dropout compensation
    Li, Dai
    Hou, Zhongsheng
    [J]. IET CONTROL THEORY AND APPLICATIONS, 2021, 15 (11): : 1533 - 1544
  • [50] Symmetric actor-critic deep reinforcement learning for cascade quadrotor flight control
    Han, Haoran
    Cheng, Jian
    Xi, Zhilong
    Lv, Maolong
    [J]. NEUROCOMPUTING, 2023, 559