A supervised Actor-Critic approach for adaptive cruise control

被引:58
|
作者
Zhao, Dongbin [1 ]
Wang, Bin [1 ]
Liu, Derong [1 ]
机构
[1] Chinese Acad Sci, Inst Automat, State Key Lab Management & Control Complex Syst, Beijing 100190, Peoples R China
基金
北京市自然科学基金; 中国国家自然科学基金;
关键词
Supervised reinforcement learning; Actor-Critic; Adaptive cruise control; Uniformly ultimate bounded; Neural networks; FEEDBACK-CONTROL; SYSTEMS; ACC;
D O I
10.1007/s00500-013-1110-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A novel supervised Actor-Critic (SAC) approach for adaptive cruise control (ACC) problem is proposed in this paper. The key elements required by the SAC algorithm namely Actor and Critic, are approximated by feed-forward neural networks respectively. The output of Actor and the state are input to Critic to approximate the performance index function. A Lyapunov stability analysis approach has been presented to prove the uniformly ultimate bounded property of the estimation errors of the neural networks. Moreover, we use the supervisory controller to pre-train Actor to achieve a basic control policy, which can improve the training convergence and success rate. We apply this method to learn an approximate optimal control policy for the ACC problem. Experimental results in several driving scenarios demonstrate that the SAC algorithm performs well, so it is feasible and effective for the ACC problem.
引用
收藏
页码:2089 / 2099
页数:11
相关论文
共 50 条
  • [31] Optimized Adaptive Nonlinear Tracking Control Using Actor-Critic Reinforcement Learning Strategy
    Wen, Guoxing
    Chen, C. L. Philip
    Ge, Shuzhi Sam
    Yang, Hongli
    Liu, Xiaoguang
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2019, 15 (09) : 4969 - 4977
  • [32] Adaptive Neural network control of a helicopter system with optimal observer and actor-critic design
    Hager, LvS
    Uren, K. R.
    van Schoor, G.
    van Rensburg, A. Janse
    [J]. NEUROCOMPUTING, 2018, 302 : 75 - 90
  • [33] Actor-Critic Reinforcement Learning for Control With Stability Guarantee
    Han, Minghao
    Zhang, Lixian
    Wang, Jun
    Pan, Wei
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (04): : 6217 - 6224
  • [34] Characterizing Motor Control of Mastication With Soft Actor-Critic
    Abdi, Amir H.
    Sagl, Benedikt
    Srungarapu, Venkata P.
    Stavness, Ian
    Prisman, Eitan
    Abolmaesumi, Purang
    Fels, Sidney
    [J]. FRONTIERS IN HUMAN NEUROSCIENCE, 2020, 14
  • [35] Variational actor-critic algorithms*,**
    Zhu, Yuhua
    Ying, Lexing
    [J]. ESAIM-CONTROL OPTIMISATION AND CALCULUS OF VARIATIONS, 2023, 29
  • [36] Quantization-Based Adaptive Actor-Critic Tracking Control With Tracking Error Constraints
    Fan, Quan-Yong
    Yang, Guang-Hong
    Ye, Dan
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (04) : 970 - 980
  • [37] Dual Variable Actor-Critic for Adaptive Safe Reinforcement Learning
    Lee, Junseo
    Heo, Jaeseok
    Kim, Dohyeong
    Lee, Gunmin
    Oh, Songhwai
    [J]. 2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2023, : 7568 - 7573
  • [38] Learning State Representation for Deep Actor-Critic Control
    Munk, Jelle
    Kober, Jens
    Babuska, Robert
    [J]. 2016 IEEE 55TH CONFERENCE ON DECISION AND CONTROL (CDC), 2016, : 4667 - 4673
  • [39] A Hessian Actor-Critic Algorithm
    Wang, Jing
    Paschalidis, Ioannis Ch
    [J]. 2014 IEEE 53RD ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2014, : 1131 - 1136
  • [40] Error controlled actor-critic
    Gao, Xingen
    Chao, Fei
    Zhou, Changle
    Ge, Zhen
    Yang, Longzhi
    Chang, Xiang
    Shang, Changjing
    Shen, Qiang
    [J]. INFORMATION SCIENCES, 2022, 612 : 62 - 74