Model-Free Recurrent Reinforcement Learning for AUV Horizontal Control

被引:7
|
作者
Huo, Yujia [1 ,2 ]
Li, Yiping [1 ]
Feng, Xisheng [1 ]
机构
[1] Chinese Acad Sci, Shenyang Inst Automat, State Key Lab Robot, Shenyang 110016, Peoples R China
[2] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
关键词
D O I
10.1088/1757-899X/428/1/012063
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, aiming at the problems of 2-DOF horizontal motion control with high precision for autonomous underwater vehicle(AUV) trajectory tracking tasks, deep reinforcement learning controllers are applied to these conditions. These control problems are considered as a POMDP (Partially Observable Markov Decision Process). Model-free reinforcement learning(RL) algorithms for continuous control mission based on Deterministic Policy Gradient(DPG) allows robots learn from received delayed rewards when interacting with environments. Recurrent neural networks LSTM (Long Short-Term Memory) are involved into the reinforcement learning algorithm. Through this deep reinforcement learning algorithm, AUVs learn from sequences of dynamic information. The horizontal trajectory tracking tasks are described by LOS method and the motion control are idealized as a SISO model. Tanh-estimators are presented as data normalization. Moreover, AUV horizontal trajectory tracking and motion control simulation results demonstrate this algorithm gets better accuracy compared with the PID method and other non-recurrent methods. Efforts show the efficiency and effectiveness of the improved deep reinforcement learning algorithm.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Model-Free Quantum Control with Reinforcement Learning
    Sivak, V. V.
    Eickbusch, A.
    Liu, H.
    Royer, B.
    Tsioutsios, I
    Devoret, M. H.
    [J]. PHYSICAL REVIEW X, 2022, 12 (01):
  • [2] Sliding mode heading control for AUV based on continuous hybrid model-free and model-based reinforcement learning
    Wang, Dianrui
    Shen, Yue
    Wan, Junhe
    Sha, Qixin
    Li, Guangliang
    Chen, Guanzhong
    He, Bo
    [J]. APPLIED OCEAN RESEARCH, 2022, 118
  • [3] Reinforcement learning based model-free optimized trajectory tracking strategy design for an AUV
    Duan, Kairong
    Fong, Simon
    Chen, C. L. Philip
    [J]. NEUROCOMPUTING, 2022, 469 : 289 - 297
  • [4] ODE-based Recurrent Model-free Reinforcement Learning for POMDPs
    Zhao, Xuanle
    Zhang, Duzhen
    Han, Liyuan
    Zhang, Tielin
    Xu, Bo
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [5] Model-free learning control of neutralization processes using reinforcement learning
    Syafiie, S.
    Tadeo, F.
    Martinez, E.
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2007, 20 (06) : 767 - 782
  • [6] Linear Quadratic Control Using Model-Free Reinforcement Learning
    Yaghmaie, Farnaz Adib
    Gustafsson, Fredrik
    Ljung, Lennart
    [J]. IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2023, 68 (02) : 737 - 752
  • [7] Model-Free Reinforcement Learning of Impedance Control in Stochastic Environments
    Stulp, Freek
    Buchli, Jonas
    Ellmer, Alice
    Mistry, Michael
    Theodorou, Evangelos A.
    Schaal, Stefan
    [J]. IEEE TRANSACTIONS ON AUTONOMOUS MENTAL DEVELOPMENT, 2012, 4 (04) : 330 - 341
  • [8] On Distributed Model-Free Reinforcement Learning Control With Stability Guarantee
    Mukherjee, Sayak
    Vu, Thanh Long
    [J]. IEEE CONTROL SYSTEMS LETTERS, 2021, 5 (05): : 1615 - 1620
  • [9] Model-Free Control for Soft Manipulators based on Reinforcement Learning
    You, Xuanke
    Zhang, Yixiao
    Chen, Xiaotong
    Liu, Xinghua
    Wang, Zhanchi
    Jiang, Hao
    Chen, Xiaoping
    [J]. 2017 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2017, : 2909 - 2915
  • [10] Model-Free Emergency Frequency Control Based on Reinforcement Learning
    Chen, Chunyu
    Cui, Mingjian
    Li, Fangxing
    Yin, Shengfei
    Wang, Xinan
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (04) : 2336 - 2346