Neural Network Model-Based Reinforcement Learning Control for AUV 3-D Path Following

被引:27
|
作者
Ma, Dongfang [1 ,2 ]
Chen, Xi [1 ,3 ]
Ma, Weihao [1 ,3 ]
Zheng, Huarong [1 ,4 ]
Qu, Fengzhong [1 ,2 ]
机构
[1] Zhejiang Univ, Inst Marine Sensing & Networking, Hangzhou 310058, Peoples R China
[2] Zhejiang Univ, Hainan Inst, Sanya 813099, Peoples R China
[3] Minist Educ, Engn Res Ctr Ocean Sensing Technol & Equipment, Zhoushan 316021, Peoples R China
[4] Key Lab Ocean Observat Imaging Testbed Zhejiang P, Zhoushan 316021, Peoples R China
来源
基金
中国国家自然科学基金;
关键词
Mathematical models; Training; Neural networks; Task analysis; Intelligent vehicles; Heuristic algorithms; Adaptation models; Path following; autonomous underwater vehicles (AUVs); reinforcement learning; neural network model; state transition function; VEHICLES; TRACKING;
D O I
10.1109/TIV.2023.3282681
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Autonomous underwater vehicles (AUVs) have become important tools in the ocean exploration and have drawn considerable attention. Precise control for AUVs is the prerequisite to effectively execute underwater tasks. However, the classical control methods such as model predictive control (MPC) rely heavily on the dynamics model of the controlled system which is difficult to obtain for AUVs. To address this issue, a new reinforcement learning (RL) framework for AUV path-following control is proposed in this article. Specifically, we propose a novel actor-model-critic (AMC) architecture integrating a neural network model with the traditional actor-critic architecture. The neural network model is designed to learn the state transition function to explore the spatio-temporal change patterns of the AUV as well as the surrounding environment. Based on the AMC architecture, a RL-based controller agent named ModelPPO is constructed to control the AUV. With the required sailing speed achieved by a traditional proportional-integral (PI) controller, ModelPPO can control the rudder and elevator fins so that the AUV follows the desired path. Finally, a simulation platform is built to evaluate the performance of the proposed method that is compared with MPC and other RL-based methods. The obtained results demonstrate that the proposed method can achieve better performance than other methods, which demonstrate the great potential of the advanced artificial intelligence methods in solving the traditional motion control problems for intelligent vehicles.
引用
收藏
页码:893 / 904
页数:12
相关论文
共 50 条
  • [21] Neural Model Extraction for Model-Based Control of a Neural Network Forward Model
    Ikemoto S.
    Takahara K.
    Kumi T.
    Hosoda K.
    SN Computer Science, 2021, 2 (1)
  • [22] Systematic Model-based Design of a Reinforcement Learning-based Neural Adaptive Cruise Control System
    Yarom, Or Aviv
    Fritz, Jannis
    Lange, Florian
    Liu-Henke, Xiaobo
    ICAART: PROCEEDINGS OF THE 14TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE - VOL 3, 2022, : 889 - 896
  • [23] Fuzzy Sliding Mode Control Based on RBF Neural Network for AUV Path Tracking
    Wang, Xiang
    Zhang, Yonglin
    Xue, Zhouzhou
    INTELLIGENT ROBOTICS AND APPLICATIONS, ICIRA 2019, PT II, 2019, 11741 : 637 - 648
  • [24] Fuzzy Neural Network Hybrid Learning Control on AUV
    Zhao, Jing
    Han, Zhaolin
    Fang, Yuanyuan
    AUTOMATION EQUIPMENT AND SYSTEMS, PTS 1-4, 2012, 468-471 : 1732 - 1735
  • [25] Fault Tolerant Control combining Reinforcement Learning and Model-based Control
    Bhan, Luke
    Quinones-Grueiro, Marcos
    Biswas, Gautam
    5TH CONFERENCE ON CONTROL AND FAULT-TOLERANT SYSTEMS (SYSTOL 2021), 2021, : 31 - 36
  • [26] Path Tracking Control and Identification of Tire Parameters using On-line Model-based Reinforcement Learning
    Kim, Taewan
    Kim, H. Jin
    2016 16TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS), 2016, : 215 - 219
  • [27] Model-Based Reinforcement Learning for Advanced Adaptive Cruise Control: A Hybrid Car Following Policy
    Yavas, Ugur
    Kumbasar, Tufan
    Ure, Naztm Kemal
    2022 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2022, : 1466 - 1472
  • [28] Path following of underactuated surface ships based on model predictive control with neural network
    Li, Ronghui
    Huang, Ji
    Pan, Xinxiang
    Hu, Qionglei
    Huang, Zhenkai
    INTERNATIONAL JOURNAL OF ADVANCED ROBOTIC SYSTEMS, 2020, 17 (04)
  • [29] Model predictive path following control of underwater vehicle based on RBF neural network
    Guo L.
    Gao J.
    Jiao H.
    Song Y.
    Chen Y.
    Pan G.
    Xibei Gongye Daxue Xuebao/Journal of Northwestern Polytechnical University, 2023, 41 (05): : 871 - 877
  • [30] Neural Network Dynamics for Model-Based Deep Reinforcement Learning with Model-Free Fine-Tuning
    Nagabandi, Anusha
    Kahn, Gregory
    Fearing, Ronald S.
    Levine, Sergey
    2018 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2018, : 7579 - 7586