Neural Network Model-Based Reinforcement Learning Control for AUV 3-D Path Following

被引:13
|
作者
Ma, Dongfang [1 ,2 ]
Chen, Xi [1 ,3 ]
Ma, Weihao [1 ,3 ]
Zheng, Huarong [1 ,4 ]
Qu, Fengzhong [1 ,2 ]
机构
[1] Zhejiang Univ, Inst Marine Sensing & Networking, Hangzhou 310058, Peoples R China
[2] Zhejiang Univ, Hainan Inst, Sanya 813099, Peoples R China
[3] Minist Educ, Engn Res Ctr Ocean Sensing Technol & Equipment, Zhoushan 316021, Peoples R China
[4] Key Lab Ocean Observat Imaging Testbed Zhejiang P, Zhoushan 316021, Peoples R China
来源
基金
中国国家自然科学基金;
关键词
Mathematical models; Training; Neural networks; Task analysis; Intelligent vehicles; Heuristic algorithms; Adaptation models; Path following; autonomous underwater vehicles (AUVs); reinforcement learning; neural network model; state transition function; VEHICLES; TRACKING;
D O I
10.1109/TIV.2023.3282681
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Autonomous underwater vehicles (AUVs) have become important tools in the ocean exploration and have drawn considerable attention. Precise control for AUVs is the prerequisite to effectively execute underwater tasks. However, the classical control methods such as model predictive control (MPC) rely heavily on the dynamics model of the controlled system which is difficult to obtain for AUVs. To address this issue, a new reinforcement learning (RL) framework for AUV path-following control is proposed in this article. Specifically, we propose a novel actor-model-critic (AMC) architecture integrating a neural network model with the traditional actor-critic architecture. The neural network model is designed to learn the state transition function to explore the spatio-temporal change patterns of the AUV as well as the surrounding environment. Based on the AMC architecture, a RL-based controller agent named ModelPPO is constructed to control the AUV. With the required sailing speed achieved by a traditional proportional-integral (PI) controller, ModelPPO can control the rudder and elevator fins so that the AUV follows the desired path. Finally, a simulation platform is built to evaluate the performance of the proposed method that is compared with MPC and other RL-based methods. The obtained results demonstrate that the proposed method can achieve better performance than other methods, which demonstrate the great potential of the advanced artificial intelligence methods in solving the traditional motion control problems for intelligent vehicles.
引用
收藏
页码:893 / 904
页数:12
相关论文
共 50 条
  • [1] Neural-Network-Based Reinforcement Learning Control for Path Following of Underactuated Ships
    Zhang Lixing
    Qiao Lei
    Chen Jianliang
    Zhang Weidong
    PROCEEDINGS OF THE 35TH CHINESE CONTROL CONFERENCE 2016, 2016, : 5786 - 5791
  • [2] Model-Based Learning Network for 3-D Localization in mmWave Communications
    Yang, Jie
    Jin, Shi
    Wen, Chao-Kai
    Guo, Jiajia
    Matthaiou, Michail
    Gao, Bo
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2021, 20 (08) : 5449 - 5466
  • [3] Path following method for AUV based on Q-Learning and RBF neural network
    Li, Zeyu
    Liu, Weidong
    Li, Le
    Zhang, Wenbo
    Guo, Liwei
    Xibei Gongye Daxue Xuebao/Journal of Northwestern Polytechnical University, 2021, 39 (03): : 477 - 483
  • [4] A NOVEL 3-D BIO-INSPIRED NEURAL NETWORK MODEL FOR THE PATH PLANNING OF AN AUV IN UNDERWATER ENVIRONMENTS
    Yan, Mingzhong
    Zhu, Daqi
    Yang, Simon X.
    INTELLIGENT AUTOMATION AND SOFT COMPUTING, 2013, 19 (04): : 555 - 566
  • [5] Efficient Neural Network Pruning Using Model-Based Reinforcement Learning
    Bencsik, Blanka
    Szemenyei, Marton
    2022 INTERNATIONAL SYMPOSIUM ON MEASUREMENT AND CONTROL IN ROBOTICS (ISMCR), 2022, : 130 - 137
  • [6] Synthesizing Neural Network Controllers with Probabilistic Model-Based Reinforcement Learning
    Higuera, Juan Camilo Gamboa
    Meger, David
    Dudek, Gregory
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 2538 - 2544
  • [7] Sliding mode heading control for AUV based on continuous hybrid model-free and model-based reinforcement learning
    Wang, Dianrui
    Shen, Yue
    Wan, Junhe
    Sha, Qixin
    Li, Guangliang
    Chen, Guanzhong
    He, Bo
    APPLIED OCEAN RESEARCH, 2022, 118
  • [8] Model-Based Reinforcement Learning For Robot Control
    Li, Xiang
    Shang, Weiwei
    Cong, Shuang
    2020 5TH INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS AND MECHATRONICS (ICARM 2020), 2020, : 300 - 305
  • [9] Control Approach Combining Reinforcement Learning and Model-Based Control
    Okawa, Yoshihiro
    Sasaki, Tomotake
    Iwane, Hidenao
    2019 12TH ASIAN CONTROL CONFERENCE (ASCC), 2019, : 1419 - 1424
  • [10] Multiple model-based reinforcement learning for nonlinear control
    Samejima, K
    Katagiri, K
    Doya, K
    Kawato, M
    ELECTRONICS AND COMMUNICATIONS IN JAPAN PART III-FUNDAMENTAL ELECTRONIC SCIENCE, 2006, 89 (09): : 54 - 69