Sliding mode heading control for AUV based on continuous hybrid model-free and model-based reinforcement learning

被引:16
|
作者
Wang, Dianrui [1 ]
Shen, Yue [1 ]
Wan, Junhe [1 ]
Sha, Qixin [1 ]
Li, Guangliang [1 ]
Chen, Guanzhong [1 ]
He, Bo [1 ]
机构
[1] Ocean Univ China, Sch Informat Sci & Engn, Qingdao 266000, Shandong, Peoples R China
基金
中国国家自然科学基金;
关键词
Autonomous underwater vehicle (AUV); Model-based reinforcement learning; Model-free reinforcement learning; Deterministic policy gradient (DPG); Sliding mode control (SMC); NONLINEAR-SYSTEMS; ADAPTIVE-CONTROL; PID CONTROL; DESIGN;
D O I
10.1016/j.apor.2021.102960
中图分类号
P75 [海洋工程];
学科分类号
0814 ; 081505 ; 0824 ; 082401 ;
摘要
For autonomous underwater vehicles (AUVs), control over AUV heading is of key importance to enable highperformance locomotion control. In this study, the heading control is achieved by using the robust sliding mode control (SMC) method. The performance of the controller can be seriously affected by its parameters. However, it is time-consuming and labor-intensive to manually adjust the parameters. Most of the existing methods rely on the accurate AUV model or prior knowledge, which are difficult to obtain. Therefore, this study is concerned with the problem of automatically tuning the SMC parameters through reinforcement learning (RL). First, an AUV dynamic model with and without current influence was successfully established. Second, a continuous hybrid Model-based Model-free (MbMf) RL method based on the deterministic policy gradient was introduced and explained. Then, the framework for tuning the parameters of SMC by the RL method was described. Finally, to demonstrate the robustness and effectiveness of our approach, extensive numerical simulations were conducted on the established AUV model. The results show that our method can automatically tune the SMC parameters. The performance is more effective than SMC with fixed parameters or SMC with a purely model-free learner.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] Reinforcement learning based model-free optimized trajectory tracking strategy design for an AUV
    Duan, Kairong
    Fong, Simon
    Chen, C. L. Philip
    NEUROCOMPUTING, 2022, 469 : 289 - 297
  • [22] Predictive representations can link model-based reinforcement learning to model-free mechanisms
    Russek, Evan M.
    Momennejad, Ida
    Botvinick, Matthew M.
    Gershman, Samuel J.
    Daw, Nathaniel D.
    PLOS COMPUTATIONAL BIOLOGY, 2017, 13 (09)
  • [23] Combining Model-Based and Model-Free Updates for Trajectory-Centric Reinforcement Learning
    Chebotar, Yevgen
    Hausman, Karol
    Zhang, Marvin
    Sukhatme, Gaurav
    Schaal, Stefan
    Levine, Sergey
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [24] Extraversion differentiates between model-based and model-free strategies in a reinforcement learning task
    Skatova, Anya
    Chan, Patricia A.
    Daw, Nathaniel D.
    FRONTIERS IN HUMAN NEUROSCIENCE, 2013, 7
  • [25] Dyna-style Model-based reinforcement learning with Model-Free Policy Optimization
    Dong, Kun
    Luo, Yongle
    Wang, Yuxin
    Liu, Yu
    Qu, Chengeng
    Zhang, Qiang
    Cheng, Erkang
    Sun, Zhiyong
    Song, Bo
    KNOWLEDGE-BASED SYSTEMS, 2024, 287
  • [26] The modulation of acute stress on model-free and model-based reinforcement learning in gambling disorder
    Wyckmans, Florent
    Banerjee, Nilosmita
    Saeremans, Melanie
    Otto, Ross
    Kornreich, Charles
    Vanderijst, Laetitia
    Gruson, Damien
    Carbone, Vincenzo
    Bechara, Antoine
    Buchanan, Tony
    Noel, Xavier
    JOURNAL OF BEHAVIORAL ADDICTIONS, 2022, 11 (03) : 831 - 844
  • [27] Model-based decision making and model-free learning
    Drummond, Nicole
    Niv, Yael
    CURRENT BIOLOGY, 2020, 30 (15) : R860 - R865
  • [28] Model-Free and Model-Based Active Learning for Regression
    O'Neill, Jack
    Delany, Sarah Jane
    MacNamee, Brian
    ADVANCES IN COMPUTATIONAL INTELLIGENCE SYSTEMS, 2017, 513 : 375 - 386
  • [29] Model-based Reinforcement Learning for Continuous Control with Posterior Sampling
    Fan, Ying
    Ming, Yifei
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [30] Reliance on model-based and model-free control in obesity
    Janssen, Lieneke K.
    Mahner, Florian P.
    Schlagenhauf, Florian
    Deserno, Lorenz
    Horstmann, Annette
    SCIENTIFIC REPORTS, 2020, 10 (01)