Adaptive Control of a Marine Vessel Based on Reinforcement Learning

被引:0
|
作者
Yin, Zhao [1 ]
He, Wei [1 ]
Sun, Changyin [2 ]
Li, Guang [3 ]
Yang, Chenguang [4 ]
机构
[1] Univ Sci & Technol Beijing, Sch Automat & Elect Engn, Beijing 100083, Peoples R China
[2] Southeast Univ, Sch Automat, Nanjing 210096, Jiangsu, Peoples R China
[3] Queen Mary Univ London, Mile End Rd, London E1 4NS, England
[4] South China Univ Technol, Coll Automat Sci & Engn, Key Lab Autonomous Syst & Networked Control, Guangzhou 510640, Guangdong, Peoples R China
基金
英国工程与自然科学研究理事会; 中国国家自然科学基金;
关键词
Reinforcement Learning; Critic Neural Networks; Actor Neural Networks; Lyapunov Method; Marine Vessel; NEURAL-NETWORK CONTROL; TRAJECTORY TRACKING; SURFACE VESSELS; CONTINUOUS-TIME; DELAY SYSTEMS;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, our main goal is to solve optimal control problem by using reinforcement learning (RL) algorithm for marine surface vessel system with known dynamic. And this algorithm is an optimal control algorithm based on policy iteration (PI), and it can obtain the suitable approximations of cost function and the optimized control policy. There are two neural networks (NNs), where critic NN aims to estimate the cost-to-go and actor NN is utilized to design suitable input controller and minimize the tracking error. A novel tuning method is given for critic NN and actor NN. The stability and convergence are proven by Lyapunov's direct method. Finally, the numerical simulations are conducted to demonstrate the feasibility and superiority of presented algorithm.
引用
收藏
页码:2735 / 2740
页数:6
相关论文
共 50 条
  • [41] Adaptive reinforcement learning system for linearization control
    Hwang, KS
    Chao, HJ
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2000, 47 (05) : 1185 - 1188
  • [42] Deep Reinforcement Learning Based Tracking Control of an Autonomous Surface Vessel in Natural Waters
    Wang, Wei
    Cao, Xiaojing
    Gonzalez-Garcia, Alejandro
    Yin, Lianhao
    Hagemann, Niklas
    Qiao, Yuanyuan
    Ratti, Carlo
    Rus, Daniela
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 3109 - 3115
  • [43] Adaptive Learning Based Tracking Control of Marine Vessels with Prescribed Performance
    Xu, Zhao
    Ge, Shuzhi Sam
    Hu, Changhua
    Hu, Jinwen
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2018, 2018
  • [44] Autonomous Control of Primary Separation Vessel using Reinforcement Learning
    Soesanto, Jansen Fajar
    Maciszewski, Bart
    Mirmontazeri, Leyli
    Romero, Sabrina
    Michonski, Mike
    Milne, Andrew
    Huang, Biao
    IFAC PAPERSONLINE, 2024, 58 (22): : 83 - 88
  • [45] Enhancing the Performance of Adaptive Iterative Learning Control with Reinforcement Learning
    Nemec, Bojan
    Simonic, Mihael
    Likar, Nejc
    Ude, Ales
    2017 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2017, : 2192 - 2199
  • [46] Incremental Reinforcement Learning Flight Control with Adaptive Learning Rate
    Liu J.-H.
    Shan J.-Y.
    Rong J.-L.
    Zheng X.
    Yuhang Xuebao/Journal of Astronautics, 2022, 43 (01): : 111 - 121
  • [47] Learning to Communicate with Reinforcement Learning for an Adaptive Traffic Control System
    Vanneste, Simon
    de Borrekens, Gauthier
    Bosmans, Stig
    Vanneste, Astrid
    Mets, Kevin
    Mercelis, Siegfried
    Latre, Steven
    Hellinckx, Peter
    ADVANCES ON P2P, PARALLEL, GRID, CLOUD AND INTERNET COMPUTING, 3PGCIC-2021, 2022, 343 : 207 - 216
  • [48] A Reinforcement Learning-Based Adaptive Learning System
    Shawky, Doaa
    Badawi, Ashraf
    INTERNATIONAL CONFERENCE ON ADVANCED MACHINE LEARNING TECHNOLOGIES AND APPLICATIONS (AMLTA2018), 2018, 723 : 221 - 231
  • [49] Adaptive Rate and Energy Harvesting Interval Control Based on Reinforcement Learning for SWIPT
    Chun, Chang-Jae
    Kang, Jae-Mo
    Kim, Il-Min
    IEEE COMMUNICATIONS LETTERS, 2018, 22 (12) : 2571 - 2574
  • [50] Model-Based OPC With Adaptive PID Control Through Reinforcement Learning
    Kim, Taeyoung
    Zhang, Shilong
    Shin, Youngsoo
    IEEE TRANSACTIONS ON SEMICONDUCTOR MANUFACTURING, 2025, 38 (01) : 48 - 56