Adaptive Optimal Surrounding Control of Multiple Unmanned Surface Vessels via Actor-Critic Reinforcement Learning

被引:1
|
作者
Lu, Renzhi [1 ,2 ,3 ,4 ]
Wang, Xiaotao [5 ]
Ding, Yiyu [5 ]
Zhang, Hai-Tao [6 ,7 ]
Zhao, Feng [8 ]
Zhu, Lijun [9 ]
He, Yong [10 ,11 ]
机构
[1] Huazhong Univ Sci & Technol, Sch Artificial Intelligence & Automat, Key Lab Image Proc & Intelligent Control, Wuhan 430074, Peoples R China
[2] Minist Educ, Key Lab Ind Internet Things & Networked Control, Chongqing 400065, Peoples R China
[3] Chongqing Univ, State Key Laboratoryof Mech Transmiss Adv Equipme, Chongqing 400044, Peoples R China
[4] Hubei Key Lab Adv Control & Intelligent Automat C, Wuhan 430074, Peoples R China
[5] Huazhong Univ Sci & Technol, Sch Artificial Intelligence & Automat, Wuhan 430074, Peoples R China
[6] Huazhong Univ Sci & Technol, Inst Artificial Intelligence, MOE Engn Res Ctr Autonomous Intelligent Unmanned, Sch Artificial Intelligence & Automat,State Key L, Wuhan 430074, Peoples R China
[7] Guangdong HUST Ind Technol Res Inst, Guangdong Prov Engn Technol Res Ctr Autonomous Un, Dongguan 523808, Peoples R China
[8] China Ship Sci Res Ctr, Wuxi 214082, Peoples R China
[9] Huazhong Univ Sci & Technol, MOE Engn Res Ctr Autonomous Intelligent Unmanned, Sch Artificial Intelligence & Automat, Wuhan 430074, Peoples R China
[10] China Univ Geosci, Sch Automat, Hubei Key Lab Adv Control & Intelligent Automat, Wuhan 430074, Peoples R China
[11] China Univ Geosci, Minist Educ, Engn Res Ctr Intelligent Technol Geoexplorat, Wuhan 430074, Peoples R China
基金
中国国家自然科学基金;
关键词
Actor-critic networks; Lyapunov functions; reinforcement learning (RL); surrounding control; unmanned surface vessels (USVs); MULTIAGENT SYSTEMS; AVOIDANCE;
D O I
10.1109/TNNLS.2024.3474289
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this article, an optimal surrounding control algorithm is proposed for multiple unmanned surface vessels (USVs), in which actor-critic reinforcement learning (RL) is utilized to optimize the merging process. Specifically, the multiple-USV optimal surrounding control problem is first transformed into the Hamilton-Jacobi-Bellman (HJB) equation, which is difficult to solve due to its nonlinearity. An adaptive actor-critic RL control paradigm is then proposed to obtain the optimal surround strategy, wherein the Bellman residual error is utilized to construct the network update laws. Particularly, a virtual controller representing intermediate transitions and an actual controller operating on a dynamics model are employed as surrounding control solutions for second-order USVs; thus, optimal surrounding control of the USVs is guaranteed. In addition, the stability of the proposed controller is analyzed by means of Lyapunov theory functions. Finally, numerical simulation results demonstrate that the proposed actor-critic RL-based surrounding controller can achieve the surrounding objective while optimizing the evolution process and obtains 9.76% and 20.85% reduction in trajectory length and energy consumption compared with the existing controller.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Adaptive Optimal Tracking Control of an Underactuated Surface Vessel Using Actor-Critic Reinforcement Learning
    Chen, Lin
    Dai, Shi-Lu
    Dong, Chao
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (06) : 7520 - 7533
  • [2] Adaptive optimal formation control for unmanned surface vehicles with guaranteed performance using actor-critic learning architecture
    Chen, Lin
    Dong, Chao
    He, Shude
    Dai, Shi-Lu
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2023, 33 (08) : 4504 - 4522
  • [3] Optimal Policy of Multiplayer Poker via Actor-Critic Reinforcement Learning
    Shi, Daming
    Guo, Xudong
    Liu, Yi
    Fan, Wenhui
    ENTROPY, 2022, 24 (06)
  • [4] Adaptive Assist-as-needed Control Based on Actor-Critic Reinforcement Learning
    Zhang, Yufeng
    Li, Shuai
    Nolan, Karen J.
    Zanotto, Damiano
    2019 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2019, : 4066 - 4071
  • [5] Actor-Critic Reinforcement Learning for Tracking Control in Robotics
    Pane, Yudha P.
    Nageshrao, Subramanya P.
    Babuska, Robert
    2016 IEEE 55TH CONFERENCE ON DECISION AND CONTROL (CDC), 2016, : 5819 - 5826
  • [6] Actor-Critic Reinforcement Learning for Control With Stability Guarantee
    Han, Minghao
    Zhang, Lixian
    Wang, Jun
    Pan, Wei
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (04) : 6217 - 6224
  • [7] REINFORCEMENT LEARNING-BASED ADAPTIVE MOTION CONTROL FOR AUTONOMOUS VEHICLES VIA ACTOR-CRITIC STRUCTURE
    Wang, Honghai
    Wei, Liangfen
    Wang, Xianchao
    He, Shuping
    DISCRETE AND CONTINUOUS DYNAMICAL SYSTEMS-SERIES S, 2024, 17 (09): : 2894 - 2911
  • [8] Dual Variable Actor-Critic for Adaptive Safe Reinforcement Learning
    Lee, Junseo
    Heo, Jaeseok
    Kim, Dohyeong
    Lee, Gunmin
    Oh, Songhwai
    2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2023, : 7568 - 7573
  • [9] Optimal Elevator Group Control via Deep Asynchronous Actor-Critic Learning
    Wei, Qinglai
    Wang, Lingxiao
    Liu, Yu
    Polycarpou, Marios M.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (12) : 5245 - 5256
  • [10] Optimized Adaptive Nonlinear Tracking Control Using Actor-Critic Reinforcement Learning Strategy
    Wen, Guoxing
    Chen, C. L. Philip
    Ge, Shuzhi Sam
    Yang, Hongli
    Liu, Xiaoguang
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2019, 15 (09) : 4969 - 4977