Optimized Backstepping Consensus Control Using Reinforcement Learning for a Class of Nonlinear Strict-Feedback-Dynamic Multi-Agent Systems

被引:65
|
作者
Wen, Guoxing [1 ,2 ]
Chen, C. L. Philip [3 ,4 ]
机构
[1] Binzhou Univ, Coll Sci, Binzhou 256600, Peoples R China
[2] Qilu Univ Technol, Sch Math & Stat, Jinan 250353, Peoples R China
[3] South China Univ Technol, Sch Comp Sci & Engn, Guangzhou 510641, Peoples R China
[4] Dalian Maritime Univ, Nav Coll, Dalian 116026, Peoples R China
基金
中国国家自然科学基金;
关键词
Optimal control; Backstepping; Artificial neural networks; Performance analysis; Nonlinear dynamical systems; Consensus control; Mathematical model; Critic-actor architecture; high-order multi-agent system (MAS); neural network (NN); optimal control; reinforcement learning (RL); ADAPTIVE OPTIMAL-CONTROL; CONTAINMENT CONTROL; TRACKING CONTROL;
D O I
10.1109/TNNLS.2021.3105548
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this article, an optimized leader-following consensus control scheme is proposed for the nonlinear strict-feedback-dynamic multi-agent system by learning from the controlling idea of optimized backstepping technique, which designs the virtual and actual controls of backstepping to be the optimized solution of corresponding subsystems so that the entire backstepping control is optimized. Since this control needs to not only ensure the optimizing system performance but also synchronize the multiple system state variables, it is an interesting and challenging topic. In order to achieve this optimized control, the neural network approximation-based reinforcement learning (RL) is performed under critic-actor architecture. In most of the existing RL-based optimal controls, since both the critic and actor RL updating laws are derived from the negative gradient of square of the Hamilton-Jacobi-Bellman (HJB) equation's approximation, which contains multiple nonlinear terms, their algorithm are inevitably intricate. However, the proposed optimized control derives the RL updating laws from the negative gradient of a simple positive function, which is correlated with the HJB equation; hence, it can be significantly simple in the algorithm. Meanwhile, it can also release two general conditions, known dynamic and persistence excitation, which are required in most of the RL-based optimal controls. Therefore, the proposed optimized scheme can be a natural selection for the high-order nonlinear multi-agent control. Finally, the effectiveness is demonstrated by both theory and simulation.
引用
收藏
页码:1524 / 1536
页数:13
相关论文
共 50 条
  • [41] Dynamic Event-Triggered Reinforcement Learning-Based Consensus Tracking of Nonlinear Multi-Agent Systems
    Xu, Bo
    Li, Yuan-Xin
    Hou, Zhongsheng
    Ahn, Choon Ki
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2023, 70 (05) : 2120 - 2132
  • [42] Nonlinear Filtering and Reinforcement Learning Based Consensus Achievement of Uncertain Multi-Agent Systems
    Borah, Kaustav Jyoti
    JOURNAL OF DYNAMIC SYSTEMS MEASUREMENT AND CONTROL-TRANSACTIONS OF THE ASME, 2024, 146 (03):
  • [43] Optimized Backstepping for Tracking Control of Strict-Feedback Systems
    Wen, Guoxing
    Ge, Shuzhi Sam
    Tu, Fangwen
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (08) : 3850 - 3862
  • [44] Distributed adaptive consensus tracking control of higher-order nonlinear strict-feedback multi-agent systems using neural networks
    Wang, Gang
    Wang, Chaoli
    Li, Lin
    Du, Qinghui
    NEUROCOMPUTING, 2016, 214 : 269 - 279
  • [45] Command-filter-based adaptive finite-time consensus control for nonlinear strict-feedback multi-agent systems with dynamic leader
    Cui, Yang
    Liu, Xiaoping
    Deng, Xin
    Wen, Guoxing
    INFORMATION SCIENCES, 2021, 565 : 17 - 31
  • [46] Adaptive Iterative Learning Backstepping Control for Nonlinear Strict-feedback Systems
    Chen, Jianyong
    PROCEEDINGS OF 2020 IEEE 9TH DATA DRIVEN CONTROL AND LEARNING SYSTEMS CONFERENCE (DDCLS'20), 2020, : 1054 - 1059
  • [47] Output-feedback optimized consensus for directed graph multi-agent systems based on reinforcement learning and subsystem error derivatives
    Li, Dongdong
    Dong, Jiuxiang
    INFORMATION SCIENCES, 2023, 649
  • [48] Robust Consensus Control for Nonlinear Multi-Agent Systems
    Liu, Xiaofeng
    Zhang, Dongxu
    PROCEEDINGS OF THE 36TH CHINESE CONTROL CONFERENCE (CCC 2017), 2017, : 8131 - 8136
  • [49] Adaptive leaderless consensus control of a class of strict-feedback nonlinear multi-agent systems with unknown control directions: A non-Nussbaum function based approach
    Ao, Wengang
    Huang, Jiangshuai
    Xue, Fangzheng
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2020, 357 (17): : 12180 - 12196
  • [50] Design of a class of nonlinear consensus protocols for multi-agent systems
    Xu, Yaojin
    Tian, Yu-Ping
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2013, 23 (13) : 1524 - 1536