Reinforcement Learning for Synchronization of Heterogeneous Multiagent Systems by Improved Q-Functions

被引:0
|
作者
Li, Jinna [1 ]
Yuan, Lin [1 ]
Cheng, Weiran [1 ]
Chai, Tianyou [2 ]
Lewis, Frank L. [3 ]
机构
[1] Liaoning Petrochem Univ, Sch Informat & Control Engn, Fushun 113001, Liaoning, Peoples R China
[2] Northeastern Univ, State Key Lab Synthet Automat Proc Ind, Shenyang 110819, Peoples R China
[3] Univ Texas Arlington, UTA Res Inst, Arlington, TX 76118 USA
基金
中国国家自然科学基金;
关键词
Synchronization; Protocols; Heuristic algorithms; Decision making; Nash equilibrium; Multi-agent systems; Games; Data-driven control; distributed control; multiagent systems (MASs); reinforcement learning (RL); synchronization; GRAPHICAL GAMES;
D O I
10.1109/TCYB.2024.3440333
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This article dedicates to investigating a methodology for enhancing adaptability to environmental changes of reinforcement learning (RL) techniques with data efficiency, by which a joint control protocol is learned using only data for multiagent systems (MASs). Thus, all followers are able to synchronize themselves with the leader and minimize their individual performance. To this end, an optimal synchronization problem of heterogeneous MASs is first formulated, and then an arbitration RL mechanism is developed for well addressing key challenges faced by the current RL techniques, that is, insufficient data and environmental changes. In the developed mechanism, an improved Q -function with an arbitration factor is designed for accommodating the fact that control protocols tend to be made by historic experiences and instinctive decision-making, such that the degree of control over agents' behaviors can be adaptively allocated by on-policy and off-policy RL techniques for the optimal multiagent synchronization problem. Finally, an arbitration RL algorithm with critic-only neural networks is proposed, and theoretical analysis and proofs of synchronization and performance optimality are provided. Simulation results verify the effectiveness of the proposed method.
引用
收藏
页码:6545 / 6558
页数:14
相关论文
共 50 条
  • [31] Optimal Synchronization Control of Multiagent Systems With Input Saturation via Off-Policy Reinforcement Learning
    Qin, Jiahu
    Li, Man
    Shi, Yang
    Ma, Qichao
    Zheng, Wei Xing
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (01) : 85 - 96
  • [32] N-learning: A reinforcement learning paradigm for multiagent systems
    Mansfield, M
    Collins, JJ
    Eaton, M
    Collins, T
    AI 2005: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2005, 3809 : 684 - 694
  • [33] Multiagent reinforcement learning method with an improved ant colony system
    Sun, RY
    Shoji, T
    Zhao, G
    2001 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS, VOLS 1-5: E-SYSTEMS AND E-MAN FOR CYBERNETICS IN CYBERSPACE, 2002, : 1612 - 1617
  • [34] Beyond Reinforcement Learning and Local View in Multiagent Systems
    Bazzan, Ana L. C.
    KUNSTLICHE INTELLIGENZ, 2014, 28 (03): : 179 - 189
  • [35] Reinforcement Learning With Task Decomposition for Cooperative Multiagent Systems
    Sun, Changyin
    Liu, Wenzhang
    Dong, Lu
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (05) : 2054 - 2065
  • [36] An Evolutionary Transfer Reinforcement Learning Framework for Multiagent Systems
    Hou, Yaqing
    Ong, Yew-Soon
    Feng, Liang
    Zurada, Jacek M.
    IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2017, 21 (04) : 601 - 615
  • [37] Prior Knowledge-Augmented Broad Reinforcement Learning Framework for Fault Diagnosis of Heterogeneous Multiagent Systems
    Guo, Li
    Ren, Yiran
    Li, Runze
    Jiang, Bin
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2024, 16 (01) : 115 - 123
  • [38] Cooperative Estimation and Robust Synchronization of Heterogeneous Multiagent Systems With Coupled Measurements
    Wu, Jingbo
    Ugrinovskii, Valery
    Allgoewer, Frank
    IEEE TRANSACTIONS ON CONTROL OF NETWORK SYSTEMS, 2018, 5 (04): : 1597 - 1607
  • [39] Distributed optimal output synchronization of heterogeneous multiagent systems with communication delays
    Zhao, Wei
    Zhang, Huaipin
    Yu, Wenwu
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2024, 34 (12) : 7821 - 7836
  • [40] Time-Varying Formation of Heterogeneous Multiagent Systems via Reinforcement Learning Subject to Switching Topologies
    Liu, Deyuan
    Liu, Hao
    Lu, Jinhu
    Lewis, Frank L.
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2023, 70 (06) : 2550 - 2560