Reinforcement Learning for Synchronization of Heterogeneous Multiagent Systems by Improved Q-Functions

被引:0
|
作者
Li, Jinna [1 ]
Yuan, Lin [1 ]
Cheng, Weiran [1 ]
Chai, Tianyou [2 ]
Lewis, Frank L. [3 ]
机构
[1] Liaoning Petrochem Univ, Sch Informat & Control Engn, Fushun 113001, Liaoning, Peoples R China
[2] Northeastern Univ, State Key Lab Synthet Automat Proc Ind, Shenyang 110819, Peoples R China
[3] Univ Texas Arlington, UTA Res Inst, Arlington, TX 76118 USA
基金
中国国家自然科学基金;
关键词
Synchronization; Protocols; Heuristic algorithms; Decision making; Nash equilibrium; Multi-agent systems; Games; Data-driven control; distributed control; multiagent systems (MASs); reinforcement learning (RL); synchronization; GRAPHICAL GAMES;
D O I
10.1109/TCYB.2024.3440333
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This article dedicates to investigating a methodology for enhancing adaptability to environmental changes of reinforcement learning (RL) techniques with data efficiency, by which a joint control protocol is learned using only data for multiagent systems (MASs). Thus, all followers are able to synchronize themselves with the leader and minimize their individual performance. To this end, an optimal synchronization problem of heterogeneous MASs is first formulated, and then an arbitration RL mechanism is developed for well addressing key challenges faced by the current RL techniques, that is, insufficient data and environmental changes. In the developed mechanism, an improved Q -function with an arbitration factor is designed for accommodating the fact that control protocols tend to be made by historic experiences and instinctive decision-making, such that the degree of control over agents' behaviors can be adaptively allocated by on-policy and off-policy RL techniques for the optimal multiagent synchronization problem. Finally, an arbitration RL algorithm with critic-only neural networks is proposed, and theoretical analysis and proofs of synchronization and performance optimality are provided. Simulation results verify the effectiveness of the proposed method.
引用
收藏
页码:6545 / 6558
页数:14
相关论文
共 50 条
  • [21] Effect of reinforcement learning on coordination of multiagent systems
    Bukkapatnam, S
    Gao, G
    NETWORK INTELLIGENCE: INTERNET-BASED MANUFACTURING, 2000, 4208 : 31 - 41
  • [22] Coordination in multiagent reinforcement learning systems by virtual reinforcement signals
    Kamal, M.
    Murata, Junichi
    INTERNATIONAL JOURNAL OF KNOWLEDGE-BASED AND INTELLIGENT ENGINEERING SYSTEMS, 2007, 11 (03) : 181 - 191
  • [23] Input-constrained optimal output synchronization of heterogeneous multiagent systems via observer-based model-free reinforcement learning
    Zhang, Tengfei
    Jia, Yingmin
    ASIAN JOURNAL OF CONTROL, 2024, 26 (01) : 98 - 113
  • [24] Scalable Reinforcement Learning for Multiagent Networked Systems
    Qu, Guannan
    Wierman, Adam
    Li, Na
    OPERATIONS RESEARCH, 2022, 70 (06) : 3601 - 3628
  • [25] Data-Driven Learning for Resilient Synchronization and Parameter Estimation of Heterogeneous Nonlinear Multiagent Systems
    Yang, Wang
    Dong, Jiuxiang
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2023, 21 (04) : 1 - 12
  • [26] Bipartite output synchronization of heterogeneous multiagent systems on signed digraphs
    Zuo, Shan
    Lewis, Frank L.
    Song, Yongduan
    Davoudi, Ali
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2018, 28 (13) : 4017 - 4031
  • [27] Synchronization of heterogeneous multiagent systems: A distributed observer and compensation framework
    Wang B.
    Zhao Y.
    Zhang L.
    Advanced Control for Applications: Engineering and Industrial Systems, 2020, 2 (02):
  • [28] Multiagent Reinforcement Social Learning toward Coordination in Cooperative Multiagent Systems
    Hao, Jianye
    Leung, Ho-Fung
    Ming, Zhong
    ACM TRANSACTIONS ON AUTONOMOUS AND ADAPTIVE SYSTEMS, 2015, 9 (04)
  • [29] Multiagent Meta-Reinforcement Learning for Optimized Task Scheduling in Heterogeneous Edge Computing Systems
    Niu, Liwen
    Chen, Xianfu
    Zhang, Ning
    Zhu, Yongdong
    Yin, Rui
    Wu, Celimuge
    Cao, Yangjie
    IEEE INTERNET OF THINGS JOURNAL, 2023, 10 (12) : 10519 - 10531
  • [30] Off-Policy Reinforcement Learning for Synchronization in Multiagent Graphical Games
    Li, Jinna
    Modares, Hamidreza
    Chai, Tianyou
    Lewis, Frank L.
    Xie, Lihua
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (10) : 2434 - 2445