Reinforcement Learning for Synchronization of Heterogeneous Multiagent Systems by Improved Q-Functions

被引:0
|
作者
Li, Jinna [1 ]
Yuan, Lin [1 ]
Cheng, Weiran [1 ]
Chai, Tianyou [2 ]
Lewis, Frank L. [3 ]
机构
[1] Liaoning Petrochem Univ, Sch Informat & Control Engn, Fushun 113001, Liaoning, Peoples R China
[2] Northeastern Univ, State Key Lab Synthet Automat Proc Ind, Shenyang 110819, Peoples R China
[3] Univ Texas Arlington, UTA Res Inst, Arlington, TX 76118 USA
基金
中国国家自然科学基金;
关键词
Synchronization; Protocols; Heuristic algorithms; Decision making; Nash equilibrium; Multi-agent systems; Games; Data-driven control; distributed control; multiagent systems (MASs); reinforcement learning (RL); synchronization; GRAPHICAL GAMES;
D O I
10.1109/TCYB.2024.3440333
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This article dedicates to investigating a methodology for enhancing adaptability to environmental changes of reinforcement learning (RL) techniques with data efficiency, by which a joint control protocol is learned using only data for multiagent systems (MASs). Thus, all followers are able to synchronize themselves with the leader and minimize their individual performance. To this end, an optimal synchronization problem of heterogeneous MASs is first formulated, and then an arbitration RL mechanism is developed for well addressing key challenges faced by the current RL techniques, that is, insufficient data and environmental changes. In the developed mechanism, an improved Q -function with an arbitration factor is designed for accommodating the fact that control protocols tend to be made by historic experiences and instinctive decision-making, such that the degree of control over agents' behaviors can be adaptively allocated by on-policy and off-policy RL techniques for the optimal multiagent synchronization problem. Finally, an arbitration RL algorithm with critic-only neural networks is proposed, and theoretical analysis and proofs of synchronization and performance optimality are provided. Simulation results verify the effectiveness of the proposed method.
引用
收藏
页码:6545 / 6558
页数:14
相关论文
共 50 条
  • [1] DROPOUT Q-FUNCTIONS for DOUBLY EFFICIENT REINFORCEMENT LEARNING
    Hiraoka, Takuya
    Imagawa, Takahisa
    Hashimoto, Taisei
    Onishi, Takashi
    Tsuruoka, Yoshimasa
    arXiv, 2021,
  • [2] Offline Reinforcement Learning via Policy Regularization and Ensemble Q-Functions
    Wang, Tao
    Xie, Shaorong
    Gao, Mingke
    Chen, Xue
    Zhang, Zhenyu
    Yu, Hang
    2022 IEEE 34TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, ICTAI, 2022, : 1167 - 1174
  • [3] Q-Transformer: Scalable Offline Reinforcement Learning via Autoregressive Q-Functions
    Chebotar, Yevgen
    Vuong, Quan
    Irpan, Alex
    Hausman, Karol
    Xia, Fei
    Lu, Yao
    Kumar, Aviral
    Yu, Tianhe
    Herzog, Alexander
    Pertsch, Karl
    Gopalakrishnan, Keerthana
    Ibarz, Julian
    Nachum, Ofir
    Sontakke, Sumedh
    Salazar, Grecia
    Tran, Huong T.
    Peralta, Jodilyn
    Tan, Clayton
    Manjunath, Deeksha
    Singht, Jaspiar
    Zitkovich, Brianna
    Jackson, Tomas
    Rao, Kanishka
    Finn, Chelsea
    Levine, Sergey
    CONFERENCE ON ROBOT LEARNING, VOL 229, 2023, 229
  • [4] The Wisdom of the Crowd: Reliable Deep Reinforcement Learning Through Ensembles of Q-Functions
    Elliott, Daniel L.
    Anderson, Charles
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (01) : 43 - 51
  • [5] Data-Based Optimal Synchronization of Heterogeneous Multiagent Systems in Graphical Games via Reinforcement Learning
    Xiong, Chunping
    Ma, Qian
    Guo, Jian
    Lewis, Frank L.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 35 (11) : 1 - 9
  • [6] Autonomous Synchronization of Heterogeneous Multiagent Systems
    Yan, Yamin
    Chen, Zhiyong
    Middleton, Richard H.
    IEEE TRANSACTIONS ON CONTROL OF NETWORK SYSTEMS, 2021, 8 (02): : 940 - 950
  • [7] An improved multiagent reinforcement learning algorithm
    Meng, XP
    Babuska, R
    Busoniu, L
    Chen, Y
    Tan, WY
    2005 IEEE/WIC/ACM International Conference on Intelligent Agent Technology, Proceedings, 2005, : 337 - 343
  • [8] Multiagent Reinforcement Learning With Unshared Value Functions
    Hu, Yujing
    Gao, Yang
    An, Bo
    IEEE TRANSACTIONS ON CYBERNETICS, 2015, 45 (04) : 647 - 662
  • [9] Coordination in multiagent reinforcement learning systems
    Kamal, MAS
    Murata, J
    KNOWLEDGE-BASED INTELLIGENT INFORMATION AND ENGINEERING SYSTEMS, PT 1, PROCEEDINGS, 2004, 3213 : 1197 - 1204
  • [10] Multiagent Reinforcement Learning With Heterogeneous Graph Attention Network
    Du, Wei
    Ding, Shifei
    Zhang, Chenglong
    Shi, Zhongzhi
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (10) : 6851 - 6860