A Decentralized Communication Framework Based on Dual-Level Recurrence for Multiagent Reinforcement Learning

被引:0
|
作者
Li, Xuesi [1 ]
Li, Jingchen [1 ]
Shi, Haobin [1 ]
Hwang, Kao-Shing [2 ]
机构
[1] Northwestern Polytech Univ, Sch Comp Sci & Engn, Xian 710129, Shaanxi, Peoples R China
[2] Natl Sun Yat Sen Univ, Dept Elect Engn, Kaohsiung 804, Taiwan
基金
中国国家自然科学基金;
关键词
Reinforcement learning; Logic gates; Training; Adaptation models; Multi-agent systems; Task analysis; Decision making; Gated recurrent network; multiagent reinforcement learning; multiagent system;
D O I
10.1109/TCDS.2023.3281878
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Designing communication channels for multiagent is a feasible method to conduct decentralized learning, especially in partially observable environments or large-scale multiagent systems. In this work, a communication model with dual-level recurrence is developed to provide a more efficient communication mechanism for the multiagent reinforcement learning field. The communications are conducted by a gated-attention-based recurrent network, in which the historical states are taken into account and regarded as the second-level recurrence. We separate communication messages from memories in the recurrent model so that the proposed communication flow can adapt changeable communication objects in the case of limited communication, and the communication results are fair to every agent. We provide a sufficient discussion about our method in both partially observable and fully observable environments. The results of several experiments suggest our method outperforms the existing decentralized communication frameworks and the corresponding centralized training method.
引用
收藏
页码:640 / 649
页数:10
相关论文
共 50 条
  • [31] Dual-level contrastive learning network for generalized zero-shot learning
    Guan, Jiaqi
    Meng, Min
    Liang, Tianyou
    Liu, Jigang
    Wu, Jigang
    VISUAL COMPUTER, 2022, 38 (9-10): : 3087 - 3095
  • [32] Voxelization of STL model based on dual-level octree
    Wang, Hong-Liang
    Guo, Rui-Feng
    Wang, Pin
    Liu, Xian
    Zhang, Yi-Wen
    Jisuanji Jicheng Zhizao Xitong/Computer Integrated Manufacturing Systems, CIMS, 2014, 20 (07): : 1553 - 1560
  • [33] Dual-level contrastive learning network for generalized zero-shot learning
    Jiaqi Guan
    Min Meng
    Tianyou Liang
    Jigang Liu
    Jigang Wu
    The Visual Computer, 2022, 38 : 3087 - 3095
  • [34] A Decentralized Approach to Intrusion Detection in Dynamic Networks of the Internet of Things Based on Multiagent Reinforcement Learning with Interagent Interaction
    M. O. Kalinin
    E. I. Tkacheva
    Automatic Control and Computer Sciences, 2023, 57 : 1025 - 1032
  • [35] Decentralized Multiagent Reinforcement Learning Based State-of-Charge Balancing Strategy for Distributed Energy Storage System
    Xiong, Zheng
    Luo, Biao
    Wang, Bing-Chuan
    Xu, Xiaodong
    Liu, Xiaodong
    Huang, Tingwen
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2024, : 12450 - 12460
  • [36] Reinforcement based Communication Topology Construction for Decentralized Learning with Non-IID Data
    Lin, Yi-Cheng
    Kuo, Jian-Jhih
    Chen, Wen-Tsuen
    Sheu, Jang-Ping
    2021 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2021,
  • [37] Decentralized Cooperative Control of Multiple Energy Storage Systems in Urban Railway Based on Multiagent Deep Reinforcement Learning
    Zhu, Feiqin
    Yang, Zhongping
    Lin, Fei
    Xin, Yue
    IEEE TRANSACTIONS ON POWER ELECTRONICS, 2020, 35 (09) : 9368 - 9379
  • [38] A Decentralized Approach to Intrusion Detection in Dynamic Networks of the Internet of Things Based on Multiagent Reinforcement Learning with Interagent Interaction
    Kalinin, M. O.
    Tkacheva, E. I.
    AUTOMATIC CONTROL AND COMPUTER SCIENCES, 2023, 57 (08) : 1025 - 1032
  • [39] CuMARL: Curiosity-Based Learning in Multiagent Reinforcement Learning
    Ningombam, Devarani Devi
    Yoo, Byunghyun
    Kim, Hyun Woo
    Song, Hwa Jeon
    Yi, Sungwon
    IEEE ACCESS, 2022, 10 : 87254 - 87265
  • [40] Decentralized multiagent reinforcement learning algorithm using a cluster-synchronized laser network
    Kotoku, Shun
    Mihana, Takatomo
    Röhm, André
    Horisaki, Ryoichi
    Physical Review E, 2024, 110 (06)