Emergent cooperation from mutual acknowledgment exchange in multi-agent reinforcement learning

被引:1
|
作者
Phan, Thomy [1 ,2 ]
Sommer, Felix [2 ]
Ritz, Fabian [2 ]
Altmann, Philipp [2 ]
Nuesslein, Jonas [2 ]
Koelle, Michael [2 ]
Belzner, Lenz [3 ]
Linnhoff-Popien, Claudia [2 ]
机构
[1] Univ Southern Calif, Los Angeles, CA 90007 USA
[2] Ludwig Maximilians Univ Munchen, Munich, Germany
[3] TH Ingolstadt, Ingolstadt, Germany
关键词
Multi-agent learning; Reinforcement learning; Mutual acknowledgments; Peer incentivization; Emergent cooperation; EVOLUTION; LEVEL;
D O I
10.1007/s10458-024-09666-5
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Peer incentivization (PI) is a recent approach where all agents learn to reward or penalize each other in a distributed fashion, which often leads to emergent cooperation. Current PI mechanisms implicitly assume a flawless communication channel in order to exchange rewards. These rewards are directly incorporated into the learning process without any chance to respond with feedback. Furthermore, most PI approaches rely on global information, which limits scalability and applicability to real-world scenarios where only local information is accessible. In this paper, we propose Mutual Acknowledgment Token Exchange (MATE), a PI approach defined by a two-phase communication protocol to exchange acknowledgment tokens as incentives to shape individual rewards mutually. All agents condition their token transmissions on the locally estimated quality of their own situations based on environmental rewards and received tokens. MATE is completely decentralized and only requires local communication and information. We evaluate MATE in three social dilemma domains. Our results show that MATE is able to achieve and maintain significantly higher levels of cooperation than previous PI approaches. In addition, we evaluate the robustness of MATE in more realistic scenarios, where agents can deviate from the protocol and communication failures can occur. We also evaluate the sensitivity of MATE w.r.t. the choice of token values.
引用
收藏
页数:36
相关论文
共 50 条
  • [21] Bottom-up Multi-agent Reinforcement Learning for Selective Cooperation
    Aotani, Takumi
    Kobayashi, Taisuke
    Sugimoto, Kenji
    2018 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2018, : 3590 - 3595
  • [22] Hierarchical reinforcement learning based on multi-agent cooperation game theory
    Tang H.
    Dong C.
    International Journal of Wireless and Mobile Computing, 2019, 16 (04): : 369 - 376
  • [23] GAMA: Graph Attention Multi-agent reinforcement learning algorithm for cooperation
    Haoqiang Chen
    Yadong Liu
    Zongtan Zhou
    Dewen Hu
    Ming Zhang
    Applied Intelligence, 2020, 50 : 4195 - 4205
  • [24] The evolution of cooperation in continuous dilemmas via multi-agent reinforcement learning
    Zhu, Congcong
    Ye, Dayong
    Zhu, Tianqing
    Zhou, Wanlei
    KNOWLEDGE-BASED SYSTEMS, 2025, 315
  • [25] PMIC: Improving Multi-Agent Reinforcement Learning with Progressive Mutual Information Collaboration
    Li, Pengyi
    Tang, Hongyao
    Yang, Tianpei
    Hao, Xiaotian
    Sang, Tong
    Zheng, Yan
    Hao, Jianye
    Taylor, Matthew E.
    Tao, Wenyuan
    Wang, Zhen
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [26] Mutual information oriented deep skill chaining for multi-agent reinforcement learning
    Xie, Zaipeng
    Ji, Cheng
    Qiao, Chentai
    Song, Wenzhan
    Li, Zewen
    Zhang, Yufeng
    Zhang, Yujing
    CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY, 2024, 9 (04) : 1014 - 1030
  • [27] Multi-agent reinforcement learning: an approach based on agents' cooperation for a common goal
    Wang, GQ
    Yu, HB
    PROCEEDINGS OF THE 8TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, VOL 1, 2004, : 336 - 339
  • [28] A fully value distributional deep reinforcement learning framework for multi-agent cooperation
    Fu, Mingsheng
    Huang, Liwei
    Li, Fan
    Qu, Hong
    Xu, Chengzhong
    NEURAL NETWORKS, 2025, 184
  • [29] Innovative Approach Towards Cooperation Models for Multi-agent Reinforcement Learning (CMMARL)
    Vidhate, Deepak A.
    Kulkarni, Parag
    SMART TRENDS IN INFORMATION TECHNOLOGY AND COMPUTER COMMUNICATIONS, SMARTCOM 2016, 2016, 628 : 468 - 478
  • [30] Partner Selection for the Emergence of Cooperation in Multi-Agent Systems Using Reinforcement Learning
    Anastassacos, Nicolas
    Hailes, Stephen
    Musolesi, Mirco
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 7047 - 7054