Co-evolution of synchronization and cooperation with multi-agent Q-learning

被引:5
|
作者
Zhu, Peican [1 ]
Cao, Zhaoheng [2 ]
Liu, Chen [3 ]
Chu, Chen [4 ]
Wang, Zhen [5 ]
机构
[1] Northwestern Polytech Univ, Sch Artificial Intelligence Opt & Elect iOPEN, Xian 710072, Peoples R China
[2] Northwestern Polytech Univ, Sch Comp Sci, Xian 710072, Peoples R China
[3] Northwestern Polytech Univ, Sch Ecol & Environm, Xian 710072, Peoples R China
[4] Yunnan Univ Finance & Econ, Sch Stat & Math, Kunming 650221, Peoples R China
[5] Northwestern Polytech Univ, Sch Cybersecur, Xian 710072, Peoples R China
基金
中国国家自然科学基金;
关键词
DILEMMA; REPUTATION; EVOLUTION; KURAMOTO; STRATEGY;
D O I
10.1063/5.0141824
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
Cooperation is a widespread phenomenon in human society and plays a significant role in achieving synchronization of various systems. However, there has been limited progress in studying the co-evolution of synchronization and cooperation. In this manuscript, we investigate how reinforcement learning affects the evolution of synchronization and cooperation. Namely, the payoff of an agent depends not only on the cooperation dynamic but also on the synchronization dynamic. Agents have the option to either cooperate or defect. While cooperation promotes synchronization among agents, defection does not. We report that the dynamic feature, which indicates the action switching frequency of the agent during interactions, promotes synchronization. We also find that cooperation and synchronization are mutually reinforcing. Furthermore, we thoroughly analyze the potential reasons for synchronization promotion due to the dynamic feature from both macro- and microperspectives. Additionally, we conduct experiments to illustrate the differences in the synchronization-promoting effects of cooperation and dynamic features.
引用
收藏
页数:8
相关论文
共 50 条
  • [41] Cooperative Multi-Agent Q-Learning Using Distributed MPC
    Esfahani, Hossein Nejatbakhsh
    Velni, Javad Mohammadpour
    [J]. IEEE CONTROL SYSTEMS LETTERS, 2024, 8 : 2193 - 2198
  • [42] Multi-Agent Coordination Method Based on Fuzzy Q-Learning
    Peng, Jun
    Liu, Miao
    Wu, Min
    Zhang, Xiaoyong
    Lin, Kuo-Chi
    [J]. 2008 7TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-23, 2008, : 5411 - +
  • [43] DVF:Multi-agent Q-learning with difference value factorization
    Huang, Anqi
    Wang, Yongli
    Sang, Jianghui
    Wang, Xiaoli
    Wang, Yupeng
    [J]. KNOWLEDGE-BASED SYSTEMS, 2024, 286
  • [44] Emotional temporal difference Q-learning signals in multi-agent system cooperation: real case studies
    Abdi, Javad
    Moshiri, Behzad
    Abdulhai, Baher
    [J]. IET INTELLIGENT TRANSPORT SYSTEMS, 2013, 7 (03) : 315 - 326
  • [45] Multi-agent collaboration system based on co-evolution method
    Xue, HT
    Shen, LC
    Zhu, HY
    Chang, WS
    [J]. PROCEEDINGS OF THE 4TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-4, 2002, : 1275 - 1279
  • [46] A Multi-Agent Adaptive Co-Evolution Method in Dynamic Environments
    Li, Yan
    Zhang, Huazhi
    Xu, Weiming
    Wang, Jianan
    Wang, Jialu
    Wang, Suyu
    [J]. MATHEMATICS, 2023, 11 (10)
  • [47] Learning multi-agent cooperation
    Rivera, Corban
    Staley, Edward
    Llorens, Ashley
    [J]. FRONTIERS IN NEUROROBOTICS, 2022, 16
  • [48] Towards Understanding Cooperative Multi-Agent Q-Learning with Value Factorization
    Wang, Jianhao
    Ren, Zhizhou
    Han, Beining
    Ye, Jianing
    Zhang, Chongjie
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [49] A Multi-Agent Q-Learning Based Rendezvous Strategy for Cognitive Radios
    Watson, Clifton L.
    Chakravarthy, Vasu D.
    Biswas, Subir
    [J]. 2017 COGNITIVE COMMUNICATIONS FOR AEROSPACE APPLICATIONS WORKSHOP (CCAA), 2017,
  • [50] Multi-agent Q-learning and regression trees for automated pricing decisions
    Sridharan, M
    Tesauro, G
    [J]. FOURTH INTERNATIONAL CONFERENCE ON MULTIAGENT SYSTEMS, PROCEEDINGS, 2000, : 447 - 448