Spatial reciprocity under reinforcement learning mechanism

被引:0
|
作者
Wang, Lu [1 ]
Shi, Xiaoqiu [1 ,2 ]
Zhou, Yang [3 ]
机构
[1] Southwest Univ Sci & Technol, Sch Mfg Sci & Engn, Mianyang 621000, Sichuan, Peoples R China
[2] Mianyang Sci & Technol City Intelligent Mfg Ind Te, Mianyang 621000, Sichuan, Peoples R China
[3] Southwest Univ Sci & Technol, Engn Technol Ctr, Mianyang 621000, Sichuan, Peoples R China
关键词
NETWORK RECIPROCITY; EVOLUTIONARY DYNAMICS; GAME;
D O I
10.1063/5.0246843
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
At present, the research on the dynamics of cooperative behavior of agents under reinforcement learning mechanism either assumes that agents have global interaction, that is, agents interact with all other agents in the population, or directly study the influence of relevant factors on cooperation evolution based on the local interaction in a network structure. It neglects to formally study how the limitation of agents that only interact with local agents affects their strategy choice. Thus, in this paper, we study the cooperative behavior of agents in a typical social decision-making environment with conflicts between individual interests and collective interests. On the one hand, a programmed game model in game theory, namely, prisoner's dilemma game, is used to capture the essence of real-world dilemmas. On the other hand, the effects of local and global strategy learning on the cooperative evolution of agents are investigated separately, and the nature of spatial reciprocity under the reinforcement learning mechanism is found. Specifically, when there is no inherent connection between the interacting agents and the learning agents within the system, the network structure has a limited effect on promoting cooperation. It is only when there is an overlap between the interacting agents and the learning agents that the spatial reciprocity effect observed in the traditional evolutionary game theory can be fully realized.
引用
收藏
页数:8
相关论文
共 50 条
  • [41] Reinforcement learning with the mechanism of short-term depression for learning rate
    Kubota, S.
    PROCEEDINGS OF THE SIXTEENTH INTERNATIONAL SYMPOSIUM ON ARTIFICIAL LIFE AND ROBOTICS (AROB 16TH '11), 2011, : 800 - 803
  • [42] Stimulus sampling as an exploration mechanism for fast reinforcement learning
    Boris B. Vladimirskiy
    Eleni Vasilaki
    Robert Urbanczik
    Walter Senn
    Biological Cybernetics, 2009, 100 : 319 - 330
  • [43] Reinforcement Learning based Scheduling Optimization Mechanism on Switches
    Lu, Xijia
    Wang, Xingwei
    Jia, Jie
    Wang, Xue
    Huang, Min
    2022 IEEE 28TH INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS, ICPADS, 2022, : 554 - 561
  • [44] Reinforcement learning using RBF networks with memory mechanism
    Ozawa, S
    Shiraga, N
    KNOWLEDGE-BASED INTELLIGENT INFORMATION AND ENGINEERING SYSTEMS, PT 1, PROCEEDINGS, 2003, 2773 : 1149 - 1156
  • [45] Stimulus sampling as an exploration mechanism for fast reinforcement learning
    Vladimirskiy, Boris B.
    Vasilaki, Eleni
    Urbanczik, Robert
    Senn, Walter
    BIOLOGICAL CYBERNETICS, 2009, 100 (04) : 319 - 330
  • [46] A Reinforcement Learning Mechanism Responsible for the Valuation of Free Choice
    Cockburn, Jeffrey
    Collins, Anne G. E.
    Frank, Michael J.
    NEURON, 2014, 83 (03) : 551 - 557
  • [47] Reinforcement Learning Based Prefetch-Control Mechanism
    Ghosh, Soma Niloy
    Sahula, Vineet
    Bhargava, Lava
    2023 IEEE ASIA PACIFIC CONFERENCE ON CIRCUITS AND SYSTEMS, APCCAS, 2024, : 110 - 114
  • [48] Individualism or Collectivism: A Reinforcement Learning Mechanism for Vaccination Decisions
    Wu, Chaohao
    Qiao, Tong
    Qiu, Hongjun
    Shi, Benyun
    Bao, Qing
    INFORMATION, 2021, 12 (02) : 1 - 16
  • [49] An adaptive work distribution mechanism based on reinforcement learning
    Huang, Zhengxing
    van der Aalst, W. M. P.
    Lu, Xudong
    Duan, Huilong
    EXPERT SYSTEMS WITH APPLICATIONS, 2010, 37 (12) : 7533 - 7541
  • [50] Recruitment-imitation mechanism for evolutionary reinforcement learning
    Lu, Shuai
    Han, Shuai
    Zhou, Wenbo
    Zhang, Junwei
    INFORMATION SCIENCES, 2021, 553 (553) : 172 - 188