Cooperative Behavior Learning Based on Social Interaction of State Conversion and Reward Exchange Among Multi-Agents

被引:2
|
作者
Zhang, Kun [1 ]
Maeda, Yoichiro [2 ]
Takahashi, Yasutake [2 ]
机构
[1] Univ Fukui, Grad Sch Engn, Dept Syst Design Engn, 3-9-1 Bunkyo, Fukui 9108507, Japan
[2] Univ Fukui, Grad Sch Engn, Dept Human & Artificial Intelligent Syst, Fukui 9108507, Japan
关键词
social interaction; behavior learning; state conversion; reward exchange; reinforcement learning;
D O I
10.20965/jaciii.2011.p0606
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In multi-agent systems, it is necessary for autonomous agents to interact with each other in order to have excellent cooperative performance. Therefore, we have studied social interaction between agents to see how they acquire cooperative behavior. We have found that sharing environmental states can improve agent cooperation through reinforcement learning, and that changing environmental states to target-related individual states improves cooperation. To further improve cooperation, we propose reward redistribution based on reward exchanges among agents. In receiving rewards from both the environment and other agents, agents learned how to adjust themselves to the environment and how to explore and strengthen cooperation in tasks that a single agent could not do alone. Agents thus cooperate best through the interaction of state conversion and reward exchange.
引用
收藏
页码:606 / 616
页数:11
相关论文
共 50 条
  • [31] Hybrid Multi-Agents and Case Based Reasoning for Aiding Green Practice in Institutions of Higher Learning
    Anthony Jnr, Bokolo
    Majid, Mazlina Abdul
    Romli, Awanis
    TEHNICKI VJESNIK-TECHNICAL GAZETTE, 2019, 26 (01): : 13 - 21
  • [32] Cooperative fault estimation for a class of heterogeneous multi-agents with stochastic nonlinearities based on finite impulse response filter
    Wu, Yutao
    Mao, Zehui
    Yan, Xing-Gang
    Jiang, Bin
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2022, 32 (08) : 4696 - 4715
  • [33] Towards an optimized design of individualized learning paths: An approach based on ontology and Multi-agents system
    El Bouhdidi, Jaber
    Ghailani, Mohamed
    Fennan, Abdelhadi
    Bouhdidi, J.E., 1600, International Journal of Computer Science Issues (IJCSI) (09) : 6 - 3
  • [34] Cooperative behavior acquisition in multi mobile robots environment by reinforcement learning based on state vector estimation
    Uchibe, E
    Asada, M
    Hosoda, K
    1998 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS 1-4, 1998, : 1558 - 1563
  • [35] HUMAN-BASED SOCIAL-INTERACTION CAN REWARD A RATS BEHAVIOR
    DAVIS, H
    PERUSSE, R
    ANIMAL LEARNING & BEHAVIOR, 1988, 16 (01): : 89 - 92
  • [36] Combined DR Pricing and Voltage Control Using Reinforcement Learning Based Multi-Agents and Load Forecasting
    Khan, Danyal Afgan
    Arshad, Ammar
    Lehtonen, Matti
    Mahmoud, Karar
    IEEE ACCESS, 2022, 10 : 130839 - 130849
  • [37] Cooperative Distance-Based Leader-Following Formation Control Using SDRE for Multi-Agents with Energy Constraints
    Babazadeh, Reza
    Selmic, Rastko
    2018 IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2018, : 508 - 514
  • [38] Decentralized Ensemble Learning Based on Sample Exchange among Multiple Agents
    Yu, Yong
    Deng, Jia
    Tang, Yanni
    Liu, Jiamou
    Chen, Wu
    BSCI '19: PROCEEDINGS OF THE 2019 ACM INTERNATIONAL SYMPOSIUM ON BLOCKCHAIN AND SECURE CRITICAL INFRASTRUCTURE, 2019, : 57 - 66
  • [39] Value-Decomposition Networks For Cooperative Multi-Agent Learning Based On Team Reward
    Sunehag, Peter
    Lever, Guy
    Gruslys, Audrunas
    Czarnecki, Wojciech Marian
    Zambaldi, Vinicius
    Jaderberg, Max
    Lanctot, Marc
    Sonnerat, Nicolas
    Leibo, Joel Z.
    Tuyls, Karl
    Graepel, Thore
    PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS (AAMAS' 18), 2018, : 2085 - 2087
  • [40] Multi-robot cooperative behavior generation based on reinforcement learning
    Li, Dong-Mei
    Chen, Wei-Dong
    Xi, Yu-Geng
    Shanghai Jiaotong Daxue Xuebao/Journal of Shanghai Jiaotong University, 2005, 39 (08): : 1331 - 1335