Expertness based cooperative Q-learning

被引:72
|
作者
Ahmadabadi, MN [1 ]
Asadpour, M
机构
[1] Univ Tehran, Dept Elect & Comp Engn, Tehran, Iran
[2] Inst Studies Theoret Phys & Math, Intelligent Syst Res Ctr, Tehran, Iran
关键词
cooperative learning; expertness; multi-agent systems; Q-learning;
D O I
10.1109/3477.979961
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
By using other agents' experiences and knowledge, a learning agent may learn faster, make fewer mistakes, and create some rules for unseen situations. These benefits would be gained if the learning agent can extract proper rules out of the other agents' knowledge for its own requirements. One possible way to do this is to have the learner assign some expertness values (intelligence level values) to the other agents and use their knowledge accordingly. In this paper, some criteria to measure the expertness of the reinforcement learning agents are introduced. Also, a new cooperative learning method, called weighted strategy sharing (WSS) is presented. In this method, each agent measures the expertness of its teammates and assigns a weight to their knowledge and learns from them accordingly. The presented methods are tested on two Hunter-Prey systems. We consider that the agents are all learning from each other and compare them with those who cooperate only with the more expert ones. Also, the effect of the communication noise, as a source of uncertainty, on the cooperative learning method is studied. Moreover, the Q-table of one of the cooperative agents is changed randomly and its effects on the presented methods are examined.
引用
收藏
页码:66 / 76
页数:11
相关论文
共 50 条
  • [21] A study on expertise of agents and its effects on cooperative Q-learning
    Araabi, Babak Nadjar
    Mastoureshgh, Sahar
    Ahmadabadi, Majid Nili
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2007, 37 (02): : 398 - 409
  • [22] Predicting and Preventing Coordination Problems in Cooperative Q-learning Systems
    Fulda, Nancy
    Ventura, Dan
    20TH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2007, : 780 - 785
  • [23] Behavior Control of Cooperative Vehicle Infrastructure System in Container Terminals Based on Q-learning
    Wu, Maopu
    Gao, Jian
    Li, Le
    Wang, Yue
    ARTIFICIAL INTELLIGENCE AND ROBOTICS, ISAIR 2022, PT II, 2022, 1701 : 240 - 246
  • [24] Multi-UAV Cooperative Task Assignment Based on Half Random Q-Learning
    Zhu, Pengxing
    Fang, Xi
    SYMMETRY-BASEL, 2021, 13 (12):
  • [25] Sequential Q-Learning With Kalman Filtering for Multirobot Cooperative Transportation
    Wang, Ying
    de Silva, Clarence W.
    IEEE-ASME TRANSACTIONS ON MECHATRONICS, 2010, 15 (02) : 261 - 268
  • [26] Evaluating cooperative-competitive dynamics with deep Q-learning
    Kopacz, Aniko
    Csato, Lehel
    Chira, Camelia
    NEUROCOMPUTING, 2023, 550
  • [27] Sparse Cooperative Multi-agent Q-learning Based on Vector Potential Field
    Liu, Liang
    Li, Longshu
    PROCEEDINGS OF THE 2009 WRI GLOBAL CONGRESS ON INTELLIGENT SYSTEMS, VOL I, 2009, : 99 - 103
  • [28] Based on switching-model cooperative control for two intersections applying Q-learning
    Zhao, Xiao-Hua
    Li, Zhen-Long
    Yu, Quan
    Zhang, Jie
    Beijing Gongye Daxue Xuebao / Journal of Beijing University of Technology, 2007, 33 (11): : 1148 - 1152
  • [29] Multi-robot Cooperative Planning by Consensus Q-learning
    Sadhu, Arup Kumar
    Konar, Amit
    Banerjee, Bonny
    Nagar, Atulya K.
    2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 4158 - 4164
  • [30] Cooperative Spectrum Sensing Using Q-Learning with Experimental Validation
    Chen, Zhe
    Qiu, Robert C.
    IEEE SOUTHEASTCON 2011: BUILDING GLOBAL ENGINEERS, 2011, : 405 - 408