Expertness based cooperative Q-learning

被引:72
|
作者
Ahmadabadi, MN [1 ]
Asadpour, M
机构
[1] Univ Tehran, Dept Elect & Comp Engn, Tehran, Iran
[2] Inst Studies Theoret Phys & Math, Intelligent Syst Res Ctr, Tehran, Iran
关键词
cooperative learning; expertness; multi-agent systems; Q-learning;
D O I
10.1109/3477.979961
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
By using other agents' experiences and knowledge, a learning agent may learn faster, make fewer mistakes, and create some rules for unseen situations. These benefits would be gained if the learning agent can extract proper rules out of the other agents' knowledge for its own requirements. One possible way to do this is to have the learner assign some expertness values (intelligence level values) to the other agents and use their knowledge accordingly. In this paper, some criteria to measure the expertness of the reinforcement learning agents are introduced. Also, a new cooperative learning method, called weighted strategy sharing (WSS) is presented. In this method, each agent measures the expertness of its teammates and assigns a weight to their knowledge and learns from them accordingly. The presented methods are tested on two Hunter-Prey systems. We consider that the agents are all learning from each other and compare them with those who cooperate only with the more expert ones. Also, the effect of the communication noise, as a source of uncertainty, on the cooperative learning method is studied. Moreover, the Q-table of one of the cooperative agents is changed randomly and its effects on the presented methods are examined.
引用
收藏
页码:66 / 76
页数:11
相关论文
共 50 条
  • [41] Cooperative Spectrum Sensing for Cognitive Radios using Distributed Q-Learning
    van den Biggelaar, Olivier
    Dricot, Jean-Michel
    De Doncker, Philippe
    Horlin, Francois
    2011 IEEE VEHICULAR TECHNOLOGY CONFERENCE (VTC FALL), 2011,
  • [42] Cooperative behavior acquisition for multi-agent systems by Q-learning
    Xie, M. C.
    Tachibana, A.
    2007 IEEE SYMPOSIUM ON FOUNDATIONS OF COMPUTATIONAL INTELLIGENCE, VOLS 1 AND 2, 2007, : 424 - +
  • [43] Distributed Cooperative Q-learning for Power Allocation in Cognitive Femtocell Networks
    Saad, Hussein
    Mohamed, Amr
    ElBatt, Tamer
    2012 IEEE VEHICULAR TECHNOLOGY CONFERENCE (VTC FALL), 2012,
  • [44] Cooperative Path Planning for Single Leader Using Q-learning Method
    Zhang, Lichuan
    Wu, Dongwei
    Ren, Ranzhen
    Xing, Runfa
    GLOBAL OCEANS 2020: SINGAPORE - U.S. GULF COAST, 2020,
  • [45] A theoretical analysis of cooperative behaviorin multi-agent Q-learning
    Waltman, Ludo
    Kaymak, Uzay
    2007 IEEE INTERNATIONAL SYMPOSIUM ON APPROXIMATE DYNAMIC PROGRAMMING AND REINFORCEMENT LEARNING, 2007, : 84 - +
  • [46] Communication-Less Cooperative Q-Learning Agents in Maze Problem
    Uwano, Fumito
    Takadama, Keiki
    INTELLIGENT AND EVOLUTIONARY SYSTEMS, IES 2016, 2017, 8 : 453 - 467
  • [47] A Cooperative Q-learning Approach for Online Power Allocation in Femtocell Networks
    Saad, Hussein
    Mohamed, Amr
    ElBatt, Tamer
    2013 IEEE 78TH VEHICULAR TECHNOLOGY CONFERENCE (VTC FALL), 2013,
  • [48] An extension of weighted strategy sharing in cooperative Q-learning for specialized agents
    Eshgh, SM
    Ahmadabadi, MN
    ICONIP'02: PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING: COMPUTATIONAL INTELLIGENCE FOR THE E-AGE, 2002, : 106 - 110
  • [49] Non-Reciprocating Sharing Methods in Cooperative Q-Learning Environments
    Cunningham, Bryan
    Cao, Yong
    2012 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON WEB INTELLIGENCE AND INTELLIGENT AGENT TECHNOLOGY (WI-IAT 2012), VOL 2, 2012, : 212 - 219
  • [50] Cooperative Deep Q-Learning Framework for Environments Providing Image Feedback
    Raghavan, Krishnan
    Narayanan, Vignesh
    Jagannathan, Sarangapani
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (07) : 9267 - 9276