Expertness based cooperative Q-learning

被引:72
|
作者
Ahmadabadi, MN [1 ]
Asadpour, M
机构
[1] Univ Tehran, Dept Elect & Comp Engn, Tehran, Iran
[2] Inst Studies Theoret Phys & Math, Intelligent Syst Res Ctr, Tehran, Iran
关键词
cooperative learning; expertness; multi-agent systems; Q-learning;
D O I
10.1109/3477.979961
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
By using other agents' experiences and knowledge, a learning agent may learn faster, make fewer mistakes, and create some rules for unseen situations. These benefits would be gained if the learning agent can extract proper rules out of the other agents' knowledge for its own requirements. One possible way to do this is to have the learner assign some expertness values (intelligence level values) to the other agents and use their knowledge accordingly. In this paper, some criteria to measure the expertness of the reinforcement learning agents are introduced. Also, a new cooperative learning method, called weighted strategy sharing (WSS) is presented. In this method, each agent measures the expertness of its teammates and assigns a weight to their knowledge and learns from them accordingly. The presented methods are tested on two Hunter-Prey systems. We consider that the agents are all learning from each other and compare them with those who cooperate only with the more expert ones. Also, the effect of the communication noise, as a source of uncertainty, on the cooperative learning method is studied. Moreover, the Q-table of one of the cooperative agents is changed randomly and its effects on the presented methods are examined.
引用
收藏
页码:66 / 76
页数:11
相关论文
共 50 条
  • [1] Multi-criteria expertness based cooperative Q-learning
    Esmat Pakizeh
    Maziar Palhang
    Mir Mohsen Pedram
    Applied Intelligence, 2013, 39 : 28 - 40
  • [2] Multi-criteria expertness based cooperative Q-learning
    Pakizeh, Esmat
    Palhang, Maziar
    Pedram, Mir Mohsen
    APPLIED INTELLIGENCE, 2013, 39 (01) : 28 - 40
  • [3] Certainty and expertness-based credit assignment for cooperative Q-learning agents with an AND-type task
    Harati, A
    Ahmadabadi, MN
    ICONIP'02: PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING: COMPUTATIONAL INTELLIGENCE FOR THE E-AGE, 2002, : 306 - 310
  • [4] Cooperative Q-Learning Based on Learning Automata
    Yang, Mao
    Tian, Yantao
    Qi, Xinyue
    2009 IEEE INTERNATIONAL CONFERENCE ON AUTOMATION AND LOGISTICS ( ICAL 2009), VOLS 1-3, 2009, : 1972 - 1977
  • [5] Cooperative Q-Learning Based on Maturity of the Policy
    Yang, Mao
    Tian, Yantao
    Liu, Xiaomei
    2009 IEEE INTERNATIONAL CONFERENCE ON MECHATRONICS AND AUTOMATION, VOLS 1-7, CONFERENCE PROCEEDINGS, 2009, : 1352 - 1356
  • [6] EFFECTS OF COMMUNICATION IN COOPERATIVE Q-LEARNING
    Darbyshire, Paul
    Wang, Dianhui
    INTERNATIONAL JOURNAL OF INNOVATIVE COMPUTING INFORMATION AND CONTROL, 2010, 6 (05): : 2113 - 2126
  • [7] Learning Automata Based Q-Learning for Content Placement in Cooperative Caching
    Yang, Zhong
    Liu, Yuanwei
    Chen, Yue
    Jiao, Lei
    IEEE TRANSACTIONS ON COMMUNICATIONS, 2020, 68 (06) : 3667 - 3680
  • [8] Expertness measuring in cooperative learning
    Ahmadabadi, MN
    Asadpur, M
    Khodaabakhsh, SH
    Nakano, E
    2000 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2000), VOLS 1-3, PROCEEDINGS, 2000, : 2261 - 2267
  • [9] Cooperative Q-learning: the knowledge sharing issue
    Ahmadabadi, MN
    Asadpour, M
    Nakano, E
    ADVANCED ROBOTICS, 2001, 15 (08) : 815 - 832
  • [10] Cooperative Q-learning based channel selection for cognitive radio networks
    Feten Slimeni
    Zied Chtourou
    Bart Scheers
    Vincent Le Nir
    Rabah Attia
    Wireless Networks, 2019, 25 : 4161 - 4171