Expertness based cooperative Q-learning

被引:72
|
作者
Ahmadabadi, MN [1 ]
Asadpour, M
机构
[1] Univ Tehran, Dept Elect & Comp Engn, Tehran, Iran
[2] Inst Studies Theoret Phys & Math, Intelligent Syst Res Ctr, Tehran, Iran
关键词
cooperative learning; expertness; multi-agent systems; Q-learning;
D O I
10.1109/3477.979961
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
By using other agents' experiences and knowledge, a learning agent may learn faster, make fewer mistakes, and create some rules for unseen situations. These benefits would be gained if the learning agent can extract proper rules out of the other agents' knowledge for its own requirements. One possible way to do this is to have the learner assign some expertness values (intelligence level values) to the other agents and use their knowledge accordingly. In this paper, some criteria to measure the expertness of the reinforcement learning agents are introduced. Also, a new cooperative learning method, called weighted strategy sharing (WSS) is presented. In this method, each agent measures the expertness of its teammates and assigns a weight to their knowledge and learns from them accordingly. The presented methods are tested on two Hunter-Prey systems. We consider that the agents are all learning from each other and compare them with those who cooperate only with the more expert ones. Also, the effect of the communication noise, as a source of uncertainty, on the cooperative learning method is studied. Moreover, the Q-table of one of the cooperative agents is changed randomly and its effects on the presented methods are examined.
引用
收藏
页码:66 / 76
页数:11
相关论文
共 50 条
  • [31] Q-LEARNING BASED THERAPY MODELING
    Jacak, Witold
    Proell, Karin
    EMSS 2009: 21ST EUROPEAN MODELING AND SIMULATION SYMPOSIUM, VOL II, 2009, : 204 - +
  • [32] Q-LEARNING
    WATKINS, CJCH
    DAYAN, P
    MACHINE LEARNING, 1992, 8 (3-4) : 279 - 292
  • [33] Deep Reinforcement Learning: From Q-Learning to Deep Q-Learning
    Tan, Fuxiao
    Yan, Pengfei
    Guan, Xinping
    NEURAL INFORMATION PROCESSING (ICONIP 2017), PT IV, 2017, 10637 : 475 - 483
  • [34] Backward Q-learning: The combination of Sarsa algorithm and Q-learning
    Wang, Yin-Hao
    Li, Tzuu-Hseng S.
    Lin, Chih-Jui
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2013, 26 (09) : 2184 - 2193
  • [35] A Framework for Improved Cooperative Learning Algorithms with Expertness (ICLAE)
    Vidhate, Deepak A.
    Kulkarni, Parag
    ADVANCED COMPUTING AND COMMUNICATION TECHNOLOGIES, 2018, 562 : 149 - 160
  • [36] A fuzzy adaptive algorithm for expertness based cooperative learning, application to herding problem
    Akbarzadeh-T, MR
    Rezaei-S, H
    Naghibi-S, MB
    NAFIPS'2003: 22ND INTERNATIONAL CONFERENCE OF THE NORTH AMERICAN FUZZY INFORMATION PROCESSING SOCIETY - NAFIPS PROCEEDINGS, 2003, : 317 - 322
  • [37] Greedy action selection and pessimistic Q-value updates in cooperative Q-learning
    Kujirai, Toshihiro
    Yokota, Takayoshi
    2018 57TH ANNUAL CONFERENCE OF THE SOCIETY OF INSTRUMENT AND CONTROL ENGINEERS OF JAPAN (SICE), 2018, : 821 - 826
  • [38] Q-learning based cooperative multi-agent system applied to coordination of overcurrent relays
    Sadeh, J.
    Rahimiyan, M.
    Journal of Applied Sciences, 2008, 8 (21) : 3924 - 3930
  • [39] Two-Step Deep Reinforcement Q-Learning based Relay Selection in Cooperative WPCNs
    Tolebi, Gulnur
    Tsiftsis, Theodoros A.
    Nauryzbayev, Galymzhan
    2023 INTERNATIONAL BALKAN CONFERENCE ON COMMUNICATIONS AND NETWORKING, BALKANCOM, 2023,
  • [40] Cooperative Q-Learning for Multiple Secondary Users in Dynamic Spectrum Access
    Venkatraman, Pavithra
    Hamdaoui, Bechir
    2011 7TH INTERNATIONAL WIRELESS COMMUNICATIONS AND MOBILE COMPUTING CONFERENCE (IWCMC), 2011, : 238 - 242