Swarm Reinforcement Learning Method Based on Hierarchical Q-Learning

被引:0
|
作者
Kuroe, Yasuaki [1 ]
Takeuchi, Kenya [1 ]
Maeda, Yutaka [1 ]
机构
[1] Kansai Univ, Fac Engn Sci, Suita, Osaka, Japan
基金
日本学术振兴会;
关键词
reinforcement learning method; partially observed Markov decision process; hierarchical Q-learning; swarm intelligence;
D O I
10.1109/SSCI50451.2021.9659877
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In last decades the reinforcement learning method has attracted a great deal of attention and many studies have been done. However, this method is basically a trial-and-error scheme and it takes much computational time to acquire optimal strategies. Furthermore, optimal strategies may not be obtained for large and complicated problems with many states. To resolve these problems we have proposed the swarm reinforcement learning method, which is developed inspired by the multi-point search optimization methods. The Swarm reinforcement learning method has been extensively studied and its effectiveness has been confirmed for several problems, especially for Markov decision processes where the agents can fully observe the states of environments. In many real-world problems, however, the agents cannot fully observe the environments and they are usually partially observable Markov decision processes (POMDPs). The purpose of this paper is to develop a swarm reinforcement learning method which can deal with POMDPs. We propose a swarm reinforcement learning method based on HQ-learning, which is a hierarchical extension of Q-learning. It is shown through experiments that the proposed method can handle POMDPs and possesses higher performance than that of the original HQ-learning.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Deep Reinforcement Learning: From Q-Learning to Deep Q-Learning
    Tan, Fuxiao
    Yan, Pengfei
    Guan, Xinping
    NEURAL INFORMATION PROCESSING (ICONIP 2017), PT IV, 2017, 10637 : 475 - 483
  • [2] A New Learning Model for Swarm Intelligence Based on Q-Learning
    Li, Fuming
    He, Xiaoxian
    Xu, Jingjing
    2010 8TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION (WCICA), 2010, : 2769 - 2775
  • [3] SASDL and RBATQ: Sparse Autoencoder With Swarm Based Deep Learning and Reinforcement Based Q-Learning for EEG Classification
    Prabhakar, Sunil Kumar
    Lee, Seong-Whan
    IEEE OPEN JOURNAL OF ENGINEERING IN MEDICINE AND BIOLOGY, 2022, 3 : 58 - 68
  • [4] Q-learning based Reinforcement Learning Approach for Lane Keeping
    Feher, Arpad
    Aradi, Szilard
    Becsi, Tamas
    2018 18TH IEEE INTERNATIONAL SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE AND INFORMATICS (CINTI), 2018, : 31 - 35
  • [5] Q-learning based on hierarchical evolutionary mechanism
    Department of Information Engineering, Meijo University, 1-501, Tenpaku, Nagoya, Aichi, 468-8502, Japan
    不详
    WSEAS Trans. Syst. Control, 2008, 3 (219-228):
  • [6] Fuzzy Q-Learning for generalization of reinforcement learning
    Berenji, HR
    FUZZ-IEEE '96 - PROCEEDINGS OF THE FIFTH IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS 1-3, 1996, : 2208 - 2214
  • [7] Deep Reinforcement Learning with Double Q-Learning
    van Hasselt, Hado
    Guez, Arthur
    Silver, David
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 2094 - 2100
  • [8] Reinforcement learning guidance law of Q-learning
    Zhang Q.
    Ao B.
    Zhang Q.
    Xi Tong Gong Cheng Yu Dian Zi Ji Shu/Systems Engineering and Electronics, 2020, 42 (02): : 414 - 419
  • [9] Hybrid MDP based integrated hierarchical Q-learning
    TARN Tzyh-Jong
    Science China(Information Sciences), 2011, 54 (11) : 2279 - 2294
  • [10] Hybrid MDP based integrated hierarchical Q-learning
    ChunLin Chen
    DaoYi Dong
    Han-Xiong Li
    Tzyh-Jong Tarn
    Science China Information Sciences, 2011, 54 : 2279 - 2294