Reinforcement learning in non-Markovian environments using automatic discovery of subgoals

被引:0
|
作者
Dung, Le Tien
Komeda, Takashi
Takagi, Motoki
机构
关键词
selected keywords relevant to the subject;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Learning time is always a critical issue in Reinforcement Learning, especially when Recurrent Neural Networks (RNNs) are used to predict Q values. By creating useful subgoals, we can speed up learning performance. In this paper, we propose a method to accelerate learning in non-Markovian environments using automatic discovery of subgoals. Once subgoals are created, sub-policies use RNNs to attain them `1hen learned RNNs are integrated into the main RNN as experts. Finally, the agent continues to learn using its new policy. Experiment results of the E maze problem and the virtual office problem show the potential of this approach.
引用
收藏
页码:2592 / 2596
页数:5
相关论文
共 50 条
  • [1] Reinforcement learning in non-Markovian environments
    Chandak, Siddharth
    Shah, Pratik
    Borkar, Vivek S.
    Dodhia, Parth
    SYSTEMS & CONTROL LETTERS, 2024, 185
  • [2] Automatic Discovery of Subgoals in Reinforcement Learning Using Strongly Connected Components
    Kazemitabar, Seyed Jalal
    Beigy, Hamid
    ADVANCES IN NEURO-INFORMATION PROCESSING, PT I, 2009, 5506 : 829 - 834
  • [3] Non-Markovian Reinforcement Learning using Fractional Dynamics
    Gupta, Gaurav
    Yin, Chenzhong
    Deshmukh, Jyotirmoy, V
    Bogdan, Paul
    2021 60TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2021, : 1542 - 1547
  • [4] Speeding up reinforcement learning using recurrent neural networks in non-Markovian environments
    Le, Tien Dung
    Komeda, Takashi
    Takagi, Motoki
    PROCEDINGS OF THE 11TH IASTED INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND SOFT COMPUTING, 2007, : 179 - 184
  • [5] Reinforcement Learning with Non-Markovian Rewards
    Gaon, Maor
    Brafman, Ronen, I
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 3980 - 3987
  • [6] Bayesian reinforcement learning in Markovian and non-Markovian tasks
    Ez-zizi, Adnane
    Farrell, Simon
    Leslie, David
    2015 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI), 2015, : 579 - 586
  • [7] K-cluster algorithm for automatic discovery of subgoals in reinforcement learning
    Wang, Ben-Nian
    Gao, Yang
    Chen, Zhao-Qian
    Xie, Jun-Yuan
    Chen, Shi-Fu
    INTERNATIONAL CONFERENCE ON COMPUTATIONAL INTELLIGENCE FOR MODELLING, CONTROL & AUTOMATION JOINTLY WITH INTERNATIONAL CONFERENCE ON INTELLIGENT AGENTS, WEB TECHNOLOGIES & INTERNET COMMERCE, VOL 1, PROCEEDINGS, 2006, : 658 - +
  • [8] APPLYING Q-LEARNING TO NON-MARKOVIAN ENVIRONMENTS
    Chizhov, Jurij
    Borisov, Arkady
    ICAART 2009: PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE, 2009, : 306 - +
  • [9] Autonomic discovery of subgoals in hierarchical reinforcement learning
    XIAO Ding
    LI Yi-tong
    SHI Chuan
    The Journal of China Universities of Posts and Telecommunications, 2014, (05) : 94 - 104
  • [10] Autonomic discovery of subgoals in hierarchical reinforcement learning
    XIAO Ding
    LI Yi-tong
    SHI Chuan
    TheJournalofChinaUniversitiesofPostsandTelecommunications, 2014, 21 (05) : 94 - 104