State abstraction in MAXQ hierarchical reinforcement learning

被引:0
|
作者
Dietterich, TG [1 ]
机构
[1] Oregon State Univ, Dept Comp Sci, Corvallis, OR 97331 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Many researchers have explored methods for hierarchical reinforcement learning (RL) with temporal abstractions, in which abstract actions are defined that can perform many primitive actions before terminating. However, little is known about learning with state abstractions, in which aspects of the state space are ignored. In previous work, we developed the MAXQ method for hierarchical RL. In this paper, we define five conditions under which state abstraction can be combined with the MAXQ value function decomposition. We prove that the MAXQ-Q learning algorithm converges under these conditions and show experimentally that state abstraction is important for the successful application of MAXQ-Q learning.
引用
收藏
页码:994 / 1000
页数:7
相关论文
共 50 条
  • [1] Hierarchical reinforcement learning with the MAXQ value function decomposition
    Dietterich, TG
    [J]. JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2000, 13 : 227 - 303
  • [2] Hierarchical Reinforcement Learning with the MAXQ Value Function Decomposition
    Dietterich, Thomas G.
    [J]. Journal of Artificial Intelligence Research, 2001, 13 (00): : 227 - 303
  • [3] Multi-agent hierarchical reinforcement learning by integrating options into MAXQ
    Shen, Jing
    Gu, Guochang
    Liu, Haibo
    [J]. FIRST INTERNATIONAL MULTI-SYMPOSIUMS ON COMPUTER AND COMPUTATIONAL SCIENCES (IMSCCS 2006), PROCEEDINGS, VOL 1, 2006, : 676 - +
  • [4] Safe state abstraction and reusable continuing subtasks in hierarchical reinforcement learning
    Hengst, Bernhard
    [J]. AI 2007: ADVANCES IN ARTIFICIAL INTELLIGENCE, PROCEEDINGS, 2007, 4830 : 58 - 67
  • [5] Language as an Abstraction for Hierarchical Deep Reinforcement Learning
    Jiang, Yiding
    Gu, Shixiang
    Murphy, Kevin
    Finn, Chelsea
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [6] Regularizing Reinforcement Learning with State Abstraction
    Akrour, Riad
    Veiga, Filipe
    Peters, Jan
    Neumann, Gerhard
    [J]. 2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 534 - 539
  • [7] Uniform State Abstraction for Reinforcement Learning
    Burden, John
    Kudenko, Daniel
    [J]. ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 1031 - 1038
  • [8] A Theory of State Abstraction for Reinforcement Learning
    Abel, David
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 9876 - 9877
  • [9] A Core Task Abstraction Approach to Hierarchical Reinforcement Learning
    Li, Zhuoru
    Narayan, Akshay
    Leong, Tze-Yun
    [J]. AAMAS'16: PROCEEDINGS OF THE 2016 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS, 2016, : 1411 - 1412
  • [10] State abstraction for programmable reinforcement learning agents
    Andre, D
    Russell, SJ
    [J]. EIGHTEENTH NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE (AAAI-02)/FOURTEENTH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE (IAAI-02), PROCEEDINGS, 2002, : 119 - 125