Safe state abstraction and reusable continuing subtasks in hierarchical reinforcement learning

被引:0
|
作者
Hengst, Bernhard [1 ]
机构
[1] NICTA, Making Sense Data Res Grp, Sydney, NSW, Australia
关键词
hierarchical reinforcement learning; state abstraction; task hierarchies; decomposition;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Hierarchical reinforcement learning methods have not been able to simultaneously abstract and reuse subtasks with discounted value functions. The contribution of this paper is to introduce two completion functions that jointly decompose the value function hierarchically to solve this problem. The significance of this result is that the benefits of hierarchical reinforcement learning can be extended to discounted value functions and to continuing (infinite horizon) reinforcement learning problems. This paper demonstrates the method with the an algorithm that discovers subtasks automatically. An example is given where the optimum policy requires a subtask never to terminate.
引用
收藏
页码:58 / 67
页数:10
相关论文
共 50 条
  • [1] State abstraction in MAXQ hierarchical reinforcement learning
    Dietterich, TG
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 12, 2000, 12 : 994 - 1000
  • [2] Language as an Abstraction for Hierarchical Deep Reinforcement Learning
    Jiang, Yiding
    Gu, Shixiang
    Murphy, Kevin
    Finn, Chelsea
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [3] Towards Safe Continuing Task Reinforcement Learning
    Calvo-Fullana, Miguel
    Chamon, Luiz F. O.
    Paternain, Santiago
    [J]. 2021 AMERICAN CONTROL CONFERENCE (ACC), 2021, : 902 - 908
  • [4] Regularizing Reinforcement Learning with State Abstraction
    Akrour, Riad
    Veiga, Filipe
    Peters, Jan
    Neumann, Gerhard
    [J]. 2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 534 - 539
  • [5] Uniform State Abstraction for Reinforcement Learning
    Burden, John
    Kudenko, Daniel
    [J]. ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 1031 - 1038
  • [6] A Theory of State Abstraction for Reinforcement Learning
    Abel, David
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 9876 - 9877
  • [7] A Core Task Abstraction Approach to Hierarchical Reinforcement Learning
    Li, Zhuoru
    Narayan, Akshay
    Leong, Tze-Yun
    [J]. AAMAS'16: PROCEEDINGS OF THE 2016 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS, 2016, : 1411 - 1412
  • [8] Towards Hierarchical Task Decomposition using Deep Reinforcement Learning for Pick and Place Subtasks
    Marzari, Luca
    Pore, Ameya
    Dall'Alba, Diego
    Aragon-Camarasa, Gerardo
    Farinelli, Alessandro
    Fiorini, Paolo
    [J]. 2021 20TH INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS (ICAR), 2021, : 640 - 645
  • [9] State abstraction for programmable reinforcement learning agents
    Andre, D
    Russell, SJ
    [J]. EIGHTEENTH NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE (AAAI-02)/FOURTEENTH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE (IAAI-02), PROCEEDINGS, 2002, : 119 - 125
  • [10] Safe Offline Reinforcement Learning Through Hierarchical Policies
    Liu, Shaofan
    Sun, Shiliang
    [J]. ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2022, PT II, 2022, 13281 : 380 - 391