Safe state abstraction and reusable continuing subtasks in hierarchical reinforcement learning

被引:0
|
作者
Hengst, Bernhard [1 ]
机构
[1] NICTA, Making Sense Data Res Grp, Sydney, NSW, Australia
关键词
hierarchical reinforcement learning; state abstraction; task hierarchies; decomposition;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Hierarchical reinforcement learning methods have not been able to simultaneously abstract and reuse subtasks with discounted value functions. The contribution of this paper is to introduce two completion functions that jointly decompose the value function hierarchically to solve this problem. The significance of this result is that the benefits of hierarchical reinforcement learning can be extended to discounted value functions and to continuing (infinite horizon) reinforcement learning problems. This paper demonstrates the method with the an algorithm that discovers subtasks automatically. An example is given where the optimum policy requires a subtask never to terminate.
引用
收藏
页码:58 / 67
页数:10
相关论文
共 50 条
  • [41] On Normative Reinforcement Learning via Safe Reinforcement Learning
    Neufeld, Emery A.
    Bartocci, Ezio
    Ciabattoni, Agata
    [J]. PRIMA 2022: PRINCIPLES AND PRACTICE OF MULTI-AGENT SYSTEMS, 2023, 13753 : 72 - 89
  • [42] Learning of deterministic exploration and temporal abstraction in reinforcement learning
    Shibata, Katsunari
    [J]. 2006 SICE-ICASE International Joint Conference, Vols 1-13, 2006, : 2212 - 2217
  • [43] A Hierarchical Deep Reinforcement Learning Framework With High Efficiency and Generalization for Fast and Safe Navigation
    Zhu, Wei
    Hayashibe, Mitsuhiro
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2023, 70 (05) : 4962 - 4971
  • [44] Multi-objective safe reinforcement learning: the relationship between multi-objective reinforcement learning and safe reinforcement learning
    Horie, Naoto
    Matsui, Tohgoroh
    Moriyama, Koichi
    Mutoh, Atsuko
    Inuzuka, Nobuhiro
    [J]. ARTIFICIAL LIFE AND ROBOTICS, 2019, 24 (03) : 352 - 359
  • [45] Safe Reinforcement Learning: A Survey
    Wang X.-S.
    Wang R.-R.
    Cheng Y.-H.
    [J]. Zidonghua Xuebao/Acta Automatica Sinica, 2023, 49 (09): : 1813 - 1835
  • [46] Multi-objective safe reinforcement learning: the relationship between multi-objective reinforcement learning and safe reinforcement learning
    Naoto Horie
    Tohgoroh Matsui
    Koichi Moriyama
    Atsuko Mutoh
    Nobuhiro Inuzuka
    [J]. Artificial Life and Robotics, 2019, 24 : 352 - 359
  • [47] Lightweight Reinforcement Learning with State Abstraction for Dynamic Spectrum Anti-Jamming Communications
    Liu, Xin
    Wang, Ximing
    Xu, Yuhua
    Du, Zhiyong
    Xu, Yifan
    Han, Hao
    [J]. IEEE Wireless Communications and Networking Conference, WCNC, 2024,
  • [48] Lightweight Reinforcement Learning with State Abstraction for Dynamic Spectrum Anti-Jamming Communications
    Liu, Xin
    Wang, Ximing
    Xu, Yuhua
    Du, Zhiyong
    Xu, Yifan
    Han, Hao
    [J]. 2024 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE, WCNC 2024, 2024,
  • [49] Hierarchical reinforcement learning algorithm based on structural state-space
    Meng, Jiang-Hua
    Zhu, Ji-Hong
    Sun, Zeng-Qi
    [J]. Kongzhi yu Juece/Control and Decision, 2007, 22 (02): : 233 - 237
  • [50] Boosting Reinforcement Learning via Hierarchical Game Playing With State Relay
    Liu, Chanjuan
    Cong, Jinmiao
    Liu, Guangyuan
    Jiang, Guifei
    Xu, Xirong
    Zhu, Enqiang
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, : 1 - 13