On Efficiency in Hierarchical Reinforcement Learning

被引:0
|
作者
Wen, Zheng [1 ]
Precup, Doina [1 ]
Ibrahimi, Morteza [1 ]
Barreto, Andre [1 ]
Van Roy, Benjamin [1 ]
Singh, Satinder [1 ]
机构
[1] DeepMind, London, England
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Hierarchical Reinforcement Learning (HRL) approaches promise to provide more efficient solutions to sequential decision making problems, both in terms of statistical as well as computational efficiency. While this has been demonstrated empirically over time in a variety of tasks, theoretical results quantifying the benefits of such methods are still few and far between. In this paper, we discuss the kind of structure in a Markov decision process which gives rise to efficient HRL methods. Specifically, we formalize the intuition that HRL can exploit well repeating "subMDPs", with similar reward and transition structure. We show that, under reasonable assumptions, a model-based Thompson sampling-style HRL algorithm that exploits this structure is statistically efficient, as established through a finite-time regret bound. We also establish conditions under which planning with structure-induced options is near-optimal and computationally efficient.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Autonomous Reinforcement Learning with Hierarchical REPS
    Daniel, Christian
    Neumann, Gerhard
    Peters, Jan
    [J]. 2013 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2013,
  • [32] Hierarchical Reinforcement Learning for Integrated Recommendation
    Xie, Ruobing
    Zhang, Shaoliang
    Wang, Rui
    Xia, Feng
    Lin, Leyu
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 4521 - 4528
  • [33] Scalable Evolutionary Hierarchical Reinforcement Learning
    Abramowitz, Sasha
    Nitschke, Geoff
    [J]. PROCEEDINGS OF THE 2022 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION, GECCO 2022, 2022, : 272 - 275
  • [34] Evaluating skills in hierarchical reinforcement learning
    Farahani, Marzieh Davoodabadi
    Mozayani, Nasser
    [J]. INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2020, 11 (10) : 2407 - 2420
  • [35] A Neural Signature of Hierarchical Reinforcement Learning
    Ribas-Fernandes, Jose J. F.
    Solway, Alec
    Diuk, Carlos
    McGuire, Joseph T.
    Barto, Andrew G.
    Niv, Yael
    Botvinick, Matthew M.
    [J]. NEURON, 2011, 71 (02) : 370 - 379
  • [36] Optimal Hierarchical Learning Path Design With Reinforcement Learning
    Li, Xiao
    Xu, Hanchen
    Zhang, Jinming
    Chang, Hua-hua
    [J]. APPLIED PSYCHOLOGICAL MEASUREMENT, 2021, 45 (01) : 54 - 70
  • [37] Learning Generalizable Locomotion Skills with Hierarchical Reinforcement Learning
    Li, Tianyu
    Lambert, Nathan
    Calandra, Roberto
    Meier, Franziska
    Rai, Akshara
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 413 - 419
  • [38] Effectively Learning Initiation Sets in Hierarchical Reinforcement Learning
    Bagaria, Akhil
    Abbatematteo, Ben
    Gottesman, Omer
    Corsaro, Matt
    Rammohan, Sreehari
    Konidaris, George
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [39] Hierarchical Reinforcement Learning Under Mixed Observability
    Hai Nguyen
    Yang, Zhihan
    Baisero, Andrea
    Ma, Xiao
    Platt, Robert
    Amato, Christopher
    [J]. ALGORITHMIC FOUNDATIONS OF ROBOTICS XV, 2023, 25 : 188 - 204
  • [40] Hierarchical reinforcement learning model for military simulations
    Sidhu, Amandeep Singh
    Chaudhari, Narendra S.
    Goh, Ghee Ming
    [J]. 2006 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORK PROCEEDINGS, VOLS 1-10, 2006, : 2572 - +