Monte Carlo Bayesian Hierarchical Reinforcement Learning

被引:0
|
作者
Ngo Anh Vien [1 ]
Hung Ngo [2 ]
Ertel, Wolfgang [3 ]
机构
[1] Univ Stuttgart, MLR Lab, Stuttgart, Germany
[2] SUPSI, USI, IDSIA, Manno, Switzerland
[3] Ravensburg Weingarten Univ Appl Sci, Weingarten, Germany
关键词
Reinforcement Learning; Bayesian RL; Hierarchical BRL;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we propose to use hierarchical action decomposition to make Bayesian model-based reinforcement learning more efficient and feasible in practice. We formulate Bayesian hierarchical reinforcement learning as a partially observable semi-Markov decision process (POSMDP). The main POSMDP task is partitioned into a hierarchy of POSMDP subtasks; lower-level subtasks get solved first, then higher-level ones. We sample from a prior belief to build an approximate model for each POSMDP, then solve using Monte Carlo Value Iteration with Macro-Actions solver. Experimental results show that our algorithm performs significantly better than that of flat BRL in terms of both reward, and especially solving time, in at least one order of magnitude.
引用
收藏
页码:1551 / 1552
页数:2
相关论文
共 50 条
  • [1] Monte Carlo Tree Search for Bayesian Reinforcement Learning
    Vien, Ngo Anh
    Ertel, Wolfgang
    [J]. 2012 11TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2012), VOL 1, 2012, : 138 - 143
  • [2] Monte-Carlo tree search for Bayesian reinforcement learning
    Ngo Anh Vien
    Ertel, Wolfgang
    Viet-Hung Dang
    Chung, TaeChoong
    [J]. APPLIED INTELLIGENCE, 2013, 39 (02) : 345 - 353
  • [3] Monte-Carlo tree search for Bayesian reinforcement learning
    Ngo Anh Vien
    Wolfgang Ertel
    Viet-Hung Dang
    TaeChoong Chung
    [J]. Applied Intelligence, 2013, 39 : 345 - 353
  • [4] Monte-Carlo utility estimates for Bayesian reinforcement learning
    Dimitrakakis, Christos
    [J]. 2013 IEEE 52ND ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2013, : 7303 - 7308
  • [5] Monte-Carlo Bayesian Reinforcement Learning Using a Compact Factored Representation
    Wu, Bo
    Feng, Yanpeng
    [J]. 2017 4TH INTERNATIONAL CONFERENCE ON INFORMATION SCIENCE AND CONTROL ENGINEERING (ICISCE), 2017, : 466 - 469
  • [6] Hierarchical Reinforcement Learning With Monte Carlo Tree Search in Computer Fighting Game
    Pinto, Ivan Pereira
    Coutinho, Luciano Reis
    [J]. IEEE TRANSACTIONS ON GAMES, 2019, 11 (03) : 290 - 295
  • [7] Monte Carlo Hierarchical Model Learning
    Menashe, Jacob
    Stone, Peter
    [J]. PROCEEDINGS OF THE 2015 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS (AAMAS'15), 2015, : 1985 - 1986
  • [8] Monte Carlo Hierarchical Model Learning
    Menashe, Jacob
    Stone, Peter
    [J]. PROCEEDINGS OF THE 2015 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS (AAMAS'15), 2015, : 771 - 779
  • [9] Hierarchical Bayesian Inverse Reinforcement Learning
    Choi, Jaedeug
    Kim, Kee-Eung
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2015, 45 (04) : 793 - 805
  • [10] On Monte Carlo Tree Search and Reinforcement Learning
    Vodopivec, Tom
    Samothrakis, Spyridon
    Ster, Branko
    [J]. JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2017, 60 : 881 - 936