Hierarchical Task Decomposition through Symbiosis in Reinforcement Learning

被引:16
|
作者
Doucette, John A. [1 ]
Lichodzijewski, Peter [1 ]
Heywood, Malcolm I. [1 ]
机构
[1] Univ Waterloo, Sch Comp Sci, Waterloo, ON N2L 3G1, Canada
关键词
Symbiosis; Reinforcement learning; Meta actions; Task decomposition; Genetic Programming; NEURAL NETWORKS;
D O I
10.1145/2330163.2330178
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Adopting a symbiotic model of evolution separates context for deploying an action from the action itself. Such a separation provides a mechanism for task decomposition in temporal sequence learning. Moreover, previously learnt policies are taken to be synonymous with meta actions (actions that are themselves policies). Should solutions to the task not be forthcoming in an initial round of evolution, then solutions from the earlier round represent the 'meta' actions for a new round of evolution. This provides the basis for evolving policy trees. A benchmarking study is performed using the Acrobot handstand task. Solutions to date from reinforcement learning have not been able to approach the performance of those established 14 years ago using an A* search and a priori knowledge regarding the Acrobot energy equations. The proposed symbiotic approach is able to match and, for the first time, better these results. Moreover, unlike previous work, solutions are tested under a broad range of Acrobot initial conditions, with hierarchical solutions providing significantly better generalization performance.
引用
收藏
页码:97 / 104
页数:8
相关论文
共 50 条
  • [1] Hierarchical Task and Motion Planning through Deep Reinforcement Learning
    Newaz, Abdullah Al Redwan
    Alam, Tauhidul
    [J]. 2021 FIFTH IEEE INTERNATIONAL CONFERENCE ON ROBOTIC COMPUTING (IRC 2021), 2021, : 100 - 105
  • [2] Towards Hierarchical Task Decomposition using Deep Reinforcement Learning for Pick and Place Subtasks
    Marzari, Luca
    Pore, Ameya
    Dall'Alba, Diego
    Aragon-Camarasa, Gerardo
    Farinelli, Alessandro
    Fiorini, Paolo
    [J]. 2021 20TH INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS (ICAR), 2021, : 640 - 645
  • [3] Hierarchical reinforcement learning for Metrical Task Systems
    de Lima, ML
    de Melo, JD
    Neto, ADD
    [J]. HIS 2005: 5TH INTERNATIONAL CONFERENCE ON HYBRID INTELLIGENT SYSTEMS, PROCEEDINGS, 2005, : 251 - 256
  • [4] Hierarchical Sub-task Decomposition for Reinforcement Learning of Multi-robot Delivery Mission
    Kawano, Hiroshi
    [J]. 2013 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2013, : 828 - 835
  • [5] Learning Task Decomposition and Exploration Shaping for Reinforcement Learning Agents
    Djurdjevic, Predrag
    Huber, Manfred
    [J]. 2008 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC), VOLS 1-6, 2008, : 365 - 372
  • [6] A Core Task Abstraction Approach to Hierarchical Reinforcement Learning
    Li, Zhuoru
    Narayan, Akshay
    Leong, Tze-Yun
    [J]. AAMAS'16: PROCEEDINGS OF THE 2016 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS, 2016, : 1411 - 1412
  • [7] Hierarchical Reinforcement Learning Explains Task Interleaving Behavior
    Gebhardt C.
    Oulasvirta A.
    Hilliges O.
    [J]. Computational Brain & Behavior, 2021, 4 (3) : 284 - 304
  • [8] Hierarchical Reinforcement Learning with the MAXQ Value Function Decomposition
    Dietterich, Thomas G.
    [J]. Journal of Artificial Intelligence Research, 2001, 13 (00): : 227 - 303
  • [9] Hierarchical reinforcement learning with the MAXQ value function decomposition
    Dietterich, TG
    [J]. JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2000, 13 : 227 - 303
  • [10] Research on task decomposition and state abstraction in reinforcement learning
    Yu Lasheng
    Jiang Zhongbin
    Liu Kang
    [J]. Artificial Intelligence Review, 2012, 38 : 119 - 127