Constructing and Evaluating Options in Reinforcement Learning

被引:0
|
作者
Farahani, Marzieh Davoodabadi [1 ]
Mozayani, Nasser [1 ]
机构
[1] Iran Univ Sci & Technol, Dept Comp Engn, Tehran, Iran
关键词
Hierarchical Reinforcement Learning; Temporal Abstraction; Option; Community Detection; Macro-Action Evaluation;
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In this paper, we propose a new subgoal based method for automatic construction of useful options. In our proposed method, subgoals are considered as border states of communities of the transition graph created after some initial agent interactions with the environment. We present a new community detection algorithm to provide an appropriate partitioning of the transition graph. Macro-actions are constructed for taking the agent from one community to other communities. In addition, we attempt to capture intuitions about features of useful macro-actions. There is a lack of a generic evaluation mechanism for evaluating each macro-action in previous research. We will propose a method for evaluating each macro-action separately. Inappropriate macro-actions are identified with this method and discarded from agent choices. Experimental results show a significant improvement in results after pruning macro-actions.
引用
收藏
页码:183 / 186
页数:4
相关论文
共 50 条
  • [1] Learning Options in Multiobjective Reinforcement Learning
    Bonini, Rodrigo Cesar
    da Silva, Felipe Leno
    Reali Costa, Anna Helena
    [J]. THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 4907 - 4908
  • [2] Reinforcement Learning for Options Trading
    Wen, Wen
    Yuan, Yuyu
    Yang, Jincui
    [J]. APPLIED SCIENCES-BASEL, 2021, 11 (23):
  • [3] Robustly Learning Composable Options in Deep Reinforcement Learning
    Bagaria, Akhil
    Senthil, Jason
    Slivinski, Matthew
    Konidaris, George
    [J]. PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2161 - 2169
  • [4] Constructing Temporal Abstractions Autonomously in Reinforcement Learning
    Bacon, Pierre-Luc
    Precup, Doina
    [J]. AI MAGAZINE, 2018, 39 (01) : 39 - 50
  • [5] Probabilistic inference for determining options in reinforcement learning
    Daniel, Christian
    van Hoof, Herke
    Peters, Jan
    Neumann, Gerhard
    [J]. MACHINE LEARNING, 2016, 104 (2-3) : 337 - 357
  • [6] Probabilistic inference for determining options in reinforcement learning
    Christian Daniel
    Herke van Hoof
    Jan Peters
    Gerhard Neumann
    [J]. Machine Learning, 2016, 104 : 337 - 357
  • [7] Evaluating the Performance of Reinforcement Learning Algorithms
    Jordan, Scott M.
    Chandak, Yash
    Cohen, Daniel
    Zhang, Mengxue
    Thomas, Philip S.
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [8] Evaluating skills in hierarchical reinforcement learning
    Marzieh Davoodabadi Farahani
    Nasser Mozayani
    [J]. International Journal of Machine Learning and Cybernetics, 2020, 11 : 2407 - 2420
  • [9] Evaluating skills in hierarchical reinforcement learning
    Farahani, Marzieh Davoodabadi
    Mozayani, Nasser
    [J]. INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2020, 11 (10) : 2407 - 2420
  • [10] Learning relational options for inductive transfer in relational reinforcement learning
    Croonenborghs, Tom
    Driessens, Kurt
    Bruynooghe, Maurice
    [J]. INDUCTIVE LOGIC PROGRAMMING, 2008, 4894 : 88 - 97