Combining reinforcement learning with symbolic planning

被引:0
|
作者
Grounds, Matthew [1 ]
Kudenko, Daniel [1 ]
机构
[1] Univ York, Dept Comp Sci, York YO10 5DD, N Yorkshire, England
来源
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
One of the major difficulties in applying Q-learning to real-world domains is the sharp increase in the number of learning steps required to converge towards an optimal policy as the size of the state space is increased. In this paper we propose a method, PLANQ-learning, that couples a Q-learner with a STRIPS planner. The planner shapes the reward function, and thus guides the Q-learner quickly to the optimal policy. We demonstrate empirically that this combination of high-level reasoning and low-level learning displays significant improvements in scaling-up behaviour as the state-space grows larger, compared to both standard Q-learning and hierarchical Q-learning methods.
引用
收藏
页码:75 / 86
页数:12
相关论文
共 50 条
  • [21] Supplementing neural reinforcement learning with symbolic methods
    Sun, R
    HYBRID NEURAL SYSTEMS, 2000, 1778 : 333 - 347
  • [22] Discovering symbolic policies with deep reinforcement learning
    Landajuela, Mikel
    Petersen, Brenden K.
    Kim, Sookyung
    Santiago, Claudio P.
    Glatt, Ruben
    Mundhenk, T. Nathan
    Pettit, Jacob F.
    Faissol, Daniel M.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [23] Contrastive Reinforcement Learning of Symbolic Reasoning Domains
    Poesia, Gabriel
    Dong, WenXin
    Goodman, Noah
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [24] Symbolic Task Inference in Deep Reinforcement Learning
    Hasanbeig, Hosein
    Jeppu, Natasha Yogananda
    Abate, Alessandro
    Melham, Tom
    Kroening, Daniel
    JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2024, 80 : 1099 - 1137
  • [25] Symbolic Task Inference in Deep Reinforcement Learning
    Hasanbeig, Hosein
    Jeppu, Natasha Yogananda
    Abate, Alessandro
    Melham, Tom
    Kroening, Daniel
    Journal of Artificial Intelligence Research, 2024, 80 : 1099 - 1137
  • [26] Combining finite element analysis and reinforcement learning for optimal grip point planning of flexible components
    Roemer, Martin
    Demircan, Fatih
    Huerkamp, Andre
    Droeder, Klaus
    PRODUCTION ENGINEERING-RESEARCH AND DEVELOPMENT, 2024,
  • [27] Safe RAN control: A Symbolic Reinforcement Learning Approach
    Nikou, Alexandros
    Mujumdar, Anusha
    Sundararajan, Vaishnavi
    Orlic, Marin
    Feljan, Aneta Vulgarakis
    2022 IEEE 17TH INTERNATIONAL CONFERENCE ON CONTROL & AUTOMATION, ICCA, 2022, : 332 - 337
  • [28] Reinforcement Learning with Symbolic Input-Output Models
    Derner, Erik
    Kubalik, Jiri
    Babska, Robert
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 3004 - 3009
  • [29] Learning Symbolic Representations for Planning with Parameterized Skills
    Ames, Barrett
    Thackston, Allison
    Konidaris, George
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 526 - 533
  • [30] Learning Symbolic Operators for Task and Motion Planning
    Silver, Tom
    Chitnis, Rohan
    Tenenbaum, Joshua
    Kaelbling, Leslie Pack
    Lozano-Perez, Tomas
    2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 3182 - 3189