Combining reinforcement learning with symbolic planning

被引:0
|
作者
Grounds, Matthew [1 ]
Kudenko, Daniel [1 ]
机构
[1] Univ York, Dept Comp Sci, York YO10 5DD, N Yorkshire, England
来源
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
One of the major difficulties in applying Q-learning to real-world domains is the sharp increase in the number of learning steps required to converge towards an optimal policy as the size of the state space is increased. In this paper we propose a method, PLANQ-learning, that couples a Q-learner with a STRIPS planner. The planner shapes the reward function, and thus guides the Q-learner quickly to the optimal policy. We demonstrate empirically that this combination of high-level reasoning and low-level learning displays significant improvements in scaling-up behaviour as the state-space grows larger, compared to both standard Q-learning and hierarchical Q-learning methods.
引用
收藏
页码:75 / 86
页数:12
相关论文
共 50 条
  • [1] Reinforcement Symbolic Learning
    Mercier, Chloe
    Alexandre, Frederic
    Vieville, Thierry
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT IV, 2021, 12894 : 608 - 612
  • [2] Integrating Symbolic Planning and Reinforcement Learning for Following Temporal Logic Specifications
    Xu, Duo
    Fekri, Faramarz
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [3] Adapting to the "Open World": The Utility of Hybrid Hierarchical Reinforcement Learning and Symbolic Planning
    Lorang, Pierrick
    Horvath, Helmut
    Kietreiber, Tobias
    Zips, Patrik
    Heitzinger, Clemens
    Scheutz, Matthias
    2024 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA 2024, 2024, : 508 - 514
  • [4] COMBINING SYMBOLIC AND NEURAL LEARNING
    SLAVLIK, JW
    MACHINE LEARNING, 1994, 14 (03) : 321 - 331
  • [5] Combining planning with reinforcement learning for multi-robot task allocation
    Strens, M
    Windelinckx, N
    ADAPTIVE AGENTS AND MULTI-AGENT SYSTEMS II: ADAPTATION AND MULTI-AGENT LEARNING, 2005, 3394 : 260 - 274
  • [6] Learning Intrinsic Symbolic Rewards in Reinforcement Learning
    Sheikh, Hassam Ullah
    Khadka, Shauharda
    Miret, Santiago
    Majumdar, Somdeb
    Phielipp, Mariano
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [7] SDRL: Interpretable and Data-efficient Deep Reinforcement Learning Leveraging Symbolic Planning
    Lyu, Daoming
    Yang, Fangkai
    Liu, Bo
    Gustafson, Steven
    ELECTRONIC PROCEEDINGS IN THEORETICAL COMPUTER SCIENCE, 2019, (306): : 354 - 354
  • [8] PEORL: Integrating Symbolic Planning and Hierarchical Reinforcement Learning for Robust Decision-Making
    Yang, Fangkai
    Lyu, Daoming
    Liu, Bo
    Gustafson, Steven
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 4860 - 4866
  • [9] SDRL: Interpretable and Data-Efficient Deep Reinforcement Learning Leveraging Symbolic Planning
    Lyu, Daoming
    Yang, Fangkai
    Liu, Bo
    Gustafson, Steven
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 2970 - 2977
  • [10] Combining Planning and Deep Reinforcement Learning in Tactical Decision Making for Autonomous Driving
    Hoel, Carl-Johan
    Driggs-Campbell, Katherine
    Wolff, Krister
    Laine, Leo
    Kochenderfer, Mykel J.
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2020, 5 (02): : 294 - 305