Inverse reinforcement learning through logic constraint inference

被引:2
|
作者
Baert, Mattijs [1 ]
Leroux, Sam [1 ]
Simoens, Pieter [1 ]
机构
[1] Univ Ghent, imec, Dept Informat Technol, IDLab, Technol pk 126, B-9052 Ghent, Belgium
关键词
Inductive logic programming; Inverse reinforcement learning; Answer set programming; Constraint inference; Constrained Markov decision process;
D O I
10.1007/s10994-023-06311-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Autonomous robots start to be integrated in human environments where explicit and implicit social norms guide the behavior of all agents. To assure safety and predictability, these artificial agents should act in accordance with the applicable social norms. However, it is not straightforward to define these rules and incorporate them in an agent's policy. Particularly because social norms are often implicit and environment specific. In this paper, we propose a novel iterative approach to extract a set of rules from observed human trajectories. This hybrid method combines the strengths of inverse reinforcement learning and inductive logic programming. We experimentally show how our method successfully induces a compact logic program which represents the behavioral constraints applicable in a Tower of Hanoi and a traffic simulator environment. The induced program is adopted as prior knowledge by a model-free reinforcement learning agent to speed up training and prevent any social norm violation during exploration and deployment. Moreover, expressing norms as a logic program provides improved interpretability, which is an important pillar in the design of safe artificial agents, as well as transferability to similar environments.
引用
收藏
页码:2593 / 2618
页数:26
相关论文
共 50 条
  • [1] Inverse reinforcement learning through logic constraint inference
    Mattijs Baert
    Sam Leroux
    Pieter Simoens
    Machine Learning, 2023, 112 : 2593 - 2618
  • [2] Inverse Reinforcement Learning with Constraint Recovery
    Das, Nirjhar
    Chattopadhyay, Arpan
    PATTERN RECOGNITION AND MACHINE INTELLIGENCE, PREMI 2023, 2023, 14301 : 179 - 188
  • [3] Guiding inference through relational reinforcement learning
    Asgharbeygi, N
    Nejati, N
    Langley, P
    Arai, S
    INDUCTIVE LOGIC PROGRAMMING, PROCEEDINGS, 2005, 3625 : 20 - 37
  • [4] Safety Constraint-Guided Reinforcement Learning with Linear Temporal Logic
    Kwon, Ryeonggu
    Kwon, Gihwon
    SYSTEMS, 2023, 11 (11):
  • [5] Learning from Demonstration for Shaping through Inverse Reinforcement Learning
    Suay, Halit Bener
    Brys, Tim
    Taylor, Matthew E.
    Chernova, Sonia
    AAMAS'16: PROCEEDINGS OF THE 2016 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS, 2016, : 429 - 437
  • [6] Spatiotemporal Costmap Inference for MPC Via Deep Inverse Reinforcement Learning
    Lee, Keuntaek
    Isele, David
    Theodorou, Evangelos A.
    Bae, Sangjae
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (02) : 3194 - 3201
  • [7] Active Task-Inference-Guided Deep Inverse Reinforcement Learning
    Memarian, Farzan
    Xu, Zhe
    Wu, Bo
    Wen, Min
    Topcu, Ufuk
    2020 59TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2020, : 1932 - 1938
  • [8] Drones Objective Inference Using Policy Error Inverse Reinforcement Learning
    Perrusquia, Adolfo
    Guo, Weisi
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (01) : 1329 - 1340
  • [9] Boosting learning and inference in Markov logic through metaheuristics
    Marenglen Biba
    Stefano Ferilli
    Floriana Esposito
    Applied Intelligence, 2011, 34 : 279 - 298
  • [10] Boosting learning and inference in Markov logic through metaheuristics
    Biba, Marenglen
    Ferilli, Stefano
    Esposito, Floriana
    APPLIED INTELLIGENCE, 2011, 34 (02) : 279 - 298