Inverse reinforcement learning through logic constraint inference

被引:2
|
作者
Baert, Mattijs [1 ]
Leroux, Sam [1 ]
Simoens, Pieter [1 ]
机构
[1] Univ Ghent, imec, Dept Informat Technol, IDLab, Technol pk 126, B-9052 Ghent, Belgium
关键词
Inductive logic programming; Inverse reinforcement learning; Answer set programming; Constraint inference; Constrained Markov decision process;
D O I
10.1007/s10994-023-06311-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Autonomous robots start to be integrated in human environments where explicit and implicit social norms guide the behavior of all agents. To assure safety and predictability, these artificial agents should act in accordance with the applicable social norms. However, it is not straightforward to define these rules and incorporate them in an agent's policy. Particularly because social norms are often implicit and environment specific. In this paper, we propose a novel iterative approach to extract a set of rules from observed human trajectories. This hybrid method combines the strengths of inverse reinforcement learning and inductive logic programming. We experimentally show how our method successfully induces a compact logic program which represents the behavioral constraints applicable in a Tower of Hanoi and a traffic simulator environment. The induced program is adopted as prior knowledge by a model-free reinforcement learning agent to speed up training and prevent any social norm violation during exploration and deployment. Moreover, expressing norms as a logic program provides improved interpretability, which is an important pillar in the design of safe artificial agents, as well as transferability to similar environments.
引用
收藏
页码:2593 / 2618
页数:26
相关论文
共 50 条
  • [31] Identifiability in inverse reinforcement learning
    Cao, Haoyang
    Cohen, Samuel N.
    Szpruch, Lukasz
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [32] A survey of inverse reinforcement learning
    Adams, Stephen
    Cody, Tyler
    Beling, Peter A.
    ARTIFICIAL INTELLIGENCE REVIEW, 2022, 55 (06) : 4307 - 4346
  • [33] Survey on Inverse Reinforcement Learning
    Zhang L.-H.
    Liu Q.
    Huang Z.-G.
    Zhu F.
    Ruan Jian Xue Bao/Journal of Software, 2023, 34 (10): : 4772 - 4803
  • [34] A survey of inverse reinforcement learning
    Stephen Adams
    Tyler Cody
    Peter A. Beling
    Artificial Intelligence Review, 2022, 55 : 4307 - 4346
  • [35] Enhancing Reinforcement Learning for Automated Driving through Virtual Lane Logic
    Fasiello, A.
    Cerrito, F.
    Razza, V.
    Canale, M.
    IFAC PAPERSONLINE, 2024, 58 (28): : 55 - 60
  • [36] Neural Logic Reinforcement Learning
    Jiang, Zhengyao
    Luo, Shan
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [37] Robot Reinforcement Learning on the Constraint Manifold
    Liu, Puze
    Tateo, Davide
    Bou-Ammar, Haitham
    Peters, Jan
    CONFERENCE ON ROBOT LEARNING, VOL 164, 2021, 164 : 1357 - 1366
  • [38] NOTES ON A CONSTRAINT-REINFORCEMENT APPROACH TO OBSERVATIONAL INFERENCE OF SATISFACTION
    SHELLY, MW
    ADELBERG, TZ
    PSYCHOLOGICAL REPORTS, 1967, 21 (02) : 459 - &
  • [39] Online Reinforcement Learning by Bayesian Inference
    Xia, Zhongpu
    Zhao, Dongbin
    2015 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2015,
  • [40] Gene Networks Inference by Reinforcement Learning
    Bonini, Rodrigo Cesar
    Martins-, David Correa, Jr.
    ADVANCES IN BIOINFORMATICS AND COMPUTATIONAL BIOLOGY, BSB 2023, 2023, 13954 : 136 - 147