Explainable Reinforcement Learning through a Causal Lens

被引:0
|
作者
Madumal, Prashan
Miller, Tim
Sonenberg, Liz
Vetere, Frank
机构
基金
澳大利亚研究理事会;
关键词
EXPLANATIONS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Prominent theories in cognitive science propose that humans understand and represent the knowledge of the world through causal relationships. In making sense of the world, we build causal models in our mind to encode cause-effect relations of events and use these to explain why new events happen by referring to counterfactuals - things that did not happen. In this paper, we use causal models to derive causal explanations of the behaviour of model-free reinforcement learning agents. We present an approach that learns a structural causal model during reinforcement learning and encodes causal relationships between variables of interest. This model is then used to generate explanations of behaviour based on counterfactual analysis of the causal model. We computationally evaluate the model in 6 domains and measure performance and task prediction accuracy. We report on a study with 120 participants who observe agents playing a real-time strategy game (Starcraft II) and then receive explanations of the agents' behaviour. We investigate: 1) participants' understanding gained by explanations through task prediction; 2) explanation satisfaction and 3) trust. Our results show that causal model explanations perform better on these measures compared to two other baseline explanation models.
引用
收藏
页码:2493 / 2500
页数:8
相关论文
共 50 条
  • [31] Efficient Reinforcement Learning with Prior Causal Knowledge
    Lu, Yangyi
    Meisami, Amirhossein
    Tewari, Ambuj
    CONFERENCE ON CAUSAL LEARNING AND REASONING, VOL 177, 2022, 177
  • [32] A causal role of estradiol in human reinforcement learning
    Veselic, Sebastijan
    Jocham, Gerhard
    Gausterer, Christian
    Wagner, Bernhard
    Ernhoefer-Ressler, Miriam
    Lanzenberger, Rupert
    Eisenegger, Christoph
    Lamm, Claus
    Vermeer, Annabel Losecaat
    HORMONES AND BEHAVIOR, 2021, 134
  • [33] CARL: A Synergistic Framework for Causal Reinforcement Learning
    Mendez-Molina, Arquimides
    Morales, Eduardo F.
    Sucar, L. Enrique
    IEEE ACCESS, 2023, 11 : 126462 - 126481
  • [34] Causal Discovery and Reinforcement Learning: A Synergistic Integration
    Mendez-Molina, Arquimides
    Morales, Eduardo F.
    Enrique Sucar, L.
    INTERNATIONAL CONFERENCE ON PROBABILISTIC GRAPHICAL MODELS, VOL 186, 2022, 186
  • [35] VERIFICATION, VALIDATION, AND CALIBRATION THROUGH A CAUSAL LENS
    Gonzales, Ronald
    Mandelli, Diego
    Wang, Congjian
    Abdo, Mohammad
    Balestra, Paolo
    Qin, Sunming
    Welker, Zachary
    Petrov, Victor
    Manera, Annalisa
    PROCEEDINGS OF 2024 VERIFICATION, VALIDATION, AND UNCERTAINTY QUANTIFICATION SYMPOSIUM, VVUQ2024, 2024,
  • [36] A User Study on Explainable Online Reinforcement Learning for Adaptive Systems
    Metzger, Andreas
    Laufer, Jan
    Feit, Felix
    Pohl, Klaus
    ACM Transactions on Autonomous and Adaptive Systems, 2024, 19 (03)
  • [37] Container Caching Optimization based on Explainable Deep Reinforcement Learning
    Jayaram, Divyashree
    Jeelani, Saad
    Ishigaki, Genya
    IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 7127 - 7132
  • [38] Sample-Based Rule Extraction for Explainable Reinforcement Learning
    Engelhardt, Raphael C.
    Lange, Moritz
    Wiskott, Laurenz
    Konen, Wolfgang
    MACHINE LEARNING, OPTIMIZATION, AND DATA SCIENCE, LOD 2022, PT I, 2023, 13810 : 330 - 345
  • [39] XPM: An Explainable Deep Reinforcement Learning Framework for Portfolio Management
    Shi, Si
    Li, Jianjun
    Li, Guohui
    Pan, Peng
    Liu, Ke
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 1661 - 1670
  • [40] Explainable Action Advising for Multi-Agent Reinforcement Learning
    Guo, Yue
    Campbell, Joseph
    Stepputtis, Simon
    Li, Ruiyu
    Hughes, Dana
    Fang, Fei
    Sycara, Katia
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 5515 - 5521