Verifiable and Interpretable Reinforcement Learning through Program Synthesis

被引:0
|
作者
Verma, Abhinav [1 ]
机构
[1] Rice Univ, 6100 Main St, Houston, TX 77005 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study the problem of generating interpretable and verifiable policies for Reinforcement Learning (RL). Unlike the popular Deep Reinforcement Learning (DRL) paradigm, in which the policy is represented by a neural network, the aim of this work is to find policies that can be represented in high-level programming languages. Such programmatic policies have several benefits, including being more easily interpreted than neural networks, and being amenable to verification by scalable symbolic methods. The generation methods for programmatic policies also provide a mechanism for systematically using domain knowledge for guiding the policy search. The interpretability and verifiability of these policies provides the opportunity to deploy RL based solutions in safety critical environments. This thesis draws on, and extends, work from both the machine learning and formal methods communities.
引用
收藏
页码:9902 / 9903
页数:2
相关论文
共 50 条
  • [21] MoET: Mixture of Expert Trees and its application to verifiable reinforcement learning
    Vasic, Marko
    Petrovic, Andrija
    Wang, Kaiyuan
    Nikolic, Mladen
    Singh, Rishabh
    Khurshid, Sarfraz
    [J]. NEURAL NETWORKS, 2022, 151 : 34 - 47
  • [22] Continuous Action Reinforcement Learning From a Mixture of Interpretable Experts
    Akrour, Riad
    Tateo, Davide
    Peters, Jan
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (10) : 6795 - 6806
  • [23] Methodology for Interpretable Reinforcement Learning Model for HVAC Energy Control
    Kotevska, Olivera
    Munk, Jeffrey
    Kurte, Kuldeep
    Du, Yan
    Amasyali, Kadir
    Smith, Robert W.
    Zandi, Helia
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2020, : 1555 - 1564
  • [24] Can Interpretable Reinforcement Learning Manage Prosperity Your Way?
    Maree, Charl
    Omlin, Christian W.
    [J]. AI, 2022, 3 (02) : 526 - 537
  • [25] Generating Interpretable Reinforcement Learning Policies using Genetic Programming
    Hein, Daniel
    Udluft, Steffen
    Runkler, Thomas A.
    [J]. PROCEEDINGS OF THE 2019 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION (GECCCO'19 COMPANION), 2019, : 23 - 24
  • [26] Optimal recovery of unsecured debt via interpretable reinforcement learning
    Mark, Michael
    Chehrazi, Naveed
    Liu, Huanxi
    Weber, Thomas A.
    [J]. MACHINE LEARNING WITH APPLICATIONS, 2022, 8
  • [27] Optimization Methods for Interpretable Differentiable Decision Trees in Reinforcement Learning
    Silva, Andrew
    Killian, Taylor
    Jimenez, Ivan Rodriguez
    Son, Sung-Hyun
    Gombolay, Matthew
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108
  • [28] Towards Interpretable Reinforcement Learning Using Attention Augmented Agents
    Mott, Alex
    Zoran, Daniel
    Chrzanowski, Mike
    Wierstra, Daan
    Rezende, Danilo J.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [29] Self-Supervised Discovering of Interpretable Features for Reinforcement Learning
    Shi, Wenjie
    Huang, Gao
    Song, Shiji
    Wang, Zhuoyuan
    Lin, Tingyu
    Wu, Cheng
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (05) : 2712 - 2724
  • [30] A formal methods approach to interpretable reinforcement learning for robotic planning
    Li, Xiao
    Serlin, Zachary
    Yang, Guang
    Belta, Calin
    [J]. SCIENCE ROBOTICS, 2019, 4 (37)