Verifiable and Interpretable Reinforcement Learning through Program Synthesis

被引:0
|
作者
Verma, Abhinav [1 ]
机构
[1] Rice Univ, 6100 Main St, Houston, TX 77005 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study the problem of generating interpretable and verifiable policies for Reinforcement Learning (RL). Unlike the popular Deep Reinforcement Learning (DRL) paradigm, in which the policy is represented by a neural network, the aim of this work is to find policies that can be represented in high-level programming languages. Such programmatic policies have several benefits, including being more easily interpreted than neural networks, and being amenable to verification by scalable symbolic methods. The generation methods for programmatic policies also provide a mechanism for systematically using domain knowledge for guiding the policy search. The interpretability and verifiability of these policies provides the opportunity to deploy RL based solutions in safety critical environments. This thesis draws on, and extends, work from both the machine learning and formal methods communities.
引用
收藏
页码:9902 / 9903
页数:2
相关论文
共 50 条
  • [1] Interpretable, Verifiable, and Robust Reinforcement Learning via Program Synthesis
    Bastani, Osbert
    Inala, Jeevana Priya
    Solar-Lezama, Armando
    [J]. XXAI - BEYOND EXPLAINABLE AI: International Workshop, Held in Conjunction with ICML 2020, July 18, 2020, Vienna, Austria, Revised and Extended Papers, 2022, 13200 : 207 - 228
  • [2] An Inductive Synthesis Framework for Verifiable Reinforcement Learning
    Zhu, He
    Xiong, Zikang
    Magill, Stephen
    Jagannathan, Suresh
    [J]. PROCEEDINGS OF THE 40TH ACM SIGPLAN CONFERENCE ON PROGRAMMING LANGUAGE DESIGN AND IMPLEMENTATION (PLDI '19), 2019, : 686 - 701
  • [3] A survey on interpretable reinforcement learning
    Glanois, Claire
    Weng, Paul
    Zimmer, Matthieu
    Li, Dong
    Yang, Tianpei
    Hao, Jianye
    Liu, Wulong
    [J]. MACHINE LEARNING, 2024, 113 (08) : 5847 - 5890
  • [4] Interpretable Control by Reinforcement Learning
    Hein, Daniel
    Limmer, Steffen
    Runkler, Thomas A.
    [J]. IFAC PAPERSONLINE, 2020, 53 (02): : 8082 - 8089
  • [5] Programmatically Interpretable Reinforcement Learning
    Verma, Abhinav
    Murali, Vijayaraghavan
    Singh, Rishabh
    Kohli, Pushmeet
    Chaudhuri, Swarat
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [6] Interpretable policy derivation for reinforcement learning based on evolutionary feature synthesis
    Zhang, Hengzhe
    Zhou, Aimin
    Lin, Xin
    [J]. COMPLEX & INTELLIGENT SYSTEMS, 2020, 6 (03) : 741 - 753
  • [7] Interpretable policy derivation for reinforcement learning based on evolutionary feature synthesis
    Hengzhe Zhang
    Aimin Zhou
    Xin Lin
    [J]. Complex & Intelligent Systems, 2020, 6 : 741 - 753
  • [8] Verifiable Reinforcement Learning via Policy Extraction
    Bastani, Osbert
    Pu, Yewen
    Solar-Lezama, Armando
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [9] Towards Interpretable Deep Reinforcement Learning Models via Inverse Reinforcement Learning
    Xie, Yuansheng
    Vosoughi, Soroush
    Hassanpour, Saeed
    [J]. 2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 5067 - 5074
  • [10] Evolving interpretable decision trees for reinforcement learning
    Costa, Vinicius G.
    Perez-Aracil, Jorge
    Salcedo-Sanz, Sancho
    Pedreira, Carlos E.
    [J]. ARTIFICIAL INTELLIGENCE, 2024, 327