Interpretable, Verifiable, and Robust Reinforcement Learning via Program Synthesis

被引:5
|
作者
Bastani, Osbert [1 ]
Inala, Jeevana Priya [2 ]
Solar-Lezama, Armando [3 ]
机构
[1] Univ Penn, Philadelphia, PA 19104 USA
[2] Microsoft Res, Redmond, WA 98052 USA
[3] MIT, Cambridge, MA 02139 USA
关键词
Interpretable reinforcement learning; Program synthesis;
D O I
10.1007/978-3-031-04083-2_11
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning is a promising strategy for automatically training policies for challenging control tasks. However, state-of-the-art deep reinforcement learning algorithms focus on training deep neural network (DNN) policies, which are black box models that are hard to interpret and reason about. In this chapter, we describe recent progress towards learning policies in the form of programs. Compared to DNNs, such programmatic policies are significantly more interpretable, easier to formally verify, and more robust. We give an overview of algorithms designed to learn programmatic policies, and describe several case studies demonstrating their various advantages.
引用
收藏
页码:207 / 228
页数:22
相关论文
共 50 条
  • [1] Verifiable and Interpretable Reinforcement Learning through Program Synthesis
    Verma, Abhinav
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 9902 - 9903
  • [2] Verifiable Reinforcement Learning via Policy Extraction
    Bastani, Osbert
    Pu, Yewen
    Solar-Lezama, Armando
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [3] An Inductive Synthesis Framework for Verifiable Reinforcement Learning
    Zhu, He
    Xiong, Zikang
    Magill, Stephen
    Jagannathan, Suresh
    [J]. PROCEEDINGS OF THE 40TH ACM SIGPLAN CONFERENCE ON PROGRAMMING LANGUAGE DESIGN AND IMPLEMENTATION (PLDI '19), 2019, : 686 - 701
  • [4] Towards Interpretable Deep Reinforcement Learning Models via Inverse Reinforcement Learning
    Xie, Yuansheng
    Vosoughi, Soroush
    Hassanpour, Saeed
    [J]. 2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 5067 - 5074
  • [5] Interpretable Enterprise Credit Rating via Reinforcement Learning
    Wang, JingQiu
    Guo, WeiYu
    [J]. ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT I, ICIC 2024, 2024, 14875 : 292 - 303
  • [6] Learning Interpretable Features via Adversarially Robust Optimization
    Khakzar, Ashkan
    Albarqouni, Shadi
    Navab, Nassir
    [J]. MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2019, PT VI, 2019, 11769 : 793 - 800
  • [7] Optimal recovery of unsecured debt via interpretable reinforcement learning
    Mark, Michael
    Chehrazi, Naveed
    Liu, Huanxi
    Weber, Thomas A.
    [J]. MACHINE LEARNING WITH APPLICATIONS, 2022, 8
  • [8] A survey on interpretable reinforcement learning
    Glanois, Claire
    Weng, Paul
    Zimmer, Matthieu
    Li, Dong
    Yang, Tianpei
    Hao, Jianye
    Liu, Wulong
    [J]. MACHINE LEARNING, 2024, 113 (08) : 5847 - 5890
  • [9] Interpretable Control by Reinforcement Learning
    Hein, Daniel
    Limmer, Steffen
    Runkler, Thomas A.
    [J]. IFAC PAPERSONLINE, 2020, 53 (02): : 8082 - 8089
  • [10] Programmatically Interpretable Reinforcement Learning
    Verma, Abhinav
    Murali, Vijayaraghavan
    Singh, Rishabh
    Kohli, Pushmeet
    Chaudhuri, Swarat
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80