Interpretable, Verifiable, and Robust Reinforcement Learning via Program Synthesis

被引:5
|
作者
Bastani, Osbert [1 ]
Inala, Jeevana Priya [2 ]
Solar-Lezama, Armando [3 ]
机构
[1] Univ Penn, Philadelphia, PA 19104 USA
[2] Microsoft Res, Redmond, WA 98052 USA
[3] MIT, Cambridge, MA 02139 USA
关键词
Interpretable reinforcement learning; Program synthesis;
D O I
10.1007/978-3-031-04083-2_11
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning is a promising strategy for automatically training policies for challenging control tasks. However, state-of-the-art deep reinforcement learning algorithms focus on training deep neural network (DNN) policies, which are black box models that are hard to interpret and reason about. In this chapter, we describe recent progress towards learning policies in the form of programs. Compared to DNNs, such programmatic policies are significantly more interpretable, easier to formally verify, and more robust. We give an overview of algorithms designed to learn programmatic policies, and describe several case studies demonstrating their various advantages.
引用
收藏
页码:207 / 228
页数:22
相关论文
共 50 条
  • [11] Robust Reinforcement Learning via Genetic Curriculum
    Song, Yeeho
    Schneider, Jeff
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022), 2022, : 5560 - 5566
  • [12] Understanding via Exploration: Discovery of Interpretable Features With Deep Reinforcement Learning
    Wei, Jiawen
    Qiu, Zhifeng
    Wang, Fangyuan
    Lin, Wenwei
    Gui, Ning
    Gui, Weihua
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (02) : 1696 - 1707
  • [13] Effective Program Debloating via Reinforcement Learning
    Heo, Kihong
    Lee, Woosuk
    Pashakhanloo, Pardis
    Naik, Mayur
    [J]. PROCEEDINGS OF THE 2018 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY (CCS'18), 2018, : 380 - 394
  • [14] Interpretable policy derivation for reinforcement learning based on evolutionary feature synthesis
    Zhang, Hengzhe
    Zhou, Aimin
    Lin, Xin
    [J]. COMPLEX & INTELLIGENT SYSTEMS, 2020, 6 (03) : 741 - 753
  • [15] Interpretable policy derivation for reinforcement learning based on evolutionary feature synthesis
    Hengzhe Zhang
    Aimin Zhou
    Xin Lin
    [J]. Complex & Intelligent Systems, 2020, 6 : 741 - 753
  • [16] Robust and Verifiable Privacy Federated Learning
    Lu Z.
    Lu S.
    Tang X.
    Wu J.
    [J]. IEEE Transactions on Artificial Intelligence, 2024, 5 (04): : 1895 - 1908
  • [17] Verifiable Learning for Robust Tree Ensembles
    Calzavara, Stefano
    Cazzaro, Lorenzo
    Pibiri, Giulio Ermanno
    Prezza, Nicola
    [J]. PROCEEDINGS OF THE 2023 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, CCS 2023, 2023, : 1850 - 1864
  • [18] Robust Reinforcement Learning via Progressive Task Sequence
    Li, Yike
    Tian, Yunzhe
    Tong, Endong
    Niu, Wenjia
    Liu, Jiqiang
    [J]. PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 455 - 463
  • [19] Robust Market Making via Adversarial Reinforcement Learning
    Spooner, Thomas
    Savani, Rahul
    [J]. PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 4590 - 4596
  • [20] PILE: Robust Privacy-Preserving Federated Learning Via Verifiable Perturbations
    Tang, Xiangyun
    Shen, Meng
    Li, Qi
    Zhu, Liehuang
    Xue, Tengfei
    Qu, Qiang
    [J]. IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2023, 20 (06) : 5005 - 5023