Towards Interpretable Deep Reinforcement Learning Models via Inverse Reinforcement Learning

被引:0
|
作者
Xie, Yuansheng [1 ]
Vosoughi, Soroush [1 ]
Hassanpour, Saeed [2 ]
机构
[1] Dartmouth Coll, Dept Comp Sci, Hanover, NH 03755 USA
[2] Dartmouth Coll, Dept Biomed Data Sci, Hanover, NH 03755 USA
关键词
Adversarial Inverse Reinforcment Learning; Natural Language Processing; Abstractive Summarization; BLACK-BOX;
D O I
10.1109/ICPR56361.2022.9956245
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Artificial Intelligence, particularly through recent advancements in deep learning (DL), has achieved exceptional performances in many tasks in fields such as natural language processing and computer vision. For certain high-stake domains, in addition to desirable performance metrics, a high level of interpretability is often required in order for AI to be reliably utilized. Unfortunately, the black box nature of DL models prevents researchers from providing explicative descriptions for a DL model's reasoning process and decisions. In this work, we propose a novel framework utilizing Adversarial Inverse Reinforcement Learning that can provide global explanations for decisions made by a Reinforcement Learning model and capture intuitive tendencies that the model follows by summarizing the model's decision-making process.
引用
收藏
页码:5067 / 5074
页数:8
相关论文
共 50 条
  • [31] Towards Knowledge Transfer in Deep Reinforcement Learning
    Glatt, Ruben
    da Silva, Felipe Leno
    Reali Costa, Anna Helena
    [J]. PROCEEDINGS OF 2016 5TH BRAZILIAN CONFERENCE ON INTELLIGENT SYSTEMS (BRACIS 2016), 2016, : 91 - 96
  • [32] From Reinforcement Learning to Deep Reinforcement Learning: An Overview
    Agostinelli, Forest
    Hocquet, Guillaume
    Singh, Sameer
    Baldi, Pierre
    [J]. BRAVERMAN READINGS IN MACHINE LEARNING: KEY IDEAS FROM INCEPTION TO CURRENT STATE, 2018, 11100 : 298 - 328
  • [33] Towards Automated Imbalanced Learning with Deep Hierarchical Reinforcement Learning
    Zha, Daochen
    Lai, Kwei-Herng
    Tan, Qiaoyu
    Ding, Sirui
    Zou, Na
    Hu, Xia Ben
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 2476 - 2485
  • [34] Optimal recovery of unsecured debt via interpretable reinforcement learning
    Mark, Michael
    Chehrazi, Naveed
    Liu, Huanxi
    Weber, Thomas A.
    [J]. MACHINE LEARNING WITH APPLICATIONS, 2022, 8
  • [35] Learning How Pedestrians Navigate: A Deep Inverse Reinforcement Learning Approach
    Fahad, Muhammad
    Chen, Zhuo
    Guo, Yi
    [J]. 2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 819 - 826
  • [36] Interpretable, Verifiable, and Robust Reinforcement Learning via Program Synthesis
    Bastani, Osbert
    Inala, Jeevana Priya
    Solar-Lezama, Armando
    [J]. XXAI - BEYOND EXPLAINABLE AI: International Workshop, Held in Conjunction with ICML 2020, July 18, 2020, Vienna, Austria, Revised and Extended Papers, 2022, 13200 : 207 - 228
  • [37] Deep sparse representation via deep dictionary learning for reinforcement learning
    Tang, Jianhao
    Li, Zhenni
    Xie, Shengli
    Ding, Shuxue
    Zheng, Shaolong
    Chen, Xueni
    [J]. 2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 2398 - 2403
  • [38] Generative Inverse Deep Reinforcement Learning for Online Recommendation
    Chen, Xiaocong
    Yao, Lina
    Sun, Aixin
    Wang, Xianzhi
    Xu, Xiwei
    Zhu, Liming
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 201 - 210
  • [39] Comfortable Driving by using Deep Inverse Reinforcement Learning
    Kishikawa, Daiko
    Arai, Sachiyo
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON AGENTS (ICA), 2019, : 38 - 43
  • [40] Risk-sensitive Inverse Reinforcement Learning via Coherent Risk Models
    Majumdar, Anirudha
    Singh, Sumeet
    Mandlekar, Ajay
    Pavone, Marco
    [J]. ROBOTICS: SCIENCE AND SYSTEMS XIII, 2017,