Towards Explainable Reinforcement Learning Using Scoring Mechanism Augmented Agents

被引:3
|
作者
Liu, Yang [1 ]
Wang, Xinzhi [1 ]
Chang, Yudong [1 ]
Jiang, Chao [1 ]
机构
[1] Shanghai Univ, Sch Comp Engn & Sci, Shanghai, Peoples R China
关键词
Deep reinforcement learning; Explainable AI; Adaptive region scoring mechanism;
D O I
10.1007/978-3-031-10986-7_44
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep reinforcement learning (DRL) is increasingly used in application areas such as medicine and finance. However, the direct mapping from state to action in DRL makes it challenging to explain why decisions are made. Existing algorithms for explaining DRL policy are posteriori, explaining to an agent after it has been trained. As a common limitation, these posteriori methods fail to improve training with the deduced knowledge. Face with that, an end-to-end trainable explanation method is proposed, in which an Adaptive Region Scoring Mechanism (ARS) is embedded into DRL system. The ARS explains the agent's action by evaluating the features of the input state that are most relevant action before DRL re-learn from task-related regions. The proposed method is validated on Atari games. Experiments demonstrate that agent using the explainable proposed mechanism outperforms the original models.
引用
收藏
页码:547 / 558
页数:12
相关论文
共 50 条
  • [21] Using Part-Based Representations for Explainable Deep Reinforcement Learning
    Kirtas, Manos
    Tsampazis, Konstantinos
    Avramelou, Loukia
    Passalis, Nikolaos
    Tefas, Anastasios
    MACHINE LEARNING AND PRINCIPLES AND PRACTICE OF KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2023, PT II, 2025, 2134 : 420 - 432
  • [22] Towards Explainable Augmented Intelligence (AI) for Crack Characterization
    Fradkin, Larissa
    Uskuplu Altinbasak, Sevda
    Darmon, Michel
    APPLIED SCIENCES-BASEL, 2021, 11 (22):
  • [23] Retrieval Augmented Reinforcement Learning
    Goyal, Anirudh
    Friesen, Abram L.
    Weber, Theophane
    Banino, Andrea
    Ke, Nan Rosemary
    Badia, Adria Puigdomenech
    Guez, Arthur
    Mirza, Mehdi
    Humphreys, Peter C.
    Konyushkova, Ksenia
    Sifre, Laurent
    Valko, Michal
    Osindero, Simon
    Lillicrap, Timothy
    Heess, Nicolas
    Blundell, Charles
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [24] Reinforcement Learning with Augmented Data
    Laskin, Michael
    Lee, Kimin
    Stooke, Adam
    Pinto, Lerrel
    Abbeel, Pieter
    Srinivas, Aravind
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [25] Towards Dependable and Explainable Machine Learning Using Automated Reasoning
    Bride, Hadrien
    Dong, Jie
    Dong, Jin Song
    Hou, Zhe
    FORMAL METHODS AND SOFTWARE ENGINEERING, ICFEM 2018, 2018, 11232 : 412 - 416
  • [26] Explainable Reinforcement Learning through a Causal Lens
    Madumal, Prashan
    Miller, Tim
    Sonenberg, Liz
    Vetere, Frank
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 2493 - 2500
  • [27] Explainable reinforcement learning for powertrain control engineering
    Laflamme, C.
    Doppler, J.
    Palvolgyi, B.
    Dominka, S.
    Viharos, Zs. J.
    Haeussler, S.
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2025, 146
  • [28] Explainable Reinforcement Learning: A Survey and Comparative Review
    Milani, Stephanie
    Topin, Nicholay
    Veloso, Manuela
    Fang, Fei
    ACM COMPUTING SURVEYS, 2024, 56 (07) : 1 - 36
  • [29] Explainable reinforcement learning for distribution network reconfiguration
    Gholizadeh, Nastaran
    Musilek, Petr
    ENERGY REPORTS, 2024, 11 : 5703 - 5715
  • [30] Explainable Reinforcement Learning via Model Transforms
    Finkelstein, Mira
    Liu, Lucy
    Schlot, Nitsan Levy
    Kolumbus, Yoav
    Parkes, David C.
    Rosenschein, Jeffrey S.
    Keren, Sarah
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,