Using ontologies to enhance human understandability of global post-hoc explanations of black-box models

被引:63
|
作者
Confalonieri, Roberto [1 ]
Weyde, Tillman [2 ]
Besold, Tarek R. [3 ]
Martin, Fermin Moscoso del Prado [4 ]
机构
[1] Free Univ Bozen Bolzano, Fac Comp Sci, I-39100 Bozen Bolzano, Italy
[2] City Univ London, Dept Comp Sci, London EC1V 0HB, England
[3] Neurocat GmbH, Rudower Chaussee 29, D-12489 Berlin, Germany
[4] Lingvist Technol OU, Tallinn, Estonia
关键词
Human-understandable explainable AI; Global explanations; Ontologies; Neural-symbolic learning and reasoning; Knowledge extraction; Concept refinement; INFORMATION-CONTENT; KNOWLEDGE; CLASSIFICATION; WEB;
D O I
10.1016/j.artint.2021.103471
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The interest in explainable artificial intelligence has grown strongly in recent years because of the need to convey safety and trust in the 'how' and 'why' of automated decision-making to users. While a plethora of approaches has been developed, only a few focus on how to use domain knowledge and how this influences the understanding of explanations by users. In this paper, we show that by using ontologies we can improve the human understandability of global post-hoc explanations, presented in the form of decision trees. In particular, we introduce TREPAN Reloaded, which builds on TREPAN, an algorithm that extracts surrogate decision trees from black-box models. TREPAN Reloaded includes ontologies, that model domain knowledge, in the process of extracting explanations to improve their understandability. We tested the understandability of the extracted explanations by humans in a user study with four different tasks. We evaluate the results in terms of response times and correctness, subjective ease of understanding and confidence, and similarity of free text responses. The results show that decision trees generated with TREPAN Reloaded, taking into account domain knowledge, are significantly more understandable throughout than those generated by standard TREPAN. The enhanced understandability of post-hoc explanations is achieved with little compromise on the accuracy with which the surrogate decision trees replicate the behaviour of the original neural network models. (C) 2021 The Author(s). Published by Elsevier B.V.
引用
收藏
页数:20
相关论文
共 50 条
  • [1] A Study on Trust in Black Box Models and Post-hoc Explanations
    El Bekri, Nadia
    Kling, Jasmin
    Huber, Marco F.
    14TH INTERNATIONAL CONFERENCE ON SOFT COMPUTING MODELS IN INDUSTRIAL AND ENVIRONMENTAL APPLICATIONS (SOCO 2019), 2020, 950 : 35 - 46
  • [2] Post-hoc explanation of black-box classifiers using confident itemsets
    Moradi, Milad
    Samwald, Matthias
    EXPERT SYSTEMS WITH APPLICATIONS, 2021, 165
  • [3] Explaining black-box classifiers using post-hoc explanations-by-example: The effect of explanations and error-rates in XAI user studies
    Kenny, Eoin M.
    Ford, Courtney
    Quinn, Molly
    Keane, Mark T.
    ARTIFICIAL INTELLIGENCE, 2021, 294
  • [4] Post-hoc Rule Based Explanations for Black Box Bayesian Optimization
    Chakraborty, Tanmay
    Wirth, Christian
    Seifert, Christin
    ARTIFICIAL INTELLIGENCE-ECAI 2023 INTERNATIONAL WORKSHOPS, PT 1, XAI3, TACTIFUL, XI-ML, SEDAMI, RAAIT, AI4S, HYDRA, AI4AI, 2023, 2024, 1947 : 320 - 337
  • [5] C-RISE: A Post-Hoc Interpretation Method of Black-Box Models for SAR ATR
    Zhu, Mingzhe
    Cheng, Jie
    Lei, Tao
    Feng, Zhenpeng
    Zhou, Xianda
    Liu, Yuanjing
    Chen, Zhihan
    REMOTE SENSING, 2023, 15 (12)
  • [6] Robust and semantic-faithful post-hoc watermarking of text generated by black-box language models
    Hao, Jifei
    Qiang, Jipeng
    Zhu, Yi
    Li, Yun
    Yuan, Yunhao
    Hu, Xiaocheng
    Ouyang, Xiaoye
    FRONTIERS OF COMPUTER SCIENCE, 2025, 19 (09)
  • [7] Learning Groupwise Explanations for Black-Box Models
    Gao, Jingyue
    Wang, Xiting
    Wang, Yasha
    Yan, Yulan
    Xie, Xing
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2396 - 2402
  • [8] Feature Importance Explanations for Temporal Black-Box Models
    Sood, Akshay
    Craven, Mark
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 8351 - 8360
  • [9] An Empirical Comparison of Interpretable Models to Post-Hoc Explanations
    Mahya, Parisa
    Fuernkranz, Johannes
    AI, 2023, 4 (02) : 426 - 436
  • [10] Considerations when learning additive explanations for black-box models
    Tan, Sarah
    Hooker, Giles
    Koch, Paul
    Gordo, Albert
    Caruana, Rich
    MACHINE LEARNING, 2023, 112 (09) : 3333 - 3359