Interpretability of deep neural networks: A review of methods, classification and hardware

被引:0
|
作者
Antamis, Thanasis [1 ]
Drosou, Anastasis [1 ]
Vafeiadis, Thanasis [1 ]
Nizamis, Alexandros [1 ]
Ioannidis, Dimosthenis [1 ]
Tzovaras, Dimitrios [1 ]
机构
[1] Ctr Res & Technol Hellas, Informat Technol Inst, Thessaloniki 57001, Greece
关键词
XAI; Deep neural networks; xDNN; Survey; BLACK-BOX; ATTENTION; RULES;
D O I
10.1016/j.neucom.2024.128204
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Artificial intelligence, and especially deep neural networks, have evolved substantially in the recent years, infiltrating numerous domains of applications, often greatly impactful to society's well-being. As a result, the need to understand how these models operate in depth and to access explanations of their decisions has become more vital than ever. Tending to this demand, the following paper aims to provide a thorough overview of the methods that have so far been developed to explain deep neural networks. Key aspects of explainability are defined and a straightforward classification of existing approaches is introduced, along with numerous examples. The task of realizing these methods on hardware is also discussed to complete the understanding of their application.
引用
收藏
页数:24
相关论文
共 50 条
  • [1] A Benchmark for Interpretability Methods in Deep Neural Networks
    Hooker, Sara
    Erhan, Dumitru
    Kindermans, Pieter-Jan
    Kim, Been
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [2] Transparency of deep neural networks for medical image analysis: A review of interpretability methods
    Salahuddin, Zohaib
    Woodruff, Henry C.
    Chatterjee, Avishek
    Lambin, Philippe
    [J]. COMPUTERS IN BIOLOGY AND MEDICINE, 2022, 140
  • [3] A Systematic Literature Review on Hardware Reliability Assessment Methods for Deep Neural Networks
    Ahmadilivani, Mohammad Hasan
    Taheri, Mahdi
    Raik, Jaan
    Daneshtalab, Masoud
    Jenihhin, Maksim
    [J]. ACM COMPUTING SURVEYS, 2024, 56 (06)
  • [4] Comparison of interpretability methods in the context of deep neural networks for radiomics application
    Marchadour, Wistan
    Badic, Bogdan
    Maison, Jonas
    Hatt, Mathieu
    Vermet, Franck
    [J]. JOURNAL OF NUCLEAR MEDICINE, 2022, 63
  • [5] Improving the Interpretability of GradCAMs in Deep Classification Networks
    Schoettl, Alfred
    [J]. 3RD INTERNATIONAL CONFERENCE ON INDUSTRY 4.0 AND SMART MANUFACTURING, 2022, 200 : 620 - 628
  • [6] New Perspective of Interpretability of Deep Neural Networks
    Kimura, Masanari
    Tanaka, Masayuki
    [J]. 2020 3RD INTERNATIONAL CONFERENCE ON INFORMATION AND COMPUTER TECHNOLOGIES (ICICT 2020), 2020, : 78 - 85
  • [7] A Demonstration of Interpretability Methods for Graph Neural Networks
    Mobaraki, Ehsan B.
    Khan, Arijit
    [J]. PROCEEDINGS OF THE 6TH ACM SIGMOD JOINT INTERNATIONAL WORKSHOP ON GRAPH DATA MANAGEMENT EXPERIENCES & SYSTEMS AND NETWORK DATA ANALYTICS, GRADES-NDA 2023, 2023,
  • [8] Interpretability Analysis of Deep Neural Networks With Adversarial Examples
    Dong, Yin-Peng
    Su, Hang
    Zhu, Jun
    [J]. Zidonghua Xuebao/Acta Automatica Sinica, 2022, 48 (01): : 75 - 86
  • [9] Optimizing for interpretability in deep neural networks with tree regularization
    Wu, Mike
    Parbhoo, Sonali
    Hughes, Michael C.
    Roth, Volker
    Doshi-Velez, Finale
    [J]. Journal of Artificial Intelligence Research, 2021, 72
  • [10] IMPROVING THE INTERPRETABILITY OF DEEP NEURAL NETWORKS WITH STIMULATED LEARNING
    Tan, Shawn
    Sim, Khe Chai
    Gales, Mark
    [J]. 2015 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING (ASRU), 2015, : 617 - 623