Evaluating Quality of Visual Explanations of Deep Learning Models for Vision Tasks

被引:0
|
作者
Yang, Yuqing [1 ,2 ]
Mahmoudpour, Saeed [1 ,2 ]
Schelkens, Peter [1 ,2 ]
Deligiannis, Nikos [1 ,2 ]
机构
[1] Vrije Univ Brussel, Dept Elect & Informat, Pl Laan 2, B-1050 Brussels, Belgium
[2] imec, Kapeldreef 75, B-3001 Leuven, Belgium
关键词
Explainable artificial intelligence; Vision Transformer; heatmaps; subjective evaluation;
D O I
10.1109/QOMEX58391.2023.10178510
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Explainable artificial intelligence (XAI) has gained considerable attention in recent years as it aims to help humans better understand machine learning decisions, making complex black-box systems more trustworthy. Visual explanation algorithms have been designed to generate heatmaps highlighting image regions that a deep neural network focuses on to make decisions. While convolutional neural network (CNN) models typically follow similar processing operations for feature encoding, the emergence of vision transformer (ViT) has introduced a new approach to machine vision decision-making. Therefore, an important question is which architecture provides more human-understandable explanations. This paper examines the explainability of deep architectures, including CNN and ViT models under different vision tasks. To this end, we first performed a subjective experiment asking humans to highlight the key visual features in images that helped them to make decisions in two different vision tasks. Next, using the human-annotated images, ground-truth heatmaps were generated that were compared against heatmaps generated by explanation methods for the deep architectures. Moreover, perturbation tests were performed for objective evaluation of the deep models' explanation heatmaps. According to the results, the explanations generated from ViT are deemed more trustworthy than those produced by other CNNs, and as the features of the input image are more dispersed, the advantage of the model becomes more evident.
引用
收藏
页码:159 / 164
页数:6
相关论文
共 50 条
  • [31] Deep Active Learning for Computer Vision Tasks: Methodologies, Applications, and Challenges
    Wu, Mingfei
    Li, Chen
    Yao, Zehuan
    APPLIED SCIENCES-BASEL, 2022, 12 (16):
  • [32] Exploration of Privacy Preserving Deep Learning Framework for Computer Vision Tasks
    Wilson, Amala
    Solh, Mashhour
    Moh, Melody
    ACMSE 2022: PROCEEDINGS OF THE 2022 ACM SOUTHEAST CONFERENCE, 2022, : 130 - 137
  • [33] Evaluating Visual Search in Glaucoma Using Deep learning
    Mishra, Anoop
    Belcher, Steven
    Anderson, David
    Khazanchi, Deepak
    AMCIS 2020 PROCEEDINGS, 2020,
  • [34] Evaluating the progress of deep learning for visual relational concepts
    Stabinger, Sebastian
    Peer, David
    Piater, Justus
    Rodriguez-Sanchez, Antonio
    JOURNAL OF VISION, 2021, 21 (11): : 11 - 23
  • [35] Evaluating Uncertainty-Based Deep Learning Explanations for Prostate Lesion Detection
    Trombley, Christopher M.
    Gulum, Mehmet Akif
    Ozen, Merve
    Esen, Enes
    Aksamoglu, Melih
    Kantardzic, Mehmed
    MACHINE LEARNING FOR HEALTHCARE CONFERENCE, VOL 182, 2022, 182 : 874 - 891
  • [36] A deep learning-based method for evaluating the quality of corporate brand packaging visual design
    Ma Y.
    Applied Mathematics and Nonlinear Sciences, 2024, 9 (01)
  • [37] Embedding deep networks into visual explanations
    Qi, Zhongang
    Khorram, Saeed
    Fuxin, Li
    ARTIFICIAL INTELLIGENCE, 2021, 292
  • [38] Visual interpretability of bioimaging deep learning models
    Rotem, Oded
    Zaritsky, Assaf
    NATURE METHODS, 2024, 21 (08) : 1394 - 1397
  • [39] An inherently interpretable deep learning model for local explanations using visual concepts
    Ullah, Mirza Ahsan
    Zia, Tehseen
    Kim, Jungeun
    Kadry, Seifedine
    PLOS ONE, 2024, 19 (10):
  • [40] Deep learning with visual explanations for leakage defect segmentation of metro shield tunnel
    Feng, Shi Jin
    Feng, Yong
    Zhang, Xiao Lei
    Chen, Yi Han
    TUNNELLING AND UNDERGROUND SPACE TECHNOLOGY, 2023, 136