Evaluating Quality of Visual Explanations of Deep Learning Models for Vision Tasks

被引:0
|
作者
Yang, Yuqing [1 ,2 ]
Mahmoudpour, Saeed [1 ,2 ]
Schelkens, Peter [1 ,2 ]
Deligiannis, Nikos [1 ,2 ]
机构
[1] Vrije Univ Brussel, Dept Elect & Informat, Pl Laan 2, B-1050 Brussels, Belgium
[2] imec, Kapeldreef 75, B-3001 Leuven, Belgium
关键词
Explainable artificial intelligence; Vision Transformer; heatmaps; subjective evaluation;
D O I
10.1109/QOMEX58391.2023.10178510
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Explainable artificial intelligence (XAI) has gained considerable attention in recent years as it aims to help humans better understand machine learning decisions, making complex black-box systems more trustworthy. Visual explanation algorithms have been designed to generate heatmaps highlighting image regions that a deep neural network focuses on to make decisions. While convolutional neural network (CNN) models typically follow similar processing operations for feature encoding, the emergence of vision transformer (ViT) has introduced a new approach to machine vision decision-making. Therefore, an important question is which architecture provides more human-understandable explanations. This paper examines the explainability of deep architectures, including CNN and ViT models under different vision tasks. To this end, we first performed a subjective experiment asking humans to highlight the key visual features in images that helped them to make decisions in two different vision tasks. Next, using the human-annotated images, ground-truth heatmaps were generated that were compared against heatmaps generated by explanation methods for the deep architectures. Moreover, perturbation tests were performed for objective evaluation of the deep models' explanation heatmaps. According to the results, the explanations generated from ViT are deemed more trustworthy than those produced by other CNNs, and as the features of the input image are more dispersed, the advantage of the model becomes more evident.
引用
收藏
页码:159 / 164
页数:6
相关论文
共 50 条
  • [1] Robustness of Deep Learning Models for Vision Tasks
    Lee, Youngseok
    Kim, Jongweon
    APPLIED SCIENCES-BASEL, 2023, 13 (07):
  • [2] LangXAI: Integrating Large Vision Models for Generating Textual Explanations to Enhance Explainability in Visual Perception Tasks
    Hung Nguyen
    Clement, Tobias
    Loc Nguyen
    Kemmerzell, Nils
    Binh Truong
    Khang Nguyen
    Abdelaal, Mohamed
    Cao, Hung
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 8754 - 8758
  • [3] Evaluating the Effectiveness of Deep Learning Models for Foundational Program Analysis Tasks
    Chen, Qian
    Yu, Chenyang
    Liu, Ruyan
    Zhang, Chi
    Wang, Yu
    Wang, Ke
    Su, Ting
    Wang, Linzhang
    PROCEEDINGS OF THE ACM ON PROGRAMMING LANGUAGES-PACMPL, 2024, 8 (OOPSLA):
  • [4] Evaluating Vision Transformer Models for Visual Quality Control in Industrial Manufacturing
    Alber, Miriam
    Hoenes, Christoph
    Baier, Patrick
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES-APPLIED DATA SCIENCE TRACK, PT X, ECML PKDD 2024, 2024, 14950 : 116 - 132
  • [5] Using Explanations to Estimate the Quality of Computer Vision Models
    Oliveira, Filipe
    Carneiro, Davide
    Pereira, Joao
    HUMAN-CENTRED TECHNOLOGY MANAGEMENT FOR A SUSTAINABLE FUTURE, VOL 2, IAMOT, 2025, : 293 - 301
  • [6] Visual Field Prediction Evaluating the Clinical Relevance of Deep Learning Models
    Eslami, Mohammad
    Kim, Julia A.
    Zhang, Miao
    Boland, Michael, V
    Wang, Mengyu
    Chang, Dolly S.
    Elze, Tobias
    OPHTHALMOLOGY SCIENCE, 2023, 3 (01):
  • [7] ViCE: Visual Counterfactual Explanations for Machine Learning Models
    Gomez, Oscar
    Holter, Steffen
    Yuan, Jun
    Bertini, Enrico
    PROCEEDINGS OF THE 25TH INTERNATIONAL CONFERENCE ON INTELLIGENT USER INTERFACES, IUI 2020, 2020, : 531 - 535
  • [8] Deep Learning for Language and Vision Tasks in Surveillance Applications
    Pastor Lopez-Monroy, A.
    Arturo Elias-Miranda, Alfredo
    Vallejo-Aldana, Daniel
    Manuel Garcia-Carmona, Juan
    Perez-Espinosa, Humberto
    COMPUTACION Y SISTEMAS, 2021, 25 (02): : 317 - 328
  • [9] Assessing the Reliability of Visual Explanations of Deep Models with Adversarial Perturbations
    Valle, Dan
    Pimentel, Tiago
    Veloso, Adrian
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [10] Evaluating the Faithfulness of Causality in Saliency-Based Explanations of Deep Learning Models for Temporal Colour Constancy
    Rizzo, Matteo
    Conati, Cristina
    Jang, Daesik
    Hu, Hui
    EXPLAINABLE ARTIFICIAL INTELLIGENCE, PT III, XAI 2024, 2024, 2155 : 125 - 142