Evaluating Quality of Visual Explanations of Deep Learning Models for Vision Tasks

被引:0
|
作者
Yang, Yuqing [1 ,2 ]
Mahmoudpour, Saeed [1 ,2 ]
Schelkens, Peter [1 ,2 ]
Deligiannis, Nikos [1 ,2 ]
机构
[1] Vrije Univ Brussel, Dept Elect & Informat, Pl Laan 2, B-1050 Brussels, Belgium
[2] imec, Kapeldreef 75, B-3001 Leuven, Belgium
关键词
Explainable artificial intelligence; Vision Transformer; heatmaps; subjective evaluation;
D O I
10.1109/QOMEX58391.2023.10178510
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Explainable artificial intelligence (XAI) has gained considerable attention in recent years as it aims to help humans better understand machine learning decisions, making complex black-box systems more trustworthy. Visual explanation algorithms have been designed to generate heatmaps highlighting image regions that a deep neural network focuses on to make decisions. While convolutional neural network (CNN) models typically follow similar processing operations for feature encoding, the emergence of vision transformer (ViT) has introduced a new approach to machine vision decision-making. Therefore, an important question is which architecture provides more human-understandable explanations. This paper examines the explainability of deep architectures, including CNN and ViT models under different vision tasks. To this end, we first performed a subjective experiment asking humans to highlight the key visual features in images that helped them to make decisions in two different vision tasks. Next, using the human-annotated images, ground-truth heatmaps were generated that were compared against heatmaps generated by explanation methods for the deep architectures. Moreover, perturbation tests were performed for objective evaluation of the deep models' explanation heatmaps. According to the results, the explanations generated from ViT are deemed more trustworthy than those produced by other CNNs, and as the features of the input image are more dispersed, the advantage of the model becomes more evident.
引用
收藏
页码:159 / 164
页数:6
相关论文
共 50 条
  • [21] Crowdsourcing and Evaluating Concept-driven Explanations of Machine Learning Models
    Mishra S.
    Rzeszotarski J.M.
    Proceedings of the ACM on Human-Computer Interaction, 2021, 5 (CSCW1)
  • [22] Causal deep learning for explainable vision-based quality inspection under visual interference
    Liang, Tianbiao
    Liu, Tianyuan
    Wang, Junliang
    Zhang, Jie
    Zheng, Pai
    JOURNAL OF INTELLIGENT MANUFACTURING, 2025, 36 (02) : 1363 - 1384
  • [23] Using model explanations to guide deep learning models towards consistent explanations for EHR data
    Watson, Matthew
    Hasan, Bashar Awwad Shiekh
    Al Moubayed, Noura
    SCIENTIFIC REPORTS, 2022, 12 (01)
  • [24] Using model explanations to guide deep learning models towards consistent explanations for EHR data
    Matthew Watson
    Bashar Awwad Shiekh Hasan
    Noura Al Moubayed
    Scientific Reports, 12
  • [25] Stochastic Integrated Explanations for Vision Models
    Barkan, Oren
    Elisha, Yehonatan
    Weill, Jonathan
    Asher, Yuval
    Eshel, Amit
    Koenigstein, Noam
    23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING, ICDM 2023, 2023, : 938 - 943
  • [26] Distributed learning of deep feature embeddings for visual recognition tasks
    Bhattacharjee, B.
    Hill, M. L.
    Wu, H.
    Chandakkar, P. S.
    Smith, J. R.
    Wegman, M. N.
    IBM JOURNAL OF RESEARCH AND DEVELOPMENT, 2017, 61 (4-5)
  • [27] Visual Data Simulation for Deep Learning in Robot Manipulation Tasks
    Surak, Miroslav
    Kosnar, Karel
    Kulich, Miroslav
    Kozak, Viktor
    Peeucil, Libor
    MODELLING AND SIMULATION FOR AUTONOMOUS SYSTEMS (MESAS 2018), 2019, 11472 : 402 - 411
  • [28] MiMVP Deep Learning Platform - Streamlining Deep Learning Model Development Workflow for Machine Vision Tasks
    Kamarudin, Nur Afiqah
    Talib, Muhammad Syukri Mohd
    Lau, Phooi Yee
    Hussin, Zarina
    Yuen, Shang Li
    Samsuri, Muhammad Hafiz
    Wong, Chin Wee
    Hon, Hock Woon
    INTERNATIONAL WORKSHOP ON ADVANCED IMAGING TECHNOLOGY, IWAIT 2024, 2024, 13164
  • [29] Evaluating deep learning models for sentiment classification
    Karakus, Betul Ay
    Talo, Muhammed
    Hallac, Ibrahim Riza
    Aydin, Galip
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2018, 30 (21):
  • [30] Deep learning-guided video compression for machine vision tasks
    Kim, Aro
    Woo, Seung-taek
    Park, Minho
    Kim, Dong-hwi
    Lim, Hanshin
    Jung, Soon-heung
    Kwak, Sangwoon
    Park, Sang-hyo
    EURASIP JOURNAL ON IMAGE AND VIDEO PROCESSING, 2024, 2024 (01)