Improving deep neural network generalization and robustness to background bias via layer-wise relevance propagation optimization

被引:0
|
作者
Pedro R. A. S. Bassi
Sergio S. J. Dertkigil
Andrea Cavalli
机构
[1] Alma Mater Studiorum - University of Bologna,Center for Biomolecular Nanotechnologies
[2] Istituto Italiano di Tecnologia,School of Medical Sciences
[3] University of Campinas (UNICAMP),undefined
[4] Istituto Italiano di Tecnologia,undefined
来源
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Features in images’ backgrounds can spuriously correlate with the images’ classes, representing background bias. They can influence the classifier’s decisions, causing shortcut learning (Clever Hans effect). The phenomenon generates deep neural networks (DNNs) that perform well on standard evaluation datasets but generalize poorly to real-world data. Layer-wise Relevance Propagation (LRP) explains DNNs’ decisions. Here, we show that the optimization of LRP heatmaps can minimize the background bias influence on deep classifiers, hindering shortcut learning. By not increasing run-time computational cost, the approach is light and fast. Furthermore, it applies to virtually any classification architecture. After injecting synthetic bias in images’ backgrounds, we compared our approach (dubbed ISNet) to eight state-of-the-art DNNs, quantitatively demonstrating its superior robustness to background bias. Mixed datasets are common for COVID-19 and tuberculosis classification with chest X-rays, fostering background bias. By focusing on the lungs, the ISNet reduced shortcut learning. Thus, its generalization performance on external (out-of-distribution) test databases significantly surpassed all implemented benchmark models.
引用
收藏
相关论文
共 49 条
  • [41] Modelling and identification of characteristic kinematic features preceding freezing of gait with convolutional neural networks and layer-wise relevance propagation
    Benjamin Filtjens
    Pieter Ginis
    Alice Nieuwboer
    Muhammad Raheel Afzal
    Joke Spildooren
    Bart Vanrumste
    Peter Slaets
    [J]. BMC Medical Informatics and Decision Making, 21
  • [42] Advancing Attribution-Based Neural Network Explainability through Relative Absolute Magnitude Layer-Wise Relevance Propagation and Multi-Component Evaluation
    Vukadin, Davor
    Afric, Petar
    Silic, Marin
    Delac, Goran
    [J]. ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2024, 15 (03)
  • [43] A spatial-temporal layer-wise relevance propagation method for improving interpretability and prediction accuracy of LSTM building energy prediction
    Li, Guannan
    Li, Fan
    Xu, Chengliang
    Fang, Xi
    [J]. ENERGY AND BUILDINGS, 2022, 271
  • [44] Effects of various information scenarios on layer-wise relevance propagation-based interpretable convolutional neural networks for air handling unit fault diagnosis
    Xiong, Chenglong
    Li, Guannan
    Yan, Ying
    Zhang, Hanyuan
    Xu, Chengliang
    Chen, Liang
    [J]. BUILDING SIMULATION, 2024,
  • [45] Identification of voxel-based texture abnormalities as new biomarkers for schizophrenia and major depressive patients using layer-wise relevance propagation on deep learning decisions
    Korda, A. I.
    Ruef, A.
    Neufang, S.
    Davatzikos, C.
    Borgwardt, S.
    Meisenzahl, E. M.
    Koutsouleris, N.
    [J]. PSYCHIATRY RESEARCH-NEUROIMAGING, 2021, 313
  • [46] Improving Robustness of Deep Neural Network Acoustic Models via Speech Separation and Joint Adaptive Training
    Narayanan, Arun
    Wang, DeLiang
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2015, 23 (01) : 92 - 101
  • [47] Four Transformer-Based Deep Learning Classifiers Embedded with an Attention U-Net-Based Lung Segmenter and Layer-Wise Relevance Propagation-Based Heatmaps for COVID-19 X-ray Scans
    Gupta, Siddharth
    Dubey, Arun K.
    Singh, Rajesh
    Kalra, Mannudeep K.
    Abraham, Ajith
    Kumari, Vandana
    Laird, John R.
    Al-Maini, Mustafa
    Gupta, Neha
    Singh, Inder
    Viskovic, Klaudija
    Saba, Luca
    Suri, Jasjit S.
    [J]. DIAGNOSTICS, 2024, 14 (14)
  • [48] LayCO: Achieving Least Lossy Accuracy for Most Efficient RRAM-Based Deep Neural Network Accelerator via Layer-Centric Co-Optimization
    Zhao, Shao-Feng
    Wang, Fang
    Liu, Bo
    Feng, Dan
    Liu, Yang
    [J]. JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2023, 38 (02) : 328 - 347
  • [49] LayCO: Achieving Least Lossy Accuracy for Most Efficient RRAM-Based Deep Neural Network Accelerator via Layer-Centric Co-Optimization
    Shao-Feng Zhao
    Fang Wang
    Bo Liu
    Dan Feng
    Yang Liu
    [J]. Journal of Computer Science and Technology, 2023, 38 : 328 - 347