Towards explainable deep visual saliency models

被引:2
|
作者
Malladi, Sai Phani Kumar [1 ]
Mukherjee, Jayanta [2 ]
Larabi, Mohamed-Chaker [3 ]
Chaudhury, Santanu [4 ]
机构
[1] Indian Inst Technol, Adv Technol Dev Ctr, Kharagpur, India
[2] IIT Kharagpur, Dept Comp Sci & Engn, Kharagpur, India
[3] Univ Poitiers, XLIM UMR CNRS 7252, Poitiers, France
[4] IIT Jodhpur, Dept Comp Sci & Engn, Jodhpur, India
关键词
Explainable saliency; Human perception; Log-Gabor filters; Color perception; ATTENTION;
D O I
10.1016/j.cviu.2023.103782
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks have shown their profound impact on achieving human-level performance in visual saliency prediction. However, it is still unclear how they learn their task and what it means in terms of understanding human visual system. In this work, we propose a framework to derive explainable saliency models from their corresponding deep architectures. Mainly, we explain a deep saliency model by under-standing its four different aspects: (1) intermediate activation maps of deep layers, (2) biologically plausible Log-Gabor (LG) filters for salient region identification, (3) positional biased behavior of Log-Gabor filters and (4) processing of color information by establishing a relevance with human visual system. We consider four state-of-the-art (SOTA) deep saliency models, namely CMRNet, UNISAL, DeepGaze IIE, and MSI-Net for their interpretation using our proposed framework. We observe that explainable models perform way better than the classical SOTA models. We also find that CMRNet transforms the input RGB space to a representation after the input layer, which is very close to YUV space of a color image. Then, we discuss about the biological consideration and relevance of our framework for its possible anatomical substratum of visual attention. We find a good correlation between components of HVS and the base operations of the proposed technique. Hence, we say that this generic explainable framework provides a new perspective to see relationship between classical methods/human visual system and DNN based ones.
引用
收藏
页数:14
相关论文
共 50 条
  • [31] TranSalNet: Towards perceptually relevant visual saliency prediction
    Lou, Jianxun
    Lin, Hanhe
    Marshall, David
    Saupe, Dietmar
    Liu, Hantao
    NEUROCOMPUTING, 2022, 494 : 455 - 467
  • [32] Deep saliency models : The quest for the loss function
    Bruckert, Alexandre
    Tavakoli, Hamed R.
    Liu, Zhi
    Christie, Marc
    Le Meur, Olivier
    NEUROCOMPUTING, 2021, 453 : 693 - 704
  • [33] Visual Saliency Guided Deep Fabric Defect Classification
    He, Yonggui
    Song, Yaoye
    Shen, Jifeng
    Yang, Wankou
    INTELLIGENCE SCIENCE AND BIG DATA ENGINEERING: VISUAL DATA ENGINEERING, PT I, 2019, 11935 : 435 - 446
  • [34] Visual Saliency Modeling with Deep Learning: A Comprehensive Review
    Abraham, Shilpa Elsa
    Kovoor, Binsu C.
    JOURNAL OF INFORMATION & KNOWLEDGE MANAGEMENT, 2023, 22 (02)
  • [35] Explainable Deep Learning Models on the Diagnosis of Pneumonia
    Yang, Yuting
    Mei, Gang
    Piccialli, Francesco
    2021 IEEE/ACM CONFERENCE ON CONNECTED HEALTH: APPLICATIONS, SYSTEMS AND ENGINEERING TECHNOLOGIES (CHASE 2021), 2021, : 134 - 138
  • [36] Explainable Deep Classification Models for Domain Generalization
    Zunino, Andrea
    Bargal, Sarah Adel
    Volpi, Riccardo
    Sameki, Mehrnoosh
    Zhang, Jianming
    Sclaroff, Stan
    Murino, Vittorio
    Saenko, Kate
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 3227 - 3236
  • [37] State of the Art of Visual Analytics for eXplainable Deep Learning
    La Rosa, B.
    Santucci, G.
    Giot, R.
    Auber, D.
    Santucci, G.
    Giot, R.
    Bertini, E.
    Giot, R.
    Angelini, M.
    COMPUTER GRAPHICS FORUM, 2023, 42 (01) : 319 - 355
  • [38] ROBUSTNESS AND REPEATABILITY OF SALIENCY MODELS SUBJECTED TO VISUAL DEGRADATIONS
    Le Meur, Olivier
    2011 18TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2011,
  • [39] Analysis of scores, datasets, and models in visual saliency prediction
    Borji, Ali
    Tavakoli, Hamed R.
    Sihite, Dicky N.
    Itti, Laurent
    2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2013, : 921 - 928
  • [40] On the usage of visual saliency models for computer generated objects
    Abid, Mona
    Da Silva, Matthieu Perreira
    Le Callet, Patrik
    2019 IEEE 21ST INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING (MMSP 2019), 2019,