Co-saliency detection with two-stage co-attention mining and individual calibration

被引:1
|
作者
Tan, Zhenshan [1 ]
Gu, Xiaodong [1 ]
Cheng, Qingrong [1 ]
机构
[1] Fudan Univ, Dept Elect Engn, Shanghai 200438, Peoples R China
关键词
Co-salient object detection; Co-attention; Edge guidance; Image fusion algorithm; DEEP; NETWORK; OBJECTS;
D O I
10.1016/j.engappai.2023.107201
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Learning-based methods have become popular in co-salient object detection (CoSOD). However, existing methods suffer from two challenging issues, including mining the inter-image co-attention and calibrating the intra-image salient objects. Moreover, the training data is insufficient. To address these challenges, we propose an end-to-end network using Two-stage Co-attention mining and Individual Calibration (TCIC) to predict the co-salient objects. Firstly, a two-stage co-attention mining architecture (TCM), including a classified co-attention module (CCM) and a focal co-attention module (FCM), is designed to model inter-image relationships. In the first stage, a CCM is applied to capture the classification interactions of multiple images, tentatively extracting the co-attention. In the second stage, we propose an FCM to adaptively suppress and aggregate multiple salient features, aiming to recalibrate the co-attention in the first stage. Secondly, considering the shape features and location information offered by the boundary features, an edge guidance module (EGM) is embedded into the individual calibration architecture (ICA) to calibrate individuals. Besides, we also adopt a co-attention transfer strategy (CTS) to keep the consistency of the co-attention during feature transfer in the decoder. Finally, TCM and ICA are integrated into a unified end-to-end framework to predict fine-grained boundary-preserving results. Besides, an image fusion algorithm (IFA) is tailored without extra pixel-level annotations for automatic generation of the composite images, aiming to supplement the training dataset. Experimental results on three prevailing testing datasets show the superiority of the proposed method in terms of various evaluation metrics.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] Co-Saliency Detection With Co-Attention Fully Convolutional Network
    Gao, Guangshuai
    Zhao, Wenting
    Liu, Qingjie
    Wang, Yunhong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2021, 31 (03) : 877 - 889
  • [2] Detecting Robust Co-Saliency with Recurrent Co-Attention Neural Network
    Li, Bo
    Sun, Zhengxing
    Tang, Lv
    Sun, Yunhan
    Shi, Jinlong
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 818 - 825
  • [3] An Information Calibration and Sliding Mining Network for Co-Saliency Object Detection
    Wei, Longsheng
    Huang, Jiu
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73
  • [4] Video co-saliency detection
    Xie, Yufeng
    Ye, Linwei
    Liu, Zhi
    Zou, Xuemei
    EIGHTH INTERNATIONAL CONFERENCE ON DIGITAL IMAGE PROCESSING (ICDIP 2016), 2016, 10033
  • [5] PREATTENTIVE CO-SALIENCY DETECTION
    Chen, Hwann-Tzong
    2010 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, 2010, : 1117 - 1120
  • [6] A Review of Co-saliency Detection
    Qian X.-L.
    Bai Z.
    Chen Y.
    Zhang D.-W.
    Shi K.-F.
    Wang F.
    Wu Q.-E.
    Wu Y.-Y.
    Wang W.
    Tien Tzu Hsueh Pao/Acta Electronica Sinica, 2019, 47 (06): : 1352 - 1365
  • [7] Revisiting Co-Saliency Detection: A Novel Approach Based on Two-Stage Multi-View Spectral Rotation Co-clustering
    Yao, Xiwen
    Han, Junwei
    Zhang, Dingwen
    Nie, Feiping
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2017, 26 (07) : 3196 - 3209
  • [8] A model of co-saliency based audio attention
    Zhao, XiaoMing
    Wang, Xinxin
    Cheng, De
    MULTIMEDIA TOOLS AND APPLICATIONS, 2020, 79 (31-32) : 23045 - 23069
  • [9] Discriminative Co-Saliency and Background Mining Transformer for Co-Salient Object Detection
    Li, Long
    Han, Junwei
    Zhang, Ni
    Liu, Nian
    Khan, Salman
    Cholakkal, Hisham
    Anwer, Rao Muhammad
    Khan, Fahad Shahbaz
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 7247 - 7256
  • [10] A model of co-saliency based audio attention
    XiaoMing Zhao
    Xinxin Wang
    De Cheng
    Multimedia Tools and Applications, 2020, 79 : 23045 - 23069