An adaptive guidance fusion network for RGB-D salient object detection

被引:1
|
作者
Sun, Haodong [1 ]
Wang, Yu [1 ]
Ma, Xinpeng [1 ]
机构
[1] Changchun Univ Sci & Technol, Sch Elect & Informat Engn, Changchun, Peoples R China
基金
美国国家科学基金会;
关键词
RGB-D salient object detection; Multi-modal features; Feature fusion; Feature refinement;
D O I
10.1007/s11760-023-02775-w
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
RGB-D salient object detection (RGB-D SOD) has currently attracted much attention for its prospect of broad application. On the basis of the "encoder-decoder" paradigm of the fully convolutional network (FCN), many FCN-based strategies have emerged and achieved huge progress, but underestimated the potential of level-specific characteristics of multi-modal features. In this paper, we propose the adaptive guided fusion network (AGFNet) to further mine the potential information between the depth image and the RGB image, and design an adaptive fusion and coarse-to-fine decoding strategy to achieve high-precision detection of salient objects. Specifically, we first use a two-stream encoder to extract the multi-level features of the RGB image and depth image but refrain from the previous practice of using depth features for each layer. Second, a simple but effective way named multi-modal selective fusion strategy is designed to fuse the multi-level features. Third, for enhancement of contextual information of each level adaptively, an adaptive cross fusion module (ACFM) fuses the features at all levels and outputs a coarse saliency map. Finally, a guided attention refinement module (GARM) utilizes the coarse saliency map to guide the final features from ACFM to realize the enhancement and obtain a refined saliency map. Our method is compared with other state-of-the-art RGB-SOD methods through extensive experiments, and the results demonstrate the superiority of our proposed AGFNet. The source code of this project is available at https://github.com/HaodongSun809/my_AGFNet.git.
引用
收藏
页码:1683 / 1693
页数:11
相关论文
共 50 条
  • [1] An adaptive guidance fusion network for RGB-D salient object detection
    Haodong Sun
    Yu Wang
    Xinpeng Ma
    [J]. Signal, Image and Video Processing, 2024, 18 : 1683 - 1693
  • [2] Adaptive fusion network for RGB-D salient object detection
    Chen, Tianyou
    Xiao, Jin
    Hu, Xiaoguang
    Zhang, Guofeng
    Wang, Shaojie
    [J]. NEUROCOMPUTING, 2023, 522 : 152 - 164
  • [3] Scale Adaptive Fusion Network for RGB-D Salient Object Detection
    Kong, Yuqiu
    Zheng, Yushuo
    Yao, Cuili
    Liu, Yang
    Wang, He
    [J]. COMPUTER VISION - ACCV 2022, PT III, 2023, 13843 : 608 - 625
  • [4] Adaptive Fusion for RGB-D Salient Object Detection
    Wang, Ningning
    Gong, Xiaojin
    [J]. IEEE ACCESS, 2019, 7 : 55277 - 55284
  • [5] Bifurcation Fusion Network for RGB-D Salient Object Detection
    Zhao, Zhi-Hua
    Chen, Li
    [J]. JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2022, 31 (12)
  • [6] Depth cue enhancement and guidance network for RGB-D salient object detection
    Li, Xiang
    Zhang, Qing
    Yan, Weiqi
    Dai, Meng
    [J]. JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2023, 95
  • [7] Heterogeneous Fusion and Integrity Learning Network for RGB-D Salient Object Detection
    Gao, Haorao
    Su, Yiming
    Wang, Fasheng
    Li, Haojie
    [J]. ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2024, 20 (07)
  • [8] Bi-Directional Progressive Guidance Network for RGB-D Salient Object Detection
    Yang, Yang
    Qin, Qi
    Luo, Yongjiang
    Liu, Yi
    Zhang, Qiang
    Han, Jungong
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (08) : 5346 - 5360
  • [9] Discriminative feature fusion for RGB-D salient object detection
    Chen, Zeyu
    Zhu, Mingyu
    Chen, Shuhan
    Lu, Lu
    Tang, Haonan
    Hu, Xuelong
    Ji, Chunfan
    [J]. COMPUTERS & ELECTRICAL ENGINEERING, 2023, 106
  • [10] Progressively Complementarity-aware Fusion Network for RGB-D Salient Object Detection
    Chen, Hao
    Li, Youfu
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 3051 - 3060