MSEDNet: Multi-scale fusion and edge-supervised network for RGB-T salient object detection

被引:11
|
作者
Peng, Daogang [1 ]
Zhou, Weiyi [1 ]
Pan, Junzhen [1 ]
Wang, Danhao [1 ]
机构
[1] Shanghai Univ Elect Power, Coll Automat Engn, 2588 Changyang Rd, Shanghai 200090, Peoples R China
关键词
RGB-T; Salient object detection; Multi-scale fusion; Edge fusion loss; SEGMENTATION;
D O I
10.1016/j.neunet.2023.12.031
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
RGB-T Salient object detection (SOD) is to accurately segment salient regions in both visible light images and thermal infrared images. However, most of existing methods for SOD neglects the critical complementarity between multiple modalities images, which is beneficial to further improve the detection accuracy. Therefore, this work introduces the MSEDNet RGB-T SOD method. We utilize an encoder to extract multi-level modalities features from both visible light images and thermal infrared images, which are subsequently categorized into high, medium, and low level. Additionally, we propose three separate feature fusion modules to comprehensively extract complementary information between different modalities during the fusion process. These modules are applied to specific feature levels: the Edge Dilation Sharpening module for low-level features, the Spatial and Channel-Aware module for mid-level features, and the Cross-Residual Fusion module for high-level features. Finally, we introduce an edge fusion loss function for supervised learning, which effectively extracts edge information from different modalities and suppresses background noise. Comparative demonstrate the superiority of the proposed MSEDNet over other state-of-the-art methods. The code and results can be found at the following link: https://github.com/Zhou-wy/MSEDNet.
引用
收藏
页码:410 / 422
页数:13
相关论文
共 50 条
  • [41] Progressive Guided Fusion Network With Multi-Modal and Multi-Scale Attention for RGB-D Salient Object Detection
    Wu, Jiajia
    Han, Guangliang
    Wang, Haining
    Yang, Hang
    Li, Qingqing
    Liu, Dongxu
    Ye, Fangjian
    Liu, Peixun
    IEEE ACCESS, 2021, 9 : 150608 - 150622
  • [42] Saliency Prototype for RGB-D and RGB-T Salient Object Detection
    Zhang, Zihao
    Wang, Jie
    Han, Yahong
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 3696 - 3705
  • [43] GOSNet: RGB-T salient object detection network based on Global Omnidirectional Scanning
    Jiang, Bochang
    Luo, Dan
    Shang, Zihan
    Liu, Sicheng
    NEUROCOMPUTING, 2025, 630
  • [44] SLMSF-Net: A Semantic Localization and Multi-Scale Fusion Network for RGB-D Salient Object Detection
    Peng, Yanbin
    Zhai, Zhinian
    Feng, Mingkun
    SENSORS, 2024, 24 (04)
  • [45] Asymmetric cross-modal activation network for RGB-T salient object detection
    Xu, Chang
    Li, Qingwu
    Zhou, Qingkai
    Jiang, Xiongbiao
    Yu, Dabing
    Zhou, Yaqin
    KNOWLEDGE-BASED SYSTEMS, 2022, 258
  • [46] Multi-scale Interactive Network for Salient Object Detection
    Pang, Youwei
    Zhao, Xiaoqi
    Zhang, Lihe
    Lu, Huchuan
    arXiv, 2020,
  • [47] Multi-Scale Cascade Network for Salient Object Detection
    Li, Xin
    Yang, Fan
    Cheng, Hong
    Chen, Junyu
    Guo, Yuxiao
    Chen, Leiting
    PROCEEDINGS OF THE 2017 ACM MULTIMEDIA CONFERENCE (MM'17), 2017, : 439 - 447
  • [48] RGB-T salient object detection via CNN feature and result saliency map fusion
    Chang Xu
    Qingwu Li
    Mingyu Zhou
    Qingkai Zhou
    Yaqin Zhou
    Yunpeng Ma
    Applied Intelligence, 2022, 52 : 11343 - 11362
  • [49] DaCFN: divide-and-conquer fusion network for RGB-T object detection
    Wang, Bofan
    Zhao, Haitao
    Zhuang, Yi
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2023, 14 (07) : 2407 - 2420
  • [50] RGB-T salient object detection via CNN feature and result saliency map fusion
    Xu, Chang
    Li, Qingwu
    Zhou, Mingyu
    Zhou, Qingkai
    Zhou, Yaqin
    Ma, Yunpeng
    APPLIED INTELLIGENCE, 2022, 52 (10) : 11343 - 11362