MSEDNet: Multi-scale fusion and edge-supervised network for RGB-T salient object detection

被引:11
|
作者
Peng, Daogang [1 ]
Zhou, Weiyi [1 ]
Pan, Junzhen [1 ]
Wang, Danhao [1 ]
机构
[1] Shanghai Univ Elect Power, Coll Automat Engn, 2588 Changyang Rd, Shanghai 200090, Peoples R China
关键词
RGB-T; Salient object detection; Multi-scale fusion; Edge fusion loss; SEGMENTATION;
D O I
10.1016/j.neunet.2023.12.031
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
RGB-T Salient object detection (SOD) is to accurately segment salient regions in both visible light images and thermal infrared images. However, most of existing methods for SOD neglects the critical complementarity between multiple modalities images, which is beneficial to further improve the detection accuracy. Therefore, this work introduces the MSEDNet RGB-T SOD method. We utilize an encoder to extract multi-level modalities features from both visible light images and thermal infrared images, which are subsequently categorized into high, medium, and low level. Additionally, we propose three separate feature fusion modules to comprehensively extract complementary information between different modalities during the fusion process. These modules are applied to specific feature levels: the Edge Dilation Sharpening module for low-level features, the Spatial and Channel-Aware module for mid-level features, and the Cross-Residual Fusion module for high-level features. Finally, we introduce an edge fusion loss function for supervised learning, which effectively extracts edge information from different modalities and suppresses background noise. Comparative demonstrate the superiority of the proposed MSEDNet over other state-of-the-art methods. The code and results can be found at the following link: https://github.com/Zhou-wy/MSEDNet.
引用
收藏
页码:410 / 422
页数:13
相关论文
共 50 条
  • [31] Multi-scale iterative refinement network for RGB-D salient object detection
    Liu, Ze-Yu
    Liu, Jian-Wei
    Zuo, Xin
    Hu, Ming-Fei
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2021, 106
  • [32] Cross-Collaboration Weighted Fusion Network for RGB-T Salient Detection
    Wang, Yumei
    Dongye, Changlei
    Zhao, Wenxiu
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT IV, ICIC 2024, 2024, 14865 : 301 - 312
  • [33] UMINet: a unified multi-modality interaction network for RGB-D and RGB-T salient object detection
    Lina Gao
    Ping Fu
    Mingzhu Xu
    Tiantian Wang
    Bing Liu
    The Visual Computer, 2024, 40 : 1565 - 1582
  • [34] Feature aggregation with transformer for RGB-T salient object detection
    Zhang, Ping
    Xu, Mengnan
    Zhang, Ziyan
    Gao, Pan
    Zhang, Jing
    NEUROCOMPUTING, 2023, 546
  • [35] Edge Preserving and Multi-Scale Contextual Neural Network for Salient Object Detection
    Wang, Xiang
    Ma, Huimin
    Chen, Xiaozhi
    You, Shaodi
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (01) : 121 - 134
  • [36] Multi-scale feature extraction and fusion with attention interaction for RGB-T
    Xing, Haijiao
    Wei, Wei
    Zhang, Lei
    Zhang, Yanning
    PATTERN RECOGNITION, 2025, 157
  • [37] RGB-D Salient Object Detection via Feature Fusion and Multi-scale Enhancement
    Wu, Peiliang
    Duan, Liangliang
    Kong, Lingfu
    COMPUTER VISION, CCCV 2015, PT II, 2015, 547 : 359 - 368
  • [38] Adaptive interactive network for RGB-T salient object detection with double mapping transformer
    Dong, Feng
    Wang, Yuxuan
    Zhu, Jinchao
    Li, Yuehua
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (20) : 59169 - 59193
  • [39] Enabling modality interactions for RGB-T salient object detection
    Zhang, Qiang
    Xi, Ruida
    Xiao, Tonglin
    Huang, Nianchang
    Luo, Yongjiang
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2022, 222
  • [40] CFRNet: Cross-Attention-Based Fusion and Refinement Network for Enhanced RGB-T Salient Object Detection
    Deng, Biao
    Liu, Di
    Cao, Yang
    Liu, Hong
    Yan, Zhiguo
    Chen, Hu
    SENSORS, 2024, 24 (22)