Dual-branch deep cross-modal interaction network for semantic segmentation with thermal images

被引:0
|
作者
Dai K. [1 ]
Chen S. [1 ]
机构
[1] School of Electronics and Information Engineering, Nanjing University of Information Science and Technology, Nanjing
基金
中国国家自然科学基金;
关键词
Cross-modal feature; Deep interaction; Semantic segmentation; Thermal images;
D O I
10.1016/j.engappai.2024.108820
中图分类号
学科分类号
摘要
Semantic segmentation using RGB (Red-Green-Blue) images and thermal datas is an indispensable component of autonomous driving. The key to RGB-Thermal (RGB and Thermal) semantic segmentation is achieving the interaction and fusion of features between RGB and thermal images. Therefore, we propose a dual-branch deep cross-modal interaction network (DCIT) based on Encoder–Decoder structure. This framework consists of two parallel networks for feature extraction from RGB and Thermal data. Specifically, in each feature extraction stage of the Encoder, we design a Cross Feature Regulation Modules (CFRM) to align and correct modality specific features by reducing the inter-modality feature differences and eliminating intra-modality noise. Then, the modality features are aggregated through Cross Modal Feature Fusion Module (CMFFM) based on cross linear attention to capture global information from modality features. Finally, Adaptive Multi-Scale Cross-positional Fusion Module (AMCFM) utilizes the fused features to integrate consistent semantic information in the Decoder stage. Our framework can improve the interaction of cross modal features. Extensive experiments on urban scene datasets demonstrate that our proposed framework outperforms other RGB-Thermal semantic segmentation methods in terms of objective metrics and subjective visual assessments. © 2024 Elsevier Ltd
引用
收藏
相关论文
共 50 条
  • [21] Graph Neural Network Enhanced Dual-Branch Network for lesion segmentation in ultrasound images
    Wang, Yaqi
    Jiang, Cunang
    Luo, Shixin
    Dai, Yu
    Zhang, Jiangxun
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 256
  • [22] CMFFN: An efficient cross-modal feature fusion network for semantic segmentation
    Zhang, Yingjian
    Li, Ning
    Jiao, Jichao
    Ai, Jiawen
    Yan, Zheng
    Zeng, Yingchao
    Zhang, Tianxiang
    Li, Qian
    Robotics and Autonomous Systems, 2025, 186
  • [23] A Dual-Branch Deep Learning Architecture for Multisensor and Multitemporal Remote Sensing Semantic Segmentation
    Bergamasco, Luca
    Bovolo, Francesca
    Bruzzone, Lorenzo
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2023, 16 : 2147 - 2162
  • [24] Deep Graph Convolutional Network with Dual-Branch and Multi-interaction
    Lou J.
    Ye H.
    Yang B.
    Li M.
    Cao F.
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2022, 35 (08): : 754 - 763
  • [25] STDBNet: Shared Trunk and Dual-Branch Network for Real-Time Semantic Segmentation
    Ren, Fenglei
    Zhou, Haibo
    Yang, Lu
    Bai, Yiwen
    Xu, Wenxue
    IEEE SIGNAL PROCESSING LETTERS, 2024, 31 : 770 - 774
  • [26] A lightweight dual-branch semantic segmentation network for enhanced obstacle detection in ship navigation
    Feng, Hui
    Liu, Wensheng
    Xu, Haixiang
    He, Jianhua
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 136
  • [27] DMANet: Dual-branch multiscale attention network for real-time semantic segmentation
    Dong, Yongsheng
    Mao, Chongchong
    Zheng, Lintao
    Wu, Qingtao
    Neurocomputing, 2025, 617
  • [28] Cross-modal attention fusion network for RGB-D semantic segmentation
    Zhao, Qiankun
    Wan, Yingcai
    Xu, Jiqian
    Fang, Lijin
    NEUROCOMPUTING, 2023, 548
  • [29] Transformer-Based Cross-Modal Information Fusion Network for Semantic Segmentation
    Duan, Zaipeng
    Huang, Xiao
    Ma, Jie
    NEURAL PROCESSING LETTERS, 2023, 55 (05) : 6361 - 6375
  • [30] Cross-Modal Hash Retrieval Model for Semantic Segmentation Network for Digital Libraries
    Tang, Siyu
    Yin, Jun
    International Journal of Advanced Computer Science and Applications, 2024, 15 (10) : 58 - 66