Multiscale deep feature selection fusion network for referring image segmentation

被引:0
|
作者
Dai, Xianwen [1 ]
Lin, Jiacheng [1 ]
Nai, Ke [1 ]
Li, Qingpeng [2 ]
Li, Zhiyong [1 ]
机构
[1] Hunan Univ, Coll Comp Sci & Elect Engn, Changsha, Peoples R China
[2] Hunan Univ, Sch Robot, Changsha, Hunan, Peoples R China
基金
中国国家自然科学基金;
关键词
Referring image segmentation; Semantic segmentation; Multi-modal fusion; Deep learning;
D O I
10.1007/s11042-023-16913-6
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Referring image segmentation has attracted extensive attention in recent years. Previous methods have explored the difficult alignment between visual and textual features, but this problem has not been effectively addressed. This leads to the problem of insufficient interaction between visual features and textual features, which affects model performance. To this end, we propose a language-aware pixel feature fusion module (LPFFM) based on self-attention mechanism to ensure that the features of the two modalities have sufficient interaction in the space and channels. Then we apply it in the shallow to deep layers of the encoder to gradually select visual features related to the text. Secondly, we propose a second selection mechanism to further select visual features that only contain the target. For this mechanism, we design an attention contrastive loss to better suppress irrelevant background information. Further, we propose a multi-scale deep features selection fusion network (MDSFNet) based on the U-net architecture. Finally, the experimental results show that our proposed method is competitive with previous methods, improving the performance by 2.87%, 3.17%, and 3.81% on three benchmark datasets, RefCOCO, RefCOCO+, and G-ref, respectively.
引用
收藏
页码:36287 / 36305
页数:19
相关论文
共 50 条
  • [31] A Review of Optical and SAR Image Deep Feature Fusion in Semantic Segmentation
    Liu, Chenfang
    Sun, Yuli
    Xu, Yanjie
    Sun, Zhongzhen
    Zhang, Xianghui
    Lei, Lin
    Kuang, Gangyao
    [J]. IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2024, 17 : 12910 - 12930
  • [32] A ViT-Based Multiscale Feature Fusion Approach for Remote Sensing Image Segmentation
    Wang, Wei
    Tang, Chen
    Wang, Xin
    Zheng, Bin
    [J]. IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2022, 19
  • [33] Dynamic feature fusion forensics network for deep image inpainting
    Ren H.
    Zhu X.
    Lu J.
    [J]. Harbin Gongye Daxue Xuebao/Journal of Harbin Institute of Technology, 2022, 54 (11): : 47 - 58
  • [34] CRMEFNet: A coupled refinement, multiscale exploration and fusion network for medical image segmentation
    Wang, Zhi
    Yu, Long
    Tian, Shengwei
    Huo, Xiangzuo
    [J]. COMPUTERS IN BIOLOGY AND MEDICINE, 2024, 171
  • [35] A Multiscale Deep Middle-level Feature Fusion Network for Hyperspectral Classification
    Li, Zhaokui
    Huang, Lin
    He, Jinrong
    [J]. REMOTE SENSING, 2019, 11 (06)
  • [36] Hierarchical Shrinkage Multiscale Network for Hyperspectral Image Classification With Hierarchical Feature Fusion
    Gao, Hongmin
    Chen, Zhonghao
    Li, Chenming
    [J]. IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2021, 14 : 5760 - 5772
  • [37] Multiscale feature fusion network for automatic port segmentation from remote sensing images
    Ju, Haoran
    Bi, Fukun
    Bian, Mingming
    Shi, Yinni
    [J]. JOURNAL OF APPLIED REMOTE SENSING, 2022, 16 (04)
  • [38] A statistical multiscale approach to image segmentation and fusion
    Cardinali, A
    Nason, GP
    [J]. 2005 7TH INTERNATIONAL CONFERENCE ON INFORMATION FUSION (FUSION), VOLS 1 AND 2, 2005, : 475 - 482
  • [39] Object Detection For Remote Sensing Image Based on Multiscale Feature Fusion Network
    Tian Tingting
    Yang Jun
    [J]. LASER & OPTOELECTRONICS PROGRESS, 2022, 59 (16)
  • [40] Underwater Image Enhancement Based on Generate Adversarial Network with Multiscale Feature Fusion
    Chen, Hui
    Wang, Shuo
    Xu, Jiachang
    Xiao, Zhexuan
    [J]. Computer Engineering and Applications, 2023, 59 (21) : 231 - 241