Efficient Depth Fusion Transformer for Aerial Image Semantic Segmentation

被引:16
|
作者
Yan, Li [1 ,2 ]
Huang, Jianming [1 ]
Xie, Hong [1 ]
Wei, Pengcheng [1 ]
Gao, Zhao [2 ]
机构
[1] Wuhan Univ, Sch Geodesy & Geomat, Wuhan 430079, Peoples R China
[2] Wuhan Univ, Sch Comp Sci, Wuhan 430072, Peoples R China
关键词
semantic segmentation; self-attention; depth fusion; transformer; RESOLUTION; RGB;
D O I
10.3390/rs14051294
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Taking depth into consideration has been proven to improve the performance of semantic segmentation through providing additional geometry information. Most existing works adopt a two-stream network, extracting features from color images and depth images separately using two branches of the same structure, which suffer from high memory and computation costs. We find that depth features acquired by simple downsampling can also play a complementary part in the semantic segmentation task, sometimes even better than the two-stream scheme with the same two branches. In this paper, a novel and efficient depth fusion transformer network for aerial image segmentation is proposed. The presented network utilizes patch merging to downsample depth input and a depth-aware self-attention (DSA) module is designed to mitigate the gap caused by difference between two branches and two modalities. Concretely, the DSA fuses depth features and color features by computing depth similarity and impact on self-attention map calculated by color feature. Extensive experiments on the ISPRS 2D semantic segmentation dataset validate the efficiency and effectiveness of our method. With nearly half the parameters of traditional two-stream scheme, our method acquires 83.82% mIoU on Vaihingen dataset outperforming other state-of-the-art methods and 87.43% mIoU on Potsdam dataset comparable to the state-of-the-art.
引用
收藏
页数:18
相关论文
共 50 条
  • [21] Semantic perceptive infrared and visible image fusion Transformer
    Yang, Xin
    Huo, Hongtao
    Li, Chang
    Liu, Xiaowen
    Wang, Wenxi
    Wang, Cheng
    [J]. PATTERN RECOGNITION, 2024, 149
  • [22] Efficient Transformer for Remote Sensing Image Segmentation
    Xu, Zhiyong
    Zhang, Weicun
    Zhang, Tianxiang
    Yang, Zhifang
    Li, Jiangyun
    [J]. REMOTE SENSING, 2021, 13 (18)
  • [23] AerialFormer: Multi-Resolution Transformer for Aerial Image Segmentation
    Hanyu, Taisei
    Yamazaki, Kashu
    Tran, Minh
    McCann, Roy A.
    Liao, Haitao
    Rainwater, Chase
    Adkins, Meredith
    Cothren, Jackson
    Le, Ngan
    [J]. REMOTE SENSING, 2024, 16 (16)
  • [24] A Self-Supervised Transformer With Feature Fusion for SAR Image Semantic Segmentation in Marine Aquaculture Monitoring
    Fan, Jianchao
    Zhou, Jianlin
    Wang, Xinzhe
    Wang, Jun
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [25] TFNet: Transformer Fusion Network for Ultrasound Image Segmentation
    Wang, Tao
    Lai, Zhihui
    Kong, Heng
    [J]. PATTERN RECOGNITION, ACPR 2021, PT I, 2022, 13188 : 314 - 325
  • [26] Evaluating Transformer-based Semantic Segmentation Networks for Pathological Image Segmentation
    Cam Nguyen
    Asad, Zuhayr
    Deng, Ruining
    Huo, Yuankai
    [J]. MEDICAL IMAGING 2022: IMAGE PROCESSING, 2022, 12032
  • [27] Seamline Determination Based on Semantic Segmentation for Aerial Image Mosaicking
    Saito, Shunta
    Arai, Ryota
    Aoki, Yoshimitsu
    [J]. IEEE ACCESS, 2015, 3 : 2847 - 2856
  • [28] EFFICIENT IMAGE SEGMENTATION FOR SEMANTIC OBJECT GENERATION
    Chen Xiaotang Yu Yinglin (Dept. of Comm. & Info. Eng.
    [J]. Journal of Electronics(China), 2002, (04) : 420 - 425
  • [29] Cross-scale sampling transformer for semantic image segmentation
    Ma, Yizhe
    Yu, Long
    Lin, Fangjian
    Tian, Shengwei
    [J]. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2023, 44 (02) : 2895 - 2907
  • [30] Refined Division Features Based on Transformer for Semantic Image Segmentation
    Li, Tianping
    Wei, Yanjun
    Liu, Meilin
    Yang, Xiaolong
    Zhang, Zhenyi
    Du, Jun
    [J]. INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2023, 2023