Tuple Perturbation-Based Contrastive Learning Framework for Multimodal Remote Sensing Image Semantic Segmentation

被引:0
|
作者
Ye, Yuanxin [1 ,2 ]
Dai, Jinkun [1 ,2 ]
Zhou, Liang [1 ,2 ]
Duan, Keyi [1 ,2 ]
Tao, Ran [3 ]
Li, Wei [3 ]
Hong, Danfeng [4 ,5 ]
机构
[1] Southwest Jiaotong Univ, Fac Geosci & Engn, Chengdu 610031, Peoples R China
[2] Southwest Jiaotong Univ, State Prov Joint Engn Lab Spatial Informat Technol, Chengdu 611756, Peoples R China
[3] Beijing Inst Technol, Sch Informat & Elect, Beijing 100081, Peoples R China
[4] Chinese Acad Sci, Aerosp Informat Res Inst, Beijing 100094, Peoples R China
[5] Univ Chinese Acad Sci, Sch Elect Elect & Commun Engn, Beijing 100049, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature extraction; Contrastive learning; Remote sensing; Optical sensors; Optical imaging; Radar polarimetry; Adaptive optics; Training; Perturbation methods; multimodal remote sensing image (RSI); negative samples; semantic segmentation; tuple perturbation;
D O I
10.1109/TGRS.2025.3542868
中图分类号
P3 [地球物理学]; P59 [地球化学];
学科分类号
0708 ; 070902 ;
摘要
Deep learning models exhibit promising potential in multimodal remote sensing image semantic segmentation (MRSISS). However, the constrained access to labeled samples for training deep learning networks significantly influences the performance of these models. To address that, self-supervised learning (SSL) methods have garnered significant interest in the remote sensing community. Accordingly, this article proposes a novel multimodal contrastive learning framework based on tuple perturbation, which includes the pretraining and fine-tuning stages. First, a tuple perturbation-based multimodal contrastive learning network (TMCNet) is designed to better explore shared and different feature representations across modalities during the pretraining stage and the tuple perturbation module is introduced to improve the network's ability to extract multimodal features by generating more complex negative samples. In the fine-tuning stage, we develop a simple and effective multimodal semantic segmentation network (MSSNet), which can reduce noise by using complementary information from various modalities to integrate multimodal features more effectively, resulting in better semantic segmentation performance. Extensive experiments have been carried out on two published multimodal image datasets including optical and synthetic aperture radar (SAR) pairs, and the results show that the proposed framework can obtain more superior performance of semantic segmentation than the current state-of-the-art methods in cases of limited labeled samples. The source code is available at https://github.com/yeyuanxin110/TMCNet-MSSNet.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] Remote Sensing Image Coding for Machines on Semantic Segmentation via Contrastive Learning
    Zhang, Junxi
    Chen, Zhenzhong
    Liu, Shan
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62
  • [2] A Mamba-Diffusion Framework for Multimodal Remote Sensing Image Semantic Segmentation
    Du, Wen-Liang
    Gu, Yang
    Zhao, Jiaqi
    Zhu, Hancheng
    Yao, Rui
    Zhou, Yong
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21
  • [3] Remote Sensing Image Semantic Segmentation Network Based on Multimodal Fusion
    Hu, Yuxiang
    Yu, Changhong
    Gao, Ming
    Computer Engineering and Applications, 60 (15): : 234 - 242
  • [4] Joint Learning of Semantic Segmentation and Height Estimation for Remote Sensing Image Leveraging Contrastive Learning
    Gao, Zhi
    Sun, Wenbo
    Lu, Yao
    Zhang, Yichen
    Song, Weiwei
    Zhang, Yongjun
    Zhai, Ruifang
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [5] SegMind: Semisupervised Remote Sensing Image Semantic Segmentation With Masked Image Modeling and Contrastive Learning Method
    Li, Zhenghong
    Chen, Hao
    Wu, Jiangjiang
    Li, Jun
    Jing, Ning
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [6] Domain Adaptation for Remote Sensing Image Semantic Segmentation: An Integrated Approach of Contrastive Learning and Adversarial Learning
    Bai, Lubin
    Du, Shihong
    Zhang, Xiuyuan
    Wang, Haoyu
    Liu, Bo
    Ouyang, Song
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [7] Simple and Efficient: A Semisupervised Learning Framework for Remote Sensing Image Semantic Segmentation
    Lu, Xiaoqiang
    Jiao, Licheng
    Liu, Fang
    Yang, Shuyuan
    Liu, Xu
    Feng, Zhixi
    Li, Lingling
    Chen, Puhua
    IEEE Transactions on Geoscience and Remote Sensing, 2022, 60
  • [8] Simple and Efficient: A Semisupervised Learning Framework for Remote Sensing Image Semantic Segmentation
    Lu, Xiaoqiang
    Jiao, Licheng
    Liu, Fang
    Yang, Shuyuan
    Liu, Xu
    Feng, Zhixi
    Li, Lingling
    Chen, Puhua
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [9] Remote Sensing Image Semantic Change Detection Boosted by Semi-Supervised Contrastive Learning of Semantic Segmentation
    Zhang, Xiuwei
    Yang, Yizhe
    Ran, Lingyan
    Chen, Liang
    Wang, Kangwei
    Yu, Lei
    Wang, Peng
    Zhang, Yanning
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62 : 1 - 13
  • [10] Category-Wise Fusion and Enhancement Learning for Multimodal Remote Sensing Image Semantic Segmentation
    Zheng, Aihua
    He, Jinbo
    Wang, Ming
    Li, Chenglong
    Luo, Bin
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60