LaST: Label-Free Self-Distillation Contrastive Learning With Transformer Architecture for Remote Sensing Image Scene Classification

被引:20
|
作者
Wang, Xuying [1 ]
Zhu, Jiawei [1 ]
Yan, Zhengliang [2 ]
Zhang, Zhaoyang [1 ]
Zhang, Yunsheng [2 ]
Chen, Yansheng [1 ]
Li, Haifeng [1 ]
机构
[1] Cent South Univ, Sch Geosci & Info Phys, Changsha 410083, Peoples R China
[2] Tianjin Zhongwei Aerosp Data Syst Technol Co Ltd, Tianjin 300301, Peoples R China
基金
中国国家自然科学基金;
关键词
Transformers; Training; Crops; Image analysis; Task analysis; Annotations; Supervised learning; Contrastive learning; remote sensing image (RSI) scene classification; self-distillation; self-supervised learning (SSL);
D O I
10.1109/LGRS.2022.3185088
中图分类号
P3 [地球物理学]; P59 [地球化学];
学科分类号
0708 ; 070902 ;
摘要
The increase in self-supervised learning (SSL), especially contrastive learning, has enabled one to train deep neural network models with unlabeled data for remote sensing image (RSI) scene classification. Nevertheless, it still suffers from the following issues: 1) the performance of the contrastive learning method is significantly impacted by the hard negative sample (HNS) issue, since the RSI scenario is complex in semantics and rich in surface features; 2) the multiscale characteristic of RSI is missed in the existing contrastive learning methods; and 3) as the backbone of a deep learning model, especially in the case of limited annotation, a convolutional neural network (CNN) does not include the adequate receptive field of convolutional kernels to capture the broad contextual information of RSI. In this regard, we propose label-free self-distillation contrastive learning with a transformer architecture (LaST). We introduce the self-distillation contrastive learning mechanism to address the HNS issue. Specifically, the LaST architecture comprises two modules: scale alignment with a multicrop module and a long-range dependence capture backbone module. In the former, we present global-local crop and scale alignment to encourage local-to-global correspondence and acquire multiscale relations. Then, the distorted views are fed into a transformer as a backbone, which is good at capturing the long-range-dependent contextual information of the RSI while maintaining the spatial smoothness of the learned features. Experiments on public datasets show that in the downstream scene classification task, LaST improves the performance of the self-supervised trained model by a maximum of 2.18% compared to the HNS-impacted contrastive learning approaches, and only 1.5% of labeled data can achieve the performance of supervised training CNNs with 10% labeled data. Moreover, this letter supports the integration of a transformer architecture and self-supervised paradigms in RSI interpretation.
引用
收藏
页数:5
相关论文
共 50 条
  • [1] Variational Self-Distillation for Remote Sensing Scene Classification
    Hu, Yutao
    Huang, Xin
    Luo, Xiaoyan
    Han, Jungong
    Cao, Xianbin
    Zhang, Jun
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [2] Vision Transformer With Contrastive Learning for Remote Sensing Image Scene Classification
    Bi, Meiqiao
    Wang, Minghua
    Li, Zhi
    Hong, Danfeng
    [J]. IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2023, 16 : 738 - 749
  • [3] Class-Aware Self-Distillation for Remote Sensing Image Scene Classification
    Wu, Bin
    Hao, Siyuan
    Wang, Wei
    [J]. IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2024, 17 : 2173 - 2188
  • [4] Remote Sensing Image Scene Classification with Noisy Label Distillation
    Zhang, Rui
    Chen, Zhenghao
    Zhang, Sanxing
    Song, Fei
    Zhang, Gang
    Zhou, Quancheng
    Lei, Tao
    [J]. REMOTE SENSING, 2020, 12 (15)
  • [5] Embedded Self-Distillation in Compact Multibranch Ensemble Network for Remote Sensing Scene Classification
    Zhao, Qi
    Ma, Yujing
    Lyu, Shuchang
    Chen, Lijiang
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [6] Remote Sensing Image Scene Classification Based on Supervised Contrastive Learning
    Guo Dongen
    Xia Ying
    Luo Xiaobo
    Feng Jiangfan
    [J]. ACTA PHOTONICA SINICA, 2021, 50 (07)
  • [7] Exploring Hybrid Contrastive Learning and Scene-to-Label Information for Multilabel Remote Sensing Image Classification
    Song, Tiecheng
    Bai, Shufen
    Yang, Feng
    Gao, Chenqiang
    Chen, Haonan
    Li, Jun
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62
  • [8] Learn by Yourself: A Feature-Augmented Self-Distillation Convolutional Neural Network for Remote Sensing Scene Image Classification
    Shi, Cuiping
    Ding, Mengxiang
    Wang, Liguo
    Pan, Haizhu
    [J]. REMOTE SENSING, 2023, 15 (23)
  • [9] Remote Sensing Image Scene Classification via Self-Supervised Learning and Knowledge Distillation
    Zhao, Yibo
    Liu, Jianjun
    Yang, Jinlong
    Wu, Zebin
    [J]. REMOTE SENSING, 2022, 14 (19)
  • [10] SIMPLE SELF-DISTILLATION LEARNING FOR NOISY IMAGE CLASSIFICATION
    Sasaya, Tenta
    Watanabe, Takashi
    Ida, Takashi
    Ono, Toshiyuki
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 795 - 799