GraphGST: Graph Generative Structure-Aware Transformer for Hyperspectral Image Classification

被引:17
|
作者
Jiang, Mengying [1 ]
Su, Yuanchao [2 ,3 ]
Gao, Lianru [3 ]
Plaza, Antonio [4 ]
Zhao, Xi-Le [5 ]
Sun, Xu [3 ]
Liu, Guizhong [1 ]
机构
[1] Xi An Jiao Tong Univ, Sch Elect & Informat Engn, Xian 710049, Peoples R China
[2] Xian Univ Sci & Technol, Coll Geomat, Dept Remote Sensing, Xian 710054, Peoples R China
[3] Chinese Acad Sci, Aerosp Informat Res Inst, Key Lab Computat Opt Imaging Technol, Beijing 100094, Peoples R China
[4] Univ Extremadura, Escuela Politecn, Dept Technol Comp & Commun, Hyperspectral Comp Lab, E-10071 Caceres, Spain
[5] Univ Elect Sci & Technol China, Res Ctr Image & Vis Comp, Sch Math Sci, Chengdu 611731, Sichuan, Peoples R China
基金
中国国家自然科学基金;
关键词
Contrastive learning (CL); graph representation learning; hyperspectral image (HSI) classification; transformer; CONVOLUTIONAL NETWORKS;
D O I
10.1109/TGRS.2023.3349076
中图分类号
P3 [地球物理学]; P59 [地球化学];
学科分类号
0708 ; 070902 ;
摘要
Transformer holds significance in deep learning (DL) research. Node embedding (NE) and positional encoding (PE) are usually two indispensable components in a Transformer. The former can excavate hidden correlations from the data, while the latter can store locational relationships between nodes. Recently, the Transformer has been applied for hyperspectral image (HSI) classification because the model can capture long-range dependencies to aggregate global features for representation learning. In an HSI, adjacent pixels tend to be homogeneous, while the NE does not identify the positional information of pixels. Therefore, PE is crucial for Transformers to understand locational relationships between pixels. However, in this area, most Transformer-based methods randomly generate PEs without considering their physical meaning, which leads to weak representations. This article proposes a new graph generative structure-aware Transformer (GraphGST) to solve the above-mentioned PE problem when implementing HSI classification. In our GraphGST, a new absolute PE (APE) is established to acquire pixels' absolute positional sequences (APSs) and is integrated into the Transformer architecture. Moreover, a generative mechanism with self-supervised learning is developed to achieve cross-view contrastive learning (CL), aiming to enhance the representation learning of the Transformer. The proposed GraphGST model can capture local-to-global correlations, and the extracted APSs can complement the spectral features of pixels to assist in NE. Several experiments with real HSIs are conducted to evaluate the effectiveness of our GraphGST. The proposed method demonstrates very competitive performance compared with other state-of-the-art (SOTA) approaches. Our source codes will be provided in the following link https://github.com/yuanchaosu/TGRS-graphGST.
引用
收藏
页码:1 / 16
页数:16
相关论文
共 50 条
  • [41] Relation Structure-Aware Heterogeneous Graph Neural Network
    Zhu, Shichao
    Zhou, Chuan
    Pan, Shirui
    Zhu, Xingquan
    Wang, Bin
    [J]. 2019 19TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2019), 2019, : 1534 - 1539
  • [42] A Structure-aware Despeckling Method of SAR Image
    Jin, Xin
    Wang, Xiaotong
    Xu, Xiaogang
    Yi, Chengtao
    [J]. FOURTH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND INFORMATION SCIENCES (CCAIS 2015), 2015, : 257 - 261
  • [43] Structure-Aware Image Resizing for Chinese Characters
    Liu, Chengdong
    Lian, Zhouhui
    Tang, Yingmin
    Xiao, Jianguo
    [J]. MULTIMEDIA MODELING (MMM 2017), PT I, 2017, 10132 : 379 - 390
  • [45] Generative Adversarial Networks for Hyperspectral Image Classification
    Zhu, Lin
    Chen, Yushi
    Ghamisi, Pedram
    Benediktsson, Jon Atli
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2018, 56 (09): : 5046 - 5063
  • [46] MGCET: MLP-mixer and Graph Convolutional Enhanced Transformer for Hyperspectral Image Classification
    Al-qaness, Mohammed A. A.
    Wu, Guoyong
    AL-Alimi, Dalal
    [J]. REMOTE SENSING, 2024, 16 (16)
  • [47] Tree Structure-Aware Few-Shot Image Classification via Hierarchical Aggregation
    Zhang, Min
    Huang, Siteng
    Li, Wenbin
    Wang, Donglin
    [J]. COMPUTER VISION, ECCV 2022, PT XX, 2022, 13680 : 453 - 470
  • [48] Retrofitting Structure-aware Transformer Language Model for End Tasks
    Fei, Hao
    Ren, Yafeng
    Ji, Donghong
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 2151 - 2161
  • [49] Structure-Aware Cross-Modal Transformer for Depth Completion
    Zhao, Linqing
    Wei, Yi
    Li, Jiaxin
    Zhou, Jie
    Lu, Jiwen
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 1016 - 1031
  • [50] Context-Aware Attentional Graph U-Net for Hyperspectral Image Classification
    Lin, Moule
    Jing, Weipeng
    Di, Donglin
    Chen, Guangsheng
    Song, Houbing
    [J]. IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2022, 19