Semantic-Guided Transformer Network for Crop Classification in Hyperspectral Images

被引:0
|
作者
Pi, Weiqiang [1 ]
Zhang, Tao [2 ]
Wang, Rongyang [1 ]
Ma, Guowei [1 ]
Wang, Yong [1 ]
Du, Jianmin [2 ]
机构
[1] Huzhou Vocat & Tech Coll, Coll Intelligent Mfg & Elevator, Huzhou 313099, Peoples R China
[2] Inner Mongolia Agr Univ, Coll Mech & Elect Engn, Hohhot 010018, Peoples R China
基金
中国国家自然科学基金;
关键词
hyperspectral image classification; transformer; deep learning; attention mechanism; convolutional neural network; ATTENTION NETWORK;
D O I
10.3390/jimaging11020037
中图分类号
TB8 [摄影技术];
学科分类号
0804 ;
摘要
The hyperspectral remote sensing images of agricultural crops contain rich spectral information, which can provide important details about crop growth status, diseases, and pests. However, existing crop classification methods face several key limitations when processing hyperspectral remote sensing images, primarily in the following aspects. First, the complex background in the images. Various elements in the background may have similar spectral characteristics to the crops, and this spectral similarity makes the classification model susceptible to background interference, thus reducing classification accuracy. Second, the differences in crop scales increase the difficulty of feature extraction. In different image regions, the scale of crops can vary significantly, and traditional classification methods often struggle to effectively capture this information. Additionally, due to the limitations of spectral information, especially under multi-scale variation backgrounds, the extraction of crop information becomes even more challenging, leading to instability in the classification results. To address these issues, a semantic-guided transformer network (SGTN) is proposed, which aims to effectively overcome the limitations of these deep learning methods and improve crop classification accuracy and robustness. First, a multi-scale spatial-spectral information extraction (MSIE) module is designed that effectively handle the variations of crops at different scales in the image, thereby extracting richer and more accurate features, and reducing the impact of scale changes. Second, a semantic-guided attention (SGA) module is proposed, which enhances the model's sensitivity to crop semantic information, further reducing background interference and improving the accuracy of crop area recognition. By combining the MSIE and SGA modules, the SGTN can focus on the semantic features of crops at multiple scales, thus generating more accurate classification results. Finally, a two-stage feature extraction structure is employed to further optimize the extraction of crop semantic features and enhance classification accuracy. The results show that on the Indian Pines, Pavia University, and Salinas benchmark datasets, the overall accuracies of the proposed model are 98.24%, 98.34%, and 97.89%, respectively. Compared with other methods, the model achieves better classification accuracy and generalization performance. In the future, the SGTN is expected to be applied to more agricultural remote sensing tasks, such as crop disease detection and yield prediction, providing more reliable technical support for precision agriculture and agricultural monitoring.
引用
收藏
页数:17
相关论文
共 50 条
  • [1] SGNet: A Transformer-Based Semantic-Guided Network for Building Change Detection
    Feng, Jiangfan
    Yang, Xinyu
    Gu, Zhujun
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2024, 17 : 9922 - 9935
  • [2] Contrastive Semantic-Guided Image Smoothing Network
    Wang, Jie
    Wang, Yongzhen
    Feng, Yidan
    Gong, Lina
    Yan, Xuefeng
    Xie, Haoran
    Wang, Fu Lee
    Wei, Mingqiang
    COMPUTER GRAPHICS FORUM, 2022, 41 (07) : 335 - 346
  • [3] Global semantic-guided network for saliency prediction
    Xie, Jiawei
    Liu, Zhi
    Li, Gongyang
    Lu, Xiaofeng
    Chen, Tao
    KNOWLEDGE-BASED SYSTEMS, 2024, 284
  • [4] Matching Thermal to Visible Face Images Using a Semantic-Guided Generative Adversarial Network
    Chen, Cunjian
    Ross, Arun
    2019 14TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG 2019), 2019, : 487 - 494
  • [5] Progressive Semantic-Guided Network for the Extraction of Raft Aquaculture Areas From Remote Sensing Images
    Lu, Yan
    Guo, Baotao
    Li, Haojie
    Cui, Binge
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21
  • [6] Semantic-Guided Attention Refinement Network for Salient Object Detection in Optical Remote Sensing Images
    Huang, Zhou
    Chen, Huaixin
    Liu, Biyuan
    Wang, Zhixi
    REMOTE SENSING, 2021, 13 (11)
  • [7] A Sub-captions Semantic-Guided Network for Image Captioning
    Tian, Wei-Dong
    Zhu, Jun-jun
    Wu, Shuang
    Zhao, Zhong-Qiu
    Zhang, Yu-Zheng
    Zhang, Tian-yu
    INTELLIGENT COMPUTING METHODOLOGIES, PT III, 2022, 13395 : 367 - 379
  • [8] Semantic Segmentation Network for Classification of Hyperspectral Images With Small Size Samples
    Ma, Li
    Li, Shuyue
    Zhou, Zhiyong
    Yao, Yafeng
    Du, Qian
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2023, 20
  • [9] SGFNet: Semantic-Guided Fusion Network for RGB-Thermal Semantic Segmentation
    WangLi, Yike
    Li, Gongyang
    Liu, Zhi
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (12) : 7737 - 7748
  • [10] SgLFT: Semantic-guided Late Fusion Transformer for video corpus moment retrieval
    Chen, Tongbao
    Wang, Wenmin
    Zhao, Minglu
    Li, Ruochen
    Jiang, Zhe
    Yu, Cheng
    NEUROCOMPUTING, 2024, 599