Weakly supervised scale adaptation data augmentation for scene classification of high-resolution remote sensing images

被引:0
|
作者
Wang L. [1 ]
Qi K. [2 ]
Yang C. [1 ,2 ]
Wu H. [1 ]
机构
[1] School of Geography and Information Engineering, China University of Geosciences (Wuhan), Wuhan
[2] National Engineering Research Center of Geographic Information System, Wuhan
[3] State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan
关键词
convolutional neural networks; data augmentation; deep learning; multi-scale; remote sensing; scene classification; weakly supervision;
D O I
10.11834/jrs.20221481
中图分类号
学科分类号
摘要
Scene classification of remote sensing images aims to assign a meaningful label to a given image. In recent years, Convolutional Neural Networks (CNNs)-based methods make a breakthrough and substantially outperform traditional methods in scene classification tasks of remote sensing images. However, obtaining features under different scales in remote sensing images is difficult due to the fixed receptive field of CNNs. This complexity seriously affects the performance of CNNs in scene classification of remote sensing images. This study proposes a method to learn the optimal scales for different scene image instances in a weakly supervised manner. A Weakly Supervised Scale Adaptive Data Augmentation Network (WSADAN) is proposed to capture feature information at different scales of remote sensing scenes, and a scale generation module and a scale fusion module are designed to improve the robustness. The scale generation module learns the optimal scale parameters based on the CNN features of the original image. The scale fusion module filters the CNN features of images with original and optimal scales to remove the noise and then deeply fuses them to exploit the correlation between features at different scales. The deeply fused multi-scale features are input into a fully connected layer to predict categories of scene images. The effectiveness of the scale generation and scale fusion modules is verified by ablation experiments. The accuracy of WSADANSGM compared with the baseline improves by 0.94% and 0.89% for the 20% and 50% training data ratios of RSSCN7 dataset, 1.27% and 0.87% for the 20% and 50% training data ratios of AID dataset, and 1.09% and 0.71% for the 10% and 20% training data ratios of NWPU dataset, respectively. Compared with WSADANSGM, WSADANSGM+SFM improves by 1.65% and 1.32% for the RSSCN7 dataset at 20% and 50% training data ratios, 1.65% and 1.26% for the AID dataset at 20% and 50% training data ratios, and 1.75% and 1.42% for the NWPU dataset at 10% and 20% training data ratios, respectively. In the experiment for scene scale change analysis, the classification accuracy of our method is higher than the baseline at any scale of image, which proves that our method can learn certain image scale information and has strong scale adaptation ability. We use three datasets for remote sensing scene classification, namely, RSSCN7, AID, and NWPU, for the experiments. On the RSSCN7 dataset, the overall accuracies are 91.65% and 94.07% with the training ratios of 20% and 50% for WSADANVGG16. For WSADAN-ResNet50, the corresponding accuracies are 92.69% and 94.82%. On the AID dataset, the overall accuracies are 92.78% and 95.18% with the training ratios of 20% and 50% for WSADAN-VGG16. For WSADAN-ResNet50, the corresponding accuracies are 93.73% and 95.88%. On the NWPU dataset, the overall accuracies are 87.01% and 90.44% with the training ratios of 10% and 20% for WSADAN-VGG16. For WSADAN-ResNet50, the corresponding accuracies are 90.71% and 92.63%. The proposed method can learn CNN features at a wider range of scales without manual multi-scale selection for different datasets. The performance of the proposed method is better than that of traditional CNNs, especially for the scene categories containing objects with large-scale variations. © 2023 Science Press. All rights reserved.
引用
收藏
页码:2815 / 2830
页数:15
相关论文
共 45 条
  • [1] Alhichri H, Alajlan N, Bazi Y, Rabczuk T., Multi-scale convolutional neural network for remote sensing scene classification, Proceedings of the 2018 IEEE International Conference on Electro/Information Technology (EIT), pp. 1-5, (2018)
  • [2] Cao R, Fang L Y, Lu T, He N J., Self-attention-based deep feature fusion for remote sensing scene classification, IEEE Geoscience and Remote Sensing Letters, 18, 1, pp. 43-47, (2021)
  • [3] Chen L J, Yang W, Xu K, Xu T., Evaluation of local features for scene classification using VHR satellite images, Proceedings of 2011 Joint Urban Remote Sensing Event, pp. 385-388, (2011)
  • [4] Chen W T, Li X J, He H X, Wang L Z., Assessing different feature sets’effects on land cover classification in complex surface-mined landscapes by ZiYuan-3 satellite imagery, Remote Sensing, 10, 1, (2018)
  • [5] Chen Y Y, Ming D P, Lv X W., Superpixel based land cover classification of VHR satellite image combining multi-scale CNN and scale parameter estimation, Earth Science Informatics, 12, 3, pp. 341-363, (2019)
  • [6] Cheng G, Guo L, Zhao T Y, Han J W, Li H H, Fang J., Automatic landslide detection from remote-sensing imagery using a scene classification method based on BoVW and pLSA, International Journal of Remote Sensing, 34, 1, pp. 45-59, (2013)
  • [7] Cheng G, Han J W, Lu X Q., Remote sensing image scene classification: benchmark and state of the art, Proceedings of the IEEE, 105, 10, pp. 1865-1883, (2017)
  • [8] Gao Y, Shi J, Li J, Wang R Y., Remote sensing scene classification with dual attention-aware network, Proceedings of the IEEE 5th International Conference on Image, Vision and Computing (ICIVC), pp. 171-175, (2020)
  • [9] Gu Y T, Wang Y T, Li Y S., A survey on deep learning-driven remote sensing image scene understanding: scene classification, scene retrieval and scene-guided object detection, Applied Sciences, 9, 10, (2019)
  • [10] Han X B, Zhong Y F, Cao L Q, Zhang L P., Pre-trained AlexNet architecture with pyramid pooling and supervision for high spatial resolution remote sensing image scene classification, Remote Sensing, 9, 8, (2017)