Semantic Segmentation Model for Remote Sensing Images Combining Super Resolution and Domain Adaption

被引:0
|
作者
Liang M. [1 ]
Wang X.-L. [1 ]
机构
[1] School of Computer Science, Shaanxi Normal University, Xi'an
来源
关键词
Adversarial learning; Domain adaptation; Remote sensing image; Semantic segmentation; Super resolution;
D O I
10.11897/SP.J.1016.2022.02619
中图分类号
学科分类号
摘要
The semantic segmentation method based on convolutional neural network rely on supervised learning with ground truth, but it cannot be well generalized to unlabeled datasets with different sources. Unsupervised domain adaptation can solve the problem of inconsistent feature distribution between unlabeled target domain data and labeled source domain data. This is due to that remote sensing images are often come from different sources, they are variable in their spatial resolution and are influenced by different imaging regions, imaging conditions and imaging times. Even images from the same region may have large differences in spectral features. The generalization of the semantic segmentation model relies on the reduction of these inter-domain differences mentioned above. Therefore, unsupervised domain adaptation methods for remote sensing image should not only reduce the differences in features between domains, but also address the problem of different spatial resolutions. This paper designs a new end-to-end semantic segmentation deep network combined with image super resolution-Semantic Segmentation Model Combining Super Resolution and Domain Adaption, which can reduce the spatial resolution difference and feature distribution difference between the low spatial resolution source domain and high spatial resolution target domain remote sensing images, and accomplish the super-resolution task for the source domain and the domain adaptation semantic segmentation task for the target domain. The SSM-SRDA model consists of three parts one is Semantic Segmentation Network with Super Resolution (SSNSR), the second is Pixel-level Domain Discriminator (PDD), and the third is Output-space Domain Discriminator (ODD). SSNSR consists of a semantic segmentation network and a super-resolution network, which share the same feature extraction network. The super-resolution network generates a high-resolution synthetic image with target image style from a low-resolution source domain image, which can eliminate spatial resolution differences and style differences to help the feature extraction module learn the same features between the source and target domains. The Feature Affinity-Loss module enhances the learning of the semantic segmentation deep network by the feature maps with more detailed structural information obtained by super-resolution. The pixel-level domain discriminator is used to reduce the pixel-level feature differences between the high-resolution target domain and the synthetic image of the source domain. High-resolution source domain synthetic images with target domain style are generated by generative adversarial learning with the super-resolution network and participate in the training of the model as additional training data. The output-space domain discriminator reduces the feature differences between source and target domain images in the output space of the semantic segmentation network. Through the adversarial learning of the semantic segmentation network and the two discriminators, SSM-SRDA aligns the feature distribution of the source domain and the target domain at the input and output stages of the segmentation network, and can be applied to the target domain datas of more different sources through domain adaptation. It is a practical and more popular model. Experiments show that SSM-SRDA is superior to the existing domain adaptive semantic segmentation methods on four sets of remote sensing image data sets, and the intersection ratio is improved by 0.7%, 1.7%, 2.2% and 3.3%, respectively. © 2022, Science Press. All right reserved.
引用
收藏
页码:2619 / 2636
页数:17
相关论文
共 46 条
  • [1] Zheng C, Zhang Y, Wang L., Semantic segmentation of remote sensing imagery using an object-based Markov random field model with auxiliary label fields, IEEE Transactions on Geoscience and Remote Sensing, 55, 5, pp. 3015-3028, (2017)
  • [2] Pan B, Shi Z, Xu X, Et al., CoinNet: Copy initialization network for multispectral imagery semantic segmentation, IEEE Geoscience and Remote Sensing Letters, 16, 5, pp. 816-820, (2019)
  • [3] Mou L, Hua Y, Zhu X X., Relation matters: Relational context-aware fully convolutional network for semantic segmentation of high-resolution aerial images, IEEE Transactions on Geoscience and Remote Sensing, 58, 11, pp. 7557-7569, (2020)
  • [4] Alam M, Wang J F, Guangpei C, Et al., Convolutional neural network for the semantic segmentation of remote sensing images, Mobile Networks and Applications, 26, 1, pp. 200-215, (2021)
  • [5] Wu Z, Wang X, Gonzalez J E, Et al., ACE: Adapting to changing environments for semantic segmentation, Proceedings of the IEEE International Conference on Computer Vision, pp. 2121-2130, (2019)
  • [6] Zhang Y, Qiu Z, Yao T, Et al., Fully convolutional adaptation networks for semantic segmentation, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6810-6818, (2018)
  • [7] Li Y, Yuan L, Vasconcelos N., Bidirectional learning for domain adaptation of semantic segmentation, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6936-6945, (2019)
  • [8] Tsai Y H, Hung W C, Schulter S, Et al., Learning to adapt structured output space for semantic segmentation, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7472-7481, (2018)
  • [9] Tasar O, Happy S L, Tarabalka Y, Alliez P., ColorMapGAN: Unsupervised domain adaptation for semantic segmentation using color mapping generative adversarial networks, IEEE Transactions on Geoscience and Remote Sensing, 58, 10, pp. 7178-7193, (2020)
  • [10] Zhang Z, Doi K, Iwasaki A, Xu G., Unsupervised domain adaptation of high-resolution aerial images via correlation alignment and self-training, IEEE Geoscience and Remote Sensing Letters, 18, 4, pp. 746-750, (2020)