Oriented Object Detection for Remote Sensing Images via Object-Wise Rotation-Invariant Semantic Representation

被引:1
|
作者
Zheng, Shangdong [1 ]
Wu, Zebin [2 ]
Du, Qian [3 ]
Xu, Yang [2 ,4 ]
Wei, Zhihui [2 ]
机构
[1] Nanjing Univ Posts & Telecommun, Sch Comp Sci, Nanjing 210023, Peoples R China
[2] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Nanjing 210094, Peoples R China
[3] Mississippi State Univ, Dept Elect & Comp Engn, Starkville, MS 39762 USA
[4] Nanjing Univ Sci & Technol, Jiangsu Key Lab Spectral Imaging & Intelligent Sen, Nanjing 210094, Peoples R China
基金
中国国家自然科学基金;
关键词
Image segmentation; Feature extraction; Task analysis; Semantics; Object detection; Remote sensing; Proposals; Oriented object detection (OOD); remote sensing images (RSIs); rotation-invariant learning; semantic representation;
D O I
10.1109/TGRS.2024.3402825
中图分类号
P3 [地球物理学]; P59 [地球化学];
学科分类号
0708 ; 070902 ;
摘要
Oriented object detection (OOD) in remote sensing images (RSIs) remains a challenging work due to an arbitrary orientation of instances. Learning rotation-invariant features is critical in modeling a fixed descriptor for instances with its rotated variants. However, most existing methods construct the descriptor from the perspectives of data or feature augmentation, but ignore the exploration of potentially useful supervision information inside the detection algorithm. In this article, we propose an object-wise rotation-invariant semantic representation (ORSR) framework, which synergizes the exploration of latent supervision, rotation-invariant learning, and guided attention mechanism into a unified network to boost the performance of OOD in RSIs. First, supervised by our constructed pseudo-ground truth of segmentation masks, a semantic segmentation branch is built along with the detection algorithm to refine the representation of backbone features. Moreover, a consistency loss function is proposed to encourage the segmentation branch to make fixed predictions for backbone features with its rotated variants. Considering that segmentation predictions remain the same affine transformations before and after rotating, we further construct a Kullback-Leibler (KL) divergence-based similarity loss function to encourage the network to model the rotation-invariant features. Finally, we separate the "object" descriptor from the segmentation predictions to extend the implicit constraint in our proposed semantic segmentation branch. The separated "object" descriptor not only involves the spatial regularizer to emphasize the high-responsive regions in the image but also can be guided by the constructed consistency loss function. We evaluate our proposed ORSR on the challenging DOTA, DIOR-R, and HRSC2016 datasets. Extensive experiments demonstrate that the proposed ORSR achieves competitive performance compared to other single-scale and multiscale (MS) detection methods.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] ROTATION-INVARIANT LATENT SEMANTIC REPRESENTATION LEARNING FOR OBJECT DETECTION IN VHR OPTICAL REMOTE SENSING IMAGES
    Yao, Xiwen
    Feng, Xiaoxu
    Cheng, Gong
    Han, Junwei
    Guo, Lei
    [J]. 2019 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM (IGARSS 2019), 2019, : 1382 - 1385
  • [2] Object detection in VHR optical remote sensing images via learning rotation-invariant HOG feature
    Cheng, Gong
    Zhou, Peicheng
    Yao, Xiwen
    Yao, Chao
    Zhang, Yanbang
    Han, Junwei
    [J]. 2016 4RTH INTERNATIONAL WORKSHOP ON EARTH OBSERVATION AND REMOTE SENSING APPLICATIONS (EORSA), 2016,
  • [3] Learning rotation-invariant binary codes for efficient object detection from remote sensing images
    Liu, Yazhou
    Xu, Hui
    Sun, Quansen
    [J]. JOURNAL OF APPLIED REMOTE SENSING, 2019, 13 (03)
  • [4] Rotation-Invariant Object Detection in Remote Sensing Images Based on Radial-Gradient Angle
    Lin, Yudong
    He, Hongjie
    Yin, Zhongke
    Chen, Fan
    [J]. IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2015, 12 (04) : 746 - 750
  • [5] Object Classification of Remote Sensing Images Based on Rotation-Invariant Discrete Hashing
    Xu, Hui
    Liu, Yazhou
    Sun, Quansen
    [J]. ADVANCES IN MULTIMEDIA INFORMATION PROCESSING - PCM 2017, PT II, 2018, 10736 : 264 - 274
  • [6] Rapid Object Detection in VHR Optical Remote Sensing Images Based on Rotation-Invariant Discrete Hashing
    Xu, Hui
    Liu, Yazhou
    Sun, Quansen
    [J]. PROCEEDINGS 2017 4TH IAPR ASIAN CONFERENCE ON PATTERN RECOGNITION (ACPR), 2017, : 91 - 96
  • [7] Learning Rotation-Invariant Convolutional Neural Networks for Object Detection in VHR Optical Remote Sensing Images
    Cheng, Gong
    Zhou, Peicheng
    Han, Junwei
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2016, 54 (12): : 7405 - 7415
  • [8] Rotation-Invariant Feature Learning for Object Detection in VHR Optical Remote Sensing Images by Double-Net
    Zhang, Zhi
    Jiang, Ruoqiao
    Mei, Shaohui
    Zhang, Shun
    Zhang, Yifan
    [J]. IEEE ACCESS, 2020, 8 : 20818 - 20827
  • [9] RiDOP: A Rotation-Invariant Detector with Simple Oriented Proposals in Remote Sensing Images
    Wei, Chongyang
    Ni, Weiping
    Qin, Yao
    Wu, Junzheng
    Zhang, Han
    Liu, Qiang
    Cheng, Kenan
    Bian, Hui
    [J]. REMOTE SENSING, 2023, 15 (03)
  • [10] Rotation-Invariant and Relation-Aware Cross-Domain Adaptation Object Detection Network for Optical Remote Sensing Images
    Chen, Ying
    Liu, Qi
    Wang, Teng
    Wang, Bin
    Meng, Xiaoliang
    [J]. REMOTE SENSING, 2021, 13 (21)