Rotation-Invariant Transformer for Point Cloud Matching

被引:43
|
作者
Yu, Hao [1 ]
Qin, Zheng [2 ]
Hou, Ji [3 ]
Saleh, Mahdi [1 ,4 ]
Li, Dongsheng [2 ]
Busam, Benjamin [1 ,4 ]
Ilic, Slobodan [1 ,5 ]
机构
[1] Tech Univ Munich, Munich, Germany
[2] NUDT, Changsha, Peoples R China
[3] Meta Real Labs, Menlo Pk, CA USA
[4] 3Dwe Ai, Munich, Germany
[5] Siemens AG, Munich, Germany
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52729.2023.00521
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The intrinsic rotation invariance lies at the core of matching point clouds with handcrafted descriptors. However, it is widely despised by recent deep matchers that obtain the rotation invariance extrinsically via data augmentation. As the finite number of augmented rotations can never span the continuous SO(3) space, these methods usually show instability when facing rotations that are rarely seen. To this end, we introduce RoITr, a Rotation-Invariant Transformer to cope with the pose variations in the point cloud matching task. We contribute both on the local and global levels. Starting from the local level, we introduce an attention mechanism embedded with Point Pair Feature (PPF)-based coordinates to describe the pose-invariant geometry, upon which a novel attention-based encoder-decoder architecture is constructed. We further propose a global transformer with rotation-invariant cross-frame spatial awareness learned by the self-attention mechanism, which significantly improves the feature distinctiveness and makes the model robust with respect to the low overlap. Experiments are conducted on both the rigid and non-rigid public benchmarks, where RoITr outperforms all the state-of-the-art models by a considerable margin in the low-overlapping scenarios. Especially when the rotations are enlarged on the challenging 3DLoMatch benchmark, RoITr surpasses the existing methods by at least 13 and 5 percentage points in terms of Inlier Ratio and Registration Recall, respectively. Code is publicly available (1).
引用
收藏
页码:5384 / 5393
页数:10
相关论文
共 50 条
  • [21] SAR image matching based on rotation-invariant description
    Chang, Yunhao
    Xu, Qing
    Xiong, Xin
    Jin, Guowang
    Hou, Huitai
    Man, Dan
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [22] SGMNet: Learning Rotation-Invariant Point Cloud Representations via Sorted Gram Matrix
    Xu, Jianyun
    Tang, Xin
    Zhu, Yushi
    Sun, Jie
    Pu, Shiliang
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 10448 - 10457
  • [23] SAR image matching based on rotation-invariant description
    Yunhao Chang
    Qing Xu
    Xin Xiong
    Guowang Jin
    Huitai Hou
    Dan Man
    Scientific Reports, 13
  • [24] Rotation-invariant pattern matching using wavelet decomposition
    Tsai, DM
    Chiang, CH
    PATTERN RECOGNITION LETTERS, 2002, 23 (1-3) : 191 - 201
  • [25] Rotation-invariant Hierarchical Segmentation on Poincare Ball for 3D Point Cloud
    Onghena, Pierre
    Gigli, Leonardo
    Velasco-Forero, Santiago
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 1757 - 1766
  • [26] Rethinking local-to-global representation learning for rotation-invariant point cloud analysis
    Wang, Zhaoxuan
    Yu, Yunlong
    Li, Xianzhi
    PATTERN RECOGNITION, 2024, 154
  • [27] RITNet: A Rotation Invariant Transformer based Network for Point Cloud Registration
    Yang, Min
    Li, Yaochen
    Wang, Su
    Yang, Shaohan
    Liu, Hujun
    2022 IEEE 34TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, ICTAI, 2022, : 616 - 621
  • [28] Using orientation codes for rotation-invariant template matching
    Ullah, F
    Kaneko, S
    PATTERN RECOGNITION, 2004, 37 (02) : 201 - 209
  • [29] Rotation-invariant fingerprint matching using radon and DCT
    Sangita Bharkad
    Manesh Kokare
    Sādhanā, 2017, 42 : 2025 - 2039
  • [30] Multispectral Image Matching Using Rotation-Invariant Distance
    Li, Qiaoliang
    Zhang, Huisheng
    Wang, Tianfu
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2011, 8 (03) : 406 - 410