Self-Supervised Learning for 3-D Point Clouds Based on a Masked Linear Autoencoder

被引:0
|
作者
Yang, Hongxin [1 ]
Wang, Ruisheng [1 ,2 ]
机构
[1] Univ Calgary, Dept Geomat Engn, Calgary, AB T2N 1N4, Canada
[2] Shenzhen Univ, Sch Architecture & Urban Planning, Shenzhen 518000, Guangdong, Peoples R China
基金
美国国家科学基金会; 加拿大自然科学与工程研究理事会;
关键词
Transformers; Three-dimensional displays; Point cloud compression; Task analysis; Data models; Standards; Memory management; Point cloud; self-attention mechanism; self-supervised learning g(SSL); transformer; NETWORK;
D O I
10.1109/TGRS.2023.3337088
中图分类号
P3 [地球物理学]; P59 [地球化学];
学科分类号
0708 ; 070902 ;
摘要
Motivated by the success of a masked autoencoder in 3-D point-cloud-based learning, this study proposes an innovative framework for self-supervised learning (SSL) on 3-D point clouds with linear complexity. In the proposed framework, every input point cloud is divided into multiple point patches, which are randomly masked at different ratios. Then, unmasked point patches are then fed to an improved transformer model, which uses an advanced linear self-attention mechanism autoencoder to learn high-level features. The pretraining objective is to recover the masked patches under the guidance of the unmasked point patches' features obtained by the designed transformer. Furthermore, a linear self-attention mechanism is designed to use three projection matrices to decompose the original scaled dot-product attention into smaller parts, using the properties of low-rank and linear decomposition to reduce the time complexity from quadratic to linear. The results of extensive experiments demonstrate that the proposed pretrained model can achieve high accuracy of 93.6% and 84.77% on the ModelNet40 and ScanObjectNN datasets, respectively, at a masking ratio of 40%. In addition, the results show that the proposed method, which uses a linear self-attention mechanism, can enhance the computational efficiency by significantly reducing the inference time and minimizing the storage memory requirements for query, key, and value (Q, K, and V) matrices compared with the existing methods. Finally, the results indicate that the proposed method can achieve state-of-the-art performance on the classification ModelNet40 dataset.
引用
收藏
页码:1 / 11
页数:11
相关论文
共 50 条
  • [41] Occlusion Guided Self-supervised Scene Flow Estimation on 3D Point Clouds
    Ouyang, Bojun
    Raviv, Dan
    2021 INTERNATIONAL CONFERENCE ON 3D VISION (3DV 2021), 2021, : 782 - 791
  • [42] Exploring Self-Supervised Learning for 3D Point Cloud Registration
    Yuan, Mingzhi
    Huang, Qiao
    Shen, Ao
    Huang, Xiaoshui
    Wang, Manning
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2025, 10 (01): : 25 - 31
  • [43] Self-Supervised Learning-Based Time Series Classification via Hierarchical Sparse Convolutional Masked-Autoencoder
    Yu, Ting
    Xu, Kele
    Wang, Xu
    Ding, Bo
    Feng, Dawei
    IEEE OPEN JOURNAL OF SIGNAL PROCESSING, 2024, 5 : 964 - 975
  • [44] Joint data and feature augmentation for self-supervised representation learning on point clouds
    Lu, Zhuheng
    Dai, Yuewei
    Li, Weiqing
    Su, Zhiyong
    GRAPHICAL MODELS, 2023, 129
  • [45] Mixed Autoencoder for Self-supervised Visual Representation Learning
    Chen, Kai
    Liu, Zhili
    Hong, Lanqing
    Xu, Hang
    Li, Zhenguo
    Yeung, Dit-Yan
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 22742 - 22751
  • [46] Self-supervised Discriminative Representation Learning by Fuzzy Autoencoder
    Yang, Wenlu
    Wang, Hongjun
    Zhang, Yinghui
    Liu, Zehao
    Li, Tianrui
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2023, 14 (01)
  • [47] SelfSwapper: Self-supervised Face Swapping via Shape Agnostic Masked AutoEncoder
    Lee, Jaeseong
    Hyung, Junha
    Jung, Sohyun
    Choo, Jaegul
    COMPUTER VISION - ECCV 2024, PT LV, 2025, 15113 : 383 - 400
  • [48] Masked autoencoder: influence of self-supervised pretraining on object segmentation in industrial images
    Anja Witte
    Sascha Lange
    Christian Lins
    Industrial Artificial Intelligence, 2 (1):
  • [49] Encoding 3D Point Contexts for Self-Supervised Spall Classification Using 3D Bridge Point Clouds
    Kasireddy, Varun
    Akinci, Burcu
    JOURNAL OF COMPUTING IN CIVIL ENGINEERING, 2023, 37 (02)
  • [50] Salt3DNet: A Self-Supervised Learning Framework for 3-D Salt Segmentation
    Yang, Liuqing
    Fomel, Sergey
    Wang, Shoudong
    Chen, Xiaohong
    Saad, Omar M.
    Chen, Yangkang
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62 : 1 - 15