Self-Supervised Learning for 3-D Point Clouds Based on a Masked Linear Autoencoder

被引:0
|
作者
Yang, Hongxin [1 ]
Wang, Ruisheng [1 ,2 ]
机构
[1] Univ Calgary, Dept Geomat Engn, Calgary, AB T2N 1N4, Canada
[2] Shenzhen Univ, Sch Architecture & Urban Planning, Shenzhen 518000, Guangdong, Peoples R China
基金
美国国家科学基金会; 加拿大自然科学与工程研究理事会;
关键词
Transformers; Three-dimensional displays; Point cloud compression; Task analysis; Data models; Standards; Memory management; Point cloud; self-attention mechanism; self-supervised learning g(SSL); transformer; NETWORK;
D O I
10.1109/TGRS.2023.3337088
中图分类号
P3 [地球物理学]; P59 [地球化学];
学科分类号
0708 ; 070902 ;
摘要
Motivated by the success of a masked autoencoder in 3-D point-cloud-based learning, this study proposes an innovative framework for self-supervised learning (SSL) on 3-D point clouds with linear complexity. In the proposed framework, every input point cloud is divided into multiple point patches, which are randomly masked at different ratios. Then, unmasked point patches are then fed to an improved transformer model, which uses an advanced linear self-attention mechanism autoencoder to learn high-level features. The pretraining objective is to recover the masked patches under the guidance of the unmasked point patches' features obtained by the designed transformer. Furthermore, a linear self-attention mechanism is designed to use three projection matrices to decompose the original scaled dot-product attention into smaller parts, using the properties of low-rank and linear decomposition to reduce the time complexity from quadratic to linear. The results of extensive experiments demonstrate that the proposed pretrained model can achieve high accuracy of 93.6% and 84.77% on the ModelNet40 and ScanObjectNN datasets, respectively, at a masking ratio of 40%. In addition, the results show that the proposed method, which uses a linear self-attention mechanism, can enhance the computational efficiency by significantly reducing the inference time and minimizing the storage memory requirements for query, key, and value (Q, K, and V) matrices compared with the existing methods. Finally, the results indicate that the proposed method can achieve state-of-the-art performance on the classification ModelNet40 dataset.
引用
收藏
页码:1 / 11
页数:11
相关论文
共 50 条
  • [21] A self-supervised learning framework based on masked autoencoder for complex wafer bin map classification
    Wang, Yi
    Ni, Dong
    Huang, Zhenyu
    Chen, Puyang
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 249
  • [22] PointUR-RL: Unified Self-Supervised Learning Method Based on Variable Masked Autoencoder for Point Cloud Reconstruction and Representation Learning
    Li, Kang
    Zhu, Qiuquan
    Wang, Haoyu
    Wang, Shibo
    Tian, He
    Zhou, Ping
    Cao, Xin
    REMOTE SENSING, 2024, 16 (16)
  • [23] Point-AGM : Attention Guided Masked Auto-Encoder for Joint Self-supervised Learning on Point Clouds
    Liu, Jie
    Yang, Mengna
    Tian, Yu
    Li, Yancui
    Song, Da
    Li, Kang
    Cao, Xin
    COMPUTER GRAPHICS FORUM, 2024, 43 (07)
  • [24] Self-Supervised Learning of Point Clouds via Orientation Estimation
    Poursaeed, Omid
    Jiang, Tianxing
    Qiao, Han
    Xu, Nayun
    Kim, Vladimir G.
    2020 INTERNATIONAL CONFERENCE ON 3D VISION (3DV 2020), 2020, : 1018 - 1028
  • [25] Self-Supervised Deep Learning on Point Clouds by Reconstructing Space
    Sauder, Jonathan
    Sievers, Bjarne
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [26] Self-Supervised Few-Shot Learning on Point Clouds
    Sharma, Charu
    Kaul, Manohar
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [27] 3D-OAE: Occlusion Auto-Encoders for Self-Supervised Learning on Point Clouds
    Zhou, Junsheng
    Wen, Xin
    Ma, Baorui
    Liu, Yu-Shen
    Gao, Yue
    Fang, Yi
    Han, Zhizhong
    2024 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2024), 2024, : 15416 - 15423
  • [28] Self-Supervised Learning for Point-Cloud Classification by a Multigrid Autoencoder
    Zhai, Ruifeng
    Song, Junfeng
    Hou, Shuzhao
    Gao, Fengli
    Li, Xueyan
    SENSORS, 2022, 22 (21)
  • [29] Implicit Autoencoder for Point-Cloud Self-Supervised Representation Learning
    Yan, Siming
    Yang, Zhenpei
    Li, Haoxiang
    Song, Chen
    Guan, Li
    Kang, Hao
    Hua, Gang
    Huang, Qixing
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 14484 - 14496
  • [30] Universal Sound Separation with Self-Supervised Audio Masked Autoencoder
    Zhao, Junqi
    Liu, Xubo
    Zhao, Jinzheng
    Yuan, Yi
    Kong, Qiuqiang
    Plumbley, Mark D.
    Wang, Wenwu
    32ND EUROPEAN SIGNAL PROCESSING CONFERENCE, EUSIPCO 2024, 2024, : 1 - 5