Self-Supervised Learning for 3-D Point Clouds Based on a Masked Linear Autoencoder

被引:0
|
作者
Yang, Hongxin [1 ]
Wang, Ruisheng [1 ,2 ]
机构
[1] Univ Calgary, Dept Geomat Engn, Calgary, AB T2N 1N4, Canada
[2] Shenzhen Univ, Sch Architecture & Urban Planning, Shenzhen 518000, Guangdong, Peoples R China
基金
美国国家科学基金会; 加拿大自然科学与工程研究理事会;
关键词
Transformers; Three-dimensional displays; Point cloud compression; Task analysis; Data models; Standards; Memory management; Point cloud; self-attention mechanism; self-supervised learning g(SSL); transformer; NETWORK;
D O I
10.1109/TGRS.2023.3337088
中图分类号
P3 [地球物理学]; P59 [地球化学];
学科分类号
0708 ; 070902 ;
摘要
Motivated by the success of a masked autoencoder in 3-D point-cloud-based learning, this study proposes an innovative framework for self-supervised learning (SSL) on 3-D point clouds with linear complexity. In the proposed framework, every input point cloud is divided into multiple point patches, which are randomly masked at different ratios. Then, unmasked point patches are then fed to an improved transformer model, which uses an advanced linear self-attention mechanism autoencoder to learn high-level features. The pretraining objective is to recover the masked patches under the guidance of the unmasked point patches' features obtained by the designed transformer. Furthermore, a linear self-attention mechanism is designed to use three projection matrices to decompose the original scaled dot-product attention into smaller parts, using the properties of low-rank and linear decomposition to reduce the time complexity from quadratic to linear. The results of extensive experiments demonstrate that the proposed pretrained model can achieve high accuracy of 93.6% and 84.77% on the ModelNet40 and ScanObjectNN datasets, respectively, at a masking ratio of 40%. In addition, the results show that the proposed method, which uses a linear self-attention mechanism, can enhance the computational efficiency by significantly reducing the inference time and minimizing the storage memory requirements for query, key, and value (Q, K, and V) matrices compared with the existing methods. Finally, the results indicate that the proposed method can achieve state-of-the-art performance on the classification ModelNet40 dataset.
引用
收藏
页码:1 / 11
页数:11
相关论文
共 50 条
  • [1] Inter-Modal Masked Autoencoder for Self-Supervised Learning on Point Clouds
    Liu, Jiaming
    Wu, Yue
    Gong, Maoguo
    Liu, Zhixiao
    Miao, Qiguang
    Ma, Wenping
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 3897 - 3908
  • [2] Masked Discrimination for Self-supervised Learning on Point Clouds
    Liu, Haotian
    Cai, Mu
    Lee, Yong Jae
    COMPUTER VISION - ECCV 2022, PT II, 2022, 13662 : 657 - 675
  • [3] Masked Autoencoder for Self-Supervised Pre-training on Lidar Point Clouds
    Hess, Georg
    Jaxing, Johan
    Svensson, Elias
    Hagerman, David
    Petersson, Christoffer
    Svensson, Lennart
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WORKSHOPS (WACVW), 2023, : 350 - 359
  • [4] A Survey on Masked Autoencoder for Visual Self-supervised Learning
    Zhang, Chaoning
    Zhang, Chenshuang
    Song, Junha
    Yi, John Seon Keun
    Kweon, In So
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 6805 - 6813
  • [5] Self-Supervised Learning Malware Traffic Classification Based on Masked Autoencoder
    Xu, Ke
    Zhang, Xixi
    Wang, Yu
    Ohtsuki, Tomoaki
    Adebisi, Bamidele
    Sari, Hikmet
    Gui, Guan
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (10): : 17330 - 17340
  • [6] Self-Supervised Learning of Local Features in 3D Point Clouds
    Thabet, Ali
    Alwassel, Humam
    Ghanem, Bernard
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, : 4048 - 4052
  • [7] ProteinMAE: masked autoencoder for protein surface self-supervised learning
    Yuan, Mingzhi
    Shen, Ao
    Fu, Kexue
    Guan, Jiaming
    Ma, Yingfan
    Qiao, Qin
    Wang, Manning
    BIOINFORMATICS, 2023, 39 (12)
  • [8] PatchMixing Masked Autoencoders for 3D Point Cloud Self-Supervised Learning
    Lin, Chengxing
    Xu, Wenju
    Zhu, Jian
    Nie, Yongwei
    Cai, Ruichu
    Xu, Xuemiao
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (10) : 9882 - 9897
  • [9] DCPoint: Global-Local Dual Contrast for Self-Supervised Representation Learning of 3-D Point Clouds
    Shi, Lu
    Zhang, Guoqing
    Cao, Qi
    Zhang, Linna
    Cen, Yigang
    Cen, Yi
    IEEE SENSORS JOURNAL, 2024, 24 (14) : 23224 - 23238
  • [10] Self-Supervised Learning on 3D Point Clouds by Learning Discrete Generative Models
    Eckart, Benjamin
    Yuan, Wentao
    Liu, Chao
    Kautz, Jan
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 8244 - 8253