TransRVNet: LiDAR Semantic Segmentation With Transformer

被引:13
|
作者
Cheng, Hui-Xian [1 ]
Han, Xian-Feng [1 ]
Xiao, Guo-Qiang [1 ]
机构
[1] Southwest Univ, Coll Comp & Informat Sci, Beibei 400715, Chongqing, Peoples R China
基金
中国国家自然科学基金;
关键词
Transformers; Point cloud compression; Three-dimensional displays; Laser radar; Semantic segmentation; Semantics; Convolutional neural networks; CNN; transformer; point cloud; range image; semantic segmentation; autonomous driving; CHALLENGE; NETWORK;
D O I
10.1109/TITS.2023.3248117
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Effective and efficient 3D semantic segmentation from large-scale LiDAR point cloud is a fundamental problem in the field of autonomous driving. In this paper, we present Transformer-Range-View Network (TransRVNet), a novel and powerful projection-based CNN-Transformer architecture to infer point-wise semantics. First, a Multi Residual Channel Interaction Attention Module (MRCIAM) is introduced to capture channel-level multi-scale feature and model intra-channel, inter-channel correlations based on attention mechanism. Then, in the encoder stage, we use a well-designed Residual Context Aggregation Module (RCAM), including a residual dilated convolution structure and a context aggregation module, to fuse information from different receptive fields while reducing the impact of missing points. Finally, a Balanced Non-square-Transformer Module (BNTM) is employed as fundamental component of decoder to achieve locally feature dependencies for more discriminative feature learning by introducing the non-square shifted window strategy. Extensive qualitative and quantitative experiments conducted on challenging SemanticKITTI and SemanticPOSS benchmarks have verified the effectiveness of our proposed technique. Our TransRVNet presents superior performance over most existing state-of-the-art approaches. The source code and trained model are available at https://github.com/huixiancheng/TransRVNet.
引用
收藏
页码:5895 / 5907
页数:13
相关论文
共 50 条
  • [1] A reversible transformer for LiDAR point cloud semantic segmentation
    Akwensi, Perpertual Hope
    Wang, Ruisheng
    [J]. 2023 20TH CONFERENCE ON ROBOTS AND VISION, CRV, 2023, : 19 - 28
  • [2] TrSeg: Transformer for semantic segmentation
    Jin, Youngsaeng
    Han, David
    Ko, Hanseok
    [J]. PATTERN RECOGNITION LETTERS, 2021, 148 : 29 - 35
  • [3] Segmenter: Transformer for Semantic Segmentation
    Strudel, Robin
    Garcia, Ricardo
    Laptev, Ivan
    Schmid, Cordelia
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 7242 - 7252
  • [4] Transformer Scale Gate for Semantic Segmentation
    Shi, Hengcan
    Hayat, Munawar
    Cai, Jianfei
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 3051 - 3060
  • [5] Pyramid Fusion Transformer for Semantic Segmentation
    Qin, Zipeng
    Liu, Jianbo
    Zhang, Xiaolin
    Tian, Maoqing
    Zhou, Aojun
    Yi, Shuai
    Li, Hongsheng
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 9630 - 9643
  • [6] SSformer: A Lightweight Transformer for Semantic Segmentation
    Shi, Wentao
    Xu, Jing
    Gao, Pan
    [J]. 2022 IEEE 24TH INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING (MMSP), 2022,
  • [7] A transformer-based real-time LiDAR semantic segmentation method for restricted mobile devices
    Liu, Chang
    Zhao, Jin
    Sun, Nianyi
    [J]. JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2024, 361 (04):
  • [8] Instant Domain Augmentation for LiDAR Semantic Segmentation
    Ryu, Kwonyoung
    Hwang, Soonmin
    Park, Jaesik
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 9350 - 9360
  • [9] Discwise Active Learning for LiDAR Semantic Segmentation
    Unal, Ozan
    Dai, Dengxin
    Unal, Ali Tamer
    Van Gool, Luc
    [J]. IEEE Robotics and Automation Letters, 2023, 8 (11) : 7671 - 7678
  • [10] Low-Latency LiDAR Semantic Segmentation
    Hori, Takahiro
    Yairi, Takehisa
    [J]. 2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 9886 - 9891