Depth- and semantics-aware multi-modal domain translation: Generating 3D panoramic color images from LiDAR point clouds

被引:3
|
作者
Cortinhal, Tiago [1 ]
Aksoy, Eren Erdal [1 ]
机构
[1] Halmstad Univ, Ctr Appl Intelligent Syst Res, Sch Informat Technol, Halmstad, Sweden
关键词
Multi-modal domain translation; Semantic perception; LiDAR;
D O I
10.1016/j.robot.2023.104583
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This work presents a new depth-and semantics-aware conditional generative model, named TITAN-Next, for cross-domain image-to-image translation in a multi-modal setup between LiDAR and camera sensors. The proposed model leverages scene semantics as a mid-level representation and is able to translate raw LiDAR point clouds to RGB-D camera images by solely relying on semantic scene segments. We claim that this is the first framework of its kind and it has practical applications in autonomous vehicles such as providing a fail-safe mechanism and augmenting available data in the target image domain. The proposed model is evaluated on the large-scale and challenging Semantic-KITTI dataset, and experimental findings show that it considerably outperforms the original TITAN-Net and other strong baselines by 23.7% margin in terms of IoU.
引用
收藏
页数:9
相关论文
共 28 条
  • [1] Semantics-aware Multi-modal Domain Translation: From LiDAR Point Clouds to Panoramic Color Images
    Cortinhal, Tiago
    Kurnaz, Fatih
    Aksoy, Eren Erdal
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2021), 2021, : 3032 - 3041
  • [2] Semantics-aware transformer for 3D reconstruction from binocular images
    Xin Jia
    Shourui Yang
    Diyi Guan
    Optoelectronics Letters, 2022, 18 : 293 - 299
  • [3] Semantics-aware transformer for 3D reconstruction from binocular images
    JIA Xin
    YANG Shourui
    GUAN Diyi
    Optoelectronics Letters, 2022, 18 (05) : 293 - 299
  • [4] Semantics-aware transformer for 3D reconstruction from binocular images
    Jia Xin
    Yang Shourui
    Guan Diyi
    OPTOELECTRONICS LETTERS, 2022, 18 (05) : 293 - 299
  • [5] Generating Adversarial Point Clouds on Multi-modal Fusion Based 3D Object Detection Model
    Wang, Huiying
    Shen, Huixin
    Zhang, Boyang
    Wen, Yu
    Meng, Dan
    INFORMATION AND COMMUNICATIONS SECURITY (ICICS 2021), PT I, 2021, 12918 : 187 - 203
  • [6] Semantics-aware LiDAR-Only Pseudo Point Cloud Generation for 3D Object Detection
    Cortinhal, Tiago
    Gouigah, Idriss
    Aksoy, Eren Erdal
    2024 35TH IEEE INTELLIGENT VEHICLES SYMPOSIUM, IEEE IV 2024, 2024, : 3220 - 3226
  • [7] 3D Environment Detection Using Multi-view Color Images and LiDAR Point Clouds
    Wu, Bo-Tai
    Li, Pei-Cian
    Chen, Jian-Hong
    Li, Yen-Ju
    Fan, Cheng
    2018 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS-TAIWAN (ICCE-TW), 2018,
  • [8] Urban 3D segmentation and modelling from street view images and LiDAR point clouds
    Babahajiani, Pouria
    Fan, Lixin
    Kamarainen, Joni-Kristian
    Gabbouj, Moncef
    MACHINE VISION AND APPLICATIONS, 2017, 28 (07) : 679 - 694
  • [9] An Integrated Model for Generating 3D Point Clouds and Segmentation from 2D Images
    Chen, Huan-Yu
    Lin, Chuen-Horng
    Lin, Yan-Yu
    Chan, Yung-Kuan
    2024 8TH INTERNATIONAL CONFERENCE ON IMAGING, SIGNAL PROCESSING AND COMMUNICATIONS, ICISPC 2024, 2024, : 94 - 101
  • [10] Urban 3D segmentation and modelling from street view images and LiDAR point clouds
    Pouria Babahajiani
    Lixin Fan
    Joni-Kristian Kämäräinen
    Moncef Gabbouj
    Machine Vision and Applications, 2017, 28 : 679 - 694