Adversarial Cross-modal Domain Adaptation for Multi-modal Semantic Segmentation in Autonomous Driving

被引:0
|
作者
Shi, Mengqi [1 ]
Cao, Haozhi [1 ]
Xie, Lihua [1 ]
Yang, Jianfei [1 ]
机构
[1] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore, Singapore
关键词
D O I
10.1109/ICARCV57592.2022.10004265
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
3D semantic segmentation is a vital problem in autonomous driving. Vehicles rely on semantic segmentation to sense the surrounding environment and identify pedestrians, roads, and other vehicles. Though many datasets are publicly available, there exists a gap between public data and real-world scenarios due to the different weathers and environments, which is formulated as the domain shift. These days, the research for Unsupervised Domain Adaptation (UDA) rises for solving the problem of domain shift and the lack of annotated datasets. This paper aims to introduce adversarial learning and cross-modal networks (2D and 3D) to boost the performance of UDA for semantic segmentation across different datasets. With this goal, we design an adversarial training scheme with a domain discriminator and render the domain-invariant feature learning. Furthermore, we demonstrate that introducing 2D modalities can contribute to the improvement of 3D modalities by our method. Experimental results show that the proposed approach improves the mIoU by 7.53% compared to the baseline and has an improvement of 3.68% for the multi-modal performance.
引用
收藏
页码:850 / 855
页数:6
相关论文
共 50 条
  • [1] Multi-modal semantic autoencoder for cross-modal retrieval
    Wu, Yiling
    Wang, Shuhui
    Huang, Qingming
    [J]. NEUROCOMPUTING, 2019, 331 : 165 - 175
  • [2] Multi-modal unsupervised domain adaptation for semantic image segmentation
    Hu, Sijie
    Bonardi, Fabien
    Bouchafa, Samia
    Sidibe, Desire
    [J]. PATTERN RECOGNITION, 2023, 137
  • [3] Adversarial unsupervised domain adaptation for 3D semantic segmentation with multi-modal learning
    Liu, Wei
    Luo, Zhiming
    Cai, Yuanzheng
    Yu, Ying
    Ke, Yang
    Marcato Junior, Jose
    Goncalves, Wesley Nunes
    Li, Jonathan
    [J]. ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2021, 176 : 211 - 221
  • [4] Adversarial Graph Attention Network for Multi-modal Cross-modal Retrieval
    Wu, Hongchang
    Guan, Ziyu
    Zhi, Tao
    zhao, Wei
    Xu, Cai
    Han, Hong
    Yang, Yarning
    [J]. 2019 10TH IEEE INTERNATIONAL CONFERENCE ON BIG KNOWLEDGE (ICBK 2019), 2019, : 265 - 272
  • [5] Unsupervised domain adaptation multi-level adversarial network for semantic segmentation based on multi-modal features
    Wang, Zeyu
    Bu, Shuhui
    Huang, Wei
    Zheng, Yuanpan
    Wu, Qinggang
    Chang, Huawen
    Zhang, Xu
    [J]. Tongxin Xuebao/Journal on Communications, 2022, 43 (12): : 157 - 171
  • [6] Cross-modal Learning for Domain Adaptation in 3D Semantic Segmentation
    Jaritz, Maximilian
    Vu, Tuan-Hung
    de Charette, Raoul
    Wirbel, Émilie
    Pérez, Patrick
    [J]. arXiv, 2021,
  • [7] Cross-Modal Learning for Domain Adaptation in 3D Semantic Segmentation
    Jaritz, Maximilian
    Tuan-Hung Vu
    de Charette, Raoul
    Wirbel, Emilie
    Perez, Patrick
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (02) : 1533 - 1544
  • [8] CA_DeepSC: Cross-Modal Alignment for Multi-Modal Semantic Communications
    Wang, Wenjun
    Liu, Minghao
    Chen, Mingkai
    [J]. IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 5871 - 5876
  • [9] Learning Cross-Modal Deep Representations for Multi-Modal MR Image Segmentation
    Li, Cheng
    Sun, Hui
    Liu, Zaiyi
    Wang, Meiyun
    Zheng, Hairong
    Wang, Shanshan
    [J]. MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2019, PT II, 2019, 11765 : 57 - 65
  • [10] Cross-Domain and Cross-Modal Knowledge Distillation in Domain Adaptation for 3D Semantic Segmentation
    Li, Miaoyu
    Zhang, Yachao
    Xie, Yuan
    Gao, Zuodong
    Li, Cuihua
    Zhang, Zhizhong
    Qu, Yanyun
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 3829 - 3837