Adversarial Cross-modal Domain Adaptation for Multi-modal Semantic Segmentation in Autonomous Driving

被引:0
|
作者
Shi, Mengqi [1 ]
Cao, Haozhi [1 ]
Xie, Lihua [1 ]
Yang, Jianfei [1 ]
机构
[1] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore, Singapore
关键词
D O I
10.1109/ICARCV57592.2022.10004265
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
3D semantic segmentation is a vital problem in autonomous driving. Vehicles rely on semantic segmentation to sense the surrounding environment and identify pedestrians, roads, and other vehicles. Though many datasets are publicly available, there exists a gap between public data and real-world scenarios due to the different weathers and environments, which is formulated as the domain shift. These days, the research for Unsupervised Domain Adaptation (UDA) rises for solving the problem of domain shift and the lack of annotated datasets. This paper aims to introduce adversarial learning and cross-modal networks (2D and 3D) to boost the performance of UDA for semantic segmentation across different datasets. With this goal, we design an adversarial training scheme with a domain discriminator and render the domain-invariant feature learning. Furthermore, we demonstrate that introducing 2D modalities can contribute to the improvement of 3D modalities by our method. Experimental results show that the proposed approach improves the mIoU by 7.53% compared to the baseline and has an improvement of 3.68% for the multi-modal performance.
引用
收藏
页码:850 / 855
页数:6
相关论文
共 50 条
  • [1] Multi-modal semantic autoencoder for cross-modal retrieval
    Wu, Yiling
    Wang, Shuhui
    Huang, Qingming
    NEUROCOMPUTING, 2019, 331 : 165 - 175
  • [2] Multi-modal unsupervised domain adaptation for semantic image segmentation
    Hu, Sijie
    Bonardi, Fabien
    Bouchafa, Samia
    Sidibe, Desire
    PATTERN RECOGNITION, 2023, 137
  • [3] Adversarial unsupervised domain adaptation for 3D semantic segmentation with multi-modal learning
    Liu, Wei
    Luo, Zhiming
    Cai, Yuanzheng
    Yu, Ying
    Ke, Yang
    Marcato Junior, Jose
    Goncalves, Wesley Nunes
    Li, Jonathan
    ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2021, 176 : 211 - 221
  • [4] Adversarial Graph Attention Network for Multi-modal Cross-modal Retrieval
    Wu, Hongchang
    Guan, Ziyu
    Zhi, Tao
    zhao, Wei
    Xu, Cai
    Han, Hong
    Yang, Yarning
    2019 10TH IEEE INTERNATIONAL CONFERENCE ON BIG KNOWLEDGE (ICBK 2019), 2019, : 265 - 272
  • [5] Unsupervised domain adaptation multi-level adversarial network for semantic segmentation based on multi-modal features
    Wang Z.
    Bu S.
    Huang W.
    Zheng Y.
    Wu Q.
    Chang H.
    Zhang X.
    Tongxin Xuebao/Journal on Communications, 2022, 43 (12): : 157 - 171
  • [6] Cross-modal Learning for Domain Adaptation in 3D Semantic Segmentation
    Jaritz, Maximilian
    Vu, Tuan-Hung
    de Charette, Raoul
    Wirbel, Émilie
    Pérez, Patrick
    arXiv, 2021,
  • [7] Cross-Modal Learning for Domain Adaptation in 3D Semantic Segmentation
    Jaritz, Maximilian
    Tuan-Hung Vu
    de Charette, Raoul
    Wirbel, Emilie
    Perez, Patrick
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (02) : 1533 - 1544
  • [8] Cross-Modal Semantic Alignment and Information Refinement for Multi-Modal Sentiment Analysis
    Ding, Meirong
    Chen, Hongye
    Zeng, Biqing
    Computer Engineering and Applications, 2024, 60 (22) : 114 - 125
  • [9] CA_DeepSC: Cross-Modal Alignment for Multi-Modal Semantic Communications
    Wang, Wenjun
    Liu, Minghao
    Chen, Mingkai
    IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 5871 - 5876
  • [10] Learning Cross-Modal Deep Representations for Multi-Modal MR Image Segmentation
    Li, Cheng
    Sun, Hui
    Liu, Zaiyi
    Wang, Meiyun
    Zheng, Hairong
    Wang, Shanshan
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2019, PT II, 2019, 11765 : 57 - 65