Deep multi-modal data analysis and fusion for robust scene understanding in CAVs

被引:1
|
作者
Papandreou, Andreas [1 ]
Kloukiniotis, Andreas [1 ]
Lalos, Aris [2 ]
Moustakas, Konstantinos [1 ]
机构
[1] Univ Patras, Dept Elect & Comp Engn, Univ Campus, Rion 26504, Greece
[2] ISI Ind Syst Inst, Patras Sci Pk Bldg, Patras, Greece
关键词
autonomous vehicles; multi-modal scene analysis; adversarial attacks;
D O I
10.1109/MMSP53017.2021.9733604
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Deep learning (DL) tends to be the integral part of Autonomous Vehicles (AVs). Therefore the development of scene analysis modules that are robust to various vulnerabilities such as adversarial inputs or cyber-attacks is becoming an imperative need for the future AV perception systems. In this paper, we deal with this issue by exploring the recent progress in Artificial Intelligence (AI) and Machine Learning (ML) to provide holistic situational awareness and eliminate the effect of the previous attacks on the scene analysis modules. We propose novel multi-modal approaches against which achieve robustness to adversarial attacks, by appropriately modifying the analysis Neural networks and by utilizing late fusion methods. More specifically, we propose a holistic approach by adding new layers to a 2D segmentation DL model enhancing its robustness to adversarial noise. Then, a novel late fusion technique has been applied, by extracting direct features from the 3D space and project them into the 2D segmented space for identifying inconsistencies. Extensive evaluation studies using the KITTI odometry dataset provide promising performance results under various types of noise.
引用
收藏
页数:6
相关论文
共 50 条
  • [41] Improving multi-modal data fusion by anomaly detection
    Jakub Simanek
    Vladimir Kubelka
    Michal Reinstein
    Autonomous Robots, 2015, 39 : 139 - 154
  • [42] Improving multi-modal data fusion by anomaly detection
    Simanek, Jakub
    Kubelka, Vladimir
    Reinstein, Michal
    AUTONOMOUS ROBOTS, 2015, 39 (02) : 139 - 154
  • [43] FUSION METHODS FOR MULTI-MODAL INDEXING OF WEB DATA
    Niaz, Usman
    Merialdo, Bernard
    2013 14TH INTERNATIONAL WORKSHOP ON IMAGE ANALYSIS FOR MULTIMEDIA INTERACTIVE SERVICES (WIAMIS), 2013,
  • [44] Strategies for Multi-Modal Scene Exploration
    Bohg, Jeannette
    Johnson-Roberson, Matthew
    Bjorkman, Marten
    Kragic, Danica
    IEEE/RSJ 2010 INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2010), 2010, : 4509 - 4515
  • [45] Dynamic Deep Multi-modal Fusion for Image Privacy Prediction
    Tonge, Ashwini
    Caragea, Cornelia
    WEB CONFERENCE 2019: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2019), 2019, : 1829 - 1840
  • [46] Robust memory-efficient data level information fusion of multi-modal biometric images
    Noore, Afzel
    Singh, Richa
    Vatsa, Mayank
    INFORMATION FUSION, 2007, 8 (04) : 337 - 346
  • [47] Deep Gated Multi-modal Fusion for Image Privacy Prediction
    Zhao, Chenye
    Caragea, Cornelia
    ACM TRANSACTIONS ON THE WEB, 2023, 17 (04)
  • [48] Guided Image Deblurring by Deep Multi-Modal Image Fusion
    Liu, Yuqi
    Sheng, Zehua
    Shen, Hui-Liang
    IEEE ACCESS, 2022, 10 : 130708 - 130718
  • [49] A deep contrastive multi-modal encoder for multi-omics data integration and analysis
    Yinghua, Ma
    Khan, Ahmad
    Heng, Yang
    Khan, Fiaz Gul
    Ali, Farman
    Al-Otaibi, Yasser D.
    Bashir, Ali Kashif
    INFORMATION SCIENCES, 2025, 700
  • [50] Deep Video Understanding with a Unified Multi-Modal Retrieval Framework
    Xie, Chen-Wei
    Sun, Siyang
    Zhao, Liming
    Wu, Jianmin
    Li, Dangwei
    Zheng, Yun
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 7055 - 7059