Navigating an Automated Driving Vehicle via the Early Fusion of Multi-Modality

被引:15
|
作者
Haris, Malik [1 ]
Glowacz, Adam [2 ]
机构
[1] Southwest Jiaotong Univ, Sch Informat Sci & Technol, Xipu Campus, Chengdu 611756, Peoples R China
[2] AGH Univ Sci & Technol, Fac Elect Engn Automat Comp Sci & Biomed Engn, Dept Automat Control & Robot, Al A Mickiewicza 30, PL-30059 Krakow, Poland
关键词
artificial intelligent; end-to-end autonomous driving; safely navigation; conditional imitation learning (CIL); conditional early fusion (CEF); situation understanding; object detection; CARLA; ROBOT;
D O I
10.3390/s22041425
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
The ability of artificial intelligence to drive toward an intended destination is a key component of an autonomous vehicle. Different paradigms are now being employed to address artificial intelligence advancement. On the one hand, modular pipelines break down the driving model into submodels, such as perception, maneuver planning and control. On the other hand, we used the end-to-end driving method to assign raw sensor data directly to vehicle control signals. The latter is less well-studied but is becoming more popular since it is easier to use. This article focuses on end-to-end autonomous driving, using RGB pictures as the primary sensor input data. The autonomous vehicle is equipped with a camera and active sensors, such as LiDAR and Radar, for safe navigation. Active sensors (e.g., LiDAR) provide more accurate depth information than passive sensors. As a result, this paper examines whether combining the RGB from the camera and active depth information from LiDAR has better results in end-to-end artificial driving than using only a single modality. This paper focuses on the early fusion of multi-modality and demonstrates how it outperforms a single modality using the CARLA simulator.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Multi-Modality Fusion Perception and Computing in Autonomous Driving
    Zhang Y.
    Zhang S.
    Zhang Y.
    Ji J.
    Duan Y.
    Huang Y.
    Peng J.
    Zhang Y.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2020, 57 (09): : 1781 - 1799
  • [2] Automated registration and fusion of the multi-modality retinal images
    Cao, Hua
    Brener, Nathan
    Thompson, Hilary
    Iyengar, S. S.
    Ye, Zhengmao
    PROCEEDINGS OF THE 40TH SOUTHEASTERN SYMPOSIUM ON SYSTEM THEORY, 2008, : 371 - +
  • [3] BEV-Guided Multi-Modality Fusion for Driving Perception
    Man, Yunze
    Gui, Liang-Yan
    Wang, Yu-Xiong
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 21960 - 21969
  • [4] Multi-Modality Sensing and Data Fusion for Multi-Vehicle Detection
    Roy, Debashri
    Li, Yuanyuan
    Jian, Tong
    Tian, Peng
    Chowdhury, Kaushik
    Ioannidis, Stratis
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 2280 - 2295
  • [5] Automated Segmentation of Corticospinal Tract in Diffusion Tensor Images via Multi-modality Multi-atlas Fusion
    Tang, Xiaoying
    Mori, Susumu
    Miller, Michael I.
    MEDICAL IMAGING 2014: BIOMEDICAL APPLICATIONS IN MOLECULAR, STRUCTURAL, AND FUNCTIONAL IMAGING, 2014, 9038
  • [6] Equivariant Multi-Modality Image Fusion
    Zhao, Zixiang
    Hai, Haowen
    Zhang, Jiangshe
    Zhang, Yulun
    Zhane, Kai
    Xu, Shuang
    Chen, Dongdong
    Timofte, Radu
    Van Gool, Luc
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 25912 - 25921
  • [7] Physiological measurements for driving drowsiness: A comparative study of multi-modality feature fusion and selection
    Wu, Yonglin
    Jiang, Xinyu
    Guo, Yao
    Zhu, Hangyu
    Dai, Chenyun
    Chen, Wei
    COMPUTERS IN BIOLOGY AND MEDICINE, 2023, 167
  • [8] Multi-modality Fusion Network for Action Recognition
    Huang, Kai
    Qin, Zheng
    Xu, Kaiping
    Ye, Shuxiong
    Wang, Guolong
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING - PCM 2017, PT II, 2018, 10736 : 139 - 149
  • [9] Fusion of multi-modality volumetric medical imagery
    Aguilar, M
    New, JR
    PROCEEDINGS OF THE FIFTH INTERNATIONAL CONFERENCE ON INFORMATION FUSION, VOL II, 2002, : 1206 - 1212
  • [10] Open-world driving scene segmentation via multi-stage and multi-modality fusion of vision-language embedding
    Niu, Yingjie
    Ding, Ming
    Zhang, Yuxiao
    Ge, Maoning
    Yang, Hanting
    Takeda, Kazuya
    2023 IEEE INTELLIGENT VEHICLES SYMPOSIUM, IV, 2023,