Navigating an Automated Driving Vehicle via the Early Fusion of Multi-Modality

被引:15
|
作者
Haris, Malik [1 ]
Glowacz, Adam [2 ]
机构
[1] Southwest Jiaotong Univ, Sch Informat Sci & Technol, Xipu Campus, Chengdu 611756, Peoples R China
[2] AGH Univ Sci & Technol, Fac Elect Engn Automat Comp Sci & Biomed Engn, Dept Automat Control & Robot, Al A Mickiewicza 30, PL-30059 Krakow, Poland
关键词
artificial intelligent; end-to-end autonomous driving; safely navigation; conditional imitation learning (CIL); conditional early fusion (CEF); situation understanding; object detection; CARLA; ROBOT;
D O I
10.3390/s22041425
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
The ability of artificial intelligence to drive toward an intended destination is a key component of an autonomous vehicle. Different paradigms are now being employed to address artificial intelligence advancement. On the one hand, modular pipelines break down the driving model into submodels, such as perception, maneuver planning and control. On the other hand, we used the end-to-end driving method to assign raw sensor data directly to vehicle control signals. The latter is less well-studied but is becoming more popular since it is easier to use. This article focuses on end-to-end autonomous driving, using RGB pictures as the primary sensor input data. The autonomous vehicle is equipped with a camera and active sensors, such as LiDAR and Radar, for safe navigation. Active sensors (e.g., LiDAR) provide more accurate depth information than passive sensors. As a result, this paper examines whether combining the RGB from the camera and active depth information from LiDAR has better results in end-to-end artificial driving than using only a single modality. This paper focuses on the early fusion of multi-modality and demonstrates how it outperforms a single modality using the CARLA simulator.
引用
收藏
页数:18
相关论文
共 50 条
  • [41] Fast and Automated Oscillation Frequency Extraction Using Bayesian Multi-Modality
    Corsaro, Enrico
    FRONTIERS IN ASTRONOMY AND SPACE SCIENCES, 2019, 6
  • [42] A Multi-modality Driven Promptable Transformer for Automated Parapneumonic Effusion Staging
    Chen, Yan
    Liu, Qing
    Xiang, Yao
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT XIII, 2024, 14437 : 248 - 259
  • [43] AFDFusion: An adaptive frequency decoupling fusion network for multi-modality image
    Wang, Chengchao
    Zhao, Zhengpeng
    Yang, Qiuxia
    Nie, Rencan
    Cao, Jinde
    Pu, Yuanyuan
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 263
  • [44] Deep learning supported disease detection with multi-modality image fusion
    Vinnarasi, F. Sangeetha Francelin
    Daniel, Jesline
    Rose, J. T. Anita
    Pugalenthi, R.
    JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY, 2021, 29 (03) : 411 - 434
  • [45] GCN-Based Multi-Modality Fusion Network for Action Recognition
    Liu, Shaocan
    Wang, Xingtao
    Xiong, Ruiqin
    Fan, Xiaopeng
    IEEE TRANSACTIONS ON MULTIMEDIA, 2025, 27 : 1242 - 1253
  • [46] CROSS-MEDIA TOPIC DETECTION: A MULTI-MODALITY FUSION FRAMEWORK
    Zhang, Yanyan
    Li, Guorong
    Chu, Lingyang
    Wang, Shuhui
    Zhang, Weigang
    Huang, Qingming
    2013 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME 2013), 2013,
  • [47] AWDF: An Adaptive Weighted Deep Fusion Architecture for Multi-modality Learning
    Xue, Qinghan
    Kolagunda, Abhishek
    Eliuk, Steven
    Wang, Xiaolong
    2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 2503 - 2512
  • [48] Operator initiated auto fusion technique with applications to multi-modality imaging
    Jaradat, H
    Tome, W
    McNutt, T
    Meyerand, E
    Paliwal, B
    MEDICAL PHYSICS, 2002, 29 (06) : 1209 - 1209
  • [49] A novel dictionary learning approach for multi-modality medical image fusion
    Zhu, Zhiqin
    Chai, Yi
    Yin, Hongpeng
    Li, Yanxia
    Liu, Zhaodong
    NEUROCOMPUTING, 2016, 214 : 471 - 482
  • [50] Multi-Modality Medical Image Fusion using Discrete Wavelet Transform
    Bhavana, V
    Krishnappa, H. K.
    PROCEEDINGS OF THE 4TH INTERNATIONAL CONFERENCE ON ECO-FRIENDLY COMPUTING AND COMMUNICATION SYSTEMS, 2015, 70 : 625 - 631