Autonomous driving enhanced: a fusion framework integrating LiDAR point clouds with monovision depth-aware transformers for robust object detection

被引:0
|
作者
Liu, Hui [1 ]
Su, Tong [2 ]
Guo, Jing [1 ]
机构
[1] Suzhou Vocat Inst Ind Technol, Sch Automot Engn, Suzhou 215104, Peoples R China
[2] Shanghai Lixin Univ Accounting & Finance, Sch Informat Management, Shanghai 201209, Peoples R China
来源
ENGINEERING RESEARCH EXPRESS | 2025年 / 7卷 / 01期
关键词
autonomous driving perception; depth-aware transformer (DAT); LiDAR point clouds; monocular depth estimation; adaptive fusion strategy; dilated convolution;
D O I
10.1088/2631-8695/ada7c7
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
In the evolving landscape of autonomous driving technology, the ability to accurately detect and localize objects in complex environments is paramount. This paper introduces an innovative object detection algorithm designed to enhance the perception capabilities of autonomous vehicles. We propose a novel fusion framework that integrates LiDAR point clouds with monocular depth estimations, utilizing a Depth-Aware Transformer (DAT) architecture. The DAT, a recent advancement in transformer models, is uniquely equipped to handle spatial hierarchies and depth cues, making it ideal for interpreting three-dimensional scenes from two-dimensional images. Our approach leverages the complementary strengths of LiDAR and monocular vision, where LiDAR provides precise depth information while the monocular camera offers rich visual textures and color information. The adaptive fusion strategy dynamically adjusts the weight given to each sensor modality based on the reliability and quality of the data in real-time, ensuring optimal performance under varying environmental conditions. We validate our method using the extensive KITTI dataset, a benchmark in autonomous driving research. Extensive experiments demonstrate that our algorithm outperforms state-of-the-art object detection models, achieving higher accuracy in object localization and classification. Moreover, our solution showcases improved robustness and generalization across diverse driving environments, thanks to the enhanced depth perception enabled by the DAT architecture. To further validate the effectiveness of our model, we conducted both comparative and ablation experiments, which confirmed the performance improvements of our approach and demonstrated the critical contributions of the DAT and Adaptive Fusion components. The proposed fusion of LiDAR and monocular depth estimation using Depth-Aware Transformers represents a significant step forward in autonomous driving perception systems. It not only advances the field of object detection but also paves the way for more sophisticated applications in autonomous navigation, where a deep understanding of the environment is crucial for safe and efficient operation.
引用
收藏
页数:16
相关论文
共 43 条
  • [1] Object detection using depth completion and camera-LiDAR fusion for autonomous driving
    Carranza-Garcia, Manuel
    Javier Galan-Sales, F.
    Maria Luna-Romera, Jose
    Riquelme, Jose C.
    INTEGRATED COMPUTER-AIDED ENGINEERING, 2022, 29 (03) : 241 - 258
  • [2] DAFormer: Depth-aware 3D Object Detection Guided by Camera Model via Transformers
    Gao, Junbin
    Ruan, Hao
    Xu, Bingrong
    Zeng, Zhigang
    2022 IEEE INTERNATIONAL CONFERENCE ON CYBORG AND BIONIC SYSTEMS, CBS, 2022, : 170 - 175
  • [3] Real time object detection using LiDAR and camera fusion for autonomous driving
    Liu, Haibin
    Wu, Chao
    Wang, Huanjie
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [4] Timely Fusion of Surround Radar/Lidar for Object Detection in Autonomous Driving Systems
    Xie, Wenjing
    Hu, Tao
    Ling, Neiwen
    Xing, Guoliang
    Xue, Chun Jason
    Guan, Nan
    2024 IEEE 30TH INTERNATIONAL CONFERENCE ON EMBEDDED AND REAL-TIME COMPUTING SYSTEMS AND APPLICATIONS, RTCSA 2024, 2024, : 31 - 36
  • [5] Real time object detection using LiDAR and camera fusion for autonomous driving
    Haibin Liu
    Chao Wu
    Huanjie Wang
    Scientific Reports, 13
  • [6] Online Camera LiDAR Fusion and Object Detection on Hybrid Data for Autonomous Driving
    Banerjee, Koyel
    Notz, Dominik
    Windelen, Johannes
    Gavarraju, Sumanth
    He, Mingkang
    2018 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2018, : 1632 - 1638
  • [7] Depth-Aware Salient Object Detection and Segmentation via Multiscale Discriminative Saliency Fusion and Bootstrap Learning
    Song, Hangke
    Liu, Zhi
    Du, Huan
    Sun, Guangling
    Le Meur, Olivier
    Ren, Tongwei
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2017, 26 (09) : 4204 - 4216
  • [8] Denoising Framework Based on Multiframe Continuous Point Clouds for Autonomous Driving LiDAR in Snowy Weather
    Yan, Xinyuan
    Yang, Junxing
    Zhu, Xinyu
    Liang, Yu
    Huang, He
    IEEE SENSORS JOURNAL, 2024, 24 (07) : 10515 - 10527
  • [9] Multimodal Object Detection and Ranging Based on Camera and Lidar Sensor Fusion for Autonomous Driving
    Khan, Danish
    Baek, Minjin
    Kim, Min Young
    Han, Dong Seog
    2022 27TH ASIA PACIFIC CONFERENCE ON COMMUNICATIONS (APCC 2022): CREATING INNOVATIVE COMMUNICATION TECHNOLOGIES FOR POST-PANDEMIC ERA, 2022, : 342 - 343
  • [10] 3D object detection based on image and LIDAR fusion for autonomous driving
    Chen G.
    Yi H.
    Mao Z.
    International Journal of Vehicle Information and Communication Systems, 2023, 8 (03) : 237 - 251