Multi-Modality Sensing and Data Fusion for Multi-Vehicle Detection

被引:30
|
作者
Roy, Debashri [1 ]
Li, Yuanyuan [1 ]
Jian, Tong [1 ]
Tian, Peng [1 ]
Chowdhury, Kaushik [1 ]
Ioannidis, Stratis [1 ]
机构
[1] Northeastern Univ, Dept Elect, Comp Engn, Boston, MA 02115 USA
基金
美国国家科学基金会;
关键词
Vehicle detection; tracking; multimodal data; fusion; latent embeddings; image; seismic; acoustic; radar; CHALLENGES; TRACKING;
D O I
10.1109/TMM.2022.3145663
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With the recent surge in autonomous driving vehicles, the need for accurate vehicle detection and tracking is critical now more than ever. Detecting vehicles from visual sensors fails in non-line-of-sight (NLOS) settings. This can be compensated by the inclusion of other modalities in a multi-domain sensing environment. We propose several deep learning based frameworks for fusing different modalities (image, radar, acoustic, seismic) through the exploitation of complementary latent embeddings, incorporating multiple state-of-the-art fusion strategies. Our proposed fusion frameworks considerably outperform unimodal detection. Moreover, fusion between image and non-image modalities improves vehicle tracking and detection under NLOS conditions. We validate our models on the real-world multimodal ESCAPE dataset, showing 33.16% improvement in vehicle detection by fusion (over visual inference alone) over test scenarios with 30-42% NLOS conditions. To demonstrate how well our framework generalizes, we also validate our models on the multimodal NuScene dataset, showing similar to 22% improvement over competing methods.
引用
收藏
页码:2280 / 2295
页数:16
相关论文
共 50 条
  • [21] An Interpretable Fusion Siamese Network for Multi-Modality Remote Sensing Ship Image Retrieval
    Xiong, Wei
    Xiong, Zhenyu
    Cui, Yaqi
    Huang, Linzhou
    Yang, Ruining
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (06) : 2696 - 2712
  • [22] Fusion Of Multi-Modality Medical Images: A Fuzzy Approach
    Kaur, Harmeet
    Kumar, Satish
    PROCEEDINGS ON 2018 IEEE 3RD INTERNATIONAL CONFERENCE ON COMPUTING, COMMUNICATION AND SECURITY (ICCCS), 2018, : 112 - 115
  • [23] Automated registration and fusion of the multi-modality retinal images
    Cao, Hua
    Brener, Nathan
    Thompson, Hilary
    Iyengar, S. S.
    Ye, Zhengmao
    PROCEEDINGS OF THE 40TH SOUTHEASTERN SYMPOSIUM ON SYSTEM THEORY, 2008, : 371 - +
  • [24] Multi-Modality Fusion Perception and Computing in Autonomous Driving
    Zhang Y.
    Zhang S.
    Zhang Y.
    Ji J.
    Duan Y.
    Huang Y.
    Peng J.
    Zhang Y.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2020, 57 (09): : 1781 - 1799
  • [25] PatchView: Multi-modality detection of security patches
    Farhi, Nitzan
    Koenigstein, Noam
    Shavitt, Yuval
    COMPUTERS & SECURITY, 2025, 151
  • [26] Plasmonic Gold Nanostars for Multi-Modality Sensing and Diagnostics
    Liu, Yang
    Yuan, Hsiangkuo
    Kersey, Farrell R.
    Register, Janna K.
    Parrott, Matthew C.
    Vo-Dinh, Tuan
    SENSORS, 2015, 15 (02): : 3706 - 3720
  • [27] Search for multi-modality data in digital libraries
    Yang, J
    Zhuang, YT
    Li, Q
    ADVANCES IN MUTLIMEDIA INFORMATION PROCESSING - PCM 2001, PROCEEDINGS, 2001, 2195 : 482 - 489
  • [28] Aquatic Skin Enabled by Multi-Modality Iontronic Sensing
    Cheng, Yu
    Guo, Chenhui
    Li, Sen
    Deng, Ka
    Tang, Jie
    Luo, Qian
    Zhang, Shiwu
    Chang, Yu
    Pan, Tingrui
    ADVANCED FUNCTIONAL MATERIALS, 2022, 32 (48)
  • [29] A Multi-vehicle Testbed for Multi-modal, Decentralized Sensing of the Environment
    Cortez, R. Andres
    Luna, Jose-Marcio
    Fierro, Rafael
    Wood, John
    2010 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2010, : 1088 - +
  • [30] OmniFuse: A general modality fusion framework for multi-modality learning on low-quality medical data
    Wu, Yixuan
    Chen, Jintai
    Hu, Lianting
    Xu, Hongxia
    Liang, Huiying
    Wu, Jian
    INFORMATION FUSION, 2025, 117