Camera-LiDAR Cross-Modality Gait Recognition

被引:0
|
作者
Guo, Wenxuan [1 ]
Liang, Yingping [2 ]
Pan, Zhiyu [1 ]
Xi, Ziheng [1 ]
Feng, Jianjiang [1 ]
Zhou, Jie [1 ]
机构
[1] Tsinghua Univ, Dept Automat, BNRist, Beijing, Peoples R China
[2] Beijing Inst Technol, Beijing, Peoples R China
来源
关键词
Gait recognition; Cross-modality; Contrastive pre-training;
D O I
10.1007/978-3-031-72754-2_25
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Gait recognition is a crucial biometric identification technique. Camera-based gait recognition has been widely applied in both research and industrial fields. LiDAR-based gait recognition has also begun to evolve most recently, due to the provision of 3D structural information. However, in certain applications, cameras fail to recognize persons, such as in low-light environments and long-distance recognition scenarios, where LiDARs work well. On the other hand, the deployment cost and complexity of LiDAR systems limit its wider application. Therefore, it is essential to consider cross-modality gait recognition between cameras and LiDARs for a broader range of applications. In this work, we propose the first cross-modality gait recognition framework between Camera and LiDAR, namely CL-Gait. It employs a two-stream network for feature embedding of both modalities. This poses a challenging recognition task due to the inherent matching between 3D and 2D data, exhibiting significant modality discrepancy. To align the feature spaces of the two modalities, i.e., camera silhouettes and LiDAR points, we propose a contrastive pre-training strategy to mitigate modality discrepancy. To make up for the absence of paired camera-LiDAR data for pre-training, we also introduce a strategy for generating data on a large scale. This strategy utilizes monocular depth estimated from single RGB images and virtual cameras to generate pseudo point clouds for contrastive pre-training. Extensive experiments show that the cross-modality gait recognition is very challenging but still contains potential and feasibility with our proposed model and pre-training strategy. To the best of our knowledge, this is the first work to address cross-modality gait recognition. The code and dataset are available at https://github.com/GWxuan/CL-Gait.
引用
收藏
页码:439 / 455
页数:17
相关论文
共 50 条
  • [21] A Cross-Modality Contrastive Learning Method for Radar Jamming Recognition
    Dong, Ganggang
    Wang, Zixuan
    Liu, Hongwei
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2025, 74
  • [22] DLFace: Deep local descriptor for cross-modality face recognition
    Peng, Chunlei
    Wang, Nannan
    Li, Jie
    Gao, Xinbo
    PATTERN RECOGNITION, 2019, 90 : 161 - 171
  • [23] Exploring Cross-Modality Affective Reactions for Audiovisual Emotion Recognition
    Mariooryad, Soroosh
    Busso, Carlos
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2013, 4 (02) : 183 - 196
  • [24] Roadside Camera-LiDAR Calibration Without Annotation
    Jin, Shaojie
    Ma, Cong
    Gao, Ying
    Hui, Fei
    Zhao, Xiangmo
    IEEE SENSORS JOURNAL, 2024, 24 (22) : 37654 - 37665
  • [25] Online Camera-LiDAR Calibration with Sensor Semantic Information
    Zhu, Yufeng
    Li, Chenghui
    Zhang, Yubo
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 4970 - 4976
  • [26] Multiple Objects Localization With Camera-LIDAR Sensor Fusion
    Hocaoglu, Gokce Sena
    Benli, Emrah
    IEEE SENSORS JOURNAL, 2025, 25 (07) : 11892 - 11905
  • [27] Cross-Modality Personalization for Retrieval
    Murrugarra-Llerena, Nils
    Kovashka, Adriana
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 6422 - 6431
  • [28] STRATEGIES IN CROSS-MODALITY MATCHING
    MILEWSKI, AE
    IACCINO, J
    PERCEPTION & PSYCHOPHYSICS, 1982, 31 (03): : 273 - 275
  • [29] Privacy-Safe Action Recognition via Cross-Modality Distillation
    Kim, Yuhyun
    Jung, Jinwook
    Noh, Hyeoncheol
    Ahn, Byungtae
    Kwon, Junghye
    Choi, Dong-Geol
    IEEE ACCESS, 2024, 12 : 125955 - 125965
  • [30] Pedestrian Recognition Using Cross-Modality Learning in Convolutional Neural Networks
    Pop, Danut Ovidiu
    Rogozan, Alexandrina
    Nashashibi, Fawzi
    Bensrhair, Abdelaziz
    IEEE INTELLIGENT TRANSPORTATION SYSTEMS MAGAZINE, 2021, 13 (01) : 210 - 224