Deep salient-Gaussian Fisher vector encoding of the spatio-temporal trajectory structures for person re-identification

被引:6
|
作者
Ksibi, Salma [1 ]
Mejdoub, Mahmoud [1 ]
Ben Amar, Chokri [1 ]
机构
[1] Univ Sfax, ENIS, REGIM Res Grp Intelligent Machines, Sfax, Tunisia
关键词
Person re-identification; Deep weighted encoding; Spatio-temporal trajectory structures; Deep spatio-temporal appearance descriptor; Deep CNN; DESCRIPTORS;
D O I
10.1007/s11042-018-6200-5
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we propose a deep spatio-temporal appearance (DSTA) descriptor for person re-identification (re-ID). The proposed descriptor is based on the deep Fisher vector (FV) encoding of the trajectory spatio-temporal structures. These have the advantage of robustly handling the misalignment in the pedestrian tracklets. The deep encoding exploits the richness of the spatio-temporal structural information around the trajectories. This is achieved by hierarchically encoding the trajectory structures leveraging a larger tracklet neighborhood scale when moving from one layer to the next one. In order to eliminate the noisy background located around the pedestrian and model the uniqueness of its identity, the deep FV encoder is further enriched towards the deep Salient-Gaussian weighted FV (deepSGFV) encoder by integrating the pedestrian Gaussian and saliency templates in the encoding process, respectively. The proposed descriptor produces competitive accuracy with respect to state-of-the art methods and especially the deep CNN ones without necessitating either pre-training or data augmentation on four challenging pedestrian video datasets: PRID2011, i-LIDS-VID, Mars and LPW. The further combination of DSTA with deep CNN boosts the current state-of-the-art methods and demonstrates their complementarity.
引用
收藏
页码:1583 / 1611
页数:29
相关论文
共 48 条
  • [41] Deep learning for Amur tiger re-identification in camera traps: A tool assisting population monitoring and spatio-temporal analysis
    Ma, Yiwen
    Tan, Mengyu
    Liu, Xiaoyan
    Zhang, Yingjie
    Xu, Zhouce
    Sun, Wanqing
    Ge, Jianping
    Feng, Limin
    ECOLOGICAL INDICATORS, 2025, 171
  • [42] DFR-ST: Discriminative feature representation with spatio-temporal cues for vehicle re-identification
    Tu, Jingzheng
    Chen, Cailian
    Huang, Xiaolin
    He, Jianping
    Guan, Xinping
    PATTERN RECOGNITION, 2022, 131
  • [43] Occluded Video-Based Person Re-Identification Based on Spatial- Temporal Trajectory Fusion
    Yun Xiao
    Song Kaili
    Zhang Xiaoguang
    Yuan Xinchao
    LASER & OPTOELECTRONICS PROGRESS, 2023, 60 (10)
  • [44] Deep Spatial-Temporal Fusion Network for Video-Based Person Re-Identification
    Chen, Lin
    Yang, Hua
    Zhu, Ji
    Zhou, Qin
    Wu, Shuang
    Gao, Zhiyong
    2017 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2017, : 1478 - 1485
  • [45] Spatio-temporal information mining and fusion feature-guided modal alignment for video-based visible-infrared person re-identification
    Zuo, Zhigang
    Li, Huafeng
    Zhang, Yafei
    Xie, Minghong
    IMAGE AND VISION COMPUTING, 2025, 157
  • [46] VIDEO BASED PERSON RE-IDENTIFICATION BY RE-RANKING ATTENTIVE TEMPORAL INFORMATION IN DEEP RECURRENT CONVOLUTIONAL NETWORKS
    Saha, Bhaswati
    Ram, K. Sai
    Mukhopadhyay, Jayanta
    Roy, Aditi
    Navelkar, Anchit
    2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2018, : 1663 - 1667
  • [47] An Improved Cross-Camera Vehicle Tracking Method: Re-Identification Feature Matching of Confidence Based on Spatio-Temporal Information
    Yu, Zhijia
    Xiang, Liangru
    Hu, Jianming
    Pei, Xin
    CICTP 2021: ADVANCED TRANSPORTATION, ENHANCED CONNECTION, 2021, : 430 - 439
  • [48] Video-Based Person Re-Identification by an End-To-End Learning Architecture with Hybrid Deep Appearance-Temporal Feature
    Sun, Rui
    Huang, Qiheng
    Xia, Miaomiao
    Zhang, Jun
    SENSORS, 2018, 18 (11)