Deep salient-Gaussian Fisher vector encoding of the spatio-temporal trajectory structures for person re-identification

被引:6
|
作者
Ksibi, Salma [1 ]
Mejdoub, Mahmoud [1 ]
Ben Amar, Chokri [1 ]
机构
[1] Univ Sfax, ENIS, REGIM Res Grp Intelligent Machines, Sfax, Tunisia
关键词
Person re-identification; Deep weighted encoding; Spatio-temporal trajectory structures; Deep spatio-temporal appearance descriptor; Deep CNN; DESCRIPTORS;
D O I
10.1007/s11042-018-6200-5
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we propose a deep spatio-temporal appearance (DSTA) descriptor for person re-identification (re-ID). The proposed descriptor is based on the deep Fisher vector (FV) encoding of the trajectory spatio-temporal structures. These have the advantage of robustly handling the misalignment in the pedestrian tracklets. The deep encoding exploits the richness of the spatio-temporal structural information around the trajectories. This is achieved by hierarchically encoding the trajectory structures leveraging a larger tracklet neighborhood scale when moving from one layer to the next one. In order to eliminate the noisy background located around the pedestrian and model the uniqueness of its identity, the deep FV encoder is further enriched towards the deep Salient-Gaussian weighted FV (deepSGFV) encoder by integrating the pedestrian Gaussian and saliency templates in the encoding process, respectively. The proposed descriptor produces competitive accuracy with respect to state-of-the art methods and especially the deep CNN ones without necessitating either pre-training or data augmentation on four challenging pedestrian video datasets: PRID2011, i-LIDS-VID, Mars and LPW. The further combination of DSTA with deep CNN boosts the current state-of-the-art methods and demonstrates their complementarity.
引用
收藏
页码:1583 / 1611
页数:29
相关论文
共 48 条
  • [21] Boosting Fisher vector based scoring functions for person re-identification
    Messelodi, Stefano
    Modena, Carla Maria
    IMAGE AND VISION COMPUTING, 2015, 44 : 44 - 58
  • [22] STCA: Utilizing a spatio-temporal cross-attention network for enhancing video person re-identification
    Bhuiyan, Amran
    Huang, Jimmy Xiangji
    IMAGE AND VISION COMPUTING, 2022, 123
  • [23] T-MAN: a neural ensemble approach for person re-identification using spatio-temporal information
    Tagore, Nirbhay Kumar
    Chattopadhyay, Pratik
    Wang, Lipo
    MULTIMEDIA TOOLS AND APPLICATIONS, 2020, 79 (37-38) : 28393 - 28409
  • [24] T-MAN: a neural ensemble approach for person re-identification using spatio-temporal information
    Nirbhay Kumar Tagore
    Pratik Chattopadhyay
    Lipo Wang
    Multimedia Tools and Applications, 2020, 79 : 28393 - 28409
  • [25] Transfer Learning of Spatio-Temporal Information using 3D-CNN for Person Re-identification
    Kansal, Kajal
    Subramanyam, A. V.
    2018 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2018, : 923 - 928
  • [26] Spatio-Temporal association rule based deep annotation-free clustering (STAR-DAC) for unsupervised person re-identification
    Raj, Sridhar
    Prasad, Munaga V. N. K.
    Balakrishnan, Ramadoss
    PATTERN RECOGNITION, 2022, 122
  • [27] Spatio-temporal inductance-pattern recognition for vehicle re-identification
    Abdulhai, B
    Tabib, SM
    TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2003, 11 (3-4) : 223 - 239
  • [28] LBP based Spatio-Temporal Covariance Descriptor for People Re-identification
    Hadjkacem, Bassem
    Ayedi, Walid
    Abid, Mohamed
    Snoussi, Hichem
    JOURNAL OF INFORMATION ASSURANCE AND SECURITY, 2016, 11 (03): : 126 - 134
  • [29] STFE: A Comprehensive Video-Based Person Re-Identification Network Based on Spatio-Temporal Feature Enhancement
    Yang, Xi
    Wang, Xian
    Liu, Liangchen
    Wang, Nannan
    Gao, Xinbo
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 7237 - 7249
  • [30] HASI: Hierarchical Attention-Aware Spatio-Temporal Interaction for Video-Based Person Re-Identification
    Chen, Si
    Da, Hui
    Wang, Da-Han
    Zhang, Xu-Yao
    Yan, Yan
    Zhu, Shunzhi
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (06) : 4973 - 4988