Action Status Based Novel Relative Feature Representations for Interaction Recognition

被引:10
|
作者
Li Yanshan [1 ,2 ]
Guo Tianyu [1 ,2 ]
Liu Xing [1 ,2 ]
Luo Wenhan [3 ]
Xie Weixin [1 ,2 ]
机构
[1] Shenzhen Univ, ATR Natl Key Lab Def Technol, Shenzhen 518000, Peoples R China
[2] Shenzhen Univ, Guangdong Key Lab Intelligent Informat Proc, Shenzhen 518000, Peoples R China
[3] Tencent, Shenzhen 518000, Peoples R China
基金
中国国家自然科学基金;
关键词
Human action analysis; Interaction recognition; Action status; Multi-stream network; Relative feature representations;
D O I
10.1049/cje.2020.00.088
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Skeleton-based action recognition has always been an important research topic in computer vision. Most of the researchers in this field currently pay more attention to actions performed by a single person while there is very little work dedicated to the identification of interactions between two people. However, the practical application of interaction recognition is actually more critical in our society considering that actions are often performed by multiple people. How to design an effective scheme to learn discriminative spatial and temporal representations for skeleton-based interaction recognition is still a challenging problem. Focusing on the characteristics of skeleton data for interactions, we first define the moving distance to distinguish the action status of the participants. Then some view-invariant relative features are proposed to fully represent the spatial and temporal relationship of the skeleton sequence. Further, a new coding method is proposed to obtain the novel relative feature representations. Finally, we design a three-stream CNN model to learn deep features for interaction recognition. We evaluate our method on SBU dataset, NTU RGB+D 60 dataset and NTU RGB+D 120 dataset. The experimental results also verify that our method is effective and exhibits great robustness compared with current state-of-the-art methods.
引用
收藏
页码:168 / 180
页数:13
相关论文
共 50 条
  • [41] EESpectrum Feature Representations for Speech Emotion Recognition
    Zhao, Ziping
    Zhao, Yiqin
    Bao, Zhongtian
    Wang, Haishuai
    Zhang, Zixing
    Li, Chao
    PROCEEDINGS OF THE JOINT WORKSHOP OF THE 4TH WORKSHOP ON AFFECTIVE SOCIAL MULTIMEDIA COMPUTING AND FIRST MULTI-MODAL AFFECTIVE COMPUTING OF LARGE-SCALE MULTIMEDIA DATA (ASMMC-MMAC'18), 2018, : 27 - 33
  • [42] Hint-based reasoning for feature recognition: status report
    Han, JH
    Regli, WC
    Brooks, S
    COMPUTER-AIDED DESIGN, 1998, 30 (13) : 1003 - 1007
  • [43] Status Recognition Technology Based on Extraction of Facial Feature Information
    Yu Guofang
    Li Jian
    ICCSSE 2009: PROCEEDINGS OF 2009 4TH INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE & EDUCATION, 2009, : 701 - 706
  • [44] Slow feature subspace: A video representation based on slow feature analysis for action recognition
    Beleza, Suzana Rita Alves
    Shimomoto, Erica K.
    Souza, Lincon S.
    Fukui, Kazuhiro
    MACHINE LEARNING WITH APPLICATIONS, 2023, 14
  • [45] Improving multi-view facial expression recognition through two novel texture-based feature representations
    Wang, Xuejian
    Fairhurst, Michael C.
    Canuto, Anne M. P.
    INTELLIGENT DATA ANALYSIS, 2020, 24 (06) : 1455 - 1476
  • [46] Feature difference and feature correlation learning mechanism for skeleton-based action recognition
    Qing, Ruxin
    Jiang, Min
    Kong, Jun
    JOURNAL OF ELECTRONIC IMAGING, 2023, 32 (01)
  • [47] A novel feature for emotion recognition in voice based applications
    Maganti, Hari Krishna
    Scherer, Stefan
    Palm, Guenther
    AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION, PROCEEDINGS, 2007, 4738 : 710 - 711
  • [48] A novel iris recognition method based on feature fusion
    Zhang, PF
    Li, DS
    Wang, Q
    PROCEEDINGS OF THE 2004 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-7, 2004, : 3661 - 3665
  • [49] Pedestrian Attribute Recognition with Part-based CNN and Combined Feature Representations
    Chen, Yiqiang
    Duffner, Stefan
    Stoian, Andrei
    Dufour, Jean-Yves
    Baskurt, Atilla
    PROCEEDINGS OF THE 13TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS (VISIGRAPP 2018), VOL 5: VISAPP, 2018, : 114 - 122
  • [50] Deep Learning-Based Acoustic Feature Representations for Dysarthric Speech Recognition
    Latha M.
    Shivakumar M.
    Manjula G.
    Hemakumar M.
    Kumar M.K.
    SN Computer Science, 4 (3)