Multi-perspective and multi-modality joint representation and recognition model for 3D action recognition

被引:48
|
作者
Gao, Z. [1 ,2 ]
Zhang, H. [1 ,2 ]
Xu, G. P. [1 ,2 ]
Xue, Y. B. [1 ,2 ]
机构
[1] Tianjin Univ Technol, Key Lab Comp Vis & Syst, Minist Educ, Tianjin 300384, Peoples R China
[2] Tianjin Univ Technol, Tianjin Key Lab Intelligence Comp & Novel Softwar, Tianjin 300384, Peoples R China
基金
中国国家自然科学基金;
关键词
3D Action recognition; Difference motion history image; Multi-perspective projection; Multi-modality feature; PHOG; MMJRR; 3-D OBJECT RETRIEVAL;
D O I
10.1016/j.neucom.2014.06.085
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we proposed multi-perspective and multi-modality discriminated and joint representation and recognition model for 3D action recognition. Specifically, for depth and RGB image sequence, we construct a novel difference motion history image, and then propose multi-perspective projections to capture the target motion process, after that, pyramid histogram of orientated gradients is extracted for each projection to describe the target motion, finally, multi-perspective and multi-modality discriminated and joint representation and recognition, model is proposed to recognize human action. Large scale experimental results on challenging and public DHA 3D and MSR-Action3D action datasets show that the performances of our difference motion history image on two modalities are much better than traditional motion history image, at the same time, our description scheme is also very robust and efficient, what is more, our proposed multi-perspective and multi-modality discriminated and joint representation and recognition model further improves the performance, which outperforms the state-of-the-art methods, and whose best performances on MSR-Action3D and DHA datasets reach 90.5% and 98.2% respectively. (C) 2014 Elsevier B.V. All rights reserved.
引用
收藏
页码:554 / 564
页数:11
相关论文
共 50 条
  • [1] Multi-modality Fusion Network for Action Recognition
    Huang, Kai
    Qin, Zheng
    Xu, Kaiping
    Ye, Shuxiong
    Wang, Guolong
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING - PCM 2017, PT II, 2018, 10736 : 139 - 149
  • [2] Multi-modality learning for human action recognition
    Ziliang Ren
    Qieshi Zhang
    Xiangyang Gao
    Pengyi Hao
    Jun Cheng
    Multimedia Tools and Applications, 2021, 80 : 16185 - 16203
  • [3] Multi-modality learning for human action recognition
    Ren, Ziliang
    Zhang, Qieshi
    Gao, Xiangyang
    Hao, Pengyi
    Cheng, Jun
    MULTIMEDIA TOOLS AND APPLICATIONS, 2021, 80 (11) : 16185 - 16203
  • [4] Multi-Modality Vertebra Recognition in Arbitrary Views Using 3D Deformable Hierarchical Model
    Cai, Yunliang
    Osman, Said
    Sharma, Manas
    Landis, Mark
    Li, Shuo
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2015, 34 (08) : 1676 - 1693
  • [5] 3D shape recognition and retrieval based on multi-modality deep learning
    Bu, Shuhui
    Wang, Lei
    Han, Pengcheng
    Liu, Zhenbao
    Li, Ke
    NEUROCOMPUTING, 2017, 259 : 183 - 193
  • [6] Human Action Recognition Via Multi-modality Information
    Gao, Zan
    Song, Jian-ming
    Zhang, Hua
    Liu, An-An
    Xue, Yan-Bing
    Xu, Guang-ping
    JOURNAL OF ELECTRICAL ENGINEERING & TECHNOLOGY, 2014, 9 (02) : 739 - 748
  • [7] Joint multi-type feature learning for multi-modality FKP recognition
    Yang, Yeping
    Fei, Lunke
    Alshehri, Adel Homoud
    Zhao, Shuping
    Sun, Weijun
    Teng, Shaohua
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 126
  • [8] Focal Channel Knowledge Distillation for Multi-Modality Action Recognition
    Gan, Lipeng
    Cao, Runze
    Li, Ning
    Yang, Man
    Li, Xiaochao
    IEEE ACCESS, 2023, 11 : 78285 - 78298
  • [9] Discriminative Multi-modality Non-negative Sparse Graph Model for Action Recognition
    Chen, Yuanbo
    Zhao, Yanyun
    Zhuang, Bojin
    Cai, Anni
    2014 IEEE VISUAL COMMUNICATIONS AND IMAGE PROCESSING CONFERENCE, 2014, : 53 - 56
  • [10] MMA: a multi-view and multi-modality benchmark dataset for human action recognition
    Zan Gao
    Tao-tao Han
    Hua Zhang
    Yan-bing Xue
    Guang-ping Xu
    Multimedia Tools and Applications, 2018, 77 : 29383 - 29404