Multimodal Multipart Learning for Action Recognition in Depth Videos

被引:76
|
作者
Shahroudy, Amir [1 ,2 ]
Ng, Tian-Tsong [2 ]
Yang, Qingxiong [3 ]
Wang, Gang [1 ]
机构
[1] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore 639798, Singapore
[2] Inst Infocomm Res, 1 Fusionopolis Way, Singapore 138632, Singapore
[3] City Univ Hong Kong, Dept Comp Sci, Hong Kong, Hong Kong, Peoples R China
基金
新加坡国家研究基金会;
关键词
Action recognition; kinect; joint sparse regression; mixed norms; structured sparsity; group feature selection; MULTITASK; FEATURES; SELECTION; TRACKING; SPARSITY; MODEL;
D O I
10.1109/TPAMI.2015.2505295
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The articulated and complex nature of human actions makes the task of action recognition difficult. One approach to handle this complexity is dividing it to the kinetics of body parts and analyzing the actions based on these partial descriptors. We propose a joint sparse regression based learning method which utilizes the structured sparsity to model each action as a combination of multimodal features from a sparse set of body parts. To represent dynamics and appearance of parts, we employ a heterogeneous set of depth and skeleton based features. The proper structure of multimodal multipart features are formulated into the learning framework via the proposed hierarchical mixed norm, to regularize the structured features of each part and to apply sparsity between them, in favor of a group feature selection. Our experimental results expose the effectiveness of the proposed learning method in which it outperforms other methods in all three tested datasets while saturating one of them by achieving perfect accuracy.
引用
收藏
页码:2123 / 2129
页数:7
相关论文
共 50 条
  • [21] An effective view and time-invariant action recognition method based on depth videos
    Liu, Zhi
    Fong, Xin
    Tian, Yingli
    2015 VISUAL COMMUNICATIONS AND IMAGE PROCESSING (VCIP), 2015,
  • [22] Action and Event Recognition in Videos by Learning From Heterogeneous Web Sources
    Niu, Li
    Xu, Xinxing
    Chen, Lin
    Duan, Lixin
    Xu, Dong
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (06) : 1290 - 1304
  • [23] Localized Multiple Kernel Learning for Realistic Human Action Recognition in Videos
    Song, Yan
    Zheng, Yan-Tao
    Tang, Sheng
    Zhou, Xiangdong
    Zhang, Yongdong
    Lin, Shouxun
    Chua, Tat-Seng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2011, 21 (09) : 1193 - 1202
  • [24] Correlation Net: Spatiotemporal multimodal deep learning for action recognition
    Yudistira, Novanto
    Kurita, Takio
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2020, 82
  • [25] Action Recognition from Videos with Complex Background via Transfer Learning
    林贤明
    李绍滋
    张洪博
    刘姝
    Journal of Donghua University(English Edition), 2010, 27 (02) : 199 - 203
  • [26] Human Action Recognition in Unconstrained Videos Using Deep Learning Techniques
    Priya, G. G. Lakshmi
    Jain, Mrinal
    Perumal, R. Srinivasa
    Mouli, P. V. S. S. R. Chandra
    INTELLIGENT COMPUTING AND COMMUNICATION, ICICC 2019, 2020, 1034 : 737 - 744
  • [27] Action Recognition in Videos through a Transfer-Learning-Based Technique
    Lopez-Lozada, Elizabeth
    Sossa, Humberto
    Rubio-Espino, Elsa
    Montiel-Perez, Jesus Yalja
    MATHEMATICS, 2024, 12 (20)
  • [28] Learnable Irrelevant Modality Dropout for Multimodal Action Recognition on Modality-Specific Annotated Videos
    Alfasly, Saghir
    Lu, Jian
    Xu, Chen
    Zou, Yuru
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 20176 - 20185
  • [29] Human Action Recognition from Depth Videos Using Pool of Multiple Projections with Greedy Selection
    Le, Chien-Quang
    Phan, Sang
    Thanh Duc Ngo
    Le, Duy-Dinh
    Satoh, Shin'ichi
    Duc Anh Duong
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2016, E99D (08) : 2161 - 2171
  • [30] PeMapNet: Action Recognition from Depth Videos Using Pyramid Energy Maps on Neural Networks
    Li, Jiahao
    Wu, Hejun
    Zhou, Xinrui
    2017 IEEE 29TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2017), 2017, : 80 - 87