Encoding Pose Features to Images With Data Augmentation for 3-D Action Recognition

被引:65
|
作者
Huynh-The, Thien [1 ,2 ]
Hua, Cam-Hao [3 ]
Kim, Dong-Seong [1 ,2 ]
机构
[1] Kumoh Natl Inst Technol, Dept IT Convergence Engn, Gumi 39177, South Korea
[2] Kumoh Natl Inst Technol, ICT Convergence Res Ctr, Gumi 39177, South Korea
[3] Kyung Hee Univ, Dept Comp Sci & Engn, Yongin 446701, South Korea
基金
新加坡国家研究基金会;
关键词
Data augmentation; deep convolutional neural networks (DCNNs); human action recognition; pose feature to image (PoF2I) encoding technique;
D O I
10.1109/TII.2019.2910876
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recently, numerous methods have been introduced for three-dimensional (3-D) action recognition using handcrafted feature descriptors coupled traditional classifiers. However, they cannot learn high-level features of a whole skeleton sequence exhaustively. In this paper, a novel encoding technique-namely, pose feature to image (PoF2I), is introduced to transform the pose features of joint-joint distance and orientation to color pixels. By concatenating the features of all skeleton frames in a sequence, a color image is generated to depict spatial joint correlations and temporal pose dynamics of an action appearance. The strategy of end-to-end fine-tuning a pretrained deep convolutional neural network, which completely capture multiple high-level features at multiscale action representation, is implemented for learning recognition models. We further propose an efficient data augmentation mechanism for informative enrichment and overfitting prevention. The experimental results on six challenging 3-D action recognition datasets demonstrate that the proposed method outperforms state-of-the-art approaches.
引用
收藏
页码:3100 / 3111
页数:12
相关论文
共 50 条
  • [21] VLAD3: Encoding Dynamics of Deep Features for Action Recognition
    Li, Yingwei
    Li, Weixin
    Mahadevan, Vijay
    Vasconcelos, Nuno
    2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 1951 - 1960
  • [22] Automatic Fault Delineation in 3-D Seismic Images With Deep Learning: Data Augmentation or Ensemble Learning?
    Li, Shizhen
    Liu, Naihao
    Li, Fangyu
    Gao, Jinghuai
    Ding, Jicai
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [23] Pose-Invariant Kinematic Features for Action Recognition
    Ramanathan, Manoj
    Yau, Wei-Yun
    Khwang, Eam Teoh
    Thalmann, Nadia Magnenat
    2017 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC 2017), 2017, : 292 - 299
  • [24] On the Benefits of 3D Pose and Tracking for Human Action Recognition
    Rajasegaran, Jathushan
    Pavlakos, Georgios
    Kanazawa, Angjoo
    Feichtenhofer, Christoph
    Malik, Jitendra
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 640 - 649
  • [25] 3D Data Sensing for Hand Pose Recognition
    Trujillo-Romero, Felipe
    Caballero-Morales, Santiago-Omar
    2013 23RD INTERNATIONAL CONFERENCE ON ELECTRONICS, COMMUNICATIONS AND COMPUTING (CONIELECOMP), 2013, : 109 - 113
  • [26] 3-D object pose estimation by shading and edge data fusion - Simulating virtual manipulation on mental images
    Nomura, Y
    Zhang, DL
    Sakaida, Y
    Fujii, S
    1996 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, PROCEEDINGS, 1996, : 866 - 871
  • [27] Microscopy - Wavelength-to-depth encoding gives 3-D images
    Powell, PN
    LASER FOCUS WORLD, 2000, 36 (12): : 52 - 54
  • [28] A data augmentation method for human action recognition using dense joint motion images
    Yao, Leiyue
    Yang, Wei
    Huang, Wei
    APPLIED SOFT COMPUTING, 2020, 97 (97)
  • [29] Pose-varied face recognition based on 3-D face model
    He, Guo-Hui
    Zhu, Bin
    Gan, Jun-Ying
    PROCEEDINGS OF 2008 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-7, 2008, : 2987 - 2992
  • [30] Combining shape and sift features for 3-D object detection and pose estimation
    Tak, Yoon-Sik
    Hwang, Eenjun
    Transactions of the Korean Institute of Electrical Engineers, 2010, 59 (02): : 429 - 435