Enhancing Robustness of Viewpoint Changes in 3D Skeleton-Based Human Action Recognition

被引:3
|
作者
Park, Jinyoon [1 ,2 ]
Kim, Chulwoong [2 ]
Kim, Seung-Chan [1 ]
机构
[1] Sungkyunkwan Univ, Dept Sport Interact Sci, Machine Learning Syst Lab, Suwon 16419, South Korea
[2] TAIIPA Taean AI Ind Promot Agcy, Taean 32154, South Korea
关键词
action recognition; machine learning; feature learning; skeletal data; data augmentation;
D O I
10.3390/math11153280
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Previous research on 3D skeleton-based human action recognition has frequently relied on a sequence-wise viewpoint normalization process, which adjusts the view directions of all segmented action sequences. This type of approach typically demonstrates robustness against variations in viewpoint found in short-term videos, a characteristic commonly encountered in public datasets. However, our preliminary investigation of complex action sequences, such as discussions or smoking, reveals its limitations in capturing the intricacies of such actions. To address these view-dependency issues, we propose a straightforward, yet effective, sequence-wise augmentation technique. This strategy enhances the robustness of action recognition models, particularly against changes in viewing direction that mainly occur within the horizontal plane (azimuth) by rotating human key points around either the z-axis or the spine vector, effectively creating variations in viewing directions. We scrutinize the robustness of this approach against real-world viewpoint variations through extensive empirical studies on multiple public datasets, including an additional set of custom action sequences. Despite the simplicity of our approach, our experimental results consistently yield improved action recognition accuracies. Compared to the sequence-wise viewpoint normalization method used with advanced deep learning models like Conv1D, LSTM, and Transformer, our approach showed a relative increase in accuracy of 34.42% for the z-axis and 10.86% for the spine vector.
引用
收藏
页数:17
相关论文
共 50 条
  • [1] Rethinking the ST-GCNs for 3D skeleton-based human action recognition
    Peng, Wei
    Shi, Jingang
    Varanka, Tuomas
    Zhao, Guoying
    NEUROCOMPUTING, 2021, 454 : 45 - 53
  • [2] 3D skeleton-based human action classification: A survey
    Lo Presti, Liliana
    La Cascia, Marco
    PATTERN RECOGNITION, 2016, 53 : 130 - 147
  • [3] 3D skeleton-based action recognition with convolutional neural networks
    Van-Nam Hoang
    Thi-Lan Le
    Thanh-Hai Tran
    Hai-Vu
    Van-Toi Nguyen
    2019 INTERNATIONAL CONFERENCE ON MULTIMEDIA ANALYSIS AND PATTERN RECOGNITION (MAPR), 2019,
  • [4] Learning Clip Representations for Skeleton-Based 3D Action Recognition
    Ke, Qiuhong
    Bennamoun, Mohammed
    An, Senjian
    Sohel, Ferdous
    Boussaid, Farid
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (06) : 2842 - 2855
  • [5] Skeleton-Based Square Grid for Human Action Recognition With 3D Convolutional Neural Network
    Ding, Wenwen
    Ding, Chongyang
    Li, Guang
    Liu, Kai
    IEEE ACCESS, 2021, 9 : 54078 - 54089
  • [6] Mix Dimension in Poincare Geometry for 3D Skeleton-based Action Recognition
    Peng, Wei
    Shi, Jingang
    Xia, Zhaoqiang
    Zhao, Guoying
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 1432 - 1440
  • [7] A Survey on 3D Skeleton-Based Action Recognition Using Learning Method
    Ren, Bin
    Liu, Mengyuan
    Ding, Runwei
    Liu, Hong
    CYBORG AND BIONIC SYSTEMS, 2024, 5
  • [8] Tripool: Graph triplet pooling for 3D skeleton-based action recognition
    Peng, Wei
    Hong, Xiaopeng
    Zhao, Guoying
    PATTERN RECOGNITION, 2021, 115
  • [9] A 3D graph convolutional networks model for 2D skeleton-based human action recognition
    Weng, Libo
    Lou, Weidong
    Shen, Xin
    Gao, Fei
    IET IMAGE PROCESSING, 2023, 17 (03) : 773 - 783
  • [10] SKELETON-BASED HUMAN ACTION RECOGNITION USING SPATIAL TEMPORAL 3D CONVOLUTIONAL NEURAL NETWORKS
    Tu, Juanhui
    Liu, Mengyuan
    Liu, Hong
    2018 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2018,