A data augmentation method for human action recognition using dense joint motion images

被引:12
|
作者
Yao, Leiyue [1 ]
Yang, Wei [2 ]
Huang, Wei [1 ]
机构
[1] Nanchang Univ, Sch Informat Engn, 999 Xuefu Rd, Nanchang 330031, Jiangxi, Peoples R China
[2] Jiangxi Univ Technol, Ctr Collaborat & Innovat, 99 ZiYang Rd, Nanchang 330098, Jiangxi, Peoples R China
基金
中国国家自然科学基金;
关键词
Human action recognition; Motion image; Action encoding; Few-shot learning; Skeleton-based action recognition; ATTENTION;
D O I
10.1016/j.asoc.2020.106713
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the development of deep learning and neural network techniques, human action recognition has made great progress in recent years. However, it remains challenging to analyse temporal information and identify human actions with few training samples. In this paper, an effective motion image called a dense joint motion image (DJMI) was proposed to transform an action to an image. Our method was compared with state-of-the-art methods, and its contributions are mainly reflected in three characteristics. First, in contrast to the current classic joint trajectory map (JTM), every pixel of the DJMI is useful and contains essential spatio-temporal information. Thus, the input parameters of the deep neural network (DNN) are reduced by an order of magnitude, and the efficiency of action recognition is improved. Second, each frame of an action video is encoded as an independent slice of the DJMI, which avoids the information loss caused by action trajectory overlap. Third, by using DJMIs, proven algorithms for graphics and images can be used to generate training samples. Compared with the original image, the generated DJMIs contain new and different spatio-temporal information, which enables DNNs to be trained well on very few samples. Our method was evaluated on three benchmark datasets, namely, Florence-3D, UTKinect-Action3D and MSR Action3D. The results showed that our method achieved a recognition speed of 37 fps with competitive accuracy on these datasets. The time efficiency and few-shot learning capability of our method enable it to be used in real-time surveillance. (C) 2020 Elsevier B.V. All rights reserved.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] Deep network for human action recognition using Weber motion
    Chaudhary, Sachin
    Murala, Subrahmanyam
    NEUROCOMPUTING, 2019, 367 : 207 - 216
  • [42] Human Action Recognition in Videos Using Hybrid Motion Features
    Liu, Si
    Liu, Jing
    Zhang, Tianzhu
    Lu, Hanqing
    ADVANCES IN MULTIMEDIA MODELING, PROCEEDINGS, 2010, 5916 : 411 - 421
  • [43] Human Action Recognition Using Ordinal Measure of Accumulated Motion
    Wonjun Kim
    Jaeho Lee
    Minjin Kim
    Daeyoung Oh
    Changick Kim
    EURASIP Journal on Advances in Signal Processing, 2010
  • [44] Action Recognition of Motion Capture Data
    Lv, Na
    Feng, Zhiquan
    Ran, Lingqiang
    Zhao, Xiuyang
    2014 7TH INTERNATIONAL CONGRESS ON IMAGE AND SIGNAL PROCESSING (CISP 2014), 2014, : 22 - 26
  • [45] Relative Dense Tracklets for Human Action Recognition
    Bilinski, Piotr
    Corvee, Etienne
    Bak, Slawomir
    Bremond, Francois
    2013 10TH IEEE INTERNATIONAL CONFERENCE AND WORKSHOPS ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG), 2013,
  • [46] Data Augmentation Method for Disconnecting Switch Images and Its Application in Automatic State Recognition
    Liu Z.
    Wang H.
    Guan M.
    Huang X.
    Wu G.
    Wang D.
    Gaodianya Jishu/High Voltage Engineering, 2020, 46 (02): : 441 - 447
  • [47] Motion boundary emphasised optical flow method for human action recognition
    Peng, Cheng
    Huang, Haozhi
    Tsoi, Ah-Chung
    Lo, Sio-Long
    Liu, Yun
    Yang, Zi-yi
    IET COMPUTER VISION, 2020, 14 (06) : 378 - 390
  • [48] Human motion analysis and action recognition
    Karahoca, Adem
    Nurullahoglu, Murat
    NEW ASPECTS OF MICROELECTRONICS, NANOELECTRONICS, OPTOELECTRONICS, 2008, : 156 - 161
  • [49] Human Action Recognition in Still Images
    Palak
    Chaudhary, Sachin
    Communications in Computer and Information Science, 2022, 1568 CCIS : 483 - 493
  • [50] Learning Motion Features from Dynamic Images of Depth Video for Human Action Recognition
    Huang, Yao
    Yang, Jianyu
    Shao, Zhanpeng
    Li, Youfu
    2021 27TH INTERNATIONAL CONFERENCE ON MECHATRONICS AND MACHINE VISION IN PRACTICE (M2VIP), 2021,