Sports action recognition algorithm based on multi-modal data recognition

被引:0
|
作者
Zhang, Lin [1 ]
机构
[1] Jilin Province Economic Management Cadre College, Changchun, China
关键词
Data fusion - Musculoskeletal system - Spatio-temporal data;
D O I
10.3233/IDT-230372
中图分类号
学科分类号
摘要
The recognition of sports action is an important research subject, which is conducive to the improvement of athletes' own level. To improve the accuracy of multi-modal data action recognition, based on the Transformer module, this study introduces a multi-head attention mechanism, fuses multi-modal data, and constructs a multi-stream structured object relationship inference network. Based on PointNet++ network and combining five different data fusion frameworks, a motion recognition model that integrates RGB data and 3D skeleton point cloud is constructed. The results showed that the Top-1 accuracy of multi-stream structured object relationship inference network was 42.5% and 42.7%, respectively, which was better than other algorithms. The accuracy of the multi-modal fusion model was improved by 15.6% and 5.1% compared with the single mode, and by 5.4% and 2.6% compared with the dual mode, which showed its superiority in the action recognition task. This showed that the fusion of multi-modal data can provide more abundant information, so as to improve the accuracy of action recognition. The accuracy of the action recognition model combining RGB data and 3D skeleton point cloud was 84.3%, 87.5%, 90.2%, 90.6% and 91.2% after the combination of different strategies, which effectively compensated for the problem of missing information in 3D skeleton point cloud and significantly improved the accuracy of action recognition. With a small amount of data, the Top-1 accuracy of the multi-stream structured object relationship inference network in this study was superior to other algorithms, showing its advantages in dealing with complex action recognition tasks. In addition, the action recognition model that fuses RGB data and 3D skeleton point cloud also achieved higher accuracy, which is better than other algorithms. This study can meet the needs of motion recognition in different scenarios and has certain reference value. © 2024 - IOS Press. All rights reserved.
引用
收藏
页码:3243 / 3257
相关论文
共 50 条
  • [41] Finding Achilles' Heel: Adversarial Attack on Multi-modal Action Recognition
    Kumar, Deepak
    Kumar, Chetan
    Seah, Chun Wei
    Xia, Siyu
    Shao, Ming
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 3829 - 3837
  • [42] Graph-Based Multi-Modal Multi-View Fusion for Facial Action Unit Recognition
    Chen, Jianrong
    Dey, Sujit
    IEEE ACCESS, 2024, 12 : 69310 - 69324
  • [43] Cross-modal learning with multi-modal model for video action recognition based on adaptive weight training
    Zhou, Qingguo
    Hou, Yufeng
    Zhou, Rui
    Li, Yan
    Wang, Jinqiang
    Wu, Zhen
    Li, Hung-Wei
    Weng, Tien-Hsiung
    CONNECTION SCIENCE, 2024, 36 (01)
  • [44] Speech recognition with multi-modal features based on neural networks
    Kim, Myung Won
    Ryu, Joung Woo
    Kim, Eun Ju
    NEURAL INFORMATION PROCESSING, PT 2, PROCEEDINGS, 2006, 4233 : 489 - 498
  • [45] Multi-modal human motion recognition based on behaviour tree
    Yang, Qin
    Zhou, Zhenhua
    INTERNATIONAL JOURNAL OF BIOMETRICS, 2024, 16 (3-4) : 381 - 398
  • [46] Empirical Mode Decomposition Based Multi-Modal Activity Recognition
    Hu, Lingyue
    Zhao, Kailong
    Zhou, Xueling
    Ling, Bingo Wing-Kuen
    Liao, Guozhao
    SENSORS, 2020, 20 (21) : 1 - 15
  • [47] Multi-Modal Pain Intensity Recognition Based on the SenseEmotion Database
    Thiam, Patrick
    Kessler, Viktor
    Amirian, Mohammadreza
    Bellmann, Peter
    Layher, Georg
    Zhang, Yan
    Velana, Maria
    Gruss, Sascha
    Walter, Steffen
    Traue, Harald C.
    Schork, Daniel
    Kim, Jonghwa
    Andre, Elisabeth
    Neumann, Heiko
    Schwenker, Friedhelm
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2021, 12 (03) : 743 - 760
  • [48] Multi-modal haptic image recognition based on deep learning
    Han, Dong
    Nie, Hong
    Chen, Jinbao
    Chen, Meng
    Deng, Zhen
    Zhang, Jianwei
    SENSOR REVIEW, 2018, 38 (04) : 486 - 493
  • [49] Multi-Modal Fusion Emotion Recognition Based on HMM and ANN
    Xu, Chao
    Cao, Tianyi
    Feng, Zhiyong
    Dong, Caichao
    CONTEMPORARY RESEARCH ON E-BUSINESS TECHNOLOGY AND STRATEGY, 2012, 332 : 541 - 550
  • [50] Geological Body Recognition Based on Multi-Modal Feature Fusion
    Fu S.
    Li C.
    Zhang H.
    Liu C.
    Li F.
    Diqiu Kexue - Zhongguo Dizhi Daxue Xuebao/Earth Science - Journal of China University of Geosciences, 2023, 48 (10): : 3743 - 3752