Sports action recognition algorithm based on multi-modal data recognition

被引:0
|
作者
Zhang, Lin [1 ]
机构
[1] Jilin Province Economic Management Cadre College, Changchun, China
关键词
Data fusion - Musculoskeletal system - Spatio-temporal data;
D O I
10.3233/IDT-230372
中图分类号
学科分类号
摘要
The recognition of sports action is an important research subject, which is conducive to the improvement of athletes' own level. To improve the accuracy of multi-modal data action recognition, based on the Transformer module, this study introduces a multi-head attention mechanism, fuses multi-modal data, and constructs a multi-stream structured object relationship inference network. Based on PointNet++ network and combining five different data fusion frameworks, a motion recognition model that integrates RGB data and 3D skeleton point cloud is constructed. The results showed that the Top-1 accuracy of multi-stream structured object relationship inference network was 42.5% and 42.7%, respectively, which was better than other algorithms. The accuracy of the multi-modal fusion model was improved by 15.6% and 5.1% compared with the single mode, and by 5.4% and 2.6% compared with the dual mode, which showed its superiority in the action recognition task. This showed that the fusion of multi-modal data can provide more abundant information, so as to improve the accuracy of action recognition. The accuracy of the action recognition model combining RGB data and 3D skeleton point cloud was 84.3%, 87.5%, 90.2%, 90.6% and 91.2% after the combination of different strategies, which effectively compensated for the problem of missing information in 3D skeleton point cloud and significantly improved the accuracy of action recognition. With a small amount of data, the Top-1 accuracy of the multi-stream structured object relationship inference network in this study was superior to other algorithms, showing its advantages in dealing with complex action recognition tasks. In addition, the action recognition model that fuses RGB data and 3D skeleton point cloud also achieved higher accuracy, which is better than other algorithms. This study can meet the needs of motion recognition in different scenarios and has certain reference value. © 2024 - IOS Press. All rights reserved.
引用
收藏
页码:3243 / 3257
相关论文
共 50 条
  • [21] Research on unified recognition model and algorithm for multi-modal gestures
    Guo Xiaopei
    Feng Zhiquan
    Sun Kaiyun
    Liu Hong
    Xie Wei
    Bi Jianping
    The Journal of China Universities of Posts and Telecommunications, 2019, 26 (02) : 30 - 42
  • [22] Research on unified recognition model and algorithm for multi-modal gestures
    Xiaopei G.
    Zhiquan F.
    Kaiyun S.
    Hong L.
    Wei X.
    Jianping B.
    Journal of China Universities of Posts and Telecommunications, 2019, 26 (02): : 30 - 42
  • [23] Multi-modal Emotion Recognition Based on Speech and Image
    Li, Yongqiang
    He, Qi
    Zhao, Yongping
    Yao, Hongxun
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING - PCM 2017, PT I, 2018, 10735 : 844 - 853
  • [24] Gesture recognition based on multi-modal feature weight
    Duan, Haojie
    Sun, Ying
    Cheng, Wentao
    Jiang, Du
    Yun, Juntong
    Liu, Ying
    Liu, Yibo
    Zhou, Dalin
    Concurrency and Computation: Practice and Experience, 2021, 33 (05)
  • [25] DCNN and DNN Based Multi-modal Depression Recognition
    Yang, Le
    Jiang, Dongmei
    Han, Wenjing
    Sahli, Hichem
    2017 SEVENTH INTERNATIONAL CONFERENCE ON AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION (ACII), 2017, : 484 - 489
  • [26] Human activity recognition based on multi-modal fusion
    Zhang, Cheng
    Zu, Tianqi
    Hou, Yibin
    He, Jian
    Yang, Shengqi
    Dong, Ruihai
    CCF TRANSACTIONS ON PERVASIVE COMPUTING AND INTERACTION, 2023, 5 (03) : 321 - 332
  • [27] Human activity recognition based on multi-modal fusion
    Cheng Zhang
    Tianqi Zu
    Yibin Hou
    Jian He
    Shengqi Yang
    Ruihai Dong
    CCF Transactions on Pervasive Computing and Interaction, 2023, 5 : 321 - 332
  • [28] Gesture recognition based on multi-modal feature weight
    Duan, Haojie
    Sun, Ying
    Cheng, Wentao
    Jiang, Du
    Yun, Juntong
    Liu, Ying
    Liu, Yibo
    Zhou, Dalin
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2021, 33 (05):
  • [29] Multi-information Complementarity Neural Networks for Multi-Modal Action Recognition
    Ding, Chuan
    Tie, Yun
    Qi, Lin
    2019 8TH INTERNATIONAL SYMPOSIUM ON NEXT GENERATION ELECTRONICS (ISNE), 2019,
  • [30] Lightweight multi-modal emotion recognition model based on modal generation
    Liu, Peisong
    Che, Manqiang
    Luo, Jiangchuan
    2022 9TH INTERNATIONAL FORUM ON ELECTRICAL ENGINEERING AND AUTOMATION, IFEEA, 2022, : 430 - 435