Human activity recognition based on multi-modal fusion

被引:0
|
作者
Zhang, Cheng [1 ]
Zu, Tianqi [1 ]
Hou, Yibin [1 ,2 ]
He, Jian [1 ,2 ]
Yang, Shengqi [1 ,2 ]
Dong, Ruihai [3 ]
机构
[1] Beijing Univ Technol, Fac Informat Technol, Beijing 100124, Peoples R China
[2] Beijing Univ Technol, Beijing Engn Res Ctr IoT Software & Syst, Beijing 100124, Peoples R China
[3] Univ Coll Dublin, Insight Ctr Data Analyt, Dublin, Ireland
关键词
Human activity recognition; Multi-modal fusion; Fall detection; Convolutional network; Wearable device;
D O I
10.1007/s42486-023-00132-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, human activity recognition (HAR) methods are developing rapidly. However, most existing methods base on single input data modality, and suffers from accuracy and robustness issues. In this paper, we present a novel multi-modal HAR architecture which fuses signals from both RGB visual data and Inertial Measurement Units (IMU) data. As for the RGB modality, the speed-weighted star RGB representation is proposed to aggregate the temporal information, and a convolutional network is employed to extract features; As for the IMU modality, Fast Fourier transform and multi-layer perceptron are employed to extract the dynamical features of IMU data. As for the feature fusion scheme, the global soft attention layer is designed to adjust the weights according to the concatenated features, and the L-softmax with soft voting is adopted to classify activities. The proposed method is evaluated on the UP-Fall dataset, the F1-scores are 0.92 and 1.00 for 11 classes classification task and fall/non-fall binary classification task respectively.
引用
收藏
页码:321 / 332
页数:12
相关论文
共 50 条
  • [41] Multi-modal fusion network with complementarity and importance for emotion recognition
    Liu, Shuai
    Gao, Peng
    Li, Yating
    Fu, Weina
    Ding, Weiping
    INFORMATION SCIENCES, 2023, 619 : 679 - 694
  • [42] Adaptive information fusion network for multi-modal personality recognition
    Bao, Yongtang
    Liu, Xiang
    Qi, Yue
    Liu, Ruijun
    Li, Haojie
    COMPUTER ANIMATION AND VIRTUAL WORLDS, 2024, 35 (03)
  • [43] Multi-Evidence and Multi-Modal Fusion Network for Ground-Based Cloud Recognition
    Liu, Shuang
    Li, Mei
    Zhang, Zhong
    Xiao, Baihua
    Durrani, Tariq S.
    REMOTE SENSING, 2020, 12 (03)
  • [44] MULTI-MODAL FUSION WITH OBSERVATION POINTS FOR SKELETON ACTION RECOGNITION
    Singh, Iqbal
    Zhu, Xiaodan
    Greenspan, Michael
    2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 1781 - 1785
  • [45] Multi-modal Emotion Recognition Based on Hypergraph
    Zong L.-L.
    Zhou J.-H.
    Xie Q.-J.
    Zhang X.-C.
    Xu B.
    Jisuanji Xuebao/Chinese Journal of Computers, 2023, 46 (12): : 2520 - 2534
  • [46] Template co-updating in multi-modal human activity recognition systems
    Franco, Annalisa
    Magnani, Antonio
    Maio, Dario
    PROCEEDINGS OF THE 35TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING (SAC'20), 2020, : 2113 - 2116
  • [47] Multi-modal recognition of worker activity for human-centered intelligent manufacturing
    Tao, Wenjin
    Leu, Ming C.
    Yin, Zhaozheng
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2020, 95 (95)
  • [48] Multi-Modal Convolutional Neural Networks for Activity Recognition
    Ha, Sojeong
    Yun, Jeong-Min
    Choi, Seungjin
    2015 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC 2015): BIG DATA ANALYTICS FOR HUMAN-CENTRIC SYSTEMS, 2015, : 3017 - 3022
  • [49] Multi-Modal Fusion Emotion Recognition Method of Speech Expression Based on Deep Learning
    Liu, Dong
    Wang, Zhiyong
    Wang, Lifeng
    Chen, Longxi
    FRONTIERS IN NEUROROBOTICS, 2021, 15
  • [50] Multi-Modal Fusion Sign Language Recognition Based on Residual Network and Attention Mechanism
    Chu Chaoqin
    Xiao Qinkun
    Zhang Yinhuan
    Xing, Liu
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2022, 36 (12)