Human activity recognition based on multi-modal fusion

被引:0
|
作者
Zhang, Cheng [1 ]
Zu, Tianqi [1 ]
Hou, Yibin [1 ,2 ]
He, Jian [1 ,2 ]
Yang, Shengqi [1 ,2 ]
Dong, Ruihai [3 ]
机构
[1] Beijing Univ Technol, Fac Informat Technol, Beijing 100124, Peoples R China
[2] Beijing Univ Technol, Beijing Engn Res Ctr IoT Software & Syst, Beijing 100124, Peoples R China
[3] Univ Coll Dublin, Insight Ctr Data Analyt, Dublin, Ireland
关键词
Human activity recognition; Multi-modal fusion; Fall detection; Convolutional network; Wearable device;
D O I
10.1007/s42486-023-00132-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, human activity recognition (HAR) methods are developing rapidly. However, most existing methods base on single input data modality, and suffers from accuracy and robustness issues. In this paper, we present a novel multi-modal HAR architecture which fuses signals from both RGB visual data and Inertial Measurement Units (IMU) data. As for the RGB modality, the speed-weighted star RGB representation is proposed to aggregate the temporal information, and a convolutional network is employed to extract features; As for the IMU modality, Fast Fourier transform and multi-layer perceptron are employed to extract the dynamical features of IMU data. As for the feature fusion scheme, the global soft attention layer is designed to adjust the weights according to the concatenated features, and the L-softmax with soft voting is adopted to classify activities. The proposed method is evaluated on the UP-Fall dataset, the F1-scores are 0.92 and 1.00 for 11 classes classification task and fall/non-fall binary classification task respectively.
引用
收藏
页码:321 / 332
页数:12
相关论文
共 50 条
  • [21] MMHAR-EnsemNet: A Multi-Modal Human Activity Recognition Model
    Das, Avigyan
    Sil, Pritam
    Singh, Pawan Kumar
    Bhateja, Vikrant
    Sarkar, Ram
    IEEE SENSORS JOURNAL, 2021, 21 (10) : 11569 - 11576
  • [22] Multi-modal Fusion
    Liu, Huaping
    Hussain, Amir
    Wang, Shuliang
    INFORMATION SCIENCES, 2018, 432 : 462 - 462
  • [23] Electromagnetic signal feature fusion and recognition based on multi-modal deep learning
    Hou C.
    Zhang X.
    Chen X.
    International Journal of Performability Engineering, 2020, 16 (06): : 941 - 949
  • [24] A Hand Features Based Fusion Recognition Network with Enhancing Multi-Modal Correlation
    Wu, Wei
    Zhang, Yuan
    Li, Yunpeng
    Li, Chuanyang
    Hao, Yan
    CMES-COMPUTER MODELING IN ENGINEERING & SCIENCES, 2024, 140 (01): : 537 - 555
  • [25] Multi-modal fusion for robust hand gesture recognition based on heterogeneous networks
    ZOU YongXiang
    CHENG Long
    HAN LiJun
    LI ZhengWei
    Science China(Technological Sciences), 2023, (11) : 3219 - 3230
  • [26] 3D shape recognition based on multi-modal information fusion
    Qi Liang
    Mengmeng Xiao
    Dan Song
    Multimedia Tools and Applications, 2021, 80 : 16173 - 16184
  • [27] Multi-modal fusion for robust hand gesture recognition based on heterogeneous networks
    YongXiang Zou
    Long Cheng
    LiJun Han
    ZhengWei Li
    Science China Technological Sciences, 2023, 66 : 3219 - 3230
  • [29] Multi-modal video event recognition based on association rules and decision fusion
    Guder, Mennan
    Cicekli, Nihan Kesim
    MULTIMEDIA SYSTEMS, 2018, 24 (01) : 55 - 72
  • [30] Multi-modal fusion for robust hand gesture recognition based on heterogeneous networks
    Zou, Yongxiang
    Cheng, Long
    Han, Lijun
    Li, Zhengwei
    SCIENCE CHINA-TECHNOLOGICAL SCIENCES, 2023, 66 (11) : 3219 - 3230