Activity Recognition from Multi-modal Sensor Data Using a Deep Convolutional Neural Network

被引:5
|
作者
Taherkhani, Aboozar [1 ]
Cosma, Georgina [1 ]
Alani, Ali A. [2 ]
McGinnity, T. M. [1 ]
机构
[1] Nottingham Trent Univ, Sch Sci & Technol, Nottingham, England
[2] Univ Diyala, Coll Sci, Comp Sci Dept, Diyala, Iraq
来源
关键词
Deep leering; Convolutional neural networks; Multi-modal data; Sensor data; Deep belief network; Smart home; Data fusion; ALGORITHM;
D O I
10.1007/978-3-030-01177-2_15
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-modal data extracted from different sensors in a smart home can be fused to build models that recognize the daily living activities of residents. This paper proposes a Deep Convolutional Neural Network to perform the activity recognition task using the multi-modal data collected from a smart residential home. The dataset contains accelerometer data (composed of three perpendicular components of acceleration and the strength of the accelerometer signal received by four receivers), video data (15 time-series related to 2D and 3D center of mass and bounding box extracted from an RGB-D camera), and Passive Infra-Red sensor data. The performance of the Deep Convolutional Neural Network is compared to the Deep Belief Network. Experimental results revealed that the Deep Convolutional Neural Network with two pairs of convolutional and max pooling layers achieved better classification accuracy than the Deep Belief Network. The Deep Belief Network uses Restricted Boltzmann Machines for pre-training the network. When training deep learning models using classes with a high number of training samples, the DBN achieved 65.97% classification accuracy, whereas the CNN achieved 75.33% accuracy. The experimental results demonstrate the challenges of dealing with multi-modal data and highlight the importance of having a suitable number of samples within each class for sufficiently training and testing deep learning models.
引用
收藏
页码:203 / 218
页数:16
相关论文
共 50 条
  • [31] Multi-sensor fusion based optimized deep convolutional neural network for boxing punch activity recognition
    Jayakumar, Brindha
    Govindarajan, Nallavan
    PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART P-JOURNAL OF SPORTS ENGINEERING AND TECHNOLOGY, 2024,
  • [32] An enhanced artificial neural network for hand gesture recognition using multi-modal features
    Uke, Shailaja N.
    Zade, Amol V.
    COMPUTER METHODS IN BIOMECHANICS AND BIOMEDICAL ENGINEERING-IMAGING AND VISUALIZATION, 2023, 11 (06): : 2278 - 2289
  • [33] On the Impact of Wireless Multimedia Network for Multi-Modal Activity Recognition
    Yamashita, Akika
    Lua, Eng Keong
    Oguchi, Masato
    2014 IEEE SYMPOSIUM ON COMPUTERS AND COMMUNICATION (ISCC), 2014,
  • [34] Convolutional Network With Twofold Feature Augmentation for Diabetic Retinopathy Recognition From Multi-Modal Images
    Hua, Cam-Hao
    Kim, Kiyoung
    Thien Huynh-The
    You, Jong In
    Yu, Seung-Young
    Le-Tien, Thuong
    Bae, Sung-Ho
    Lee, Sungyoung
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2021, 25 (07) : 2686 - 2697
  • [35] Multi-modal aggression identification using Convolutional Neural Network and Binary Particle Swarm Optimization
    Kumari, Kirti
    Singh, Jyoti Prakash
    Dwivedi, Yogesh K.
    Rana, Nripendra P.
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2021, 118 : 187 - 197
  • [36] Virtual Multi-modal Object Detection and Classification with Deep Convolutional Neural Networks
    Mitsakos, Nikolaos
    Papadakis, Manos
    WAVELETS AND SPARSITY XVIII, 2019, 11138
  • [37] Multi-modal wound classification using wound image and location by deep neural network
    Anisuzzaman, D. M.
    Patel, Yash
    Rostami, Behrouz
    Niezgoda, Jeffrey
    Gopalakrishnan, Sandeep
    Yu, Zeyun
    SCIENTIFIC REPORTS, 2022, 12 (01)
  • [38] Multi-modal speech emotion detection using optimised deep neural network classifier
    Padman, Sweta Nishant
    Magare, Dhiraj
    COMPUTER METHODS IN BIOMECHANICS AND BIOMEDICAL ENGINEERING-IMAGING AND VISUALIZATION, 2023, 11 (05): : 2020 - 2038
  • [39] A Deep GRU-BiLSTM Network for Multi-modal Emotion Recognition from Text
    Yacoubi, Ibtissem
    Ferjaoui, Radhia
    Ben Khalifa, Anouar
    2024 IEEE 7TH INTERNATIONAL CONFERENCE ON ADVANCED TECHNOLOGIES, SIGNAL AND IMAGE PROCESSING, ATSIP 2024, 2024, : 138 - 143
  • [40] Multi-Modal Brain Tumor Detection Using Deep Neural Network and Multiclass SVM
    Maqsood, Sarmad
    Damasevicius, Robertas
    Maskeliunas, Rytis
    MEDICINA-LITHUANIA, 2022, 58 (08):