Fusion of Appearance and Motion Features for Daily Activity Recognition from Egocentric Perspective

被引:2
|
作者
Lye, Mohd Haris [1 ]
AlDahoul, Nouar [1 ,2 ]
Abdul Karim, Hezerul [1 ]
机构
[1] Multimedia Univ, Fac Engn, Cyberjaya 63100, Selangor, Malaysia
[2] NYU, Comp Sci, POB 1291888, Abu Dhabi, U Arab Emirates
关键词
activities of daily living; convolutional neural network; egocentric vision; feature fusion; optical flow; DESCRIPTORS;
D O I
10.3390/s23156804
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Vidos from a first-person or egocentric perspective offer a promising tool for recognizing various activities related to daily living. In the egocentric perspective, the video is obtained from a wearable camera, and this enables the capture of the person's activities in a consistent viewpoint. Recognition of activity using a wearable sensor is challenging due to various reasons, such as motion blur and large variations. The existing methods are based on extracting handcrafted features from video frames to represent the contents. These features are domain-dependent, where features that are suitable for a specific dataset may not be suitable for others. In this paper, we propose a novel solution to recognize daily living activities from a pre-segmented video clip. The pre-trained convolutional neural network (CNN) model VGG16 is used to extract visual features from sampled video frames and then aggregated by the proposed pooling scheme. The proposed solution combines appearance and motion features extracted from video frames and optical flow images, respectively. The methods of mean and max spatial pooling (MMSP) and max mean temporal pyramid (TPMM) pooling are proposed to compose the final video descriptor. The feature is applied to a linear support vector machine (SVM) to recognize the type of activities observed in the video clip. The evaluation of the proposed solution was performed on three public benchmark datasets. We performed studies to show the advantage of aggregating appearance and motion features for daily activity recognition. The results show that the proposed solution is promising for recognizing activities of daily living. Compared to several methods on three public datasets, the proposed MMSP-TPMM method produces higher classification performance in terms of accuracy (90.38% with LENA dataset, 75.37% with ADL dataset, 96.08% with FPPA dataset) and average per-class precision (AP) (58.42% with ADL dataset and 96.11% with FPPA dataset).
引用
收藏
页数:20
相关论文
共 50 条
  • [1] Deep appearance and motion learning for egocentric activity recognition
    Wang, Xuanhan
    Gao, Lianli
    Song, Jingkuan
    Zhen, Xiantong
    Sebe, Nicu
    Shen, Heng Tao
    NEUROCOMPUTING, 2018, 275 : 438 - 447
  • [2] RECOGNIZING ACTIVITIES FROM EGOCENTRIC IMAGES WITH APPEARANCE AND MOTION FEATURES
    Chen, Yanhua
    Pei, Mingtao
    Nie, Zhengang
    2021 IEEE 31ST INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2021,
  • [3] Daily Activity Recognition Combining Gaze Motion and Visual Features
    Shiga, Yuki
    Dengel, Andreas
    Toyama, Takumi
    Kise, Koichi
    Utsumi, Yuzuko
    PROCEEDINGS OF THE 2014 ACM INTERNATIONAL JOINT CONFERENCE ON PERVASIVE AND UBIQUITOUS COMPUTING (UBICOMP'14 ADJUNCT), 2014, : 1103 - 1111
  • [4] Egocentric Daily Activity Recognition via Multitask Clustering
    Yan, Yan
    Ricci, Elisa
    Liu, Gaowen
    Sebe, Nicu
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2015, 24 (10) : 2984 - 2995
  • [5] Activity recognition using an egocentric perspective of everyday objects
    Surie, Dipak
    Pederson, Thomas
    Lagriffoul, Fabien
    Janlert, Lars-Erik
    Sjolie, Daniel
    UBIQUITOUS INTELLIGENCE AND COMPUTING, PROCEEDINGS, 2007, 4611 : 246 - +
  • [6] Multiple Features Fusion System for Motion Recognition
    Hua, Jiang
    Ju, Zhaojie
    Chen, Disi
    Zhou, Dalin
    Zhao, Haoyi
    Jiang, Du
    Li, Gongfa
    INTELLIGENT ROBOTICS AND APPLICATIONS, ICIRA 2019, PT IV, 2019, 11743 : 445 - 455
  • [7] Egocentric activity recognition using two-stage decision fusion
    Arabacı, Mehmet Ali
    Surer, Elif
    Temizel, Alptekin
    Neural Computing and Applications, 2024, 36 (36) : 22889 - 22903
  • [8] Robust Anomaly Detection via Fusion of Appearance and Motion Features
    Chen, Zhu
    Li, Weihai
    Fei, Chi
    Liu, Bin
    Yu, Nenghai
    2018 IEEE INTERNATIONAL CONFERENCE ON VISUAL COMMUNICATIONS AND IMAGE PROCESSING (IEEE VCIP), 2018,
  • [9] Action Recognition by Time Series of Retinotopic Appearance and Motion Features
    Barrett, Daniel Paul
    Siskind, Jeffrey Mark
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2016, 26 (12) : 2250 - 2263
  • [10] A perspective on human activity recognition from inertial motion data
    Gomaa, Walid
    Khamis, Mohamed A.
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (28): : 20463 - 20568