Recognizing Personal Locations From Egocentric Videos

被引:28
|
作者
Furnari, Antonino [1 ]
Farinella, Giovanni Maria [1 ]
Battiato, Sebastiano [1 ]
机构
[1] Univ Catania, Dept Math & Comp Sci, I-95124 Catania, Italy
关键词
Context-aware computing; egocentric dataset; egocentric vision; first person vision; personal location recognition; CONTEXT; CLASSIFICATION; RECOGNITION; SCENE; SHAPE;
D O I
10.1109/THMS.2016.2612002
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Contextual awareness in wearable computing allows for construction of intelligent systems, which are able to interact with the user in a more natural way. In this paper, we study how personal locations arising from the user's daily activities can be recognized from egocentric videos. We assume that few training samples are available for learning purposes. Considering the diversity of the devices available on the market, we introduce a benchmark dataset containing egocentric videos of eight personal locations acquired by a user with four different wearable cameras. To make our analysis useful in real-world scenarios, we propose a method to reject negative locations, i.e., those not belonging to any of the categories of interest for the end-user. We assess the performances of the main state-of-the-art representations for scene and object classification on the considered task, as well as the influence of device-specific factors such as the field of view and the wearing modality. Concerning the different device-specific factors, experiments revealed that the best results are obtained using a head-mounted wide-angular device. Our analysis shows the effectiveness of using representations based on convolutional neural networks, employing basic transfer learning techniques and an entropy-based rejection algorithm.
引用
收藏
页码:6 / 18
页数:13
相关论文
共 50 条
  • [1] Recognizing Personal Contexts from Egocentric Images
    Furnari, Antonino
    Farinella, Giovanni M.
    Battiato, Sebastiano
    2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOP (ICCVW), 2015, : 393 - 401
  • [2] Recognizing Camera Wearer from Hand Gestures in Egocentric Videos
    Thapar, Daksh
    Nigam, Aditya
    Arora, Chetan
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 2095 - 2103
  • [3] Recognizing Micro-Actions and Reactions from Paired Egocentric Videos
    Yonetani, Ryo
    Kitani, Kris M.
    Sato, Yoichi
    2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 2629 - 2638
  • [4] Jointly Recognizing Object Fluents and Tasks in Egocentric Videos
    Liu, Yang
    Wei, Ping
    Zhu, Song-Chun
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 2943 - 2951
  • [5] Together Recognizing, Localizing and Summarizing Actions in Egocentric Videos
    Sahu, Abhimanyu
    Chowdhury, Ananda S.
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 4330 - 4340
  • [6] Anonymizing Egocentric Videos
    Thapar, Daksh
    Nigam, Aditya
    Arora, Chetan
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 2300 - 2309
  • [7] Head Motion Signatures from Egocentric Videos
    Poleg, Yair
    Arora, Chetan
    Peleg, Shmuel
    COMPUTER VISION - ACCV 2014, PT III, 2015, 9005 : 315 - 329
  • [8] Generic Action Recognition from Egocentric Videos
    Singh, Suriya
    Arora, Chetan
    Jawahar, C. V.
    2015 FIFTH NATIONAL CONFERENCE ON COMPUTER VISION, PATTERN RECOGNITION, IMAGE PROCESSING AND GRAPHICS (NCVPRIPG), 2015,
  • [9] Learning Navigation Subroutines from Egocentric Videos
    Kumar, Ashish
    Gupta, Saurabh
    Malik, Jitendra
    CONFERENCE ON ROBOT LEARNING, VOL 100, 2019, 100
  • [10] Market basket analysis from egocentric videos
    Santarcangelo, Vito
    Farinella, Giovanni Maria
    Furnari, Antonino
    Battiato, Sebastiano
    PATTERN RECOGNITION LETTERS, 2018, 112 : 83 - 90