Approaching the Real-World: Supporting Activity Recognition Training with Virtual IMU Data

被引:22
|
作者
Kwon, Hyeokhyen [1 ]
Wang, Bingyao [2 ]
Abowd, Gregory D. [3 ]
Ploetz, Thomas [1 ]
机构
[1] Georgia Inst Technol, Sch Interact Comp, Atlanta, GA 30332 USA
[2] Georgia Inst Technol, Coll Comp, Atlanta, GA 30332 USA
[3] Northeastern Univ, Dept Elect & Comp Engn, Boston, MA 02115 USA
关键词
Activity Recognition; Data Collection; Machine Learning;
D O I
10.1145/3478096
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recently, IMUTube introduced a paradigm change for bootstrapping human activity recognition (HAR) systems for wearables. The key idea is to utilize videos of activities to support training activity recognizers based on inertial measurement units (IMUs). This system retrieves video from public repositories and subsequently generates virtual IMU data from this. The ultimate vision for such a system is to make large amounts of weakly labeled videos accessible for model training in HAR and, as such, to overcome one of the most pressing issues in the field: the lack of significant amounts of labeled sample data. In this paper we present the first in-detail exploration of IMUTube in a realistic assessment scenario: the analysis of free-weight gym exercises. We make significant progress towards a flexible, fully-functional IMUTube system by extending it such that it can handle a range of artifacts that are common in unrestricted online videos, including various forms of video noise, non-human poses, body part occlusions, and extreme camera and human motion. By overcoming these real-world challenges, we are able to generate high-quality virtual IMU data, which allows us to employ IMUTube for practical analysis tasks. We show that HAR systems trained by incorporating virtual sensor data generated by IMUTube significantly outperform baseline models trained only with real IMU data. In doing so we demonstrate the practical utility of IMUTube and the progress made towards the final vision of the new bootstrapping paradigm.
引用
收藏
页数:32
相关论文
共 50 条
  • [1] Challenges in data collection in real-world environments for activity recognition
    Lameski, Petre
    Dimitrievski, Ace
    Zdravevski, Eftim
    Trajkovik, Vladimir
    Koceski, Saso
    PROCEEDINGS OF 18TH INTERNATIONAL CONFERENCE ON SMART TECHNOLOGIES (IEEE EUROCON 2019), 2019,
  • [2] Approaching Real-World Navigation Using Object Recognition Network
    Zheng, Zejia
    Weng, Juyang
    2015 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2015,
  • [3] Approaching Camera-based Real-World Navigation Using Object Recognition
    Zheng, Zejia
    He, Xie
    Weng, Juyang
    INNS CONFERENCE ON BIG DATA 2015 PROGRAM, 2015, 53 : 428 - 436
  • [4] Real-World Battles with Real-World Data
    Brown, Jeffrey
    Bate, Andrew
    Platt, Robert
    Raebel, Marsha
    Sauer, Brian
    Trifiro, Gianluca
    PHARMACOEPIDEMIOLOGY AND DRUG SAFETY, 2017, 26 : 254 - 255
  • [5] Real-world study: from real-world data to real-world evidence
    Wen, Yi
    TRANSLATIONAL BREAST CANCER RESEARCH, 2020, 1
  • [6] Real-world virtual reality
    Hirose, M
    TELECOMMUNICATION, TELEIMMERSION AND TELEXISTENCE II, 2005, : 89 - +
  • [7] Real-world Data for Virtual Reality Experiences: Interpreting Excavations
    Antlej, Kaja
    Bykersma, Matthew
    Mortimer, Michael
    Vickers-Rich, Pat
    Rich, Thomas
    Horan, Ben
    2018 3RD DIGITAL HERITAGE INTERNATIONAL CONGRESS (DIGITALHERITAGE) HELD JOINTLY WITH 2018 24TH INTERNATIONAL CONFERENCE ON VIRTUAL SYSTEMS & MULTIMEDIA (VSMM 2018), 2018, : 338 - 345
  • [8] Translating real-world evidence/real-world data
    Ravenstijn, Paulien
    CTS-CLINICAL AND TRANSLATIONAL SCIENCE, 2024, 17 (05):
  • [9] Real-World Skills in a Virtual World: An Innovative Activity for Teaching Developmental Psychopathology
    Koch, Mary Kate
    Mendle, Jane
    TEACHING OF PSYCHOLOGY, 2023, 50 (04) : 427 - 432
  • [10] Video-based face recognition on real-world data
    Stallkamp, Johannes
    Ekenel, Hazim K.
    Stiefelhagen, Rainer
    2007 IEEE 11TH INTERNATIONAL CONFERENCE ON COMPUTER VISION, VOLS 1-6, 2007, : 313 - 320