Egocentric action anticipation from untrimmed videos

被引:0
|
作者
Rodin, Ivan [1 ]
Furnari, Antonino [1 ,2 ]
Farinella, Giovanni Maria [1 ,2 ]
机构
[1] Univ Catania, Catania, Italy
[2] Univ Catania, Next Vis srl Spinoff, Catania, Italy
关键词
computer vision; pattern recognition;
D O I
10.1049/cvi2.12342
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Egocentric action anticipation involves predicting future actions performed by the camera wearer from egocentric video. Although the task has recently gained attention in the research community, current approaches often assume that input videos are 'trimmed', meaning that a short video sequence is sampled a fixed time before the beginning of the action. However, trimmed action anticipation has limited applicability in real-world scenarios, where it is crucial to deal with 'untrimmed' video inputs and the exact moment of action initiation cannot be assumed at test time. To address these limitations, an untrimmed action anticipation task is proposed, which, akin to temporal action detection, assumes that the input video is untrimmed at test time, while still requiring predictions to be made before actions take place. The authors introduce a benchmark evaluation procedure for methods designed to address this novel task and compare several baselines on the EPIC-KITCHENS-100 dataset. Through our experimental evaluation, testing a variety of models, the authors aim to better understand their performance in untrimmed action anticipation. Our results reveal that the performance of current models designed for trimmed action anticipation is limited, emphasising the need for further research in this area.
引用
收藏
页数:11
相关论文
共 50 条
  • [41] Temporal Action Localization in Untrimmed Videos via Multi-stage CNNs
    Shou, Zheng
    Wang, Dongang
    Chang, Shih-Fu
    2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 1049 - 1058
  • [42] Unsupervised Segmentation of Action Segments in Egocentric Videos using Gaze
    Hipiny
    Ujir, H.
    Minoi, J. L.
    Juan, S. F. Samson
    Khairuddin, M. A.
    Sunar, M. S.
    2017 IEEE INTERNATIONAL CONFERENCE ON SIGNAL AND IMAGE PROCESSING APPLICATIONS (ICSIPA), 2017, : 351 - 356
  • [43] Rethinking Online Action Detection in Untrimmed Videos: A Novel Online Evaluation Protocol
    Baptista-Rios, Marcos
    Lopez-Sastre, Roberto J.
    Caba Heilbron, Fabian
    Van Gemert, Jan C.
    Acevedo-Rodriguez, F. Javier
    Maldonado-Bascon, Saturnino
    IEEE ACCESS, 2020, 8 : 5139 - 5146
  • [44] What If We Could Not See? Counterfactual Analysis for Egocentric Action Anticipation
    Zhang, Tianyu
    Min, Weiqing
    Yang, Jiahao
    Liu, Tao
    Jiang, Shuqiang
    Rui, Yong
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 1316 - 1322
  • [45] The Blessings of Unlabeled Background in Untrimmed Videos
    Liu, Yuan
    Chen, Jingyuan
    Chen, Zhenfang
    Deng, Bing
    Huang, Jianqiang
    Zhang, Hanwang
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 6172 - 6181
  • [46] Anonymizing Egocentric Videos
    Thapar, Daksh
    Nigam, Aditya
    Arora, Chetan
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 2300 - 2309
  • [47] A Proposal-Based Solution to Spatio-Temporal Action Detection in Untrimmed Videos
    Gleason, Joshua
    Ranjan, Rajeev
    Schwarcz, Steven
    Castillo, Carlos D.
    Chen, Jun-Cheng
    Chellappa, Rama
    2019 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2019, : 141 - 150
  • [48] Temporal Cricket Stroke Localization from Untrimmed Highlight Videos
    Gupta, Arpan
    Balan, Sakthi M.
    ELEVENTH INDIAN CONFERENCE ON COMPUTER VISION, GRAPHICS AND IMAGE PROCESSING (ICVGIP 2018), 2018,
  • [49] Head Motion Signatures from Egocentric Videos
    Poleg, Yair
    Arora, Chetan
    Peleg, Shmuel
    COMPUTER VISION - ACCV 2014, PT III, 2015, 9005 : 315 - 329
  • [50] Learning Navigation Subroutines from Egocentric Videos
    Kumar, Ashish
    Gupta, Saurabh
    Malik, Jitendra
    CONFERENCE ON ROBOT LEARNING, VOL 100, 2019, 100