Play and rewind: Context-aware video temporal action proposals

被引:17
|
作者
Gao, Lianli [1 ]
Li, Tao [1 ]
Song, Jingkuan [1 ]
Zhao, Zhou [2 ]
Shen, Heng Tao [1 ]
机构
[1] Univ Elect Sci & Technol China, Chengdu, Peoples R China
[2] Zhejiang Univ, Hangzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
Temporal action proposal generation and detection; Deep learning; Untrimmed video analysis; ACTION RECOGNITION;
D O I
10.1016/j.patcog.2020.107477
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we investigate the problem of Temporal Action Proposal (TAP) generation, which plays a fundamental role in large-scale untrimmed video analysis but remains largely unsolved. Most of the prior works proposed the temporal actions by predicting the temporal boundaries or actionness scores of video units. Nevertheless, context information among surrounding video units has not been adequately explored, which may result in severe loss of information. In this work, we propose a context-aware temporal action proposal network which makes full use of the contextual information in two aspects: 1) To generate initial proposals, we design a Bi-directional Parallel LSTMs to extract the visual features of a video unit by considering its contextual information. Therefore, the prediction of temporal boundaries and actionness scores will be more accurate because it knows what happened in the past and what will happen in the future; and 2) To refine the initial proposals, we design an action-attention based reranking network which considers both surrounding proposal and initial actionness scores to assign true action proposals with high confidence scores. Extensive experiments are conducted on two challenging datasets for both temporal action proposal generation and detection tasks, demonstrating the effectiveness of the proposed approach. In particular, on THUMOS'14 dataset, our method significantly surpasses state-of-the-art methods by 7.73% on AR@50. Our code is released at: https://github.com/Rheelt/TAPG. (C) 2020 Elsevier Ltd. All rights reserved.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Context-Aware Video Object Proposals
    Geng, Wenjing
    Wu, Gangshan
    2016 IEEE 22ND INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS (ICPADS), 2016, : 1203 - 1206
  • [2] Adaptive video object proposals by a context-aware model
    Wenjing Geng
    Chunlong Zhang
    Gangshan Wu
    Multimedia Tools and Applications, 2018, 77 : 10589 - 10614
  • [3] Adaptive video object proposals by a context-aware model
    Geng, Wenjing
    Zhang, Chunlong
    Wu, Gangshan
    MULTIMEDIA TOOLS AND APPLICATIONS, 2018, 77 (09) : 10589 - 10614
  • [4] Spatial-Temporal Context-Aware Online Action Detection and Prediction
    Huang, Jingjia
    Li, Nannan
    Li, Thomas
    Liu, Shan
    Li, Ge
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2020, 30 (08) : 2650 - 2662
  • [5] Context-Aware Surveillance Video Summarization
    Zhang, Shu
    Zhu, Yingying
    Roy-Chowdhury, Amit K.
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2016, 25 (11) : 5469 - 5478
  • [6] Context-aware Video Surveillance System
    An, Tae-Ki
    Kim, Moon-Hyun
    JOURNAL OF ELECTRICAL ENGINEERING & TECHNOLOGY, 2012, 7 (01) : 115 - 123
  • [7] Context-Aware Memory Attention Network for Video-Based Action Recognition
    Koh, Thean Chun
    Yeo, Chai Kiat
    Vaitesswar, U. S.
    Jing, Xuan
    2022 IEEE 14TH IMAGE, VIDEO, AND MULTIDIMENSIONAL SIGNAL PROCESSING WORKSHOP (IVMSP), 2022,
  • [8] Separately Guided Context-Aware Network for Weakly Supervised Temporal Action Detection
    Li, Bairong
    Pan, Yifan
    Liu, Ruixin
    Zhu, Yuesheng
    NEURAL PROCESSING LETTERS, 2023, 55 (05) : 6269 - 6288
  • [9] Separately Guided Context-Aware Network for Weakly Supervised Temporal Action Detection
    Bairong Li
    Yifan Pan
    Ruixin Liu
    Yuesheng Zhu
    Neural Processing Letters, 2023, 55 : 6269 - 6288
  • [10] A Novel Action Saliency and Context-Aware Network for Weakly-Supervised Temporal Action Localization
    Zhao, Yibo
    Zhang, Hua
    Gao, Zan
    Gao, Wenjie
    Wang, Meng
    Chen, Shengyong
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 8253 - 8266