Action-Aware Network with Upper and Lower Limit Loss for Weakly-Supervised Temporal Action Localization

被引:0
|
作者
Mingwen Bi
Jiaqi Li
Xinliang Liu
Qingchuan Zhang
Zhenghong Yang
机构
[1] China Agricultural University,College of Information and Electrical Engineering
[2] Ministry of Agriculture and Rural Affairs,Key Laboratory of Agricultural Informatization Standardization
[3] Beijing Technology and Business University,National Engineering Research Center for Agri
来源
Neural Processing Letters | 2023年 / 55卷
关键词
Weakly-supervised learning; Temporal action localization; Upper and lower limit loss; Action-aware network;
D O I
暂无
中图分类号
学科分类号
摘要
Weakly-supervised temporal action localization aims to detect the temporal boundaries of action instances in untrimmed videos only by relying on video-level action labels. The main challenge of the research is to accurately segment the action from the background in the absence of frame-level labels. Previous methods consider the action-related context in the background as the main factor restricting the segmentation performance. Most of them take action labels as pseudo-labels for context and suppress context frames in class activation sequences using the attention mechanism. However, this only applies to fixed shots or videos with a single theme. For videos with frequent scene switching and complicated themes, such as casual shots of unexpected events and secret shots, the strong randomness and weak continuity of the action cause the assumption not to be valid. In addition, the wrong pseudo-labels will enhance the weight of context frames, which will affect the segmentation performance. To address above issues, in this paper, we define a new video frame division standard (action instance, action-related context, no-action background), propose an Action-aware Network with Upper and Lower loss AUL-Net, which limits the activation of context to a reasonable range through a two-branch weight-sharing framework with a three-branch attention mechanism, so that the model has wider applicability while accurately suppressing context and background. We conducted extensive experiments on the self-built food safety video dataset FS-VA, and the results show that our method outperforms the state-of-the-art model.
引用
收藏
页码:4307 / 4324
页数:17
相关论文
共 50 条
  • [41] Entropy guided attention network for weakly-supervised action localization
    Cheng, Yi
    Sun, Ying
    Fan, Hehe
    Zhuo, Tao
    Lim, Joo-Hwee
    Kankanhalli, Mohan
    PATTERN RECOGNITION, 2022, 129
  • [42] W-ART: ACTION RELATION TRANSFORMER FOR WEAKLY-SUPERVISED TEMPORAL ACTION LOCALIZATION
    Li, Mengzhu
    Wu, Hongjun
    Liu, Yongcheng
    Liu, Hongzhe
    Xu, Cheng
    Li, Xuewei
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 2195 - 2199
  • [43] Semantic and Temporal Contextual Correlation Learning for Weakly-Supervised Temporal Action Localization
    Fu, Jie
    Gao, Junyu
    Xu, Changsheng
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (10) : 12427 - 12443
  • [44] Action-aware Masking Network with Group-based Attention for Temporal Action Localization
    Kang, Tae-Kyung
    Lee, Gun-Hee
    Jin, Kyung-Min
    Lee, Seong-Whan
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 6047 - 6056
  • [45] CoLA: Weakly-Supervised Temporal Action Localization with Snippet Contrastive Learning
    Zhang, Can
    Cao, Meng
    Yang, Dongming
    Chen, Jie
    Zou, Yuexian
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 16005 - 16014
  • [46] Diffusion-based framework for weakly-supervised temporal action localization
    Zou, Yuanbing
    Zhao, Qingjie
    Sarker, Prodip Kumar
    Li, Shanshan
    Wang, Lei
    Liu, Wangwang
    Pattern Recognition, 2025, 160
  • [47] Adversarial Seeded Sequence Growing for Weakly-Supervised Temporal Action Localization
    Zhang, Chengwei
    Xu, Yunlu
    Cheng, Zhanzhan
    Niu, Yi
    Pu, Shiliang
    Wu, Fei
    Zou, Futai
    PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA (MM'19), 2019, : 738 - 746
  • [48] Dual-Evidential Learning for Weakly-supervised Temporal Action Localization
    Chen, Mengyuan
    Gao, Junyu
    Yang, Shicai
    Xu, Changsheng
    COMPUTER VISION - ECCV 2022, PT IV, 2022, 13664 : 192 - 208
  • [49] Learning Background Suppression Model for Weakly-supervised Temporal Action Localization
    Liu, Mengxue
    Gao, Xiangjun
    Ge, Fangzhen
    Liu, Huaiyu
    Li, Wenjing
    IAENG International Journal of Computer Science, 2021, 48 (04):
  • [50] Unleashing the Potential of Adjacent Snippets for Weakly-supervised Temporal Action Localization
    Liu, Qinying
    Wang, Zilei
    Chen, Ruoxi
    Li, Zhilin
    Proceedings - IEEE International Conference on Multimedia and Expo, 2023, 2023-July : 1032 - 1037