Weakly Supervised Actor-Action Segmentation via Robust Multi-Task Ranking

被引:17
|
作者
Yan, Yan [1 ]
Xu, Chenliang [2 ]
Cai, Dawen [3 ]
Corso, Jason J. [1 ]
机构
[1] Univ Michigan, Dept Elect Engn & Comp Sci, Ann Arbor, MI 48109 USA
[2] Univ Rochester, Dept Comp Sci, Rochester, NY 14627 USA
[3] Univ Michigan, Dept Cell & Dev Biol, Biophys, Ann Arbor, MI 48109 USA
关键词
D O I
10.1109/CVPR.2017.115
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Fine-grained activity understanding in videos has attracted considerable recent attention with a shift from action classification to detailed actor and action understanding that provides compelling results for perceptual needs of cutting-edge autonomous systems. However, current methods for detailed understanding of actor and action have significant limitations: they require large amounts of finely labeled data, and they fail to capture any internal relationship among actors and actions. To address these issues, in this paper, we propose a novel, robust multi-task ranking model for weakly-supervised actor-action segmentation where only video-level tags are given for training samples. Our model is able to share useful information among different actors and actions while learning a ranking matrix to select representative supervoxels for actors and actions respectively. Final segmentation results are generated by a conditional random field that considers various ranking scores for video parts. Extensive experimental results on the Actor-Action Dataset (A2D) demonstrate that the proposed approach outperforms the state-of-the-art weakly supervised methods and performs as well as the topperforming fully supervised method.
引用
收藏
页码:1022 / 1031
页数:10
相关论文
共 50 条
  • [21] Multi-task Supervised Learning via Cross-learning
    Cervino, Juan
    Andres Bazerque, Juan
    Calvo-Fullana, Miguel
    Ribeiro, Alejandro
    29TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2021), 2021, : 1381 - 1385
  • [22] MULTI-TASK SELF-SUPERVISED LEARNING FOR ROBUST SPEECH RECOGNITION
    Ravanelli, Mirco
    Zhong, Jianyuan
    Pascual, Santiago
    Swietojanski, Pawel
    Monteiro, Joao
    Trmal, Jan
    Bengio, Yoshua
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 6989 - 6993
  • [23] Stock Ranking with Multi-Task Learning
    Ma, Tao
    Tan, Ying
    EXPERT SYSTEMS WITH APPLICATIONS, 2022, 199
  • [24] Semantic Segmentation via Multi-task, Multi-domain Learning
    Fourure, Damien
    Emonet, Remi
    Fromont, Elisa
    Muselet, Damien
    Tremeau, Alain
    Wolf, Christian
    STRUCTURAL, SYNTACTIC, AND STATISTICAL PATTERN RECOGNITION, S+SSPR 2016, 2016, 10029 : 333 - 343
  • [25] Multi-task Heterogeneous Framework for Semi-supervised Medical Image Segmentation
    Cao, Jinghan
    Fan, Huijie
    Fu, Shengpeng
    Xu, Ling
    Chen, Xi'ai
    Lin, Sen
    INTELLIGENT ROBOTICS AND APPLICATIONS, ICIRA 2024, PT II, 2025, 15202 : 77 - 88
  • [26] SELF-SUPERVISED MULTI-TASK LEARNING FOR SEMANTIC SEGMENTATION OF URBAN SCENES
    Santiago, Jonathan Gonzalez
    Schenkel, Fabian
    Middelmann, Wolfgang
    IMAGE AND SIGNAL PROCESSING FOR REMOTE SENSING XXVII, 2021, 11862
  • [27] MULTI-TASK CURRICULUM LEARNING FOR SEMI-SUPERVISED MEDICAL IMAGE SEGMENTATION
    Wang, Kaiping
    Zhan, Bo
    Luo, Yanmei
    Zhou, Jiliu
    Wu, Xi
    Wang, Yan
    2021 IEEE 18TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI), 2021, : 925 - 928
  • [28] Weakly Supervised Multi-Task Representation Learning for Human Activity Analysis Using Wearables
    Sheng, Taoran
    Huber, Manfred
    PROCEEDINGS OF THE ACM ON INTERACTIVE MOBILE WEARABLE AND UBIQUITOUS TECHNOLOGIES-IMWUT, 2020, 4 (02):
  • [29] Robust Trust Region for Weakly Supervised Segmentation
    Marin, Dmitrii
    Boykov, Yuri
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 6588 - 6598
  • [30] Robust Stuttering Detection via Multi-task and Adversarial Learning
    Sheikh, Shakeel A.
    Sahidullah, Md
    Hirsch, Fabrice
    Ouni, Slim
    2022 30TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2022), 2022, : 190 - 194