JSSE: Joint Sequential Semantic Encoder for Zero-Shot Event Recognition

被引:1
|
作者
Madapana N. [1 ]
Wachs J.P. [1 ]
机构
[1] Purdue University, School of Industrial Engineering, West Lafayette, 47906, IN
来源
基金
美国国家科学基金会; 美国医疗保健研究与质量局; 美国国家卫生研究院;
关键词
Action and gesture recognition; activity; semantic descriptors; transfer learning; zero-shot learning (ZSL);
D O I
10.1109/TAI.2022.3208860
中图分类号
学科分类号
摘要
Zero-shot learning (ZSL) is a paradigm in transfer learning that aims to recognize unknown categories by having a mere description of them. The problem of ZSL has been thoroughly studied in the domain of static object recognition; however, ZSL for dynamic events (zero-shot event recognition, ZSER) such as activities and gestures has hardly been investigated. In this context, this article addresses ZSER by relying on semantic attributes of events to transfer the learned knowledge from seen classes to unseen ones. First, we utilized the Amazon Mechanical Turk platform to create the first attribute-based gesture dataset, referred to as zero shot gestural learning (ZSGL), comprising the categories present in MSRC and Italian gesture datasets. Overall, our ZSGL dataset consisted of 26 categories, 65 discriminative attributes, and 16 attribute annotations and 400 examples per category. We used trainable recurrent networks and 3-D convolutional neural networks (CNNs) to learn the spatiotemporal features. Next, we propose a simple yet effective end-to-end approach for ZSER, referred to as joint sequential semantic encoder (JSSE), to explore temporal patterns, to efficiently represent events in the latent space, and to simultaneously optimize for both the semantic and classification tasks. We evaluate our model on ZSGL and two action datasets (UCF and HMDB), and compared the performance of JSSE against several existing baselines under four experimental conditions: 1) within-category, 2) across-category, 3) closed-set, and 4) open-set. Results show that JSSE considerably outperforms (p< 0.05) other approaches and performs favorably for both the datasets under all experimental conditions. © 2020 IEEE.
引用
收藏
页码:1472 / 1483
页数:11
相关论文
共 50 条
  • [1] JSE: Joint Semantic Encoder for zero-shot gesture learning
    Naveen Madapana
    Juan Wachs
    Pattern Analysis and Applications, 2022, 25 : 679 - 692
  • [2] JS']JSE: Joint Semantic Encoder for zero-shot gesture learning
    Madapana, Naveen
    Wachs, Juan
    PATTERN ANALYSIS AND APPLICATIONS, 2022, 25 (03) : 679 - 692
  • [3] Marginalized Latent Semantic Encoder for Zero-Shot Learning
    Ding, Zhengming
    Liu, Hongfu
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 6184 - 6192
  • [4] Joint Visual and Semantic Optimization for zero-shot learning
    Wu, Hanrui
    Yan, Yuguang
    Chen, Sentao
    Huang, Xiangkang
    Wu, Qingyao
    Ng, Michael K.
    KNOWLEDGE-BASED SYSTEMS, 2021, 215 (215)
  • [5] Zero-Shot Object Recognition by Semantic Manifold Distance
    Fu, Zhenyong
    Xiang, Tao
    Kodirov, Elyor
    Gong, Shaogang
    2015 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2015, : 2635 - 2644
  • [6] Global Semantic Descriptors for Zero-Shot Action Recognition
    Estevam, Valter
    Laroca, Rayson
    Pedrini, Helio
    Menotti, David
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 1843 - 1847
  • [7] SEMANTIC EMBEDDING SPACE FOR ZERO-SHOT ACTION RECOGNITION
    Xu, Xun
    Hospedales, Timothy
    Gong, Shaogang
    2015 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2015, : 63 - 67
  • [8] Learning complementary semantic information for zero-shot recognition
    Hu, Xiaoming
    Wang, Zilei
    Li, Junjie
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2023, 115
  • [9] Zero-Shot Image Recognition Algorithm via Semantic Auto-Encoder Combining Relation Network
    Lin K.
    Li H.
    Bai J.
    Li A.
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2019, 32 (03): : 214 - 224
  • [10] JOINT PROJECTION AND SUBSPACE LEARNING FOR ZERO-SHOT RECOGNITION
    Liu, Guangzhen
    Guan, Jiechao
    Zhang, Manli
    Zhang, Jianhong
    Wang, Zihao
    Lu, Zhiwu
    2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, : 1228 - 1233