Distinguishing Unseen from Seen for Generalized Zero-shot Learning

被引:23
|
作者
Su, Hongzu [1 ]
Li, Jingjing [1 ,2 ]
Chen, Zhi [3 ]
Zhu, Lei [4 ]
Lu, Ke [1 ]
机构
[1] Univ Elect Sci & Technol China, Chengdu, Sichuan, Peoples R China
[2] UESTC Guangdong, Inst Elect & Informat Engn, Shenzhen, Guangdong, Peoples R China
[3] Univ Queensland, Brisbane, Qld, Australia
[4] Shandong Normal Univ, Jinan, Shandong, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52688.2022.00773
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Generalized zero-shot learning (GZSL) aims to recognize samples whose categories may not have been seen at training. Recognizing unseen classes as seen ones or vice versa often leads to poor performance in GZSL. Therefore, distinguishing seen and unseen domains is naturally an effective yet challenging solution for GZSL. In this paper, we present a novel method which leverages both visual and semantic modalities to distinguish seen and unseen categories. Specifically, our method deploys two variational autoencoders to generate latent representations for visual and semantic modalities in a shared latent space, in which we align latent representations of both modalities by Wasserstein distance and reconstruct two modalities with the representations of each other. In order to learn a clearer boundary between seen and unseen classes, we propose a two-stage training strategy which takes advantage of seen and unseen semantic descriptions and searches a threshold to separate seen and unseen visual samples. At last, a seen expert and an unseen expert are used for final classification. Extensive experiments on five widely used benchmarks verify that the proposed method can significantly improve the results of GZSL. For instance, our method correctly recognizes more than 99% samples when separating domains and improves the final classification accuracy from 72.6% to 82.9% on AWA1.
引用
收藏
页码:7875 / 7884
页数:10
相关论文
共 50 条
  • [1] Seen and unseen classes oriented augmented class synthesis for zero-shot learning
    Chen, Yuan
    Zhou, Yuan
    INFORMATION SCIENCES, 2025, 689
  • [2] Learning domain invariant unseen features for generalized zero-shot classification
    Li, Xiao
    Fang, Min
    Li, Haikun
    Wu, Jinqiao
    KNOWLEDGE-BASED SYSTEMS, 2020, 206
  • [3] Generalized zero-shot learning for classifying unseen wafer map patterns
    Kim, Han Kyul
    Shim, Jaewoong
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 133
  • [4] Zero-shot classification with unseen prototype learning
    Ji, Zhong
    Cui, Biying
    Yu, Yunlong
    Pang, Yanwei
    Zhang, Zhongfei
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (17): : 12307 - 12317
  • [5] Zero-shot classification with unseen prototype learning
    Zhong Ji
    Biying Cui
    Yunlong Yu
    Yanwei Pang
    Zhongfei Zhang
    Neural Computing and Applications, 2023, 35 : 12307 - 12317
  • [6] Infer unseen from seen: Relation regularized zero-shot visual dialog
    Zhang, Zefan
    Li, Shun
    Ji, Yi
    Liu, Chunping
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2023, 97
  • [7] Generalized zero-shot domain adaptation with target unseen class prototype learning
    Li, Xiao
    Fang, Min
    Chen, Bo
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (20): : 17793 - 17807
  • [8] Generalized zero-shot domain adaptation with target unseen class prototype learning
    Xiao Li
    Min Fang
    Bo Chen
    Neural Computing and Applications, 2022, 34 : 17793 - 17807
  • [9] Zero-VAE-GAN: Generating Unseen Features for Generalized and Transductive Zero-Shot Learning
    Gao, Rui
    Hou, Xingsong
    Qin, Jie
    Chen, Jiaxin
    Liu, Li
    Zhu, Fan
    Zhang, Zhao
    Shao, Ling
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 3665 - 3680
  • [10] Learning unseen visual prototypes for zero-shot classification
    Li, Xiao
    Fang, Min
    Feng, Dazheng
    Li, Haikun
    Wu, Jinqiao
    KNOWLEDGE-BASED SYSTEMS, 2018, 160 : 176 - 187