Distinguishing Unseen from Seen for Generalized Zero-shot Learning

被引:23
|
作者
Su, Hongzu [1 ]
Li, Jingjing [1 ,2 ]
Chen, Zhi [3 ]
Zhu, Lei [4 ]
Lu, Ke [1 ]
机构
[1] Univ Elect Sci & Technol China, Chengdu, Sichuan, Peoples R China
[2] UESTC Guangdong, Inst Elect & Informat Engn, Shenzhen, Guangdong, Peoples R China
[3] Univ Queensland, Brisbane, Qld, Australia
[4] Shandong Normal Univ, Jinan, Shandong, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52688.2022.00773
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Generalized zero-shot learning (GZSL) aims to recognize samples whose categories may not have been seen at training. Recognizing unseen classes as seen ones or vice versa often leads to poor performance in GZSL. Therefore, distinguishing seen and unseen domains is naturally an effective yet challenging solution for GZSL. In this paper, we present a novel method which leverages both visual and semantic modalities to distinguish seen and unseen categories. Specifically, our method deploys two variational autoencoders to generate latent representations for visual and semantic modalities in a shared latent space, in which we align latent representations of both modalities by Wasserstein distance and reconstruct two modalities with the representations of each other. In order to learn a clearer boundary between seen and unseen classes, we propose a two-stage training strategy which takes advantage of seen and unseen semantic descriptions and searches a threshold to separate seen and unseen visual samples. At last, a seen expert and an unseen expert are used for final classification. Extensive experiments on five widely used benchmarks verify that the proposed method can significantly improve the results of GZSL. For instance, our method correctly recognizes more than 99% samples when separating domains and improves the final classification accuracy from 72.6% to 82.9% on AWA1.
引用
收藏
页码:7875 / 7884
页数:10
相关论文
共 50 条
  • [31] From Zero-shot Learning to Conventional Supervised Classification: Unseen Visual Data Synthesis
    Long, Yang
    Liu, Li
    Shao, Ling
    Shen, Fumin
    Ding, Guiguang
    Han, Jungong
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 6165 - 6174
  • [32] Generalized zero-shot classification via iteratively generating and selecting unseen samples
    Li, Xiao
    Fang, Min
    Chen, Bo
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2021, 92
  • [33] Towards Zero-Shot Learning with Fewer Seen Class Examples
    Verma, Vinay Kumar
    Mishra, Ashish
    Pandey, Anubha
    Murthy, Hema A.
    Rai, Piyush
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WACV 2021, 2021, : 2240 - 2250
  • [34] From Classical to Generalized Zero-Shot Learning: A Simple Adaptation Process
    Le Cacheux, Yannick
    Le Borgne, Herve
    Crucianu, Michel
    MULTIMEDIA MODELING, MMM 2019, PT II, 2019, 11296 : 465 - 477
  • [35] Learning Multiple Criteria Calibration for Generalized Zero-shot Learning
    Lu, Ziqian
    Lu, Zhe-Ming
    Yu, Yunlong
    He, Zewei
    Luo, Hao
    Zheng, Yangming
    KNOWLEDGE-BASED SYSTEMS, 2024, 300
  • [36] Synthetic Sample Selection for Generalized Zero-Shot Learning
    Gowda, Shreyank N.
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW, 2023, : 58 - 67
  • [37] Discriminative comparison classifier for generalized zero-shot learning
    Hou, Mingzhen
    Xia, Wei
    Zhang, Xiangdong
    Gao, Quanxue
    NEUROCOMPUTING, 2020, 414 (414) : 10 - 17
  • [38] Data-Free Generalized Zero-Shot Learning
    Tang, Bowen
    Zhang, Jing
    Yan, Long
    Yu, Qian
    Sheng, Lu
    Xu, Dong
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 6, 2024, : 5108 - 5117
  • [39] Generalized Zero-Shot Learning with Noisy Labeled Data
    Xu, Liqing
    Liu, Xueliang
    Jiang, Yishun
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT XI, 2024, 14435 : 289 - 300
  • [40] Transferable Contrastive Network for Generalized Zero-Shot Learning
    Jiang, Huajie
    Wang, Ruiping
    Shan, Shiguang
    Chen, Xilin
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 9764 - 9773