Invariant and consistent: Unsupervised representation learning for few-shot visual recognition

被引:1
|
作者
Wu, Heng [1 ]
Zhao, Yifan [2 ]
Li, Jia [1 ]
机构
[1] Beihang Univ, Sch Comp Sci & Engn, State Key Lab Virtual Real Technol & Syst, Beijing 100191, Peoples R China
[2] Peking Univ, Sch Comp Sci, Beijing 100871, Peoples R China
基金
中国国家自然科学基金;
关键词
Few -shot visual recognition; Unsupervised learning; Geometric invariance; Pairwise consistency;
D O I
10.1016/j.neucom.2022.11.073
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Few-shot visual recognition aims to identify novel unseen classes with few labels while learning gener-alized prior knowledge from base classes. Recent ideas propose to explore this problem in an unsuper-vised setting, i.e., without any labels in base classes, which reduces the heavy consumption of manual annotations. In this paper, we build upon a self-supervised insight and propose a novel unsupervised learning approach that joints Invariant and Consistent (InCo) representation for the few-shot task. For the invariant representation operation, we present a geometric invariance module to construct the rota-tion prediction of each instance, which learns the intra-instance variance and improves the feature dis-crimination. To further build consistency representation of inter-instance, we propose a pairwise consistency module from two contrastive learning aspects: a holistic contrastive learning with historical training queues, and a local contrastive learning for enhancing the representation of current training samples. Moreover, to better facilitate contrastive learning among features, we introduce an asymmetric convolutional architecture to encode high-quality representations. Comprehensive experiments on 4 public benchmarks demonstrate the utility of our approach and the superiority compared to existing approaches.(c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:1 / 14
页数:14
相关论文
共 50 条
  • [1] Cooperative density-aware representation learning for few-shot visual recognition
    Zheng, Zijun
    Feng, Xiang
    Yu, Huiqun
    Gao, Mengqi
    NEUROCOMPUTING, 2022, 471 : 208 - 218
  • [2] Unsupervised Deep Representation Learning and Few-Shot Classification of PolSAR Images
    Zhang, Lamei
    Zhang, Siyu
    Zou, Bin
    Dong, Hongwei
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [3] VISUAL TEMPO CONTRASTIVE LEARNING FOR FEW-SHOT ACTION RECOGNITION
    Wang, Guangge
    Ye, Weirong
    Wang, Xiao
    Jin, Rongrong
    Wang, Hanzi
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 1096 - 1100
  • [4] Unsupervised meta-learning for few-shot learning
    Xu, Hui
    Wang, Jiaxing
    Li, Hao
    Ouyang, Deqiang
    Shao, Jie
    PATTERN RECOGNITION, 2021, 116
  • [5] Few-shot learning for ear recognition
    Zhang, Jie
    Yu, Wen
    Yang, Xudong
    Deng, Fang
    PROCEEDINGS OF 2019 INTERNATIONAL CONFERENCE ON IMAGE, VIDEO AND SIGNAL PROCESSING (IVSP 2019), 2019, : 50 - 54
  • [6] Knowledge Graph enhanced Multimodal Learning for Few-shot Visual Recognition
    Han, Mengya
    Zhan, Yibing
    Yu, Baosheng
    Luo, Yong
    Du, Bo
    Tao, Dacheng
    2022 IEEE 24TH INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING (MMSP), 2022,
  • [7] Few-shot and Many-shot Fusion Learning in Mobile Visual Food Recognition
    Zhao, Heng
    Yap, Kim-Hui
    Kot, Alex C.
    Duan, Lingyu
    Cheung, Ngai-Man
    2019 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2019,
  • [8] Class Representation Networks for Few-Shot Learning
    Zhai, Yongping
    Wang, Junhua
    PROCEEDINGS OF 2020 IEEE 11TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING AND SERVICE SCIENCE (ICSESS 2020), 2020, : 133 - 137
  • [9] Few-shot Food Recognition via Multi-view Representation Learning
    Jiang, Shuqiang
    Min, Weiqing
    Lyu, Yongqiang
    Liu, Linhu
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2020, 16 (03)
  • [10] Representation based meta-learning for few-shot spoken intent recognition
    Mittal, Ashish
    Bharadwaj, Samarth
    Khare, Shreya
    Chemmengath, Saneem
    Sankaranarayanan, Karthik
    Kingsbury, Brian
    INTERSPEECH 2020, 2020, : 4283 - 4287