Realistic Evaluation of Transductive Few-Shot Learning

被引:0
|
作者
Veilleux, Olivier [1 ]
Boudiaf, Malik [1 ]
Piantanida, Pablo [2 ]
Ben Ayed, Ismail [1 ]
机构
[1] ETS Montreal, Montreal, PQ, Canada
[2] Univ Paris Saclay, Cent Supelec, CNRS, L2S, Paris, France
基金
加拿大自然科学与工程研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transductive inference is widely used in few-shot learning, as it leverages the statistics of the unlabeled query set of a few-shot task, typically yielding substantially better performances than its inductive counterpart. The current few-shot benchmarks use perfectly class-balanced tasks at inference. We argue that such an artificial regularity is unrealistic, as it assumes that the marginal label probability of the testing samples is known and fixed to the uniform distribution. In fact, in realistic scenarios, the unlabeled query sets come with arbitrary and unknown label marginals. We introduce and study the effect of arbitrary class distributions within the query sets of few-shot tasks at inference, removing the class-balance artefact. Specifically, we model the marginal probabilities of the classes as Dirichlet-distributed random variables, which yields a principled and realistic sampling within the simplex. This leverages the current few-shot benchmarks, building testing tasks with arbitrary class distributions. We evaluate experimentally state-of-the-art transductive methods over 3 widely used data sets, and observe, surprisingly, substantial performance drops, even below inductive methods in some cases. Furthermore, we propose a generalization of the mutual-information loss, based on ff-divergences, which can handle effectively class-distribution variations. Empirically, we show that our transductive ff-divergence optimization outperforms state-of-the-art methods across several data sets, models and few-shot settings. Our code is publicly available at https://github.com/oveilleux/Realistic_Transductive_Few_Shot.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Transductive distribution calibration for few-shot learning
    Li, Gang
    Zheng, Changwen
    Su, Bing
    Neurocomputing, 2022, 500 : 604 - 615
  • [2] Transductive Information Maximization For Few-Shot Learning
    Boudiaf, Malik
    Masud, Ziko Imtiaz
    Rony, Jerome
    Dolz, Jose
    Piantanida, Pablo
    Ben Ayed, Ismail
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [3] Transductive distribution calibration for few-shot learning
    Li, Gang
    Zheng, Changwen
    Su, Bing
    NEUROCOMPUTING, 2022, 500 : 604 - 615
  • [4] EASE: Unsupervised Discriminant Subspace Learning for Transductive Few-Shot Learning
    Zhu, Hao
    Koniusz, Piotr
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 9068 - 9078
  • [5] Relation fusion propagation network for transductive few-shot learning
    Huang, Yixiang
    Hao, Hongyu
    Ge, Weichao
    Cao, Yang
    Wu, Ming
    Zhang, Chuang
    Guo, Jun
    PATTERN RECOGNITION, 2024, 151
  • [6] Few-Shot Learning with Localization in Realistic Settings
    Wertheimer, Davis
    Hariharan, Bharath
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 6551 - 6560
  • [7] A transductive learning method to leverage graph structure for few-shot learning
    Wang, Yaning
    Liu, Zijian
    Luo, Yang
    Luo, Chunbo
    PATTERN RECOGNITION LETTERS, 2022, 159 : 189 - 195
  • [8] ADAPTIVE ANCHOR LABEL PROPAGATION FOR TRANSDUCTIVE FEW-SHOT LEARNING
    Lazarou, Michalis
    Avrithis, Yannis
    Ren, Guangyu
    Stathaki, Tania
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 331 - 335
  • [9] Transductive clustering optimization learning for few-shot image classification
    Wang, Yi
    Bian, Xiong
    Zhu, Songhao
    JOURNAL OF ELECTRONIC IMAGING, 2023, 32 (04)
  • [10] Transductive Relation-Propagation Network for Few-shot Learning
    Ma, Yuqing
    Bai, Shihao
    An, Shan
    Liu, Wei
    Liu, Aishan
    Zhen, Xiantong
    Liu, Xianglong
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 804 - 810