Knowledge Distillation Meets Open-Set Semi-supervised Learning

被引:2
|
作者
Yang, Jing [1 ]
Zhu, Xiatian [2 ,3 ]
Bulat, Adrian [2 ]
Martinez, Brais [2 ]
Tzimiropoulos, Georgios [2 ,4 ]
机构
[1] Univ Nottingham, Nottingham, England
[2] Samsung AI Ctr, Cambridge, England
[3] Univ Surrey, Guildford, England
[4] Queen Mary Univ London, London, England
关键词
Knowledge distillation; Structured representational knowledge; Open-set semi-supervised learning; Out-of-distribution;
D O I
10.1007/s11263-024-02192-7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing knowledge distillation methods mostly focus on distillation of teacher's prediction and intermediate activation. However, the structured representation, which arguably is one of the most critical ingredients of deep models, is largely overlooked. In this work, we propose a novel semantic representational distillation (SRD) method dedicated for distilling representational knowledge semantically from a pretrained teacher to a target student. The key idea is that we leverage the teacher's classifier as a semantic critic for evaluating the representations of both teacher and student and distilling the semantic knowledge with high-order structured information over all feature dimensions. This is accomplished by introducing a notion of cross-network logit computed through passing student's representation into teacher's classifier. Further, considering the set of seen classes as a basis for the semantic space in a combinatorial perspective, we scale SRD to unseen classes for enabling effective exploitation of largely available, arbitrary unlabeled training data. At the problem level, this establishes an interesting connection between knowledge distillation with open-set semi-supervised learning (SSL). Extensive experiments show that our SRD outperforms significantly previous state-of-the-art knowledge distillation methods on both coarse object classification and fine face recognition tasks, as well as less studied yet practically crucial binary network distillation. Under more realistic open-set SSL settings we introduce, we reveal that knowledge distillation is generally more effective than existing out-of-distribution sample detection, and our proposed SRD is superior over both previous distillation and SSL competitors. The source code is available at https://github.com/jingyang2017/SRD_ossl.
引用
收藏
页码:315 / 334
页数:20
相关论文
共 50 条
  • [21] SOAMC: A Semi-Supervised Open-Set Recognition Algorithm for Automatic Modulation Classification
    Di, Chengliang
    Ji, Jinwei
    Sun, Chao
    Liang, Linlin
    ELECTRONICS, 2024, 13 (21)
  • [22] LaRW: boosting open-set semi-supervised learning with label-guided re-weighting
    Ouyang, Jihong
    Mao, Dong
    Meng, Qingyi
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (15) : 46419 - 46437
  • [23] LaRW: boosting open-set semi-supervised learning with label-guided re-weighting
    Jihong Ouyang
    Dong Mao
    Qingyi Meng
    Multimedia Tools and Applications, 2024, 83 : 46419 - 46437
  • [24] Partial Optimal Transport Based Out-of-Distribution Detection for Open-Set Semi-Supervised Learning
    Ren, Yilong
    Feng, Chuanwen
    Xie, Xike
    Zhou, S. Kevin
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 4851 - 4859
  • [25] ProSub: Probabilistic Open-Set Semi-supervised Learning with Subspace-Based Out-of-Distribution Detection
    Wallin, Erik
    Svensson, Lennart
    Kahle, Fredrik
    Hammarstrand, Lars
    COMPUTER VISION - ECCV 2024, PT LXI, 2025, 15119 : 129 - 147
  • [26] Open-Set Semi-supervised Medical Image Classification with Learnable Prototypes and Outlier Filter
    He, Along
    Li, Tao
    Zhao, Yitian
    Zhao, Junyong
    Fu, Huazhu
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2024, PT XI, 2024, 15011 : 492 - 501
  • [27] Trash to Treasure: Harvesting OOD Data with Cross-Modal Matching for Open-Set Semi-Supervised Learning
    Huang, Junkai
    Fang, Chaowei
    Chen, Weikai
    Chai, Zhenhua
    Wei, Xiaolin
    Wei, Pengxu
    Lin, Liang
    Li, Guanbin
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 8290 - 8299
  • [28] Knowledge Distillation for Semi-supervised Domain Adaptation
    Orbes-Arteainst, Mauricio
    Cardoso, Jorge
    Sorensen, Lauge
    Igel, Christian
    Ourselin, Sebastien
    Modat, Marc
    Nielsen, Mads
    Pai, Akshay
    OR 2.0 CONTEXT-AWARE OPERATING THEATERS AND MACHINE LEARNING IN CLINICAL NEUROIMAGING, 2019, 11796 : 68 - 76
  • [29] Open-set Recognition with Supervised Contrastive Learning
    Kodama, Yuto
    Wang, Yinan
    Kawakami, Rei
    Naemura, Takeshi
    PROCEEDINGS OF 17TH INTERNATIONAL CONFERENCE ON MACHINE VISION APPLICATIONS (MVA 2021), 2021,
  • [30] When Semi-supervised Learning Meets Ensemble Learning
    Zhou, Zhi-Hua
    MULTIPLE CLASSIFIER SYSTEMS, PROCEEDINGS, 2009, 5519 : 529 - 538