Knowledge Distillation Meets Open-Set Semi-supervised Learning

被引:2
|
作者
Yang, Jing [1 ]
Zhu, Xiatian [2 ,3 ]
Bulat, Adrian [2 ]
Martinez, Brais [2 ]
Tzimiropoulos, Georgios [2 ,4 ]
机构
[1] Univ Nottingham, Nottingham, England
[2] Samsung AI Ctr, Cambridge, England
[3] Univ Surrey, Guildford, England
[4] Queen Mary Univ London, London, England
关键词
Knowledge distillation; Structured representational knowledge; Open-set semi-supervised learning; Out-of-distribution;
D O I
10.1007/s11263-024-02192-7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing knowledge distillation methods mostly focus on distillation of teacher's prediction and intermediate activation. However, the structured representation, which arguably is one of the most critical ingredients of deep models, is largely overlooked. In this work, we propose a novel semantic representational distillation (SRD) method dedicated for distilling representational knowledge semantically from a pretrained teacher to a target student. The key idea is that we leverage the teacher's classifier as a semantic critic for evaluating the representations of both teacher and student and distilling the semantic knowledge with high-order structured information over all feature dimensions. This is accomplished by introducing a notion of cross-network logit computed through passing student's representation into teacher's classifier. Further, considering the set of seen classes as a basis for the semantic space in a combinatorial perspective, we scale SRD to unseen classes for enabling effective exploitation of largely available, arbitrary unlabeled training data. At the problem level, this establishes an interesting connection between knowledge distillation with open-set semi-supervised learning (SSL). Extensive experiments show that our SRD outperforms significantly previous state-of-the-art knowledge distillation methods on both coarse object classification and fine face recognition tasks, as well as less studied yet practically crucial binary network distillation. Under more realistic open-set SSL settings we introduce, we reveal that knowledge distillation is generally more effective than existing out-of-distribution sample detection, and our proposed SRD is superior over both previous distillation and SSL competitors. The source code is available at https://github.com/jingyang2017/SRD_ossl.
引用
收藏
页码:315 / 334
页数:20
相关论文
共 50 条
  • [1] SCOMatch: Alleviating Overtrusting in Open-Set Semi-supervised Learning
    Wane, Zerun
    Xiang, Liuyu
    Huang, Lang
    Mao, Jiafeng
    Xiao, Ling
    Yamasaki, Toshihiko
    COMPUTER VISION - ECCV 2024, PT LI, 2025, 15109 : 217 - 233
  • [2] Mutual Filter Teaching for Open-Set Semi-Supervised Learning
    Li, Xiaokun
    Yi, Rumeng
    Huang, Yaping
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 7700 - 7708
  • [3] Closed loop networks for open-set semi-supervised learning
    Ouyang, Jihong
    Meng, Qingyi
    Li, Ximing
    Zhang, Zhengjie
    Li, Changchun
    Wang, Wenting
    INFORMATION SCIENCES, 2025, 699
  • [4] Open-Set Semi-Supervised Object Detection
    Liu, Yen-Cheng
    Ma, Chih-Yao
    Dai, Xiaoliang
    Tian, Junjiao
    Vajda, Peter
    He, Zijian
    Kira, Zsolt
    COMPUTER VISION - ECCV 2022, PT XXX, 2022, 13690 : 143 - 159
  • [5] OpenMatch: Open-set Consistency Regularization for Semi-supervised Learning with Outliers
    Saito, Kuniaki
    Kim, Donghyun
    Saenko, Kate
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [6] DeCAB: Debiased Semi-supervised Learning for Imbalanced Open-Set Data
    Huang, Xiaolin
    Li, Mengke
    Lu, Yang
    Wang, Hanzi
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT IX, 2024, 14433 : 104 - 119
  • [7] Exploration and Exploitation of Unlabeled Data for Open-Set Semi-supervised Learning
    Zhao, Ganlong
    Li, Guanbin
    Qin, Yipeng
    Zhang, Jinjin
    Chai, Zhenhua
    Wei, Xiaolin
    Lin, Liang
    Yu, Yizhou
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (12) : 5888 - 5904
  • [8] Robust Semi-Supervised Learning by Wisely Leveraging Open-Set Data
    Yang, Yang
    Jiang, Nan
    Xu, Yi
    Zhan, De-Chuan
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (12) : 8334 - 8347
  • [9] OSSGAN: Open-Set Semi-Supervised Image Generation
    Katsumata, Kai
    Duc Minh Vo
    Nakayama, Hideki
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 11175 - 11183
  • [10] ANEDL: Adaptive Negative Evidential Deep Learning for Open-Set Semi-supervised Learning
    Yu, Yang
    Deng, Danruo
    Liu, Furui
    Dou, Qi
    Jin, Yueming
    Chen, Guangyong
    Heng, Pheng Ann
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 15, 2024, : 16587 - 16595