Knowledge Distillation Meets Open-Set Semi-supervised Learning

被引:2
|
作者
Yang, Jing [1 ]
Zhu, Xiatian [2 ,3 ]
Bulat, Adrian [2 ]
Martinez, Brais [2 ]
Tzimiropoulos, Georgios [2 ,4 ]
机构
[1] Univ Nottingham, Nottingham, England
[2] Samsung AI Ctr, Cambridge, England
[3] Univ Surrey, Guildford, England
[4] Queen Mary Univ London, London, England
关键词
Knowledge distillation; Structured representational knowledge; Open-set semi-supervised learning; Out-of-distribution;
D O I
10.1007/s11263-024-02192-7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing knowledge distillation methods mostly focus on distillation of teacher's prediction and intermediate activation. However, the structured representation, which arguably is one of the most critical ingredients of deep models, is largely overlooked. In this work, we propose a novel semantic representational distillation (SRD) method dedicated for distilling representational knowledge semantically from a pretrained teacher to a target student. The key idea is that we leverage the teacher's classifier as a semantic critic for evaluating the representations of both teacher and student and distilling the semantic knowledge with high-order structured information over all feature dimensions. This is accomplished by introducing a notion of cross-network logit computed through passing student's representation into teacher's classifier. Further, considering the set of seen classes as a basis for the semantic space in a combinatorial perspective, we scale SRD to unseen classes for enabling effective exploitation of largely available, arbitrary unlabeled training data. At the problem level, this establishes an interesting connection between knowledge distillation with open-set semi-supervised learning (SSL). Extensive experiments show that our SRD outperforms significantly previous state-of-the-art knowledge distillation methods on both coarse object classification and fine face recognition tasks, as well as less studied yet practically crucial binary network distillation. Under more realistic open-set SSL settings we introduce, we reveal that knowledge distillation is generally more effective than existing out-of-distribution sample detection, and our proposed SRD is superior over both previous distillation and SSL competitors. The source code is available at https://github.com/jingyang2017/SRD_ossl.
引用
收藏
页码:315 / 334
页数:20
相关论文
共 50 条
  • [41] Ensemble Knowledge Distillation for Federated Semi-Supervised Image Classification
    Shang, Ertong
    Liu, Hui
    Zhang, Jingyang
    Zhao, Runqi
    Du, Junzhao
    TSINGHUA SCIENCE AND TECHNOLOGY, 2025, 30 (01): : 112 - 123
  • [42] Robust semi-supervised learning in open environments
    Guo, Lan-Zhe
    Jia, Lin-Han
    Shao, Jie-Jing
    Li, Yu-Feng
    FRONTIERS OF COMPUTER SCIENCE, 2025, 19 (08)
  • [43] Promote knowledge mining towards open-world semi-supervised learning
    Zhao, Tianhao
    Lin, Yutian
    Wu, Yu
    Du, Bo
    PATTERN RECOGNITION, 2024, 149
  • [44] Supervised Contrastive Learning for Open-Set Hyperspectral Image Classification
    Li, Zhaokui
    Bi, Ke
    Wang, Yan
    Fang, Zhuoqun
    Zhang, Jinen
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2023, 20
  • [45] GFD-SSL: generative federated knowledge distillation-based semi-supervised learning
    Karami, Ali
    Ramezani, Reza
    Baraani Dastjerdi, Ahmad
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (12) : 5509 - 5529
  • [46] Semi-Supervised Learning with Mutual Distillation for Monocular Depth Estimation
    Baek, Jongbeom
    Kim, Gyeongnyeon
    Kim, Seungryong
    2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022), 2022, : 4562 - 4569
  • [47] Cluster-aware Semi-supervised Learning: Relational Knowledge Distillation Provably Learns Clustering
    Dong, Yijun
    Miller, Kevin
    Lei, Qi
    Ward, Rachel
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [48] Expert-Guided Knowledge Distillation for Semi-Supervised Vessel Segmentation
    Shen, Ning
    Xu, Tingfa
    Huang, Shiqi
    Mu, Feng
    Li, Jianan
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2023, 27 (11) : 5542 - 5553
  • [49] Semi-supervised Campus Network Intrusion Detection Based on Knowledge Distillation
    Chen, Junjun
    Guo, Qiang
    Fu, Zhongnan
    Shang, Qun
    Ma, Hao
    Wang, Nai
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [50] Self Pseudo Entropy Knowledge Distillation for Semi-Supervised Semantic Segmentation
    Lu, Xiaoqiang
    Jiao, Licheng
    Li, Lingling
    Liu, Fang
    Liu, Xu
    Yang, Shuyuan
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (08) : 7359 - 7372