The Staged Knowledge Distillation in Video Classification: Harmonizing Student Progress by a Complementary Weakly Supervised Framework

被引:2
|
作者
Wang, Chao [1 ]
Tang, Zheng [2 ]
机构
[1] China Acad Railway Sci, Beijing 100081, Peoples R China
[2] NVIDIA, Redmond, WA 98052 USA
关键词
Training; Uncertainty; Correlation; Generators; Data models; Task analysis; Computational modeling; Knowledge distillation; weakly supervised learning; teacher-student architecture; substage learning process; video classification; label-efficient learning;
D O I
10.1109/TCSVT.2023.3294977
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In the context of label-efficient learning on video data, the distillation method and the structural design of the teacher-student architecture have a significant impact on knowledge distillation. However, the relationship between these factors has been overlooked in previous research. To address this gap, we propose a new weakly supervised learning framework for knowledge distillation in video classification that is designed to improve the efficiency and accuracy of the student model. Our approach leverages the concept of substage-based learning to distill knowledge based on the combination of student substages and the correlation of corresponding substages. We also employ the progressive cascade training method to address the accuracy loss caused by the large capacity gap between the teacher and the student. Additionally, we propose a pseudo-label optimization strategy to improve the initial data label. To optimize the loss functions of different distillation substages during the training process, we introduce a new loss method based on feature distribution. We conduct extensive experiments on both real and simulated data sets, demonstrating that our proposed approach outperforms existing distillation methods in terms of knowledge distillation for video classification tasks. Our proposed substage-based distillation approach has the potential to inform future research on label-efficient learning for video data.
引用
收藏
页码:6646 / 6660
页数:15
相关论文
共 50 条
  • [21] Teacher-student collaborative knowledge distillation for image classification
    Chuanyun Xu
    Wenjian Gao
    Tian Li
    Nanlan Bai
    Gang Li
    Yang Zhang
    Applied Intelligence, 2023, 53 : 1997 - 2009
  • [22] Ensemble Knowledge Distillation for Federated Semi-Supervised Image Classification
    Shang, Ertong
    Liu, Hui
    Zhang, Jingyang
    Zhao, Runqi
    Du, Junzhao
    TSINGHUA SCIENCE AND TECHNOLOGY, 2025, 30 (01): : 112 - 123
  • [23] Weakly Supervised Referring Expression Grounding via Target-Guided Knowledge Distillation
    Mi, Jinpeng
    Tang, Song
    Ma, Zhiyuan
    Liu, Dan
    Li, Qingdu
    Zhang, Jianwei
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2023), 2023, : 8299 - 8305
  • [24] Weakly Supervised Exaggeration Transfer for Caricature Generation With Cross-Modal Knowledge Distillation
    Tong, Shuo
    Liu, Han
    He, Yuxin
    Du, Chenxiao
    Wang, Wenqing
    Guo, Runyuan
    Liu, Jingyun
    IEEE COMPUTER GRAPHICS AND APPLICATIONS, 2024, 44 (04) : 98 - 112
  • [25] Weakly Supervised Referring Expression Grounding via Dynamic Self-Knowledge Distillation
    Mi, Jinpeng
    Chen, Zhiqian
    Zhang, Jianwei
    2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, IROS, 2023, : 1254 - 1260
  • [26] Spatial likelihood voting with self-knowledge distillation for weakly supervised object detection
    Chen, Ze
    Fu, Zhihang
    Huang, Jianqiang
    Tao, Mingyuan
    Jiang, Rongxin
    Tian, Xiang
    Chen, Yaowu
    Hua, Xian-Sheng
    IMAGE AND VISION COMPUTING, 2021, 116
  • [27] Weakly supervised text classification framework for noisy-labeled imbalanced
    Zhang, Wenxin
    Zhou, Yaya
    Liu, Shuhui
    Zhang, Yupei
    Shang, Xuequn
    NEUROCOMPUTING, 2024, 610
  • [28] A weakly supervised framework for real-world point cloud classification
    Deng, An
    Wu, Yunchao
    Zhang, Peng
    Lu, Zhuheng
    Li, Weiqing
    Su, Zhiyong
    COMPUTERS & GRAPHICS-UK, 2022, 102 : 78 - 88
  • [29] Uninformed Teacher-Student for hard-samples distillation in weakly supervised mitosis localization
    Fernandez-Martin, Claudio
    Silva-Rodriguez, Julio
    Kiraz, Umay
    Morales, Sandra
    Janssen, Emiel A. M.
    Naranjo, Valery
    COMPUTERIZED MEDICAL IMAGING AND GRAPHICS, 2024, 112
  • [30] Enhance Weakly-Supervised Aspect Detection with External Knowledge (Student Abstract)
    Zheng, Zhuoming
    Cai, Yi
    Li, Liuwu
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 13119 - 13120