The Staged Knowledge Distillation in Video Classification: Harmonizing Student Progress by a Complementary Weakly Supervised Framework

被引:2
|
作者
Wang, Chao [1 ]
Tang, Zheng [2 ]
机构
[1] China Acad Railway Sci, Beijing 100081, Peoples R China
[2] NVIDIA, Redmond, WA 98052 USA
关键词
Training; Uncertainty; Correlation; Generators; Data models; Task analysis; Computational modeling; Knowledge distillation; weakly supervised learning; teacher-student architecture; substage learning process; video classification; label-efficient learning;
D O I
10.1109/TCSVT.2023.3294977
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In the context of label-efficient learning on video data, the distillation method and the structural design of the teacher-student architecture have a significant impact on knowledge distillation. However, the relationship between these factors has been overlooked in previous research. To address this gap, we propose a new weakly supervised learning framework for knowledge distillation in video classification that is designed to improve the efficiency and accuracy of the student model. Our approach leverages the concept of substage-based learning to distill knowledge based on the combination of student substages and the correlation of corresponding substages. We also employ the progressive cascade training method to address the accuracy loss caused by the large capacity gap between the teacher and the student. Additionally, we propose a pseudo-label optimization strategy to improve the initial data label. To optimize the loss functions of different distillation substages during the training process, we introduce a new loss method based on feature distribution. We conduct extensive experiments on both real and simulated data sets, demonstrating that our proposed approach outperforms existing distillation methods in terms of knowledge distillation for video classification tasks. Our proposed substage-based distillation approach has the potential to inform future research on label-efficient learning for video data.
引用
收藏
页码:6646 / 6660
页数:15
相关论文
共 50 条
  • [31] Complementary branch fusing class and semantic knowledge for robust weakly supervised semantic segmentation
    Han, Woojung
    Kang, Seil
    Choo, Kyobin
    Hwang, Seong Jae
    PATTERN RECOGNITION, 2024, 157
  • [32] Better and Faster: Knowledge Transfer from Multiple Self-supervised Learning Tasks via Graph Distillation for Video Classification
    Zhang, Chenrui
    Peng, Yuxin
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 1135 - 1141
  • [33] VARIATIONAL STUDENT: LEARNING COMPACT AND SPARSER NETWORKS IN KNOWLEDGE DISTILLATION FRAMEWORK
    Hegde, Srinidhi
    Prasad, Ranjitha
    Hebbalaguppe, Ramya
    Kumar, Vishwajeet
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 3247 - 3251
  • [34] A Novel Knowledge Distillation Method for Self-Supervised Hyperspectral Image Classification
    Chi, Qiang
    Lv, Guohua
    Zhao, Guixin
    Dong, Xiangjun
    REMOTE SENSING, 2022, 14 (18)
  • [35] Weakly-Supervised Domain Adaptation of Deep Regression Trackers via Reinforced Knowledge Distillation
    Dunnhofer, Matteo
    Martinel, Niki
    Micheloni, Christian
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (03) : 5016 - 5023
  • [36] Weakly supervised object localization via knowledge distillation based on foreground-background contrast
    Ma, Siteng
    Hou, Biao
    Li, Zhihao
    Wu, Zitong
    Guo, Xianpeng
    Yang, Chen
    Jiao, Licheng
    NEUROCOMPUTING, 2024, 576
  • [37] Cross-task Knowledge Transfer for Extremely Weakly Supervised Text Classification
    Park, Seongmin
    Kim, Kyungho
    Lee, Jihwa
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 5329 - 5341
  • [38] A weakly supervised knowledge attentive network for aspect-level sentiment classification
    Bai, Qingchun
    Xiao, Jun
    Zhou, Jie
    JOURNAL OF SUPERCOMPUTING, 2023, 79 (05): : 5403 - 5420
  • [39] A weakly supervised knowledge attentive network for aspect-level sentiment classification
    Qingchun Bai
    Jun Xiao
    Jie Zhou
    The Journal of Supercomputing, 2023, 79 : 5403 - 5420
  • [40] An Effective Semi-Supervised Learning Framework for Temporal Student Classification
    Vo Thi Ngoc Chau
    Nguyen Hua Phung
    PROCEEDINGS OF 2019 6TH NATIONAL FOUNDATION FOR SCIENCE AND TECHNOLOGY DEVELOPMENT (NAFOSTED) CONFERENCE ON INFORMATION AND COMPUTER SCIENCE (NICS), 2019, : 363 - 369