The Staged Knowledge Distillation in Video Classification: Harmonizing Student Progress by a Complementary Weakly Supervised Framework

被引:2
|
作者
Wang, Chao [1 ]
Tang, Zheng [2 ]
机构
[1] China Acad Railway Sci, Beijing 100081, Peoples R China
[2] NVIDIA, Redmond, WA 98052 USA
关键词
Training; Uncertainty; Correlation; Generators; Data models; Task analysis; Computational modeling; Knowledge distillation; weakly supervised learning; teacher-student architecture; substage learning process; video classification; label-efficient learning;
D O I
10.1109/TCSVT.2023.3294977
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In the context of label-efficient learning on video data, the distillation method and the structural design of the teacher-student architecture have a significant impact on knowledge distillation. However, the relationship between these factors has been overlooked in previous research. To address this gap, we propose a new weakly supervised learning framework for knowledge distillation in video classification that is designed to improve the efficiency and accuracy of the student model. Our approach leverages the concept of substage-based learning to distill knowledge based on the combination of student substages and the correlation of corresponding substages. We also employ the progressive cascade training method to address the accuracy loss caused by the large capacity gap between the teacher and the student. Additionally, we propose a pseudo-label optimization strategy to improve the initial data label. To optimize the loss functions of different distillation substages during the training process, we introduce a new loss method based on feature distribution. We conduct extensive experiments on both real and simulated data sets, demonstrating that our proposed approach outperforms existing distillation methods in terms of knowledge distillation for video classification tasks. Our proposed substage-based distillation approach has the potential to inform future research on label-efficient learning for video data.
引用
收藏
页码:6646 / 6660
页数:15
相关论文
共 50 条
  • [1] Weakly Supervised Cross-lingual Semantic Relation Classification via Knowledge Distillation
    Vyas, Yogarshi
    Carpuat, Marine
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 5285 - 5296
  • [2] Bi-directional Weakly Supervised Knowledge Distillation for Whole Slide Image Classification
    Qu, Linhao
    Luo, Xiaoyuan
    Wang, Manning
    Song, Zhijian
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [3] Overlooked Video Classification in Weakly Supervised Video Anomaly Detection
    Tan, Weijun
    Yao, Qi
    Liu, Jingfeng
    2024 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WORKSHOPS, WACVW 2024, 2024, : 212 - 220
  • [4] Improving Weakly Supervised Visual Grounding by Contrastive Knowledge Distillation
    Wang, Liwei
    Huang, Jing
    Li, Yin
    Xu, Kun
    Yang, Zhengyuan
    Yu, Dong
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 14085 - 14095
  • [5] Multi-Label Image Classification via Knowledge Distillation from Weakly-Supervised Detection
    Liu, Yongcheng
    Sheng, Lu
    Shao, Jing
    Yan, Junjie
    Xiang, Shiming
    Pan, Chunhong
    PROCEEDINGS OF THE 2018 ACM MULTIMEDIA CONFERENCE (MM'18), 2018, : 700 - 708
  • [6] Self-supervised knowledge distillation for complementary label learning
    Liu, Jiabin
    Li, Biao
    Lei, Minglong
    Shi, Yong
    NEURAL NETWORKS, 2022, 155 : 318 - 327
  • [7] Video Classification via Weakly Supervised Sequence Modeling
    Liu, Jingjing
    Chen, Chao
    Zhu, Yan
    Liu, Wei
    Metaxas, Dimitris N.
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2016, 152 : 79 - 87
  • [8] Weakly Supervised Classification of Hyperspectral Image Based on Complementary Learning
    Huang, Lingbo
    Chen, Yushi
    He, Xin
    REMOTE SENSING, 2021, 13 (24)
  • [9] Weakly Supervised Dense Video Captioning via Jointly Usage of Knowledge Distillation and Cross-modal Matching
    Wu, Bofeng
    Niu, Guocheng
    Yu, Jun
    Xiao, Xinyan
    Zhang, Jian
    Wu, Hua
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 1157 - 1164
  • [10] Knowledge Consistency Distillation for Weakly Supervised One Step Person Search
    Li, Zongyi
    Shi, Yuxuan
    Ling, Hefei
    Chen, Jiazhong
    Wang, Runsheng
    Zhao, Chengxin
    Wang, Qian
    Huang, Shijuan
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (11) : 11695 - 11708