Ensemble Knowledge Distillation for Learning Improved and Efficient Networks

被引:13
|
作者
Asif, Umar [1 ]
Tang, Jianbin [1 ]
Harrer, Stefan [1 ]
机构
[1] IBM Res Australia, Southbank, Vic, Australia
关键词
D O I
10.3233/FAIA200188
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Ensemble models comprising of deep Convolutional Neural Networks (CNN) have shown significant improvements in model generalization but at the cost of large computation and memory requirements. In this paper, we present a framework for learning compact CNN models with improved classification performance and model generalization. For this, we propose a CNN architecture of a compact student model with parallel branches which are trained using ground truth labels and information from high capacity teacher networks in an ensemble learning fashion. Our framework provides two main benefits: i) Distilling knowledge from different teachers into the student network promotes heterogeneity in learning features at different branches of the student network and enables the network to learn diverse solutions to the target problem. ii) Coupling the branches of the student network through ensembling encourages collaboration and improves the quality of the final predictions by reducing variance in the network outputs. Experiments on the well established CIFAR-10 and CIFAR-100 datasets show that our Ensemble Knowledge Distillation (EKD) improves classification accuracy and model generalization especially in situations with limited training data. Experiments also show that our EKD based compact networks outperform in terms of mean accuracy on the test datasets compared to other knowledge distillation based methods.
引用
收藏
页码:953 / 960
页数:8
相关论文
共 50 条
  • [31] An improved hybrid structure learning strategy for Bayesian networks based on ensemble learning
    Gao, Wenlong
    Zeng, Zhimei
    Ma, Xiaojie
    Ke, Yongsong
    Zhi, Minqian
    INTELLIGENT DATA ANALYSIS, 2023, 27 (04) : 1103 - 1120
  • [32] VARIATIONAL STUDENT: LEARNING COMPACT AND SPARSER NETWORKS IN KNOWLEDGE DISTILLATION FRAMEWORK
    Hegde, Srinidhi
    Prasad, Ranjitha
    Hebbalaguppe, Ramya
    Kumar, Vishwajeet
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 3247 - 3251
  • [33] Ensemble2Net: Learning from Ensemble Teacher Networks via Knowledge Transfer
    Chang, Chih-Lin
    Li, Shiou-Chi
    Huang, Jen-Wei
    2021 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2021), 2021,
  • [34] Efficient learning for distillation of DNN by self distillation
    Takagi J.
    Hattori M.
    IEEJ Transactions on Electronics, Information and Systems, 2019, 139 (12) : 1509 - 1516
  • [35] Knowledge distillation for secondary pulmonary tuberculosis classification ensemble
    Zhou, Qinghua
    Zhu, Hengde
    Zhang, Xin
    Zhang, Yudong
    COMPANION PROCEEDINGS OF THE 14TH IEEE/ACM INTERNATIONAL CONFERENCE ON UTILITY AND CLOUD COMPUTING (UCC'21 COMPANION), 2021,
  • [36] Learn From the Past: Experience Ensemble Knowledge Distillation
    Wang, Chaofei
    Zhang, Shaowei
    Song, Shiji
    Huang, Gao
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 4736 - 4743
  • [37] Adversarial-Based Ensemble Feature Knowledge Distillation
    Mingwen Shao
    Shunhang Li
    Zilu Peng
    Yuantao Sun
    Neural Processing Letters, 2023, 55 : 10315 - 10329
  • [38] Ensemble Modeling with Contrastive Knowledge Distillation for Sequential Recommendation
    Du, Hanwen
    Yuan, Huanhuan
    Zhao, Pengpeng
    Zhuang, Fuzhen
    Liu, Guanfeng
    Zhao, Lei
    Liu, Yanchi
    Sheng, Victor S.
    PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 58 - 67
  • [39] Adversarial-Based Ensemble Feature Knowledge Distillation
    Shao, Mingwen
    Li, Shunhang
    Peng, Zilu
    Sun, Yuantao
    NEURAL PROCESSING LETTERS, 2023, 55 (08) : 10315 - 10329
  • [40] Interpretable Respiratory Sound Analysis with Ensemble Knowledge Distillation
    Wang, Cheng
    Li, Jianqiang
    Chen, Jie
    Zhang, Heng
    Wang, Li
    Liu, Zun
    2021 6TH IEEE INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS AND MECHATRONICS (ICARM 2021), 2021, : 19 - 24