Ensemble Knowledge Distillation for Learning Improved and Efficient Networks

被引:13
|
作者
Asif, Umar [1 ]
Tang, Jianbin [1 ]
Harrer, Stefan [1 ]
机构
[1] IBM Res Australia, Southbank, Vic, Australia
关键词
D O I
10.3233/FAIA200188
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Ensemble models comprising of deep Convolutional Neural Networks (CNN) have shown significant improvements in model generalization but at the cost of large computation and memory requirements. In this paper, we present a framework for learning compact CNN models with improved classification performance and model generalization. For this, we propose a CNN architecture of a compact student model with parallel branches which are trained using ground truth labels and information from high capacity teacher networks in an ensemble learning fashion. Our framework provides two main benefits: i) Distilling knowledge from different teachers into the student network promotes heterogeneity in learning features at different branches of the student network and enables the network to learn diverse solutions to the target problem. ii) Coupling the branches of the student network through ensembling encourages collaboration and improves the quality of the final predictions by reducing variance in the network outputs. Experiments on the well established CIFAR-10 and CIFAR-100 datasets show that our Ensemble Knowledge Distillation (EKD) improves classification accuracy and model generalization especially in situations with limited training data. Experiments also show that our EKD based compact networks outperform in terms of mean accuracy on the test datasets compared to other knowledge distillation based methods.
引用
收藏
页码:953 / 960
页数:8
相关论文
共 50 条
  • [41] A Sample-Efficient OPF Learning Method Based on Annealing Knowledge Distillation
    Dong, Ziheng
    Hou, Kai
    Liu, Zeyu
    Yu, Xiaodan
    Jia, Hongjie
    Zhang, Chi
    IEEE ACCESS, 2022, 10 : 99724 - 99733
  • [42] Learning Efficient and Accurate Detectors With Dynamic Knowledge Distillation in Remote Sensing Imagery
    Zhang, Yidan
    Yan, Zhiyuan
    Sun, Xian
    Diao, Wenhui
    Fu, Kun
    Wang, Lei
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [43] Efficient-PrototypicalNet with self knowledge distillation for few-shot learning
    Lim, Jit Yan
    Lim, Kian Ming
    Ooi, Shih Yin
    Lee, Chin Poo
    NEUROCOMPUTING, 2021, 459 : 327 - 337
  • [44] Simultaneous Learning Knowledge Distillation for Image Restoration: Efficient Model Compression for Drones
    Zhang, Yongheng
    DRONES, 2025, 9 (03)
  • [45] Efficient online learning with improved LSTM neural networks
    Mirza, Ali H.
    Kerpicci, Mine
    Kozat, Suleyman S.
    DIGITAL SIGNAL PROCESSING, 2020, 102
  • [46] Knowledge Distillation based Cooperative Reinforcement Learning for Connectivity Preservation in UAV Networks
    Wang, Zhihong
    Wei, Yunkai
    Wu, Fan
    2021 6TH INTERNATIONAL CONFERENCE ON UK-CHINA EMERGING TECHNOLOGIES (UCET 2021), 2021, : 171 - 176
  • [47] Inplace knowledge distillation with teacher assistant for improved training of flexible deep neural networks
    Ozerov, Alexey
    Duong, Ngoc Q. K.
    29TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2021), 2021, : 1356 - 1360
  • [48] Progressive Ensemble Distillation: Building Ensembles for Efficient Inference
    Dennis, Don Kurian
    Shetty, Abhishek
    Sevekari, Anish
    Koishida, Kazuhito
    Smith, Virginia
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [49] Communication-Efficient Federated Learning in Drone-Assisted IoT Networks: Path Planning and Enhanced Knowledge Distillation Techniques
    Gad, Gad
    Farrag, Aya
    Fadlullah, Zubair Md
    Fouda, Mostafa M.
    2023 IEEE 34TH ANNUAL INTERNATIONAL SYMPOSIUM ON PERSONAL, INDOOR AND MOBILE RADIO COMMUNICATIONS, PIMRC, 2023,
  • [50] Self-Decoupling and Ensemble Distillation for Efficient Segmentation
    Liu, Yuang
    Zhang, Wei
    Wang, Jun
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 2, 2023, : 1772 - 1780