Model Selection - Knowledge Distillation Framework for Model Compression

被引:0
|
作者
Chen, Renhai [1 ]
Yuan, Shimin [1 ]
Wang, Shaobo [1 ]
Li, Zhenghan [1 ]
Xing, Meng [1 ]
Feng, Zhiyong [1 ]
机构
[1] Tianjin Univ, Shenzhen Res Inst, Coll Intelligence & Comp, Tianjin, Peoples R China
基金
中国国家自然科学基金;
关键词
model selection; model compression; knowledge distillation;
D O I
10.1109/SSCI50451.2021.9659861
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The significant increase in the computation and parameter storage costs of CNNs promotes its development in various applications and restricts its deployment in edge devices as well. Therefore, many neural network pruning methods has been proposed for neural network compression and acceleration. However, there are two major limitations to these methods: First, prevailing methods usually design single pruning criteria for the primitive network and fail to consider the diversity of potential optimal sub-network structure. Second, these methods utilize traditional training method to train the sub-network, which is not enough to develop the expression ability of the sub-network under the current task.In this paper, we propose Model Selection - Knowledge Distillation (MS-KD) framework to solve the above problems. Specifically, we develop multiple pruning criteria for the primitive network, and the potential optimal structure is obtained through model selection.Furthermore, instead of traditional training methods, we use knowledge distillation to train the learned sub-network and make full use of the structure advantages of the sub-network.To validate our approach, we conduct extensive experiments on prevalent image classification datasets.The results demonstrate that our MS-KD framework outperforms the existing methods under a wide range of data sets, models, and inference costs.
引用
收藏
页数:6
相关论文
共 50 条
  • [21] DISCOVER THE EFFECTIVE STRATEGY FOR FACE RECOGNITION MODEL COMPRESSION BY IMPROVED KNOWLEDGE DISTILLATION
    Wang, Mengjiao
    Liu, Rujie
    Abe, Narishige
    Uchida, Hidetsugu
    Matsunami, Tomoaki
    Yamada, Shigefumi
    2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2018, : 2416 - 2420
  • [22] Spirit Distillation: A Model Compression Method with Multi-domain Knowledge Transfer
    Wu, Zhiyuan
    Jiang, Yu
    Zhao, Minghao
    Cui, Chupeng
    Yang, Zongmin
    Xue, Xinhui
    Qi, Hong
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT I, 2021, 12815 : 553 - 565
  • [23] Explanation Guided Knowledge Distillation for Pre-trained Language Model Compression
    Yang, Zhao
    Zhang, Yuanzhe
    Sui, Dianbo
    Ju, Yiming
    Zhao, Jun
    Liu, Kang
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2024, 23 (02)
  • [24] Knowledge Distillation for Sequence Model
    Huang, Mingkun
    You, Yongbin
    Chen, Zhehuai
    Qian, Yanmin
    Yu, Kai
    19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 3703 - 3707
  • [25] Simultaneous Learning Knowledge Distillation for Image Restoration: Efficient Model Compression for Drones
    Zhang, Yongheng
    DRONES, 2025, 9 (03)
  • [26] AUGMENTING KNOWLEDGE DISTILLATION WITH PEER-TO-PEER MUTUAL LEARNING FOR MODEL COMPRESSION
    Niyaz, Usma
    Bathula, Deepti R.
    2022 IEEE INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (IEEE ISBI 2022), 2022,
  • [27] Model compression via pruning and knowledge distillation for person re-identification
    Haonan Xie
    Wei Jiang
    Hao Luo
    Hongyan Yu
    Journal of Ambient Intelligence and Humanized Computing, 2021, 12 : 2149 - 2161
  • [28] Model Compression by Iterative Pruning with Knowledge Distillation and Its Application to Speech Enhancement
    Wei, Zeyuan
    Li, Hao
    Zhang, Xueliang
    INTERSPEECH 2022, 2022, : 941 - 945
  • [29] Attention-Fused CNN Model Compression with Knowledge Distillation for Brain Tumor Segmentation
    Xu, Pengcheng
    Kim, Kyungsang
    Liu, Huafeng
    Li, Quanzheng
    MEDICAL IMAGE UNDERSTANDING AND ANALYSIS, MIUA 2022, 2022, 13413 : 328 - 338
  • [30] Compression of Time Series Classification Model MC-MHLF using Knowledge Distillation
    Gengyo, Akari
    Tamura, Keiichi
    2021 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2021, : 22 - 27