Spirit Distillation: A Model Compression Method with Multi-domain Knowledge Transfer

被引:6
|
作者
Wu, Zhiyuan [1 ]
Jiang, Yu [1 ,2 ]
Zhao, Minghao [1 ]
Cui, Chupeng [1 ]
Yang, Zongmin [1 ]
Xue, Xinhui [1 ]
Qi, Hong [1 ,2 ]
机构
[1] Jilin Univ, Coll Comp Sci & Technol, Changchun, Peoples R China
[2] Jilin Univ, Minist Educ, Key Lab Symbol Computat & Knowledge Engn, Changchun, Peoples R China
基金
中国国家自然科学基金;
关键词
Knowledge transfer; Knowledge distillation; Multi-domain; Model compression; Few-shot learning;
D O I
10.1007/978-3-030-82136-4_45
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent applications pose requirements of both cross-domain knowledge transfer and model compression to machine learning models due to insufficient training data and limited computational resources. In this paper, we propose a new knowledge distillation model, named Spirit Distillation (SD), which is a model compression method with multi-domain knowledge transfer. The compact student network mimics out a representation equivalent to the front part of the teacher network, through which the general knowledge can be transferred from the source domain (teacher) to the target domain (student). To further improve the robustness of the student, we extend SD to Enhanced Spirit Distillation (ESD) in exploiting a more comprehensive knowledge by introducing the proximity domainwhich is similar to the target domain for feature extraction. Persuasive experiments are conducted on Cityscapes semantic segmentation with the prior knowledge transferred fromCOCO2017 and KITTI. Results demonstrate that our method can boost mIOU and high-precision accuracy by 1.4% and 8.2% respectively with 78.2% segmentation variance, and can gain a precise compact network with only 41.8% FLOPs.
引用
收藏
页码:553 / 565
页数:13
相关论文
共 50 条
  • [31] Domain attention model for multi-domain sentiment classification
    Yuan, Zhigang
    Wu, Sixing
    Wu, Fangzhao
    Liu, Junxin
    Huang, Yongfeng
    KNOWLEDGE-BASED SYSTEMS, 2018, 155 : 1 - 10
  • [32] An environment for multi-domain ontology development and knowledge acquisition
    Si, JX
    Cao, CG
    Wang, H
    Gu, F
    Feng, QZ
    Zhang, CX
    Zeng, QT
    Tian, W
    Zheng, YF
    ENGINEERING AND DEPLOYMENT OF COOPERATIVE INFORMATION SYSTEMS, PROCEEDINGS, 2002, 2480 : 104 - 116
  • [33] Knowledge Extraction From National Standards for Natural Resources: A Method for Multi-Domain Texts
    Ban, Taiyu
    Wang, Xiangyu
    Wang, Xin
    Zhu, Jiarun
    Chen, Lvzhou
    Fan, Yizhan
    JOURNAL OF DATABASE MANAGEMENT, 2023, 34 (01)
  • [34] Multi-domain speech compression based on wavelet packet transform
    Wu, XD
    Li, YM
    Chen, HY
    ELECTRONICS LETTERS, 1998, 34 (02) : 154 - 155
  • [35] Multi-domain boundary element method with dissipation
    Xiaobo Chen
    Wenyang Duan
    Journal of Marine Science and Application, 2012, 11 (1) : 18 - 23
  • [36] An Object-oriented Modular Design Model Supported by Integrated Multi-domain Knowledge
    Li, X.
    Huang, Y. Q.
    ADVANCES IN MATERIALS MANUFACTURING SCIENCE AND TECHNOLOGY XIV, 2012, 697-698 : 785 - +
  • [37] Multi-domain Boundary Element Method with Dissipation
    Chen, Xiaobo
    Duan, Wenyang
    JOURNAL OF MARINE SCIENCE AND APPLICATION, 2012, 11 (01) : 18 - 23
  • [38] A wavelet optimized adaptive multi-domain method
    Hesthaven, JS
    Jameson, LM
    JOURNAL OF COMPUTATIONAL PHYSICS, 1998, 145 (01) : 280 - 296
  • [39] A domain renumbering algorithm for multi-domain boundary face method
    Zhang, Jianming
    Lu, Chenjun
    Li, Yuan
    Han, Lei
    Wang, Pan
    Li, Guangyao
    ENGINEERING ANALYSIS WITH BOUNDARY ELEMENTS, 2014, 44 : 19 - 27
  • [40] Multi-domain spoken language understanding with transfer learning
    Jeong, Minwoo
    Lee, Gary Geunbae
    SPEECH COMMUNICATION, 2009, 51 (05) : 412 - 424