Spirit Distillation: A Model Compression Method with Multi-domain Knowledge Transfer

被引:6
|
作者
Wu, Zhiyuan [1 ]
Jiang, Yu [1 ,2 ]
Zhao, Minghao [1 ]
Cui, Chupeng [1 ]
Yang, Zongmin [1 ]
Xue, Xinhui [1 ]
Qi, Hong [1 ,2 ]
机构
[1] Jilin Univ, Coll Comp Sci & Technol, Changchun, Peoples R China
[2] Jilin Univ, Minist Educ, Key Lab Symbol Computat & Knowledge Engn, Changchun, Peoples R China
基金
中国国家自然科学基金;
关键词
Knowledge transfer; Knowledge distillation; Multi-domain; Model compression; Few-shot learning;
D O I
10.1007/978-3-030-82136-4_45
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent applications pose requirements of both cross-domain knowledge transfer and model compression to machine learning models due to insufficient training data and limited computational resources. In this paper, we propose a new knowledge distillation model, named Spirit Distillation (SD), which is a model compression method with multi-domain knowledge transfer. The compact student network mimics out a representation equivalent to the front part of the teacher network, through which the general knowledge can be transferred from the source domain (teacher) to the target domain (student). To further improve the robustness of the student, we extend SD to Enhanced Spirit Distillation (ESD) in exploiting a more comprehensive knowledge by introducing the proximity domainwhich is similar to the target domain for feature extraction. Persuasive experiments are conducted on Cityscapes semantic segmentation with the prior knowledge transferred fromCOCO2017 and KITTI. Results demonstrate that our method can boost mIOU and high-precision accuracy by 1.4% and 8.2% respectively with 78.2% segmentation variance, and can gain a precise compact network with only 41.8% FLOPs.
引用
收藏
页码:553 / 565
页数:13
相关论文
共 50 条
  • [41] A Multi-Domain Camera Model Identification Feature Restoration Network to Counter AI Compression Attacks
    Zhang, Jinkai
    Han, Zijuan
    Liu, Yunxia
    Yang, Yang
    2024 ASIA PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE, APSIPA ASC, 2024,
  • [42] Multi-Domain Simulation Model of a Wheel Loader
    Saha, Rohit
    Hwang, Long-Kung
    Kumar, Mahesh Madurai
    Zhao, Yunfeng
    Yu, Chen
    Ransijn, Bob
    SAE INTERNATIONAL JOURNAL OF COMMERCIAL VEHICLES, 2016, 9 (02) : 252 - 259
  • [43] A TRACED ROLES MODEL FOR MULTI-DOMAIN AUTHORIZATION
    Benjumea, Andres
    Agudo, Isaac
    INTERNATIONAL JOURNAL ON INFORMATION TECHNOLOGIES AND SECURITY, 2009, 1 (04): : 55 - 64
  • [44] MULTI-DOMAIN TLM MODEL FOR INTRAVASCULAR ULTRASOUND
    Borji, Rafik
    Franchek, Matthew A.
    PROCEEDINGS OF THE ASME DYNAMIC SYSTEMS AND CONTROL CONFERENCE 2009, PTS A AND B, 2010, : 697 - 704
  • [45] Automatic Support for Multi-Domain Model Management
    Torres, Weslley
    van den Brand, Mark G. J.
    Serebrenik, Alexander
    2020 IEEE INTERNATIONAL CONFERENCE ON SOFTWARE MAINTENANCE AND EVOLUTION (ICSME 2020), 2020, : 830 - 833
  • [46] Multi-domain Neural Network Language Model
    Alumae, Tanel
    14TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2013), VOLS 1-5, 2013, : 2181 - 2185
  • [47] Research on Expression Method of a Unified Constraint Multi-domain Model for Complex Products
    Chen Guojin
    Su Shaohui
    Gong Youping
    Zhu Miaofen
    LIFE SYSTEM MODELING AND INTELLIGENT COMPUTING, PT II, 2010, 98 : 177 - 183
  • [48] A uniform human knowledge interface to the multi-domain knowledge bases in the National Knowledge Infrastructure
    Feng, QG
    Cao, CN
    Si, JX
    Zheng, YF
    APPLICATIONS AND INNOVATIONS IN INTELLIGENT SYSTEMS X, 2003, : 163 - 176
  • [49] Compression of Acoustic Model via Knowledge Distillation and Pruning
    Li, Chenxing
    Zhu, Lei
    Xu, Shuang
    Gao, Peng
    Xu, Bo
    2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 2785 - 2790
  • [50] Utilizing online content as domain knowledge in a multi-domain dynamic dialogue system
    Wootton, Craig
    McTear, Michael
    Anderson, Terry
    INTERSPEECH 2007: 8TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION, VOLS 1-4, 2007, : 693 - 696