Private Model Compression via Knowledge Distillation

被引:0
|
作者
Wang, Ji [1 ]
Bao, Weidong [1 ]
Sun, Lichao [2 ]
Zhu, Xiaomin [1 ,3 ]
Cao, Bokai [4 ]
Yu, Philip S. [2 ,5 ]
机构
[1] Natl Univ Def Technol, Coll Syst Engn, Changsha, Hunan, Peoples R China
[2] Univ Illinois, Dept Comp Sci, Chicago, IL USA
[3] Natl Univ Def Technol, State Key Lab High Performance Comp, Changsha, Hunan, Peoples R China
[4] Facebook Inc, Menlo Pk, CA USA
[5] Tsinghua Univ, Inst Data Sci, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The soaring demand for intelligent mobile applications calls for deploying powerful deep neural networks (DNNs) on mobile devices. However, the outstanding performance of DNNs notoriously relies on increasingly complex models, which in turn is associated with an increase in computational expense far surpassing mobile devices' capacity. What is worse, app service providers need to collect and utilize a large volume of users' data, which contain sensitive information, to build the sophisticated DNN models. Directly deploying these models on public mobile devices presents prohibitive privacy risk. To benefit from the on-device deep learning without the capacity and privacy concerns, we design a private model compression framework RONA. Following the knowledge distillation paradigm, we jointly use hint learning, distillation learning, and self learning to train a compact and fast neural network. The knowledge distilled from the cumbersome model is adaptively bounded and carefully perturbed to enforce differential privacy. We further propose an elegant query sample selection method to reduce the number of queries and control the privacy loss. A series of empirical evaluations as well as the implementation on an Android mobile device show that RONA can not only compress cumbersome models efficiently but also provide a strong privacy guarantee. For example, on SVHN, when a meaningful (9.83, 10(-6))-differential privacy is guaranteed, the compact model trained by RONA can obtain 20x compression ratio and 19x speed-up with merely 0.97% accuracy loss.
引用
收藏
页码:1190 / +
页数:9
相关论文
共 50 条
  • [31] Spirit Distillation: A Model Compression Method with Multi-domain Knowledge Transfer
    Wu, Zhiyuan
    Jiang, Yu
    Zhao, Minghao
    Cui, Chupeng
    Yang, Zongmin
    Xue, Xinhui
    Qi, Hong
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT I, 2021, 12815 : 553 - 565
  • [32] Explanation Guided Knowledge Distillation for Pre-trained Language Model Compression
    Yang, Zhao
    Zhang, Yuanzhe
    Sui, Dianbo
    Ju, Yiming
    Zhao, Jun
    Liu, Kang
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2024, 23 (02)
  • [33] Simultaneous Learning Knowledge Distillation for Image Restoration: Efficient Model Compression for Drones
    Zhang, Yongheng
    DRONES, 2025, 9 (03)
  • [34] AUGMENTING KNOWLEDGE DISTILLATION WITH PEER-TO-PEER MUTUAL LEARNING FOR MODEL COMPRESSION
    Niyaz, Usma
    Bathula, Deepti R.
    2022 IEEE INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (IEEE ISBI 2022), 2022,
  • [35] Model Compression by Iterative Pruning with Knowledge Distillation and Its Application to Speech Enhancement
    Wei, Zeyuan
    Li, Hao
    Zhang, Xueliang
    INTERSPEECH 2022, 2022, : 941 - 945
  • [36] Accumulation Knowledge Distillation for Conditional GAN Compression
    Gao, Tingwei
    Long, Rujiao
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 1294 - 1303
  • [37] Adaptive Contrastive Knowledge Distillation for BERT Compression
    Guo, Jinyang
    Liu, Jiaheng
    Wang, Zining
    Ma, Yuqing
    Gong, Ruihao
    Xu, Ke
    Liu, Xianglong
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 8941 - 8953
  • [38] Incremental event detection via an improved knowledge distillation based model
    Lin, Yi
    Xu, Changhua
    Yu, Hang
    Tian, Pinzhuo
    Luo, Xiangfeng
    NEUROCOMPUTING, 2023, 551
  • [39] Attention-Fused CNN Model Compression with Knowledge Distillation for Brain Tumor Segmentation
    Xu, Pengcheng
    Kim, Kyungsang
    Liu, Huafeng
    Li, Quanzheng
    MEDICAL IMAGE UNDERSTANDING AND ANALYSIS, MIUA 2022, 2022, 13413 : 328 - 338
  • [40] Compression of Time Series Classification Model MC-MHLF using Knowledge Distillation
    Gengyo, Akari
    Tamura, Keiichi
    2021 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2021, : 22 - 27