Private Model Compression via Knowledge Distillation

被引:0
|
作者
Wang, Ji [1 ]
Bao, Weidong [1 ]
Sun, Lichao [2 ]
Zhu, Xiaomin [1 ,3 ]
Cao, Bokai [4 ]
Yu, Philip S. [2 ,5 ]
机构
[1] Natl Univ Def Technol, Coll Syst Engn, Changsha, Hunan, Peoples R China
[2] Univ Illinois, Dept Comp Sci, Chicago, IL USA
[3] Natl Univ Def Technol, State Key Lab High Performance Comp, Changsha, Hunan, Peoples R China
[4] Facebook Inc, Menlo Pk, CA USA
[5] Tsinghua Univ, Inst Data Sci, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The soaring demand for intelligent mobile applications calls for deploying powerful deep neural networks (DNNs) on mobile devices. However, the outstanding performance of DNNs notoriously relies on increasingly complex models, which in turn is associated with an increase in computational expense far surpassing mobile devices' capacity. What is worse, app service providers need to collect and utilize a large volume of users' data, which contain sensitive information, to build the sophisticated DNN models. Directly deploying these models on public mobile devices presents prohibitive privacy risk. To benefit from the on-device deep learning without the capacity and privacy concerns, we design a private model compression framework RONA. Following the knowledge distillation paradigm, we jointly use hint learning, distillation learning, and self learning to train a compact and fast neural network. The knowledge distilled from the cumbersome model is adaptively bounded and carefully perturbed to enforce differential privacy. We further propose an elegant query sample selection method to reduce the number of queries and control the privacy loss. A series of empirical evaluations as well as the implementation on an Android mobile device show that RONA can not only compress cumbersome models efficiently but also provide a strong privacy guarantee. For example, on SVHN, when a meaningful (9.83, 10(-6))-differential privacy is guaranteed, the compact model trained by RONA can obtain 20x compression ratio and 19x speed-up with merely 0.97% accuracy loss.
引用
收藏
页码:1190 / +
页数:9
相关论文
共 50 条
  • [21] Multi-Granularity Structural Knowledge Distillation for Language Model Compression
    Liu, Chang
    Tao, Chongyang
    Feng, Jiazhan
    Zhao, Dongyan
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 1001 - 1011
  • [22] Efficient Neural Data Compression for Machine Type Communications via Knowledge Distillation
    Hussien, Mostafa
    Xu, Yi Tian
    Wu, Di
    Liu, Xue
    Dudek, Gregory
    2022 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2022), 2022, : 1169 - 1174
  • [23] Model Conversion via Differentially Private Data-Free Distillation
    Liu, Bochao
    Wang, Pengju
    Li, Shikun
    Zeng, Dan
    Ge, Shiming
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 2187 - 2195
  • [24] Distilling a Powerful Student Model via Online Knowledge Distillation
    Li, Shaojie
    Lin, Mingbao
    Wang, Yan
    Wu, Yongjian
    Tian, Yonghong
    Shao, Ling
    Ji, Rongrong
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (11) : 8743 - 8752
  • [25] Compressing Visual-linguistic Model via Knowledge Distillation
    Fang, Zhiyuan
    Wang, Jianfeng
    Hu, Xiaowei
    Wang, Lijuan
    Yang, Yezhou
    Liu, Zicheng
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 1408 - 1418
  • [26] Graph-Based Model Compression for HSR Bogies Fault Diagnosis at IoT Edge via Adversarial Knowledge Distillation
    Wan, Wenqing
    Chen, Jinglong
    Xie, Jingsong
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (02) : 1787 - 1796
  • [27] Differentially Private Knowledge Distillation for Mobile Analytics
    Lyu, Lingjuan
    Chen, Chi-Hua
    PROCEEDINGS OF THE 43RD INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '20), 2020, : 1809 - 1812
  • [28] Mitigating carbon footprint for knowledge distillation based deep learning model compression
    Rafat, Kazi
    Islam, Sadia
    Mahfug, Abdullah Al
    Hossain, Md. Ismail
    Rahman, Fuad
    Momen, Sifat
    Rahman, Shafin
    Mohammed, Nabeel
    PLOS ONE, 2023, 18 (05):
  • [29] Joint structured pruning and dense knowledge distillation for efficient transformer model compression
    Cui, Baiyun
    Li, Yingming
    Zhang, Zhongfei
    NEUROCOMPUTING, 2021, 458 : 56 - 69
  • [30] DISCOVER THE EFFECTIVE STRATEGY FOR FACE RECOGNITION MODEL COMPRESSION BY IMPROVED KNOWLEDGE DISTILLATION
    Wang, Mengjiao
    Liu, Rujie
    Abe, Narishige
    Uchida, Hidetsugu
    Matsunami, Tomoaki
    Yamada, Shigefumi
    2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2018, : 2416 - 2420