Private Model Compression via Knowledge Distillation

被引:0
|
作者
Wang, Ji [1 ]
Bao, Weidong [1 ]
Sun, Lichao [2 ]
Zhu, Xiaomin [1 ,3 ]
Cao, Bokai [4 ]
Yu, Philip S. [2 ,5 ]
机构
[1] Natl Univ Def Technol, Coll Syst Engn, Changsha, Hunan, Peoples R China
[2] Univ Illinois, Dept Comp Sci, Chicago, IL USA
[3] Natl Univ Def Technol, State Key Lab High Performance Comp, Changsha, Hunan, Peoples R China
[4] Facebook Inc, Menlo Pk, CA USA
[5] Tsinghua Univ, Inst Data Sci, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The soaring demand for intelligent mobile applications calls for deploying powerful deep neural networks (DNNs) on mobile devices. However, the outstanding performance of DNNs notoriously relies on increasingly complex models, which in turn is associated with an increase in computational expense far surpassing mobile devices' capacity. What is worse, app service providers need to collect and utilize a large volume of users' data, which contain sensitive information, to build the sophisticated DNN models. Directly deploying these models on public mobile devices presents prohibitive privacy risk. To benefit from the on-device deep learning without the capacity and privacy concerns, we design a private model compression framework RONA. Following the knowledge distillation paradigm, we jointly use hint learning, distillation learning, and self learning to train a compact and fast neural network. The knowledge distilled from the cumbersome model is adaptively bounded and carefully perturbed to enforce differential privacy. We further propose an elegant query sample selection method to reduce the number of queries and control the privacy loss. A series of empirical evaluations as well as the implementation on an Android mobile device show that RONA can not only compress cumbersome models efficiently but also provide a strong privacy guarantee. For example, on SVHN, when a meaningful (9.83, 10(-6))-differential privacy is guaranteed, the compact model trained by RONA can obtain 20x compression ratio and 19x speed-up with merely 0.97% accuracy loss.
引用
收藏
页码:1190 / +
页数:9
相关论文
共 50 条
  • [41] AD-KD: Attribution-Driven Knowledge Distillation for Language Model Compression
    Wu, Siyue
    Chen, Hongzhan
    Quan, Xiaojun
    Wang, Qifan
    Wang, Rui
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 8449 - 8465
  • [42] HRKD: Hierarchical Relational Knowledge Distillation for Cross-domain Language Model Compression
    Done, Chenhe
    Li, Yaliang
    Shen, Ying
    Qui, Minghui
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 3126 - 3136
  • [43] PKDGAN: Private Knowledge Distillation With Generative Adversarial Networks
    Zhuo, Cheng
    Gao, Di
    Liu, Liangwei
    IEEE TRANSACTIONS ON BIG DATA, 2024, 10 (06) : 775 - 788
  • [44] Knowledge Distillation via Information Matching
    Zhu, Honglin
    Jiang, Ning
    Tang, Jialiang
    Huang, Xinlei
    NEURAL INFORMATION PROCESSING, ICONIP 2023, PT IV, 2024, 14450 : 405 - 417
  • [45] Collaborative knowledge distillation via filter knowledge transfer
    Gou, Jianping
    Hu, Yue
    Sun, Liyuan
    Wang, Zhi
    Ma, Hongxing
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 238
  • [46] Lifelong Compression Mixture Model via Knowledge Relationship Graph
    Ye, Fei
    Bors, Adrian G.
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 9, 2023, : 10900 - 10908
  • [47] Knowledge Distillation for Sequence Model
    Huang, Mingkun
    You, Yongbin
    Chen, Zhehuai
    Qian, Yanmin
    Yu, Kai
    19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 3703 - 3707
  • [48] Combining Weight Pruning and Knowledge Distillation For CNN Compression
    Aghli, Nima
    Ribeiro, Eraldo
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 3185 - 3192
  • [49] Few Sample Knowledge Distillation for Efficient Network Compression
    Li, Tianhong
    Li, Jianguo
    Liu, Zhuang
    Zhang, Changshui
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, : 14627 - 14635
  • [50] Differentially Private Model Compression
    Mireshghallah, Fatemehsadat
    Backurs, Arturs
    Inan, Huseyin A.
    Wutschitz, Lukas
    Kulkarni, Janardhan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,