Dual discriminator adversarial distillation for data-free model compression

被引:12
|
作者
Zhao, Haoran [1 ]
Sun, Xin [1 ,2 ]
Dong, Junyu [1 ]
Manic, Milos [3 ]
Zhou, Huiyu [4 ]
Yu, Hui [5 ]
机构
[1] Ocean Univ China, Coll Informat Sci & Engn, Qingdao, Peoples R China
[2] Tech Univ Munich, Dept Aerosp & Geodesy, Munich, Germany
[3] Virginia Commonwealth Univ, Coll Engn, Richmond, VA USA
[4] Univ Leicester, Sch Informat, Leicester, Leics, England
[5] Univ Portsmouth, Sch Creat Technol, Portsmouth, Hants, England
基金
中国国家自然科学基金;
关键词
Deep neural networks; Image classification; Model compression; Knowledge distillation; Data-free; KNOWLEDGE; NETWORK; RECOGNITION;
D O I
10.1007/s13042-021-01443-0
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Knowledge distillation has been widely used to produce portable and efficient neural networks which can be well applied on edge devices for computer vision tasks. However, almost all top-performing knowledge distillation methods need to access the original training data, which usually has a huge size and is often unavailable. To tackle this problem, we propose a novel data-free approach in this paper, named Dual Discriminator Adversarial Distillation (DDAD) to distill a neural network without the need of any training data or meta-data. To be specific, we use a generator to create samples through dual discriminator adversarial distillation, which mimics the original training data. The generator not only uses the pre-trained teacher's intrinsic statistics in existing batch normalization layers but also obtains the maximum discrepancy from the student model. Then the generated samples are used to train the compact student network under the supervision of the teacher. The proposed method obtains an efficient student network which closely approximates its teacher network, without using the original training data. Extensive experiments are conducted to demonstrate the effectiveness of the proposed approach on CIFAR, Caltech101 and ImageNet datasets for classification tasks. Moreover, we extend our method to semantic segmentation tasks on several public datasets such as CamVid, NYUv2, Cityscapes and VOC 2012. To the best of our knowledge, this is the first work on generative model based data-free knowledge distillation on large-scale datasets such as ImageNet, Cityscapes and VOC 2012. Experiments show that our method outperforms all baselines for data-free knowledge distillation.
引用
收藏
页码:1213 / 1230
页数:18
相关论文
共 50 条
  • [41] Effectiveness of Arbitrary Transfer Sets for Data-free Knowledge Distillation
    Nayak, Gaurav Kumar
    Mopuri, Konda Reddy
    Chakraborty, Anirban
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2021), 2021, : 1429 - 1437
  • [42] Frequency Domain Distillation for Data-Free Quantization of Vision Transformer
    Nan, Gongrui
    Chao, Fei
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT VIII, 2024, 14432 : 205 - 216
  • [43] Unpacking the Gap Box Against Data-Free Knowledge Distillation
    Wang, Yang
    Qian, Biao
    Liu, Haipeng
    Rui, Yong
    Wang, Meng
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (09) : 6280 - 6291
  • [44] Layer-Wise Data-Free CNN Compression
    Horton, Maxwell
    Jin, Yanzi
    Farhadi, Ali
    Rastegari, Mohammad
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2019 - 2026
  • [45] DFDS: Data-Free Dual Substitutes Hard-Label Black-Box Adversarial Attack
    Jiang, Shuliang
    He, Yusheng
    Zhang, Rui
    Kang, Zi
    Xia, Hui
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT III, KSEM 2024, 2024, 14886 : 274 - 285
  • [46] Augmented Geometric Distillation for Data-Free Incremental Person ReID
    Lu, Yichen
    Wang, Mei
    Deng, Weihong
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 7319 - 7328
  • [47] Data-free Knowledge Distillation based on GNN for Node Classification
    Zeng, Xinfeng
    Liu, Tao
    Zeng, Ming
    Wu, Qingqiang
    Wang, Meihong
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, DASFAA 2024, PT 2, 2025, 14851 : 243 - 258
  • [48] FedAlign: Federated Model Alignment via Data-Free Knowledge Distillation for Machine Fault Diagnosis
    Sun, Wenjun
    Yan, Ruqiang
    Jin, Ruibing
    Zhao, Rui
    Chen, Zhenghua
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73 : 1 - 12
  • [49] Data-free Universal Adversarial Perturbation and Black-box Attack
    Zhang, Chaoning
    Benz, Philipp
    Karjauv, Adil
    Kweon, In So
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 7848 - 7857
  • [50] D3K: Dynastic Data-Free Knowledge Distillation
    Li, Xiufang
    Sun, Qigong
    Jiao, Licheng
    Liu, Fang
    Liu, Xu
    Li, Lingling
    Chen, Puhua
    Zuo, Yi
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 8358 - 8371