De-confounded Data-free Knowledge Distillation for Handling Distribution Shifts

被引:2
|
作者
Wang, Yuzheng [1 ]
Yang, Dingkang [1 ]
Chen, Zhaoyu [1 ]
Liu, Yang [1 ]
Liu, Siao [1 ]
Zhang, Wenqiang [2 ]
Zhang, Lihua [1 ]
Qi, Lizhe [1 ,2 ,3 ]
机构
[1] Fudan Univ, Acad Engn & Technol, Shanghai Engn Res Ctr AI & Robot, Shanghai, Peoples R China
[2] Fudan Univ, Acad Engn & Technol, Engn Res Ctr AI & Robot, Minist Educ, Shanghai, Peoples R China
[3] Green Ecol Smart Technol Sch Enterprise Joint Res, Shanghai, Peoples R China
来源
2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2024年
关键词
CAUSAL INFERENCE;
D O I
10.1109/CVPR52733.2024.01199
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Data-Free Knowledge Distillation (DFKD) is a promising task to train high-performance small models to enhance actual deployment without relying on the original training data. Existing methods commonly avoid relying on private data by utilizing synthetic or sampled data. However, a long-overlooked issue is that the severe distribution shifts between their substitution and original data, which manifests as huge differences in the quality of images and class proportions. The harmful shifts are essentially the confounder that significantly causes performance bottlenecks. To tackle the issue, this paper proposes a novel perspective with causal inference to disentangle the student models from the impact of such shifts. By designing a customized causal graph, we first reveal the causalities among the variables in the DFKD task. Subsequently, we propose a Knowledge Distillation Causal Intervention ( KDCI) framework based on the backdoor adjustment to de-confound the confounder. KDCI can be flexibly combined with most existing state-of-the-art baselines. Experiments in combination with six representative DFKD methods demonstrate the effectiveness of our KDCI, which can obviously help existing methods under almost all settings, e.g., improving the baseline by up to 15.54% accuracy on the CIFAR-100 dataset.
引用
收藏
页码:12615 / 12625
页数:11
相关论文
共 50 条
  • [41] Memory efficient data-free distillation for continual learning
    Li, Xiaorong
    Wang, Shipeng
    Sun, Jian
    Xu, Zongben
    PATTERN RECOGNITION, 2023, 144
  • [42] FedAlign: Federated Model Alignment via Data-Free Knowledge Distillation for Machine Fault Diagnosis
    Sun, Wenjun
    Yan, Ruqiang
    Jin, Ruibing
    Zhao, Rui
    Chen, Zhenghua
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73 : 1 - 12
  • [43] IFHE: Intermediate-Feature Heterogeneity Enhancement for Image Synthesis in Data-Free Knowledge Distillation
    Chen, Yi
    Liu, Ning
    Ren, Ao
    Yang, Tao
    Liu, Duo
    2023 60TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC, 2023,
  • [44] DFRD: Data-Free Robustness Distillation for Heterogeneous Federated Learning
    Luo, Kangyang
    Wang, Shuai
    Fu, Yexuan
    Li, Xiang
    Lan, Yunshi
    Gao, Ming
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [45] Dual discriminator adversarial distillation for data-free model compression
    Zhao, Haoran
    Sun, Xin
    Dong, Junyu
    Manic, Milos
    Zhou, Huiyu
    Yu, Hui
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2022, 13 (05) : 1213 - 1230
  • [46] Dual discriminator adversarial distillation for data-free model compression
    Haoran Zhao
    Xin Sun
    Junyu Dong
    Milos Manic
    Huiyu Zhou
    Hui Yu
    International Journal of Machine Learning and Cybernetics, 2022, 13 : 1213 - 1230
  • [47] Frequency Domain Distillation for Data-Free Quantization of Vision Transformer
    Nan, Gongrui
    Chao, Fei
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT VIII, 2024, 14432 : 205 - 216
  • [48] Model Conversion via Differentially Private Data-Free Distillation
    Liu, Bochao
    Wang, Pengju
    Li, Shikun
    Zeng, Dan
    Ge, Shiming
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 2187 - 2195
  • [49] Augmented Geometric Distillation for Data-Free Incremental Person ReID
    Lu, Yichen
    Wang, Mei
    Deng, Weihong
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 7319 - 7328
  • [50] Data-Free Low-Bit Quantization via Dynamic Multi-teacher Knowledge Distillation
    Huang, Chong
    Lin, Shaohui
    Zhang, Yan
    Li, Ke
    Zhang, Baochang
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT VIII, 2024, 14432 : 28 - 41