De-confounded Data-free Knowledge Distillation for Handling Distribution Shifts

被引:2
|
作者
Wang, Yuzheng [1 ]
Yang, Dingkang [1 ]
Chen, Zhaoyu [1 ]
Liu, Yang [1 ]
Liu, Siao [1 ]
Zhang, Wenqiang [2 ]
Zhang, Lihua [1 ]
Qi, Lizhe [1 ,2 ,3 ]
机构
[1] Fudan Univ, Acad Engn & Technol, Shanghai Engn Res Ctr AI & Robot, Shanghai, Peoples R China
[2] Fudan Univ, Acad Engn & Technol, Engn Res Ctr AI & Robot, Minist Educ, Shanghai, Peoples R China
[3] Green Ecol Smart Technol Sch Enterprise Joint Res, Shanghai, Peoples R China
关键词
CAUSAL INFERENCE;
D O I
10.1109/CVPR52733.2024.01199
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Data-Free Knowledge Distillation (DFKD) is a promising task to train high-performance small models to enhance actual deployment without relying on the original training data. Existing methods commonly avoid relying on private data by utilizing synthetic or sampled data. However, a long-overlooked issue is that the severe distribution shifts between their substitution and original data, which manifests as huge differences in the quality of images and class proportions. The harmful shifts are essentially the confounder that significantly causes performance bottlenecks. To tackle the issue, this paper proposes a novel perspective with causal inference to disentangle the student models from the impact of such shifts. By designing a customized causal graph, we first reveal the causalities among the variables in the DFKD task. Subsequently, we propose a Knowledge Distillation Causal Intervention ( KDCI) framework based on the backdoor adjustment to de-confound the confounder. KDCI can be flexibly combined with most existing state-of-the-art baselines. Experiments in combination with six representative DFKD methods demonstrate the effectiveness of our KDCI, which can obviously help existing methods under almost all settings, e.g., improving the baseline by up to 15.54% accuracy on the CIFAR-100 dataset.
引用
收藏
页码:12615 / 12625
页数:11
相关论文
共 50 条
  • [1] Momentum Adversarial Distillation: Handling Large Distribution Shifts in Data-Free Knowledge Distillation
    Do, Kien
    Le, Hung
    Dung Nguyen
    Dang Nguyen
    Harikumar, Haripriya
    Truyen Tran
    Rana, Santu
    Venkatesh, Svetha
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [2] Conditional generative data-free knowledge distillation
    Yu, Xinyi
    Yan, Ling
    Yang, Yang
    Zhou, Libo
    Ou, Linlin
    IMAGE AND VISION COMPUTING, 2023, 131
  • [3] Data-free Knowledge Distillation for Object Detection
    Chawla, Akshay
    Yin, Hongxu
    Molchanov, Pavlo
    Alvarez, Jose
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WACV 2021, 2021, : 3288 - 3297
  • [4] Data-Free Network Quantization With Adversarial Knowledge Distillation
    Choi, Yoojin
    Choi, Jihwan
    El-Khamy, Mostafa
    Lee, Jungwon
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, : 3047 - 3057
  • [5] ROBUSTNESS AND DIVERSITY SEEKING DATA-FREE KNOWLEDGE DISTILLATION
    Han, Pengchao
    Park, Jihong
    Wang, Shiqiang
    Liu, Yejun
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 2740 - 2744
  • [6] Data-free knowledge distillation in neural networks for regression
    Kang, Myeonginn
    Kang, Seokho
    EXPERT SYSTEMS WITH APPLICATIONS, 2021, 175
  • [7] Empirical Study of Data-Free Iterative Knowledge Distillation
    Shah, Het
    Vaswani, Ashwin
    Dash, Tirtharaj
    Hebbalaguppe, Ramya
    Srinivasan, Ashwin
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT III, 2021, 12893 : 546 - 557
  • [8] Data-free Knowledge Distillation for Reusing Recommendation Models
    Wang, Cheng
    Sun, Jiacheng
    Dong, Zhenhua
    Zhu, Jieming
    Li, Zhenguo
    Li, Ruixuan
    Zhang, Rui
    PROCEEDINGS OF THE 17TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, RECSYS 2023, 2023, : 386 - 395
  • [9] Variational Data-Free Knowledge Distillation for Continual Learning
    Li, Xiaorong
    Wang, Shipeng
    Sun, Jian
    Xu, Zongben
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (10) : 12618 - 12634
  • [10] Data-Free Knowledge Distillation for Heterogeneous Federated Learning
    Zhu, Zhuangdi
    Hong, Junyuan
    Zhou, Jiayu
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139