Boosting Accuracy and Robustness of Student Models via Adaptive Adversarial Distillation

被引:10
|
作者
Huang, Bo [1 ,2 ]
Chen, Mingyang [1 ,2 ]
Wang, Yi [3 ]
Lu, Junda [4 ]
Cheng, Minhao [2 ]
Wang, Wei [1 ,2 ]
机构
[1] Hong Kong Univ Sci & Technol Guangzhou, Guangzhou, Peoples R China
[2] Hong Kong Univ Sci & Technol, Hong Kong, Peoples R China
[3] Dongguan Univ Technol, Dongguan, Peoples R China
[4] Macquarie Univ, Sydney, NSW, Australia
关键词
D O I
10.1109/CVPR52729.2023.02363
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Distilled student models in teacher-student architectures are widely considered for computational-effective deployment in real-time applications and edge devices. However, there is a higher risk of student models to encounter adversarial attacks at the edge. Popular enhancing schemes such as adversarial training have limited performance on compressed networks. Thus, recent studies concern about adversarial distillation (AD) that aims to inherit not only prediction accuracy but also adversarial robustness of a robust teacher model under the paradigm of robust optimization. In the min-max framework of AD, existing AD methods generally use fixed supervision information from the teacher model to guide the inner optimization for knowledge distillation which often leads to an overcorrection towards model smoothness. In this paper, we propose an adaptive adversarial distillation (AdaAD) that involves the teacher model in the knowledge optimization process in a way interacting with the student model to adaptively search for the inner results. Comparing with state-of-the-art methods, the proposed AdaAD can significantly boost both the prediction accuracy and adversarial robustness of student models in most scenarios. In particular, the ResNet-18 model trained by AdaAD achieves top-rank performance (54.23% robust accuracy) on RobustBench under AutoAttack.
引用
收藏
页码:24668 / 24677
页数:10
相关论文
共 50 条
  • [41] Toward a Better Tradeoff Between Accuracy and Robustness for Image Classification via Adversarial Feature Diversity
    Xue, Wei
    Wang, Yonghao
    Wang, Yuchi
    Wang, Yue
    Du, Mingyang
    Zheng, Xiao
    IEEE JOURNAL ON MINIATURIZATION FOR AIR AND SPACE SYSTEMS, 2024, 5 (04): : 254 - 264
  • [42] Robustness meets accuracy in adversarial training for graph autoencoder
    Zhou, Xianchen
    Hu, Kun
    Wang, Hongxia
    NEURAL NETWORKS, 2023, 157 : 114 - 124
  • [43] Towards Better Accuracy and Robustness with Localized Adversarial Training
    Rothberg, Eitan
    Chen, Tingting
    Ji, Hao
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 10017 - 10018
  • [44] Evaluating Accuracy and Adversarial Robustness of Quanvolutional Neural Networks
    Sooksatra, Korn
    Rivas, Pablo
    Orduz, Javier
    2021 INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE AND COMPUTATIONAL INTELLIGENCE (CSCI 2021), 2021, : 152 - 157
  • [45] Boosting adversarial transferability in vision-language models via multimodal feature heterogeneity
    Chen, Long
    Chen, Yuling
    Ouyang, Zhi
    Dou, Hui
    Zhang, Yangwen
    Sang, Haiwei
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [46] SegPGD: An Effective and Efficient Adversarial Attack for Evaluating and Boosting Segmentation Robustness
    Gu, Jindong
    Zhao, Hengshuang
    Tresp, Volker
    Torr, Philip H. S.
    COMPUTER VISION, ECCV 2022, PT XXIX, 2022, 13689 : 308 - 325
  • [47] Adversarial Robustness of Deep Sensor Fusion Models
    Wang, Shaojie
    Wu, Tong
    Chakrabarti, Ayan
    Vorobeychik, Yevgeniy
    2022 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2022), 2022, : 1371 - 1380
  • [48] Adversarial Robustness of Phishing Email Detection Models
    Gholampour, Parisa Mehdi
    Verma, Rakesh M.
    PROCEEDINGS OF THE 9TH ACM INTERNATIONAL WORKSHOP ON SECURITY AND PRIVACY ANALYTICS, IWSPA 2023, 2023, : 67 - 76
  • [49] On the Robustness of Semantic Segmentation Models to Adversarial Attacks
    Arnab, Anurag
    Miksik, Ondrej
    Torr, Philip H. S.
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 888 - 897
  • [50] Improving adversarial robustness through a curriculum-guided reliable distillation
    Li, Jiawen
    Fang, Kun
    Huang, Xiaolin
    Yang, Jie
    COMPUTERS & SECURITY, 2023, 133