Boosting Accuracy and Robustness of Student Models via Adaptive Adversarial Distillation

被引:10
|
作者
Huang, Bo [1 ,2 ]
Chen, Mingyang [1 ,2 ]
Wang, Yi [3 ]
Lu, Junda [4 ]
Cheng, Minhao [2 ]
Wang, Wei [1 ,2 ]
机构
[1] Hong Kong Univ Sci & Technol Guangzhou, Guangzhou, Peoples R China
[2] Hong Kong Univ Sci & Technol, Hong Kong, Peoples R China
[3] Dongguan Univ Technol, Dongguan, Peoples R China
[4] Macquarie Univ, Sydney, NSW, Australia
关键词
D O I
10.1109/CVPR52729.2023.02363
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Distilled student models in teacher-student architectures are widely considered for computational-effective deployment in real-time applications and edge devices. However, there is a higher risk of student models to encounter adversarial attacks at the edge. Popular enhancing schemes such as adversarial training have limited performance on compressed networks. Thus, recent studies concern about adversarial distillation (AD) that aims to inherit not only prediction accuracy but also adversarial robustness of a robust teacher model under the paradigm of robust optimization. In the min-max framework of AD, existing AD methods generally use fixed supervision information from the teacher model to guide the inner optimization for knowledge distillation which often leads to an overcorrection towards model smoothness. In this paper, we propose an adaptive adversarial distillation (AdaAD) that involves the teacher model in the knowledge optimization process in a way interacting with the student model to adaptively search for the inner results. Comparing with state-of-the-art methods, the proposed AdaAD can significantly boost both the prediction accuracy and adversarial robustness of student models in most scenarios. In particular, the ResNet-18 model trained by AdaAD achieves top-rank performance (54.23% robust accuracy) on RobustBench under AutoAttack.
引用
收藏
页码:24668 / 24677
页数:10
相关论文
共 50 条
  • [21] Facial attributes: Accuracy and adversarial robustness
    Rozsa, Andras
    Gunther, Manuel
    Rudd, Ethan M.
    Boult, Terrance E.
    PATTERN RECOGNITION LETTERS, 2019, 124 : 100 - 108
  • [22] Towards Interpreting Vulnerability of Object Detection Models via Adversarial Distillation
    Zhang, Yaoyuan
    Tan, Yu-an
    Lu, Mingfeng
    Liu, Lu
    Zhang, Quanxing
    Li, Yuanzhang
    Wang, Dianxin
    APPLIED CRYPTOGRAPHY AND NETWORK SECURITY WORKSHOPS, ACNS 2022, 2022, 13285 : 53 - 65
  • [23] Towards interpreting vulnerability of object detection models via adversarial distillation
    Zhang, Yaoyuan
    Tan, Yu-an
    Lu, Mingfeng
    Liu, Lu
    Wang, Dianxin
    Zhang, Quanxing
    Li, Yuanzhang
    JOURNAL OF INFORMATION SECURITY AND APPLICATIONS, 2023, 73
  • [24] Learning Differentially Private Diffusion Models via Stochastic Adversarial Distillation
    Liu, Bochao
    Wang, Pengju
    Ge, Shiming
    COMPUTER VISION-ECCV 2024, PT VII, 2025, 15065 : 55 - 71
  • [25] Diverse Knowledge Distillation (DKD): A Solution for Improving The Robustness of Ensemble Models Against Adversarial Attacks
    Mirzaeian, Ali
    Kosecka, Jana
    Homayoun, Houman
    Mohsenin, Tinoosh
    Sasan, Avesta
    PROCEEDINGS OF THE 2021 TWENTY SECOND INTERNATIONAL SYMPOSIUM ON QUALITY ELECTRONIC DESIGN (ISQED 2021), 2021, : 319 - 324
  • [26] Improving adversarial robustness of deep neural networks via adaptive margin evolution
    Ma, Linhai
    Liang, Liang
    NEUROCOMPUTING, 2023, 551
  • [27] An Adaptive Model Ensemble Adversarial Attack for Boosting Adversarial Transferability
    Chen, Bin
    Yin, Jiali
    Chen, Shukai
    Chen, Bohao
    Liu, Ximeng
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 4466 - 4475
  • [28] GAAT: Group Adaptive Adversarial Training to Improve the Trade-Off Between Robustness and Accuracy
    Qian, Yaguan
    Liang, Xiaoyu
    Kang, Ming
    Wang, Bin
    Gu, Zhaoquan
    Wang, Xing
    Wu, Chunming
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2022, 36 (13)
  • [29] Boosting Adversarial Robustness using Feature Level Stochastic Smoothing
    Addepalli, Sravanti
    Jain, Samyak
    Sriramanan, Gaurang
    Babu, R. Venkatesh
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 93 - 102
  • [30] Boosting Barely Robust Learners: A New Perspective on Adversarial Robustness
    Blum, Avrim
    Montasser, Omar
    Shakhnarovich, Greg
    Zhang, Hongyang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,