Revisiting Adversarial Robustness Distillation from the Perspective of Robust Fairness

被引:0
|
作者
Yue, Xinli [1 ]
Mou, Ningping [1 ]
Wang, Qian [1 ]
Zhao, Lingchen [1 ]
机构
[1] Wuhan Univ, Sch Cyber Sci & Engn, Key Lab Aerosp Informat Secur & Trusted Comp, Minist Educ, Wuhan 430072, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Adversarial Robustness Distillation (ARD) aims to transfer the robustness of large teacher models to small student models, facilitating the attainment of robust performance on resource-limited devices. However, existing research on ARD primarily focuses on the overall robustness of student models, overlooking the crucial aspect of robust fairness. Specifically, these models may demonstrate strong robustness on some classes of data while exhibiting high vulnerability on other classes. Unfortunately, the "buckets effect" implies that the robustness of the deployed model depends on the classes with the lowest level of robustness. In this paper, we first investigate the inheritance of robust fairness during ARD and reveal that student models only partially inherit robust fairness from teacher models. We further validate this issue through fine-grained experiments with various model capacities and find that it may arise due to the gap in capacity between teacher and student models, as well as the existing methods treating each class equally during distillation. Based on these observations, we propose Fair Adversarial Robustness Distillation (Fair-ARD), a novel framework for enhancing the robust fairness of student models by increasing the weights of difficult classes, and design a geometric perspective-based method to quantify the difficulty of different classes for determining the weights. Extensive experiments show that Fair-ARD surpasses both state-of-the-art ARD methods and existing robust fairness algorithms in terms of robust fairness (e.g., the worst-class robustness under AutoAttack is improved by at most 12.3% and 5.3% using ResNet18 on CIFAR10, respectively), while also slightly improving overall robustness. Our code is available at: https://github.com/NISP-official/Fair-ARD.
引用
收藏
页数:12
相关论文
共 50 条
  • [31] Revisiting Fairness in MPC: Polynomial Number of Parties and General Adversarial Structures
    Dachman-Soled, Dana
    THEORY OF CRYPTOGRAPHY, TCC 2020, PT II, 2020, 12551 : 595 - 620
  • [32] Enhancing Adversarial Robustness via Stochastic Robust Framework
    Sun, Zhenjiang
    Li, Yuanbo
    Hu, Cong
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT IV, 2024, 14428 : 187 - 198
  • [33] Transferring Adversarial Robustness Through Robust Representation Matching
    Vaishnavi, Pratik
    Eykholt, Kevin
    Rahmati, Amir
    PROCEEDINGS OF THE 31ST USENIX SECURITY SYMPOSIUM, 2022, : 2083 - 2098
  • [34] Adversarial Knowledge Distillation For Robust Spoken Language Understanding
    Wang, Ye
    Ling, Baishun
    Wang, Yanmeng
    Xue, Junhao
    Wang, Shaojun
    Xiao, Jing
    INTERSPEECH 2022, 2022, : 2708 - 2712
  • [35] Adversarial robustness via robust low rank representations
    Awasthi, Pranjal
    Jain, Himanshu
    Rawat, Ankit Singh
    Vijayaraghavan, Aravindan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS (NEURIPS 2020), 2020, 33
  • [36] Enhanced Accuracy and Robustness via Multi-teacher Adversarial Distillation
    Zhao, Shiji
    Yu, Jie
    Sun, Zhenlong
    Zhang, Bo
    Wei, Xingxing
    COMPUTER VISION - ECCV 2022, PT IV, 2022, 13664 : 585 - 602
  • [37] Boosting Accuracy and Robustness of Student Models via Adaptive Adversarial Distillation
    Huang, Bo
    Chen, Mingyang
    Wang, Yi
    Lu, Junda
    Cheng, Minhao
    Wang, Wei
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 24668 - 24677
  • [38] Improving adversarial robustness through a curriculum-guided reliable distillation
    Li, Jiawen
    Fang, Kun
    Huang, Xiaolin
    Yang, Jie
    COMPUTERS & SECURITY, 2023, 133
  • [39] Fairness is essential for robustness: fair adversarial training by identifying and augmenting hard examples
    Mou, Ningping
    Yue, Xinli
    Zhao, Lingchen
    Wang, Qian
    FRONTIERS OF COMPUTER SCIENCE, 2025, 19 (03)
  • [40] Fairness-aware training of face attribute classifiers via adversarial robustness
    Zeng, Huimin
    Yue, Zhenrui
    Kou, Ziyi
    Zhang, Yang
    Shang, Lanyu
    Wang, Dong
    KNOWLEDGE-BASED SYSTEMS, 2023, 264