Revisiting Adversarial Robustness Distillation from the Perspective of Robust Fairness

被引:0
|
作者
Yue, Xinli [1 ]
Mou, Ningping [1 ]
Wang, Qian [1 ]
Zhao, Lingchen [1 ]
机构
[1] Wuhan Univ, Sch Cyber Sci & Engn, Key Lab Aerosp Informat Secur & Trusted Comp, Minist Educ, Wuhan 430072, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Adversarial Robustness Distillation (ARD) aims to transfer the robustness of large teacher models to small student models, facilitating the attainment of robust performance on resource-limited devices. However, existing research on ARD primarily focuses on the overall robustness of student models, overlooking the crucial aspect of robust fairness. Specifically, these models may demonstrate strong robustness on some classes of data while exhibiting high vulnerability on other classes. Unfortunately, the "buckets effect" implies that the robustness of the deployed model depends on the classes with the lowest level of robustness. In this paper, we first investigate the inheritance of robust fairness during ARD and reveal that student models only partially inherit robust fairness from teacher models. We further validate this issue through fine-grained experiments with various model capacities and find that it may arise due to the gap in capacity between teacher and student models, as well as the existing methods treating each class equally during distillation. Based on these observations, we propose Fair Adversarial Robustness Distillation (Fair-ARD), a novel framework for enhancing the robust fairness of student models by increasing the weights of difficult classes, and design a geometric perspective-based method to quantify the difficulty of different classes for determining the weights. Extensive experiments show that Fair-ARD surpasses both state-of-the-art ARD methods and existing robust fairness algorithms in terms of robust fairness (e.g., the worst-class robustness under AutoAttack is improved by at most 12.3% and 5.3% using ResNet18 on CIFAR10, respectively), while also slightly improving overall robustness. Our code is available at: https://github.com/NISP-official/Fair-ARD.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Revisiting Adversarial Robustness Distillation: Robust Soft Labels Make Student Better
    Zi, Bojia
    Zhao, Shihao
    Ma, Xingjun
    Jiang, Yu-Gang
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 16423 - 16432
  • [2] Boosting Barely Robust Learners: A New Perspective on Adversarial Robustness
    Blum, Avrim
    Montasser, Omar
    Shakhnarovich, Greg
    Zhang, Hongyang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [3] Adversarial Robustness for Latent Models: Revisiting the Robust-Standard Accuracies Tradeoff
    Javanmard, Adel
    Mehrabi, Mohammad
    OPERATIONS RESEARCH, 2024, 72 (03) : 1016 - 1030
  • [4] Individual Fairness Revisited: Transferring Techniques from Adversarial Robustness
    Yeom, Samuel
    Fredrikson, Matt
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 437 - 443
  • [5] Revisiting Residual Networks for Adversarial Robustness
    Huang, Shihua
    Lu, Zhichao
    Deb, Kalyanmoy
    Boddeti, Vishnu Naresh
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 8202 - 8211
  • [6] An Adversarial Perspective on Accuracy, Robustness, Fairness, and Privacy: Multilateral-Tradeoffs in Trustworthy ML
    Gittens, Alex
    Yener, Bulent
    Yung, Moti
    IEEE Access, 2022, 10 : 120850 - 120865
  • [7] An Adversarial Perspective on Accuracy, Robustness, Fairness, and Privacy: Multilateral-Tradeoffs in Trustworthy ML
    Gittens, Alex
    Yener, Bulent
    Yung, Moti
    IEEE ACCESS, 2022, 10 : 120850 - 120865
  • [8] Revisiting model fairness via adversarial examples
    Zhang, Tao
    Zhu, Tianqing
    Li, Jing
    Zhou, Wanlei
    Yu, Philip S.
    KNOWLEDGE-BASED SYSTEMS, 2023, 277
  • [9] On the Adversarial Robustness of Robust Estimators
    Lai, Lifeng
    Bayraktar, Erhan
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2020, 66 (08) : 5097 - 5109
  • [10] On Adversarial Robustness of Demographic Fairness in Face Attribute Recognition
    Zeng, Huimin
    Yue, Zhenrui
    Shang, Lanyu
    Zhang, Yang
    Wang, Dong
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 527 - 535