Revisiting Adversarial Robustness Distillation from the Perspective of Robust Fairness

被引:0
|
作者
Yue, Xinli [1 ]
Mou, Ningping [1 ]
Wang, Qian [1 ]
Zhao, Lingchen [1 ]
机构
[1] Wuhan Univ, Sch Cyber Sci & Engn, Key Lab Aerosp Informat Secur & Trusted Comp, Minist Educ, Wuhan 430072, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Adversarial Robustness Distillation (ARD) aims to transfer the robustness of large teacher models to small student models, facilitating the attainment of robust performance on resource-limited devices. However, existing research on ARD primarily focuses on the overall robustness of student models, overlooking the crucial aspect of robust fairness. Specifically, these models may demonstrate strong robustness on some classes of data while exhibiting high vulnerability on other classes. Unfortunately, the "buckets effect" implies that the robustness of the deployed model depends on the classes with the lowest level of robustness. In this paper, we first investigate the inheritance of robust fairness during ARD and reveal that student models only partially inherit robust fairness from teacher models. We further validate this issue through fine-grained experiments with various model capacities and find that it may arise due to the gap in capacity between teacher and student models, as well as the existing methods treating each class equally during distillation. Based on these observations, we propose Fair Adversarial Robustness Distillation (Fair-ARD), a novel framework for enhancing the robust fairness of student models by increasing the weights of difficult classes, and design a geometric perspective-based method to quantify the difficulty of different classes for determining the weights. Extensive experiments show that Fair-ARD surpasses both state-of-the-art ARD methods and existing robust fairness algorithms in terms of robust fairness (e.g., the worst-class robustness under AutoAttack is improved by at most 12.3% and 5.3% using ResNet18 on CIFAR10, respectively), while also slightly improving overall robustness. Our code is available at: https://github.com/NISP-official/Fair-ARD.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] Adversarial Robustness of Neural Networks from the Perspective of Lipschitz Calculus: A Survey
    Zuehlke, Monty-maximilian
    Kudenko, Daniel
    ACM COMPUTING SURVEYS, 2025, 57 (06)
  • [22] Revisiting Hilbert-Schmidt Information Bottleneck for Adversarial Robustness
    Wang, Zifeng
    Jian, Tong
    Masoomi, Aria
    Ioannidis, Stratis
    Dy, Jennifer
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [23] Improving Robust Fairness via Balance Adversarial Training
    Sun, Chunyu
    Xu, Chenye
    Yao, Chengyuan
    Liang, Siyuan
    Wu, Yichao
    Liang, Ding
    Liu, Xianglong
    Liu, Aishan
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 12, 2023, : 15161 - 15169
  • [24] Revisiting single-step adversarial training for robustness and generalization
    Li, Zhuorong
    Yu, Daiwei
    Wu, Minghui
    Chan, Sixian
    Yu, Hongchuan
    Han, Zhike
    PATTERN RECOGNITION, 2024, 151
  • [25] Fairness-Aware Regression Robust to Adversarial Attacks
    Jin, Yulu
    Lai, Lifeng
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2023, 71 : 4092 - 4105
  • [26] Revisiting the Adversarial Robustness-Accuracy Tradeoff in Robot Learning
    Lechner, Mathias
    Amini, Alexander
    Rus, Daniela
    Henzinger, Thomas A.
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (03) : 1595 - 1602
  • [27] On Adversarial Robustness: A Neural Architecture Search perspective
    Devaguptapu, Chaitanya
    Agarwal, Devansh
    Mittal, Gaurav
    Gopalani, Pulkit
    Balasubramanian, Vineeth N.
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2021), 2021, : 152 - 161
  • [28] A Federated Framework for Edge Computing Devices with Collaborative Fairness and Adversarial Robustness
    Yang, Hailin
    Huang, Yanhong
    Shi, Jianqi
    Yang, Yang
    JOURNAL OF GRID COMPUTING, 2023, 21 (03)
  • [29] A Federated Framework for Edge Computing Devices with Collaborative Fairness and Adversarial Robustness
    Hailin Yang
    Yanhong Huang
    Jianqi Shi
    Yang Yang
    Journal of Grid Computing, 2023, 21
  • [30] Causal Adversarial Perturbations for Individual Fairness and Robustness in Heterogeneous Data Spaces
    Ehyaei, Ahmad-Reza
    Mohammadi, Kiarash
    Karimi, Amir-Hossein
    Samadi, Samira
    Farnadi, Golnoosh
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 10, 2024, : 11847 - 11855