Robust Distillation via Untargeted and Targeted Intermediate Adversarial Samples

被引:1
|
作者
Dong, Junhao [1 ,2 ]
Koniusz, Piotr [3 ,4 ]
Chen, Junxi [5 ]
Wang, Z. Jane [6 ]
Ong, Yew-Soon [1 ,2 ]
机构
[1] Nanyang Technol Univ, Singapore, Singapore
[2] ASTAR, IHPC, CFAR, Singapore, Singapore
[3] Australian Natl Univ, Canberra, ACT, Australia
[4] Data61 CSIRO, Eveleigh, NSW, Australia
[5] Sun Yat Sen Univ, Guangzhou, Peoples R China
[6] Univ British Columbia, Vancouver, BC, Canada
基金
新加坡国家研究基金会;
关键词
D O I
10.1109/CVPR52733.2024.02686
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Adversarially robust knowledge distillation aims to compress large-scale models into lightweight models while preserving adversarial robustness and natural performance on a given dataset. Existing methods typically align probability distributions of natural and adversarial samples between teacher and student models, but they overlook intermediate adversarial samples along the "adversarial path" formed by the multi-step gradient ascent of a sample towards the decision boundary. Such paths capture rich information about the decision boundary. In this paper, we propose a novel adversarially robust knowledge distillation approach by incorporating such adversarial paths into the alignment process. Recognizing the diverse impacts of intermediate adversarial samples (ranging from benign to noisy), we propose an adaptive weighting strategy to selectively emphasize informative adversarial samples, thus ensuring efficient utilization of lightweight model capacity. Moreover, we propose a dual-branch mechanism exploiting two following insights: (i) complementary dynamics of adversarial paths obtained by targeted and untargeted adversarial learning, and (ii) inherent differences between the gradient ascent path from class c(i) towards the nearest class boundary and the gradient descent path from a specific class c(j) towards the decision region of c(i) (i not equal j). Comprehensive experiments demonstrate the effectiveness of our method on lightweight models under various settings.
引用
收藏
页码:28432 / 28442
页数:11
相关论文
共 50 条
  • [1] Untargeted, Targeted and Universal Adversarial Attacks and Defenses on Time Series
    Rathore, Pradeep
    Basak, Arghya
    Nistala, Harsha
    Runkana, Venkataramana
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [2] Guided Adversarial Contrastive Distillation for Robust Students
    Bai, Tao
    Zhao, Jun
    Wen, Bihan
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 9643 - 9655
  • [3] UNTARGETED ADVERSARIAL ATTACK VIA EXPANDING THE SEMANTIC GAP
    Wu, Aming
    Han, Yahong
    Zhang, Quanxin
    Kuang, Xiaohui
    2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, : 514 - 519
  • [4] Knowledge Distillation with Adversarial Samples Supporting Decision Boundary
    Heo, Byeongho
    Lee, Minsik
    Yun, Sangdoo
    Choi, Jin Young
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 3771 - 3778
  • [5] Adversarial Knowledge Distillation For Robust Spoken Language Understanding
    Wang, Ye
    Ling, Baishun
    Wang, Yanmeng
    Xue, Junhao
    Wang, Shaojun
    Xiao, Jing
    INTERSPEECH 2022, 2022, : 2708 - 2712
  • [6] Curriculum Optimisation via Evolutionary Computation, for a Neural Learner Robust to Categorical Adversarial Samples
    Leu, George
    Tang, Jiangjun
    Lakshika, Erandi
    Merrick, Kathryn
    Barlow, Michael
    2017 FOURTH ASIAN CONFERENCE ON DEFENCE TECHNOLOGY - JAPAN (ACDT), 2017, : 11 - 16
  • [7] Robust Regularization with Adversarial Labelling of Perturbed Samples
    Guo, Xiaohui
    Zhang, Richong
    Zheng, Yaowei
    Mao, Yongyi
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2490 - 2496
  • [8] Revisiting Adversarial Robustness Distillation from the Perspective of Robust Fairness
    Yue, Xinli
    Mou, Ningping
    Wang, Qian
    Zhao, Lingchen
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [9] Learning Ordered Top-k Adversarial Attacks via Adversarial Distillation
    Zhang, Zekun
    Wu, Tianfu
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, : 3364 - 3373
  • [10] Robust Adversarial Classification via Abstaining
    Al Makdah, Abed AlRahman
    Katewa, Vaibhav
    Pasqualetti, Fabio
    2021 60TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2021, : 763 - 768