UNMASK: Adversarial Detection and Defense Through Robust Feature Alignment

被引:15
|
作者
Freitas, Scott [1 ]
Chen, Shang-Tse [2 ]
Wang, Zijie J. [1 ]
Chau, Duen Horng [1 ]
机构
[1] Georgia Tech, Atlanta, GA 30332 USA
[2] Natl Taiwan Univ, Taipei, Taiwan
关键词
deep learning; adversarial defense; robust features; adversarial detection;
D O I
10.1109/BigData50022.2020.9378303
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent research has demonstrated that deep learning architectures are vulnerable to adversarial attacks, highlighting the vital need for defensive techniques to detect and mitigate these attacks before they occur. We present UNMASK, an adversarial detection and defense framework based on robust feature alignment. UNMASK combats adversarial attacks by extracting robust features (e.g., beak, wings, eyes) from an image (e.g., "bird") and comparing them to the expected features of the classification. For example, if the extracted features for a "bird" image are wheel, saddle and frame, the model may be under attack. UNMASK detects such attacks and defends the model by rectifying the misclassification, re-classifying the image based on its robust features. Our extensive evaluation shows that UNMASK detects up to 96.75% of attacks, and defends the model by correctly classifying up to 93% of adversarial images produced by the current strongest attack, Projected Gradient Descent, in the gray-box setting. UNMASK provides significantly better protection than adversarial training across 8 attack vectors, averaging 31.18% higher accuracy. We open source the code repository and data with this paper: https://github.com/safreita1/unmask.
引用
收藏
页码:1081 / 1088
页数:8
相关论文
共 50 条
  • [31] Robust Training with Feature-Based Adversarial Example
    Fu, Xuanming
    Yang, Zhengfeng
    Xue, Hao
    Wang, Jianlin
    Zeng, Zhenbing
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2957 - 2963
  • [32] Towards robust adversarial defense on perturbed graphs with noisy labels
    Li, Ding
    Xia, Hui
    Hu, Chunqiang
    Zhang, Rui
    Du, Yu
    Feng, Xiaolong
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 269
  • [33] An adversarial defense algorithm based on robust U-net
    Zhang, Dian
    Dong, Yunwei
    Yang, Hongji
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (15) : 45575 - 45601
  • [35] An adversarial defense algorithm based on robust U-net
    Dian Zhang
    Yunwei Dong
    Hongji Yang
    Multimedia Tools and Applications, 2024, 83 : 45575 - 45601
  • [36] Towards Robust Ensemble Defense Against Adversarial Examples Attack
    Mani, Nag
    Moh, Melody
    Moh, Teng-Sheng
    2019 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2019,
  • [37] Mean Teacher DETR with Masked Feature Alignment: A Robust Domain Adaptive Detection Transformer Framework
    Weng, Weixi
    Yuan, Chun
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 6, 2024, : 5912 - 5920
  • [38] Controllable Invariance through Adversarial Feature Learning
    Xie, Qizhe
    Dai, Zihang
    Du, Yulun
    Hovy, Eduard
    Neubig, Graham
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [39] Boosting Adversarial Transferability Through Intermediate Feature
    He, Chenghai
    Li, Xiaoqian
    Zhang, Xiaohang
    Zhang, Kai
    Li, Hailing
    Xiong, Gang
    Li, Xuan
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT V, 2023, 14258 : 28 - 39
  • [40] Detection of Adversarial DDoS Attacks Using Symmetric Defense Generative Adversarial Networks
    Shieh, Chin-Shiuh
    Thanh-Tuan Nguyen
    Lin, Wan-Wei
    Lai, Wei Kuang
    Horng, Mong-Fong
    Miu, Denis
    ELECTRONICS, 2022, 11 (13)