UNMASK: Adversarial Detection and Defense Through Robust Feature Alignment

被引:15
|
作者
Freitas, Scott [1 ]
Chen, Shang-Tse [2 ]
Wang, Zijie J. [1 ]
Chau, Duen Horng [1 ]
机构
[1] Georgia Tech, Atlanta, GA 30332 USA
[2] Natl Taiwan Univ, Taipei, Taiwan
来源
2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA) | 2020年
关键词
deep learning; adversarial defense; robust features; adversarial detection;
D O I
10.1109/BigData50022.2020.9378303
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent research has demonstrated that deep learning architectures are vulnerable to adversarial attacks, highlighting the vital need for defensive techniques to detect and mitigate these attacks before they occur. We present UNMASK, an adversarial detection and defense framework based on robust feature alignment. UNMASK combats adversarial attacks by extracting robust features (e.g., beak, wings, eyes) from an image (e.g., "bird") and comparing them to the expected features of the classification. For example, if the extracted features for a "bird" image are wheel, saddle and frame, the model may be under attack. UNMASK detects such attacks and defends the model by rectifying the misclassification, re-classifying the image based on its robust features. Our extensive evaluation shows that UNMASK detects up to 96.75% of attacks, and defends the model by correctly classifying up to 93% of adversarial images produced by the current strongest attack, Projected Gradient Descent, in the gray-box setting. UNMASK provides significantly better protection than adversarial training across 8 attack vectors, averaging 31.18% higher accuracy. We open source the code repository and data with this paper: https://github.com/safreita1/unmask.
引用
收藏
页码:1081 / 1088
页数:8
相关论文
共 50 条
  • [41] Robust Image Alignment for Tampering Detection
    Battiato, Sebastiano
    Farinella, Giovanni Maria
    Messina, Enrico
    Puglisi, Giovanni
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2012, 7 (04) : 1105 - 1117
  • [42] Detection defense against adversarial attacks with saliency map
    Ye, Dengpan
    Chen, Chuanxi
    Liu, Changrui
    Wang, Hao
    Jiang, Shunzhi
    INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2022, 37 (12) : 10193 - 10210
  • [43] Defense Against Adversarial Attacks on Audio DeepFake Detection
    Kawa, Piotr
    Plata, Marcin
    Syga, Piotr
    INTERSPEECH 2023, 2023, : 5276 - 5280
  • [44] Class-Disentanglement and Applications in Adversarial Detection and Defense
    Yang, Kaiwen
    Zhou, Tianyi
    Zhang, Yonggang
    Tian, Xinmei
    Tao, Dacheng
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [45] Feature partitioning for robust tree ensembles and their certification in adversarial scenarios
    Stefano Calzavara
    Claudio Lucchese
    Federico Marcuzzi
    Salvatore Orlando
    EURASIP Journal on Information Security, 2021
  • [46] Feature partitioning for robust tree ensembles and their certification in adversarial scenarios
    Calzavara, Stefano
    Lucchese, Claudio
    Marcuzzi, Federico
    Orlando, Salvatore
    EURASIP JOURNAL ON INFORMATION SECURITY, 2021, 2021 (01)
  • [47] Towards robust classification detection for adversarial examples
    Liu, Huangxiaolie
    Zhang, Dong
    Chen, Huijun
    INTERNATIONAL CONFERENCE FOR INTERNET TECHNOLOGY AND SECURED TRANSACTIONS (ICITST-2020), 2020, : 23 - 29
  • [48] Adversarial intensity awareness for robust object detection
    Cheng, Jikang
    Huang, Baojin
    Fang, Yan
    Han, Zhen
    Wang, Zhongyuan
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2025, 251
  • [49] Adversarial Discriminative Attention for Robust Anomaly Detection
    Kimura, Daiki
    Chaudhury, Subhajit
    Narita, Minori
    Munawar, Asim
    Tachibana, Ryuki
    2020 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2020, : 2161 - 2170
  • [50] ROBUST DETECTION OF ADVERSARIAL ATTACKS ON MEDICAL IMAGES
    Li, Xin
    Zhu, Dongxiao
    2020 IEEE 17TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI 2020), 2020, : 1154 - 1158