UNMASK: Adversarial Detection and Defense Through Robust Feature Alignment

被引:15
|
作者
Freitas, Scott [1 ]
Chen, Shang-Tse [2 ]
Wang, Zijie J. [1 ]
Chau, Duen Horng [1 ]
机构
[1] Georgia Tech, Atlanta, GA 30332 USA
[2] Natl Taiwan Univ, Taipei, Taiwan
关键词
deep learning; adversarial defense; robust features; adversarial detection;
D O I
10.1109/BigData50022.2020.9378303
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent research has demonstrated that deep learning architectures are vulnerable to adversarial attacks, highlighting the vital need for defensive techniques to detect and mitigate these attacks before they occur. We present UNMASK, an adversarial detection and defense framework based on robust feature alignment. UNMASK combats adversarial attacks by extracting robust features (e.g., beak, wings, eyes) from an image (e.g., "bird") and comparing them to the expected features of the classification. For example, if the extracted features for a "bird" image are wheel, saddle and frame, the model may be under attack. UNMASK detects such attacks and defends the model by rectifying the misclassification, re-classifying the image based on its robust features. Our extensive evaluation shows that UNMASK detects up to 96.75% of attacks, and defends the model by correctly classifying up to 93% of adversarial images produced by the current strongest attack, Projected Gradient Descent, in the gray-box setting. UNMASK provides significantly better protection than adversarial training across 8 attack vectors, averaging 31.18% higher accuracy. We open source the code repository and data with this paper: https://github.com/safreita1/unmask.
引用
收藏
页码:1081 / 1088
页数:8
相关论文
共 50 条
  • [1] Robust feature learning for adversarial defense via hierarchical feature alignment
    Zhang, Xiaoqin
    Wang, Jinxin
    Wang, Tao
    Jiang, Runhua
    Xu, Jiawei
    Zhao, Li
    INFORMATION SCIENCES, 2021, 560 : 256 - 270
  • [2] Boosting transferability of targeted adversarial examples with non-robust feature alignment
    Zhu, Hegui
    Sui, Xiaoyan
    Ren, Yuchen
    Jia, Yanmeng
    Zhang, Libo
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 227
  • [3] ROLDEF: RObust Layered DEFense for Intrusion Detection Against Adversarial Attacks
    Gungor, Onat
    Rosing, Tajana
    Alcsanli, Bans
    2024 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION, DATE, 2024,
  • [4] Adaptive feature alignment for adversarial training
    Zhao, Kai
    Wang, Tao
    Zhang, Ruixin
    Shen, Wei
    PATTERN RECOGNITION LETTERS, 2024, 186 : 184 - 190
  • [5] Stylized Pairing for Robust Adversarial Defense
    Guan, Dejian
    Zhao, Wentao
    Liu, Xiao
    APPLIED SCIENCES-BASEL, 2022, 12 (18):
  • [6] Adversarial anchor-guided feature refinement for adversarial defense
    Lee, Hakmin
    Ro, Yong Man
    IMAGE AND VISION COMPUTING, 2023, 136
  • [7] Attack-less adversarial training for a robust adversarial defense
    Ho, Jiacang
    Lee, Byung-Gook
    Kang, Dae-Ki
    APPLIED INTELLIGENCE, 2022, 52 (04) : 4364 - 4381
  • [8] Attack-less adversarial training for a robust adversarial defense
    Jiacang Ho
    Byung-Gook Lee
    Dae-Ki Kang
    Applied Intelligence, 2022, 52 : 4364 - 4381
  • [9] Defense against Adversarial Patch Attacks for Aerial Image Semantic Segmentation by Robust Feature Extraction
    Wang, Zhen
    Wang, Buhong
    Zhang, Chuanlei
    Liu, Yaohui
    REMOTE SENSING, 2023, 15 (06)
  • [10] Shielding Object Detection: Enhancing Adversarial Defense through Ensemble Methods
    Peng, Ziwen
    Chen, Xi
    Huang, Wei
    Kong, Xianglong
    Li, Jianpeng
    Xue, Song
    2024 5TH INFORMATION COMMUNICATION TECHNOLOGIES CONFERENCE, ICTC 2024, 2024, : 88 - 97