Orthogonal Deep Models as Defense Against Black-Box Attacks

被引:5
|
作者
Jalwana, Mohammad A. A. K. [1 ]
Akhtar, Naveed [1 ]
Bennamoun, Mohammed [1 ]
Mian, Ajmal [1 ]
机构
[1] Univ Western Australia, Dept Comp Sci & Software Engn, Perth, WA 6009, Australia
基金
澳大利亚研究理事会;
关键词
Deep learning; adversarial examples; adversarial perturbations; orthogonal models; robust deep learning;
D O I
10.1109/ACCESS.2020.3005961
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning has demonstrated state-of-the-art performance for a variety of challenging computer vision tasks. On one hand, this has enabled deep visual models to pave the way for a plethora of critical applications like disease prognostics and smart surveillance. On the other, deep learning has also been found vulnerable to adversarial attacks, which calls for new techniques to defend deep models against these attacks. Among the attack algorithms, the black-box schemes are of serious practical concern since they only need publicly available knowledge of the targeted model. We carefully analyze the inherent weakness of deep models in black-box settings where the attacker may develop the attack using a model similar to the targeted model. Based on our analysis, we introduce a novel gradient regularization scheme that encourages the internal representation of a deep model to be orthogonal to another, even if the architectures of the two models are similar. Our unique constraint allows a model to concomitantly endeavour for higher accuracy while maintaining near orthogonal alignment of gradients with respect to a reference model. Detailed empirical study verifies that controlled misalignment of gradients under our orthogonality objective significantly boosts a model's robustness against transferable black-box adversarial attacks. In comparison to regular models, the orthogonal models are significantly more robust to a range of l(p) norm bounded perturbations. We verify the effectiveness of our technique on a variety of large-scale models.
引用
收藏
页码:119744 / 119757
页数:14
相关论文
共 50 条
  • [1] Boundary Defense Against Black-box Adversarial Attacks
    Aithal, Manjushree B.
    Li, Xiaohua
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2349 - 2356
  • [2] An Adaptive Black-Box Defense Against Trojan Attacks (TROJDEF)
    Liu, Guanxiong
    Khreishah, Abdallah
    Sharadgah, Fatima
    Khalil, Issa
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (04) : 5367 - 5381
  • [3] An Adaptive Black-box Defense against Trojan Attacks on Text Data
    Alsharadgah, Fatima
    Khreishah, Abdallah
    Al-Ayyoub, Mahmoud
    Jararweh, Yaser
    Liu, Guanxiong
    Khalil, Issa
    Almutiry, Muhannad
    Saeed, Nasir
    2021 EIGHTH INTERNATIONAL CONFERENCE ON SOCIAL NETWORK ANALYSIS, MANAGEMENT AND SECURITY (SNAMS), 2021, : 155 - 162
  • [4] Black-Box Adversarial Attacks against Audio Forensics Models
    Jiang, Yi
    Ye, Dengpan
    SECURITY AND COMMUNICATION NETWORKS, 2022, 2022
  • [5] Random Noise Defense Against Query-Based Black-Box Attacks
    Qin, Zeyu
    Fan, Yanbo
    Zha, Hongyuan
    Wu, Baoyuan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [6] Towards Lightweight Black-Box Attacks Against Deep Neural Networks
    Sun, Chenghao
    Zhang, Yonggang
    Wan Chaoqun
    Wang, Qizhou
    Li, Ya
    Liu, Tongliang
    Han, Bo
    Tian, Xinmei
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [7] Ensemble adversarial black-box attacks against deep learning systems
    Hang, Jie
    Han, Keji
    Chen, Hui
    Li, Yun
    PATTERN RECOGNITION, 2020, 101
  • [8] Efficient Label Contamination Attacks Against Black-Box Learning Models
    Zhao, Mengchen
    An, Bo
    Gao, Wei
    Zhang, Teng
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 3945 - 3951
  • [9] Black-box adversarial attacks against image quality assessment models
    Ran, Yu
    Zhang, Ao-Xiang
    Li, Mingjie
    Tang, Weixuan
    Wang, Yuan-Gen
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 260
  • [10] Efficient Black-Box Adversarial Attacks for Deep Driving Maneuver Classification Models
    Sarker, Ankur
    Shen, Haiying
    Sen, Tanmoy
    Mendelson, Quincy
    2021 IEEE 18TH INTERNATIONAL CONFERENCE ON MOBILE AD HOC AND SMART SYSTEMS (MASS 2021), 2021, : 536 - 544