Orthogonal Deep Models as Defense Against Black-Box Attacks

被引:5
|
作者
Jalwana, Mohammad A. A. K. [1 ]
Akhtar, Naveed [1 ]
Bennamoun, Mohammed [1 ]
Mian, Ajmal [1 ]
机构
[1] Univ Western Australia, Dept Comp Sci & Software Engn, Perth, WA 6009, Australia
基金
澳大利亚研究理事会;
关键词
Deep learning; adversarial examples; adversarial perturbations; orthogonal models; robust deep learning;
D O I
10.1109/ACCESS.2020.3005961
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning has demonstrated state-of-the-art performance for a variety of challenging computer vision tasks. On one hand, this has enabled deep visual models to pave the way for a plethora of critical applications like disease prognostics and smart surveillance. On the other, deep learning has also been found vulnerable to adversarial attacks, which calls for new techniques to defend deep models against these attacks. Among the attack algorithms, the black-box schemes are of serious practical concern since they only need publicly available knowledge of the targeted model. We carefully analyze the inherent weakness of deep models in black-box settings where the attacker may develop the attack using a model similar to the targeted model. Based on our analysis, we introduce a novel gradient regularization scheme that encourages the internal representation of a deep model to be orthogonal to another, even if the architectures of the two models are similar. Our unique constraint allows a model to concomitantly endeavour for higher accuracy while maintaining near orthogonal alignment of gradients with respect to a reference model. Detailed empirical study verifies that controlled misalignment of gradients under our orthogonality objective significantly boosts a model's robustness against transferable black-box adversarial attacks. In comparison to regular models, the orthogonal models are significantly more robust to a range of l(p) norm bounded perturbations. We verify the effectiveness of our technique on a variety of large-scale models.
引用
收藏
页码:119744 / 119757
页数:14
相关论文
共 50 条
  • [21] B3: Backdoor Attacks against Black-box Machine Learning Models
    Gong, Xueluan
    Chen, Yanjiao
    Yang, Wenbin
    Huang, Huayang
    Wang, Qian
    ACM TRANSACTIONS ON PRIVACY AND SECURITY, 2023, 26 (04)
  • [22] Stateful Defenses for Machine Learning Models Are Not Yet Secure Against Black-box Attacks
    Feng, Ryan
    Hooda, Ashish
    Mangaokar, Neal
    Fawaz, Kassem
    Jha, Somesh
    Prakash, Atul
    PROCEEDINGS OF THE 2023 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, CCS 2023, 2023, : 786 - 800
  • [23] IN DEFENSE OF BLACK-BOX EPIDEMIOLOGY
    SAVITZ, DA
    EPIDEMIOLOGY, 1994, 5 (05) : 550 - 552
  • [24] Simple Black-Box Adversarial Attacks on Deep Neural Networks
    Narodytska, Nina
    Kasiviswanathan, Shiva
    2017 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2017, : 1310 - 1318
  • [25] Black-box Attacks Against Neural Binary Function Detection
    Bundt, Joshua
    Davinroy, Michael
    Agadakos, Ioannis
    Oprea, Alina
    Robertson, William
    PROCEEDINGS OF THE 26TH INTERNATIONAL SYMPOSIUM ON RESEARCH IN ATTACKS, INTRUSIONS AND DEFENSES, RAID 2023, 2023, : 1 - 16
  • [26] Heuristic Black-Box Adversarial Attacks on Video Recognition Models
    Wei, Zhipeng
    Chen, Jingjing
    Wei, Xingxing
    Jiang, Linxi
    Chua, Tat-Seng
    Zhou, Fengfeng
    Jiang, Yu-Gang
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 12338 - 12345
  • [27] Black-Box Adversarial Attacks Against Deep Learning Based Malware Binaries Detection with GAN
    Yuan, Junkun
    Zhou, Shaofang
    Lin, Lanfen
    Wang, Feng
    Cui, Jia
    ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 2536 - 2542
  • [28] Stealthy Black-Box Attacks on Deep Learning Non-Intrusive Load Monitoring Models
    Wang, Junfei
    Srikantha, Pirathayini
    IEEE TRANSACTIONS ON SMART GRID, 2021, 12 (04) : 3479 - 3492
  • [29] Drift Detection for Black-Box Deep Learning Models
    Piano, Luca
    Garcea, Fabio
    Cavallone, Andrea
    Vazquez, Ignacio Aparicio
    Morra, Lia
    Lamberti, Fabrizio
    IT PROFESSIONAL, 2024, 26 (02) : 24 - 31
  • [30] Data reduction for black-box adversarial attacks against deep neural networks based on side-channel attacks
    Zhou, Hanxun
    Liu, Zhihui
    Hu, Yufeng
    Zhang, Shuo
    Kang, Longyu
    Feng, Yong
    Wang, Yan
    Guo, Wei
    Zou, Cliff C.
    COMPUTERS & SECURITY, 2025, 153