Orthogonal Deep Models as Defense Against Black-Box Attacks

被引:5
|
作者
Jalwana, Mohammad A. A. K. [1 ]
Akhtar, Naveed [1 ]
Bennamoun, Mohammed [1 ]
Mian, Ajmal [1 ]
机构
[1] Univ Western Australia, Dept Comp Sci & Software Engn, Perth, WA 6009, Australia
基金
澳大利亚研究理事会;
关键词
Deep learning; adversarial examples; adversarial perturbations; orthogonal models; robust deep learning;
D O I
10.1109/ACCESS.2020.3005961
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning has demonstrated state-of-the-art performance for a variety of challenging computer vision tasks. On one hand, this has enabled deep visual models to pave the way for a plethora of critical applications like disease prognostics and smart surveillance. On the other, deep learning has also been found vulnerable to adversarial attacks, which calls for new techniques to defend deep models against these attacks. Among the attack algorithms, the black-box schemes are of serious practical concern since they only need publicly available knowledge of the targeted model. We carefully analyze the inherent weakness of deep models in black-box settings where the attacker may develop the attack using a model similar to the targeted model. Based on our analysis, we introduce a novel gradient regularization scheme that encourages the internal representation of a deep model to be orthogonal to another, even if the architectures of the two models are similar. Our unique constraint allows a model to concomitantly endeavour for higher accuracy while maintaining near orthogonal alignment of gradients with respect to a reference model. Detailed empirical study verifies that controlled misalignment of gradients under our orthogonality objective significantly boosts a model's robustness against transferable black-box adversarial attacks. In comparison to regular models, the orthogonal models are significantly more robust to a range of l(p) norm bounded perturbations. We verify the effectiveness of our technique on a variety of large-scale models.
引用
收藏
页码:119744 / 119757
页数:14
相关论文
共 50 条
  • [41] Fuzzing-based hard-label black-box attacks against machine learning models
    Qin, Yi
    Yue, Chuan
    COMPUTERS & SECURITY, 2022, 117
  • [42] 3D-Adv: Black-Box Adversarial Attacks against Deep Learning Models through 3D Sensors
    Yang, Kaichen
    Lin, Xuan-Yi
    Sun, Yixin
    Ho, Tsung-Yi
    Jin, Yier
    2021 58TH ACM/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2021, : 547 - 552
  • [43] Universal Certified Defense for Black-Box Models Based on Random Smoothing
    Li Q.
    Chen J.
    Zhang Z.-J.
    He K.
    Du R.-Y.
    Wang X.-X.
    Jisuanji Xuebao/Chinese Journal of Computers, 2024, 47 (03): : 690 - 702
  • [44] Black-Box Data Poisoning Attacks on Crowdsourcing
    Chen, Pengpeng
    Yang, Yongqiang
    Yang, Dingqi
    Sun, Hailong
    Chen, Zhijun
    Lin, Peng
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 2975 - 2983
  • [45] Procedural Noise Adversarial Examples for Black-Box Attacks on Deep Convolutional Networks
    Co, Kenneth T.
    Munoz-Gonzalez, Luis
    de Maupeou, Sixte
    Lupu, Emil C.
    PROCEEDINGS OF THE 2019 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY (CCS'19), 2019, : 275 - 289
  • [46] Black-box Evolutionary Search for Adversarial Examples against Deep Image Classifiers in Non-Targeted Attacks
    Prochazka, Stepan
    Neruda, Roman
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [47] Toward Visual Distortion in Black-Box Attacks
    Li, Nannan
    Chen, Zhenzhong
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 6156 - 6167
  • [48] Provably Efficient Black-Box Action Poisoning Attacks Against Reinforcement Learning
    Liu, Guanlin
    Lai, Lifeng
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [49] Cheating your apps: Black-box adversarial attacks on deep learning apps
    Cao, Hongchen
    Li, Shuai
    Zhou, Yuming
    Fan, Ming
    Zhao, Xuejiao
    Tang, Yutian
    JOURNAL OF SOFTWARE-EVOLUTION AND PROCESS, 2024, 36 (04)
  • [50] Resiliency of SNN on Black-Box Adversarial Attacks
    Paudel, Bijay Raj
    Itani, Aashish
    Tragoudas, Spyros
    20TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2021), 2021, : 799 - 806