Projan: A probabilistic trojan attack on deep neural networks

被引:0
|
作者
Saremi, Mehrin [1 ]
Khalooei, Mohammad [2 ]
Rastgoo, Razieh [3 ]
Sabokrou, Mohammad [4 ,5 ]
机构
[1] Semnan Univ, Farzanegan Campus, Semnan 3513119111, Iran
[2] Amirkabir Univ Technol, Dept Comp Engn, Tehran, Iran
[3] Semnan Univ, Fac Elect & Comp Engn, Semnan 3513119111, Iran
[4] Inst Res Fundamental Sci, Tehran, Iran
[5] Okinawa Inst Sci & Technol, Onna, Okinawa, Japan
关键词
Deep learning; AI security; Trojan attack; Backdoor attack; Probabilistic trojan attack;
D O I
10.1016/j.knosys.2024.112565
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks have gained popularity due to their outstanding performance across various domains. However, because of their lack of explainability, they are vulnerable to some kinds of threats including the trojan or backdoor attack, in which an adversary can train the model to respond to a crafted peculiar input pattern (also called trigger) according to their will.<br /> Several trojan attack and defense methods have been proposed in the literature. Many of the defense methods are based on the assumption that the possibly existing trigger must be able to affect the model's behavior, making it output a certain class label for all inputs. In this work, we propose an alternative attack method that violates this assumption. Instead of a single trigger that works on all inputs, a few triggers are generated that will affect only some of the inputs. At attack time, the adversary will need to try more than one trigger to succeed, which might be possible in some real-world situations.<br /> Our experiments on MNIST and CIFAR-10 datasets show that such an attack can be implemented successfully, reaching an attack success rate similar to baseline methods called BadNet and N-to-One. We also tested wide range of defense methods and verified that in general, this kind of backdoor is more difficult for defense algorithms to detect. The code is available at https://github.com/programehr/Projan.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] Backdoor Attack on Deep Neural Networks in Perception Domain
    Mo, Xiaoxing
    Zhang, Leo Yu
    Sun, Nan
    Luo, Wei
    Gao, Shang
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [22] Adaptive Backdoor Attack against Deep Neural Networks
    He, Honglu
    Zhu, Zhiying
    Zhang, Xinpeng
    CMES-COMPUTER MODELING IN ENGINEERING & SCIENCES, 2023, 136 (03): : 2617 - 2633
  • [23] One Pixel Attack for Fooling Deep Neural Networks
    Su, Jiawei
    Vargas, Danilo Vasconcellos
    Sakurai, Kouichi
    IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2019, 23 (05) : 828 - 841
  • [24] Cocktail Universal Adversarial Attack on Deep Neural Networks
    Li, Shaoxin
    Li, Xiaofeng
    Che, Xin
    Li, Xintong
    Zhang, Yong
    Chu, Lingyang
    COMPUTER VISION - ECCV 2024, PT LXV, 2025, 15123 : 396 - 412
  • [25] POSTER: Practical Fault Attack on Deep Neural Networks
    Breier, Jakub
    Hou, Xiaolu
    Jap, Dirmanto
    Ma, Lei
    Bhasin, Shivam
    Liu, Yang
    PROCEEDINGS OF THE 2018 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY (CCS'18), 2018, : 2204 - 2206
  • [26] Patch Based Backdoor Attack on Deep Neural Networks
    Manna, Debasmita
    Tripathy, Somanath
    INFORMATION SYSTEMS SECURITY, ICISS 2024, 2025, 15416 : 422 - 440
  • [27] A probabilistic framework for mutation testing in deep neural networks
    Tambon, Florian
    Khomh, Foutse
    Antoniol, Giuliano
    INFORMATION AND SOFTWARE TECHNOLOGY, 2023, 155
  • [28] Probabilistic Forecasting of Symbol Sequences with Deep Neural Networks
    Hauser, Michael
    Fu, Yiwei
    Li, Yue
    Phoha, Shashi
    Ray, Asok
    2017 AMERICAN CONTROL CONFERENCE (ACC), 2017, : 3147 - 3152
  • [29] ADMM Attack: An Enhanced Adversarial Attack for Deep Neural Networks with Undetectable Distortions
    Zhao, Pu
    Xu, Kaidi
    Liu, Sijia
    Wang, Yanzhi
    Lin, Xue
    24TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC 2019), 2019, : 499 - 505
  • [30] Trojan Attack on Deep Generative Models in Autonomous Driving
    Ding, Shaohua
    Tian, Yulong
    Xu, Fengyuan
    Li, Qun
    Zhong, Sheng
    SECURITY AND PRIVACY IN COMMUNICATION NETWORKS, SECURECOMM, PT I, 2019, 304 : 299 - 318