SPECTRE Defending Against Backdoor Attacks Using Robust Statistics

被引:0
|
作者
Hayase, Jonathan [1 ]
Kong, Weihao [1 ]
Somani, Raghav [1 ]
Oh, Sewoong [1 ]
机构
[1] Univ Washington, Paul G Allen Sch Comp Sci & Engn, Seattle, WA 98195 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Modern machine learning increasingly requires training on a large collection of data from multiple sources, not all of which can be trusted. A particularly concerning scenario is when a small fraction of poisoned data changes the behavior of the trained model when triggered by an attacker-specified watermark. Such a compromised model will be deployed unnoticed as the model is accurate otherwise. There have been promising attempts to use the intermediate representations of such a model to separate corrupted examples from clean ones. However, these defenses work only when a certain spectral signature of the poisoned examples is large enough for detection. There is a wide range of attacks that cannot be protected against by the existing defenses. We propose a novel defense algorithm using robust covariance estimation to amplify the spectral signature of corrupted data. This defense provides a clean model, completely removing the backdoor, even in regimes where previous methods have no hope of detecting the poisoned examples.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] PTB: Robust physical backdoor attacks against deep neural networks in real world
    Xue, Mingfu
    He, Can
    Wu, Yinghao
    Sun, Shichang
    Zhang, Yushu
    Wang, Jian
    Liu, Weiqiang
    COMPUTERS & SECURITY, 2022, 118
  • [32] exLCL for defense against spectre attacks
    Wang S.
    Zhao Y.
    Lü Z.
    Qinghua Daxue Xuebao/Journal of Tsinghua University, 2021, 61 (11): : 1221 - 1227
  • [33] PerVK : A Robust Personalized Federated Framework to Defend Against Backdoor Attacks for IoT Applications
    Wang, Yongkang
    Zhai, Di-Hua
    Xia, Yuanqing
    Liu, Danyang
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2024, 20 (03) : 4930 - 4939
  • [34] Robust and Transferable Backdoor Attacks Against Deep Image Compression With Selective Frequency Prior
    Yu, Yi
    Wang, Yufei
    Yang, Wenhan
    Guo, Lanqing
    Lu, Shijian
    Duan, Ling-Yu
    Tan, Yap-Peng
    Kot, Alex C.
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2025, 47 (03) : 1674 - 1693
  • [35] EmoBack: Backdoor Attacks Against Speaker Identification Using Emotional Prosody
    Schoof, Coen
    Koffas, Stefanos
    Conti, Mauro
    Picek, Stjepan
    PROCEEDINGS OF THE 2024 WORKSHOP ON ARTIFICIAL INTELLIGENCE AND SECURITY, AISEC 2024, 2024, : 137 - 148
  • [36] On the Effectiveness of Adversarial Training Against Backdoor Attacks
    Gao, Yinghua
    Wu, Dongxian
    Zhang, Jingfeng
    Gan, Guanhao
    Xia, Shu-Tao
    Niu, Gang
    Sugiyama, Masashi
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (10) : 14878 - 14888
  • [37] Verifying Neural Networks Against Backdoor Attacks
    Pham, Long H.
    Sun, Jun
    COMPUTER AIDED VERIFICATION (CAV 2022), PT I, 2022, 13371 : 171 - 192
  • [38] Backdoor attacks against distributed swarm learning
    Chen, Kongyang
    Zhang, Huaiyuan
    Feng, Xiangyu
    Zhang, Xiaoting
    Mi, Bing
    Jin, Zhiping
    ISA TRANSACTIONS, 2023, 141 : 59 - 72
  • [39] RAB: Provable Robustness Against Backdoor Attacks
    Weber, Maurice
    Xu, Xiaojun
    Karlas, Bojan
    Zhang, Ce
    Li, Bo
    2023 IEEE SYMPOSIUM ON SECURITY AND PRIVACY, SP, 2023, : 1311 - 1328
  • [40] Defending Deep Learning Based Anomaly Detection Systems Against White-Box Adversarial Examples and Backdoor Attacks
    Alrawashdeh, Khaled
    Goldsmith, Stephen
    PROCEEDINGS OF THE 2020 IEEE INTERNATIONAL SYMPOSIUM ON TECHNOLOGY AND SOCIETY (ISTAS), 2021, : 294 - 301