Efficient Label Contamination Attacks Against Black-Box Learning Models

被引:0
|
作者
Zhao, Mengchen [1 ]
An, Bo [1 ]
Gao, Wei [2 ]
Zhang, Teng [2 ]
机构
[1] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore, Singapore
[2] Nanjing Univ, Natl Key Lab Novel Software Technol, Nanjing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Label contamination attack (LCA) is an important type of data poisoning attack where an attacker manipulates the labels of training data to make the learned model beneficial to him. Existing work on LCA assumes that the attacker has full knowledge of the victim learning model, whereas the victim model is usually a black-box to the attacker. In this paper, we develop a Projected Gradient Ascent (PGA) algorithm to compute LCAs on a family of empirical risk minimizations and show that an attack on one victim model can also be effective on other victim models. This makes it possible that the attacker designs an attack against a substitute model and transfers it to a black-box victim model. Based on the observation of the transferability, we develop a defense algorithm to identify the data points that are most likely to be attacked. Empirical studies show that PGA significantly outperforms existing baselines and linear learning models are better substitute models than nonlinear ones.
引用
收藏
页码:3945 / 3951
页数:7
相关论文
共 50 条
  • [1] Query-efficient label-only attacks against black-box machine learning models
    Ren, Yizhi
    Zhou, Qi
    Wang, Zhen
    Wu, Ting
    Wu, Guohua
    Choo, Kim-Kwang Raymond
    COMPUTERS & SECURITY, 2020, 90
  • [2] Fuzzing-based hard-label black-box attacks against machine learning models
    Qin, Yi
    Yue, Chuan
    COMPUTERS & SECURITY, 2022, 117
  • [3] Provably Efficient Black-Box Action Poisoning Attacks Against Reinforcement Learning
    Liu, Guanlin
    Lai, Lifeng
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [4] Practical Black-Box Attacks against Machine Learning
    Papernot, Nicolas
    McDaniel, Patrick
    Goodfellow, Ian
    Jha, Somesh
    Celik, Z. Berkay
    Swami, Ananthram
    PROCEEDINGS OF THE 2017 ACM ASIA CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY (ASIA CCS'17), 2017, : 506 - 519
  • [5] Black-Box Adversarial Attacks against Audio Forensics Models
    Jiang, Yi
    Ye, Dengpan
    SECURITY AND COMMUNICATION NETWORKS, 2022, 2022
  • [6] Orthogonal Deep Models as Defense Against Black-Box Attacks
    Jalwana, Mohammad A. A. K.
    Akhtar, Naveed
    Bennamoun, Mohammed
    Mian, Ajmal
    IEEE ACCESS, 2020, 8 : 119744 - 119757
  • [7] Automatic Selection Attacks Framework for Hard Label Black-Box Models
    Liu, Xiaolei
    Li, Xiaoyu
    Zheng, Desheng
    Bai, Jiayu
    Peng, Yu
    Zhang, Shibin
    IEEE INFOCOM 2022 - IEEE CONFERENCE ON COMPUTER COMMUNICATIONS WORKSHOPS (INFOCOM WKSHPS), 2022,
  • [8] B3: Backdoor Attacks against Black-box Machine Learning Models
    Gong, Xueluan
    Chen, Yanjiao
    Yang, Wenbin
    Huang, Huayang
    Wang, Qian
    ACM TRANSACTIONS ON PRIVACY AND SECURITY, 2023, 26 (04)
  • [9] Stateful Defenses for Machine Learning Models Are Not Yet Secure Against Black-box Attacks
    Feng, Ryan
    Hooda, Ashish
    Mangaokar, Neal
    Fawaz, Kassem
    Jha, Somesh
    Prakash, Atul
    PROCEEDINGS OF THE 2023 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, CCS 2023, 2023, : 786 - 800
  • [10] Black-box adversarial attacks against image quality assessment models
    Ran, Yu
    Zhang, Ao-Xiang
    Li, Mingjie
    Tang, Weixuan
    Wang, Yuan-Gen
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 260