Label-Only Membership Inference Attacks

被引:0
|
作者
Choquette-Choo, Christopher A. [1 ,2 ]
Tramer, Florian [3 ]
Carlini, Nicholas [4 ]
Papernot, Nicolas [1 ,2 ]
机构
[1] Univ Toronto, Toronto, ON, Canada
[2] Vector Inst, Toronto, ON, Canada
[3] Stanford Univ, Stanford, CA 94305 USA
[4] Google, Mountain View, CA 94043 USA
基金
加拿大自然科学与工程研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Membership inference is one of the simplest privacy threats faced by machine learning models that are trained on private sensitive data. In this attack, an adversary infers whether a particular point was used to train the model, or not, by observing the model's predictions. Whereas current attack methods all require access to the model's predicted confidence score, we introduce a labelonly attack that instead evaluates the robustness of the model's predicted (hard) labels under perturbations of the input, to infer membership. Our label-only attack is not only as-effective as attacks requiring access to confidence scores, it also demonstrates that a class of defenses against membership inference, which we call "confidence masking" because they obfuscate the confidence scores to thwart attacks, are insufficient to prevent the leakage of private information. Our experiments show that training with differential privacy or strong l(2) regularization are the only current defenses that meaningfully decrease leakage of private information, even for points that are outliers of the training distribution.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Membership inference attacks against compression models
    Jin, Yong
    Lou, Weidong
    Gao, Yanghua
    COMPUTING, 2023, 105 (11) : 2419 - 2442
  • [32] Membership inference attacks against compression models
    Yong Jin
    Weidong Lou
    Yanghua Gao
    Computing, 2023, 105 : 2419 - 2442
  • [33] Membership Inference Attacks Against Recommender Systems
    Zhang, Minxing
    Ren, Zhaochun
    Wang, Zihan
    Ren, Pengjie
    Chen, Zhumin
    Hu, Pengfei
    Zhang, Yang
    CCS '21: PROCEEDINGS OF THE 2021 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2021, : 864 - 879
  • [34] VAE-Based Membership Cleanser Against Membership Inference Attacks
    Hu, Li
    Yan, Hongyang
    Peng, Yun
    Hu, Haibo
    Wang, Shaowei
    Li, Jin
    IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2025, 22 (02) : 1253 - 1264
  • [35] Unstoppable Attack: Label-Only Model Inversion Via Conditional Diffusion Model
    Liu, Rongke
    Wang, Dong
    Ren, Yizhi
    Wang, Zhen
    Guo, Kaitian
    Qin, Qianqian
    Liu, Xiaolei
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 3958 - 3973
  • [36] Membership Inference Attacks on Deep Regression Models for Neuroimaging
    Gupta, Umang
    Stripelis, Dimitris
    Lam, Pradeep K.
    Thompson, Paul M.
    Ambite, Jose Luis
    Steeg, Greg Ver
    MEDICAL IMAGING WITH DEEP LEARNING, VOL 143, 2021, 143 : 228 - 251
  • [37] Mitigating Membership Inference Attacks in Machine Learning as a Service
    Bouhaddi, Myria
    Adi, Kamel
    2023 IEEE INTERNATIONAL CONFERENCE ON CYBER SECURITY AND RESILIENCE, CSR, 2023, : 262 - 268
  • [38] Deepmarking: Leveraging Adversarial Noise for Membership Inference Attacks
    Jelstrup, Malthe Andreas Lejbolle
    Bigdeli, Siavash Arjomand
    2024 IEEE INTERNATIONAL CONFERENCE ON COMPUTATIONAL PHOTOGRAPHY, ICCP 2024, 2024,
  • [39] Resisting membership inference attacks through knowledge distillation
    Zheng, Junxiang
    Cao, Yongzhi
    Wang, Hanpin
    NEUROCOMPUTING, 2021, 452 : 114 - 126
  • [40] A Survey on Membership Inference Attacks Against Machine Learning
    Bai, Yang
    Chen, Ting
    Fan, Mingyu
    International Journal of Network Security, 2021, 23 (04) : 685 - 697