PRIV-ML: Analyzing Privacy Loss in Iterative Machine Learning With Differential Privacy

被引:0
|
作者
Thantharate, Pratik [1 ]
Todurkar, Divya Ananth [1 ]
Anurag, T. [1 ]
机构
[1] Univ Missouri, Kansas City, MO 65211 USA
关键词
Differential Privacy; Privacy Loss Quantification; Iterative Composition; Machine Learning; privacy budget; Formal Verification; AI Fairness; Trustworthy AI; Privacy Risks; Data Sharing; Cloud Computing;
D O I
10.1109/Cloud-Summit61220.2024.00024
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Differential privacy offers rigorous protections for emerging paradigms like federated machine learning, decentralized analytics, and web3 applications. The parameters. (epsilon) and d (delta) are crucial in balancing privacy and utility by bounding the maximum divergence between outputs on neighboring datasets. However, quantifying cumulative privacy loss over long-running processes involving iterative model queries, validations, tuning, and multi-party computations remains an open challenge restricting adoption. This paper proposes a comprehensive methodology to evaluate end-to-end differential privacy guarantees across complex Machine Learning (ML) workflows. We develop a heuristic algorithm that maintains a privacy budget depleted per operation based on computed data sensitivity and noise calibration. By tracking tight stochastic bounds on the cumulative privacy loss random variable using advanced composition theorems, our approach can formally verify guarantees over iterative workflows. Simulations demonstrate the technique quantifying privacy loss across 950 successive histogram queries under (. = 1, d = 10-5)-differential privacy while sustaining utility with an average error of only 4.5% compared to non-private histograms, underscoring the importance of formally tracking cumulative privacy loss. Our framework provides a practical solution for measurable privacy-preserving machine learning pipelines without degrading accuracy or utility. By interfacing with diverse mechanisms and adapting noise to empirical sensitivities, we facilitate precise reasoning of privacy risks throughout model life cycles. We also analyze privacy parameter implications across application domains. This paper lays a rigorous foundation for developing trustworthy AI systems that protect sensitive data.
引用
收藏
页码:107 / 112
页数:6
相关论文
共 50 条
  • [1] Preserving User Privacy for Machine Learning: Local Differential Privacy or Federated Machine Learning?
    Zheng, Huadi
    Hu, Haibo
    Han, Ziyang
    IEEE INTELLIGENT SYSTEMS, 2020, 35 (04) : 5 - 14
  • [2] How Differential Privacy Reinforces Privacy of Machine Learning Models?
    Ben Hamida, Sana
    Mrabet, Hichem
    Jemai, Abderrazak
    ADVANCES IN COMPUTATIONAL COLLECTIVE INTELLIGENCE, ICCCI 2022, 2022, 1653 : 661 - 673
  • [3] Quantum machine learning with differential privacy
    William M. Watkins
    Samuel Yen-Chi Chen
    Shinjae Yoo
    Scientific Reports, 13
  • [4] Quantum machine learning with differential privacy
    Watkins, William M.
    Chen, Samuel Yen-Chi
    Yoo, Shinjae
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [5] How to DP-fy ML: A Practical Guide to Machine Learning with Differential Privacy
    Ponomareva, Natalia
    Hazimeh, Hussein
    Kurakin, Alex
    Xu, Zheng
    Denison, Carson
    McMahan, H. Brendan
    Vassilvitskii, Sergei
    Chien, Steve
    Thakurta, Abhradeep
    JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2023, 77 : 1113 - 1201
  • [6] How to DP-fy ML: A Practical Tutorial to Machine Learning with Differential Privacy
    Ponomareva, Natalia
    Vassilvitskii, Sergei
    Xu, Zheng
    McMahan, Brendan
    Kurakin, Alexey
    Zhang, Chiyuan
    PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 5823 - 5824
  • [7] How to DP-fy ML: A Practical Guide to Machine Learning with Differential Privacy
    Ponomareva N.
    Hazimeh H.
    Kurakin A.
    Xu Z.
    Denison C.
    McMahan H.B.
    Vassilvitskii S.
    Chien S.
    Thakurta A.
    Journal of Artificial Intelligence Research, 2023, 77 : 1113 - 1201
  • [8] Privacy-preserving quantum machine learning using differential privacy
    Senekane, Makhamisa
    Mafu, Mhlambululi
    Taele, Benedict Molibeli
    2017 IEEE AFRICON, 2017, : 1432 - 1435
  • [9] Privacy Risk in Machine Learning: Analyzing the Connection to Overfitting
    Yeom, Samuel
    Giacomelli, Irene
    Fredrikson, Matt
    Jha, Somesh
    IEEE 31ST COMPUTER SECURITY FOUNDATIONS SYMPOSIUM (CSF 2018), 2018, : 268 - 282
  • [10] Enhancing correlated big data privacy using differential privacy and machine learning
    Biswas, Sreemoyee
    Fole, Anuja
    Khare, Nilay
    Agrawal, Pragati
    JOURNAL OF BIG DATA, 2023, 10 (01)