Privacy-Preserving Techniques in Generative AI and Large Language Models: A Narrative Review

被引:3
|
作者
Feretzakis, Georgios [1 ]
Papaspyridis, Konstantinos [2 ]
Gkoulalas-Divanis, Aris [3 ]
Verykios, Vassilios S. [1 ]
机构
[1] Hellen Open Univ, Sch Sci & Technol, Patras 26335, Greece
[2] Univ Toronto, Comp Sci, Toronto, ON M5S 2E4, Canada
[3] Merat Healthcare, Dublin D02 NY19, Ireland
关键词
privacy-preserving techniques; generative AI; large language models (LLMs); differential privacy; federated learning; homomorphic encryption; secure multi-party computation; model inversion; membership inference; privacy-enhancing technologies; post-quantum cryptography; CHALLENGES;
D O I
10.3390/info15110697
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Generative AI, including large language models (LLMs), has transformed the paradigm of data generation and creative content, but this progress raises critical privacy concerns, especially when models are trained on sensitive data. This review provides a comprehensive overview of privacy-preserving techniques aimed at safeguarding data privacy in generative AI, such as differential privacy (DP), federated learning (FL), homomorphic encryption (HE), and secure multi-party computation (SMPC). These techniques mitigate risks like model inversion, data leakage, and membership inference attacks, which are particularly relevant to LLMs. Additionally, the review explores emerging solutions, including privacy-enhancing technologies and post-quantum cryptography, as future directions for enhancing privacy in generative AI systems. Recognizing that achieving absolute privacy is mathematically impossible, the review emphasizes the necessity of aligning technical safeguards with legal and regulatory frameworks to ensure compliance with data protection laws. By discussing the ethical and legal implications of privacy risks in generative AI, the review underscores the need for a balanced approach that considers performance, scalability, and privacy preservation. The findings highlight the need for ongoing research and innovation to develop privacy-preserving techniques that keep pace with the scaling of generative AI, especially in large language models, while adhering to regulatory and ethical standards.
引用
收藏
页数:25
相关论文
共 50 条
  • [1] GENERATIVE AI FOR SECURE AND PRIVACY-PRESERVING MOBILE CROWDSENSING
    Yang, Yaoqi
    Zhang, Bangning
    Guo, Daoxing
    Du, Hongyang
    Xiong, Zehui
    Niyato, Dusit
    Han, Zhu
    IEEE WIRELESS COMMUNICATIONS, 2024, 31 (06) : 29 - 38
  • [2] Feasibility and Prospect of Privacy-preserving Large Language Models in Radiology
    Cai, Wenli
    RADIOLOGY, 2023, 309 (01)
  • [3] Local large language models for privacy-preserving accelerated review of historic echocardiogram reports
    Vaid, Akhil
    Duong, Son Q.
    Lampert, Joshua
    Kovatch, Patricia
    Freeman, Robert
    Argulian, Edgar
    Croft, Lori
    Lerakis, Stamatios
    Goldman, Martin
    Khera, Rohan
    Nadkarni, Girish N.
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024,
  • [4] Privacy-preserving large language models for structured medical information retrieval
    Wiest, Isabella Catharina
    Ferber, Dyke
    Zhu, Jiefu
    van Treeck, Marko
    Meyer, Sonja K.
    Juglan, Radhika
    Carrero, Zunamys I.
    Paech, Daniel
    Kleesiek, Jens
    Ebert, Matthias P.
    Truhn, Daniel
    Kather, Jakob Nikolas
    NPJ DIGITAL MEDICINE, 2024, 7 (01):
  • [5] Privacy-preserving distributed clustering using generative models
    Merugu, S
    Ghosh, J
    THIRD IEEE INTERNATIONAL CONFERENCE ON DATA MINING, PROCEEDINGS, 2003, : 211 - 218
  • [6] Selective privacy-preserving framework for large language models fine-tuning
    Wang, Teng
    Zhai, Lindong
    Yang, Tengfei
    Luo, Zhucheng
    Liu, Shuanggen
    INFORMATION SCIENCES, 2024, 678
  • [7] InferDPT: Privacy-preserving Inference for Black-box Large Language Models
    Tong, Meng
    Chen, Kejiang
    Zhang, Jie
    Qi, Yuang
    Zhang, Weiming
    Yu, Nenghai
    Zhang, Tianwei
    Zhang, Zhikun
    arXiv, 2023,
  • [8] A review of privacy-preserving techniques for deep learning
    Boulemtafes, Amine
    Derhab, Abdelouahid
    Challal, Yacine
    NEUROCOMPUTING, 2020, 384 : 21 - 45
  • [9] Foundation Models, Generative AI, and Large Language Models
    Ross, Angela
    McGrow, Kathleen
    Zhi, Degui
    Rasmy, Laila
    CIN-COMPUTERS INFORMATICS NURSING, 2024, 42 (05) : 377 - 387
  • [10] Detection of suicidality from medical text using privacy-preserving large language models
    Wiest, Isabella Catharina
    Verhees, Falk Gerrik
    Ferber, Dyke
    Zhu, Jiefu
    Bauer, Michael
    Lewitzka, Ute
    Pfennig, Andrea
    Mikolas, Pavol
    Kather, Jakob Nikolas
    BRITISH JOURNAL OF PSYCHIATRY, 2024, 225 (06) : 532 - 537