Distributed learning: a reliable privacy-preserving strategy to change multicenter collaborations using AI

被引:24
|
作者
Kirienko, Margarita [1 ,2 ]
Sollini, Martina [2 ,3 ]
Ninatti, Gaia [2 ]
Loiacono, Daniele [4 ]
Giacomello, Edoardo [4 ]
Gozzi, Noemi [3 ]
Amigoni, Francesco [4 ]
Mainardi, Luca [4 ]
Lanzi, Pier Luca [4 ]
Chiti, Arturo [2 ,3 ]
机构
[1] Fdn IRCCS Ist Nazl Tumori, Milan, Italy
[2] Humanitas Univ, Dept Biomed Sci, Milan, Italy
[3] IRCCS Humanitas Res Hosp, Milan, Italy
[4] Politecn Milan, DEIB, Milan, Italy
关键词
Machine learning; Clinical trial; Privacy; Ethics; Distributed learning; Federated learning;
D O I
10.1007/s00259-021-05339-7
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
Purpose The present scoping review aims to assess the non-inferiority of distributed learning over centrally and locally trained machine learning (ML) models in medical applications. Methods We performed a literature search using the term "distributed learning" OR "federated learning" in the PubMed/MEDLINE and EMBASE databases. No start date limit was used, and the search was extended until July 21, 2020. We excluded articles outside the field of interest; guidelines or expert opinion, review articles and meta-analyses, editorials, letters or commentaries, and conference abstracts; articles not in the English language; and studies not using medical data. Selected studies were classified and analysed according to their aim(s). Results We included 26 papers aimed at predicting one or more outcomes: namely risk, diagnosis, prognosis, and treatment side effect/adverse drug reaction. Distributed learning was compared to centralized or localized training in 21/26 and 14/26 selected papers, respectively. Regardless of the aim, the type of input, the method, and the classifier, distributed learning performed close to centralized training, but two experiments focused on diagnosis. In all but 2 cases, distributed learning outperformed locally trained models. Conclusion Distributed learning resulted in a reliable strategy for model development; indeed, it performed equally to models trained on centralized datasets. Sensitive data can get preserved since they are not shared for model development. Distributed learning constitutes a promising solution for ML-based research and practice since large, diverse datasets are crucial for success.
引用
收藏
页码:3791 / 3804
页数:14
相关论文
共 50 条
  • [21] Privacy-Preserving Lane Change Prediction using Deep Learning Models
    Qasemabadi, Armin Nejadhossein
    Mozaffari, Saeed
    Ahmadi, Majid
    Alirezaee, Shahpour
    2024 5TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE, ROBOTICS AND CONTROL, AIRC 2024, 2024, : 46 - 49
  • [22] Privacy-preserving distributed clustering
    Erkin, Zekeriya
    Veugen, Thijs
    Toft, Tomas
    Lagendijk, Reginald L.
    EURASIP JOURNAL ON INFORMATION SECURITY, 2013, (01):
  • [23] A Privacy-Preserving Distributed Control Strategy in Islanded AC Microgrids
    Wang, Ziqiang
    Ma, Meiling
    Zhou, Quan
    Xiong, Linyun
    Wang, Lingling
    Wang, Jinming
    Wang, Jie
    IEEE TRANSACTIONS ON SMART GRID, 2022, 13 (05) : 3369 - 3382
  • [24] Lightweight Crypto-Assisted Distributed Differential Privacy for Privacy-Preserving Distributed Learning
    Lyu, Lingjuan
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [25] Privacy-Preserving Distributed IDS Using Incremental Learning for IoT Health Systems
    Tabassum, Aliya
    Erbad, Aiman
    Mohamed, Amr
    Guizani, Mohsen
    IEEE ACCESS, 2021, 9 : 14271 - 14283
  • [26] Privacy-preserving explainable AI: a survey
    Nguyen, Thanh Tam
    Huynh, Thanh Trung
    Ren, Zhao
    Nguyen, Thanh Toan
    Nguyen, Phi Le
    Yin, Hongzhi
    Nguyen, Quoc Viet Hung
    SCIENCE CHINA-INFORMATION SCIENCES, 2025, 68 (01)
  • [27] Privacy-Preserving AI for Future Networks
    Perino, Diego
    Katevas, Kleomenis
    Lutu, Andra
    Marin, Eduard
    Kourtellis, Nicolas
    COMMUNICATIONS OF THE ACM, 2022, 65 (04) : 52 - 53
  • [28] Privacy-preserving explainable AI: a survey
    Thanh Tam NGUYEN
    Thanh Trung HUYNH
    Zhao REN
    Thanh Toan NGUYEN
    Phi Le NGUYEN
    Hongzhi YIN
    Quoc Viet Hung NGUYEN
    Science China(Information Sciences), 2025, 68 (01) : 23 - 56
  • [29] DATA-WEIGHTED ENSEMBLE LEARNING FOR PRIVACY-PRESERVING DISTRIBUTED LEARNING
    Xie, Liyang
    Plis, Sergey
    Sarwate, Anand D.
    2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 2309 - 2313
  • [30] Privacy-preserving distributed clustering using generative models
    Merugu, S
    Ghosh, J
    THIRD IEEE INTERNATIONAL CONFERENCE ON DATA MINING, PROCEEDINGS, 2003, : 211 - 218