4.7 Review

Distributed learning: a reliable privacy-preserving strategy to change multicenter collaborations using AI

期刊

出版社

SPRINGER
DOI: 10.1007/s00259-021-05339-7

关键词

Machine learning; Clinical trial; Privacy; Ethics; Distributed learning; Federated learning

资金

  1. AIRC (Italian Association for Cancer Research) [2019-23959]

向作者/读者索取更多资源

This scoping review aimed to assess the non-inferiority of distributed learning compared to centrally and locally trained machine learning models in medical applications. The study found that distributed learning performed close to centralized training and outperformed locally trained models in most cases, indicating its potential significance for ML-based research and practice.
Purpose The present scoping review aims to assess the non-inferiority of distributed learning over centrally and locally trained machine learning (ML) models in medical applications. Methods We performed a literature search using the term distributed learning OR federated learning in the PubMed/MEDLINE and EMBASE databases. No start date limit was used, and the search was extended until July 21, 2020. We excluded articles outside the field of interest; guidelines or expert opinion, review articles and meta-analyses, editorials, letters or commentaries, and conference abstracts; articles not in the English language; and studies not using medical data. Selected studies were classified and analysed according to their aim(s). Results We included 26 papers aimed at predicting one or more outcomes: namely risk, diagnosis, prognosis, and treatment side effect/adverse drug reaction. Distributed learning was compared to centralized or localized training in 21/26 and 14/26 selected papers, respectively. Regardless of the aim, the type of input, the method, and the classifier, distributed learning performed close to centralized training, but two experiments focused on diagnosis. In all but 2 cases, distributed learning outperformed locally trained models. Conclusion Distributed learning resulted in a reliable strategy for model development; indeed, it performed equally to models trained on centralized datasets. Sensitive data can get preserved since they are not shared for model development. Distributed learning constitutes a promising solution for ML-based research and practice since large, diverse datasets are crucial for success.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据