4.4 Review

Lack of Transparency and Potential Bias in Artificial Intelligence Data Sets and Algorithms A Scoping Review

期刊

JAMA DERMATOLOGY
卷 157, 期 11, 页码 1362-1369

出版社

AMER MEDICAL ASSOC
DOI: 10.1001/jamadermatol.2021.3129

关键词

-

资金

  1. National Institutes of Health (NIH) [T32 5T32AR007422-38]
  2. Melanoma Research Alliance
  3. Cancer Center Support grant from the NIH/National Cancer Institute [P30 CA008748]
  4. Charina Fund
  5. National Science Foundation [1942926]
  6. NIH [U01MH098953]
  7. Silicon Valley Foundation
  8. Chan-Zuckerberg Initiative
  9. Div Of Information & Intelligent Systems
  10. Direct For Computer & Info Scie & Enginr [1942926] Funding Source: National Science Foundation

向作者/读者索取更多资源

This scoping review identified issues in data sets used for developing and testing clinical AI algorithms for skin disease, including lack of characterization and transparency, nonstandard disease labels, and inability to fully assess patient diversity.
IMPORTANCE Clinical artificial intelligence (AI) algorithms have the potential to improve clinical care, but fair, generalizable algorithms depend on the clinical data on which they are trained and tested. OBJECTIVE To assess whether data sets used for training diagnostic AI algorithms addressing skin disease are adequately described and to identify potential sources of bias in these data sets. DATA SOURCES In this scoping review, PubMed was used to search for peer-reviewed research articles published between January 1, 2015, and November 1, 2020, with the following paired search terms: deep learning and dermatology, artificial intelligence and dermatology, deep learning and dermatologist, and artificial intelligence and dermatologist. STUDY SELECTION Studies that developed or tested an existing deep learning algorithm for triage, diagnosis, or monitoring using clinical or dermoscopic images of skin disease were selected, and the articles were independently reviewed by 2 investigators to verify that they met selection criteria. CONSENSUS PROCESS Data set audit criteria were determined by consensus of all authors after reviewing existing literature to highlight data set transparency and sources of bias. RESULTS A total of 70 unique studies were included. Among these studies, 1 065 291 images were used to develop or test AI algorithms, of which only 257 372 (24.2%) were publicly available. Only 14 studies (20.0%) included descriptions of patient ethnicity or race in at least 1 data set used. Only 7 studies (10.0%) included any information about skin tone in at least 1 data set used. Thirty-six of the 56 studies developing new AI algorithms for cutaneous malignant neoplasms (64.3%) met the gold standard criteria for disease labeling. Public data sets were cited more often than private data sets, suggesting that public data sets contribute more to new development and benchmarks. CONCLUSIONS AND RELEVANCE This scoping review identified 3 issues in data sets that are used to develop and test clinical AI algorithms for skin disease that should be addressed before clinical translation: (1) sparsity of data set characterization and lack of transparency, (2) nonstandard and unverified disease labels, and (3) inability to fully assess patient diversity used for algorithm development and testing.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.4
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据