4.8 Article

AI for radiographic COVID-19 detection selects shortcuts over signal

期刊

NATURE MACHINE INTELLIGENCE
卷 3, 期 7, 页码 610-619

出版社

NATURE PORTFOLIO
DOI: 10.1038/s42256-021-00338-7

关键词

-

资金

  1. National Science Foundation [CAREER DBI-1552309]
  2. National Institutes of Health [R35 GM 128638, R01 AG061132]

向作者/读者索取更多资源

Recent deep learning systems to detect COVID-19 from chest radiographs may rely on confounding factors rather than medical pathology, leading to accuracy issues when tested in new hospitals. The approach to obtain training data for these AI systems introduces a nearly ideal scenario for learning spurious shortcuts, raising concerns in medical-imaging AI. Evaluation of models on external data is insufficient to ensure reliance on medically relevant pathology, highlighting the importance of explainable AI for clinical deployment of machine-learning healthcare models.
Artificial intelligence (AI) researchers and radiologists have recently reported AI systems that accurately detect COVID-19 in chest radiographs. However, the robustness of these systems remains unclear. Using state-of-the-art techniques in explainable AI, we demonstrate that recent deep learning systems to detect COVID-19 from chest radiographs rely on confounding factors rather than medical pathology, creating an alarming situation in which the systems appear accurate, but fail when tested in new hospitals. We observe that the approach to obtain training data for these AI systems introduces a nearly ideal scenario for AI to learn these spurious 'shortcuts'. Because this approach to data collection has also been used to obtain training data for the detection of COVID-19 in computed tomography scans and for medical imaging tasks related to other diseases, our study reveals a far-reaching problem in medical-imaging AI. In addition, we show that evaluation of a model on external data is insufficient to ensure AI systems rely on medically relevant pathology, because the undesired 'shortcuts' learned by AI systems may not impair performance in new hospitals. These findings demonstrate that explainable AI should be seen as a prerequisite to clinical deployment of machine-learning healthcare models. The urgency of the developing COVID-19 epidemic has led to a large number of novel diagnostic approaches, many of which use machine learning. DeGrave and colleagues use explainable AI techniques to analyse a selection of these approaches and find that the methods frequently learn to identify features unrelated to the actual disease.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据