4.6 Article

Multi-Radiologist User Study for Artificial Intelligence-Guided Grading of COVID-19 Lung Disease Severity on Chest Radiographs

期刊

ACADEMIC RADIOLOGY
卷 28, 期 4, 页码 572-576

出版社

ELSEVIER SCIENCE INC
DOI: 10.1016/j.acra.2021.01.016

关键词

COVID-19; Chest radiograph; Artificial intelligence; Computer-assisted diagnosis

资金

  1. National Institute of Biomedical Imaging and Bioengineering (NIBIB), National Institutes of Health [P41EB015896]

向作者/读者索取更多资源

This study aimed to investigate whether an artificial intelligence system could improve the consistency of radiologist reporting of COVID-19 pneumonia on chest radiographs. The results showed that the use of an AI system led to improved interrater agreement among radiologists, with disease severity categories significantly associated with clinical outcomes.
Rationale and Objectives: Radiographic findings of COVID-19 pneumonia can be used for patient risk stratification; however, radiologist reporting of disease severity is inconsistent on chest radiographs (CXRs). We aimed to see if an artificial intelligence (AI) system could help improve radiologist interrater agreement. Materials and Methods: We performed a retrospective multi-radiologist user study to evaluate the impact of an AI system, the PXS score model, on the grading of categorical COVID-19 lung disease severity on 154 chest radiographs into four ordinal grades (normal/minimal, mild, moderate, and severe). Four radiologists (two thoracic and two emergency radiologists) independently interpreted 154 CXRs from 154 unique patients with COVID-19 hospitalized at a large academic center, before and after using the AI system (median washout time interval was 16 days). Three different thoracic radiologists assessed the same 154 CXRs using an updated version of the AI system trained on more imaging data. Radiologist interrater agreement was evaluated using Cohen and Fleiss kappa where appropriate. The lung disease severity categories were associated with clinical outcomes using a previously published outcomes dataset using Fisher's exact test and Chi-square test for trend. Results: Use of the AI system improved radiologist interrater agreement (Fleiss K = 0.40 to 0.66, before and after use of the system). The Fleiss K for three radiologists using the updated AI system was 0.74. Severity categories were significantly associated with subsequent intubation or death within 3 days. Conclusion: An AI system used at the time of CXR study interpretation can improve the interrater agreement of radiologists.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据