4.6 Article

Benchmarking the symptom-checking capabilities of ChatGPT for a broad range of diseases

出版社

OXFORD UNIV PRESS
DOI: 10.1093/jamia/ocad245

关键词

symptom checking; ChatGPT; benchmarking; learning health system; medical training

向作者/读者索取更多资源

This study evaluates the symptom-checking accuracy of ChatGPT for a broad range of diseases using the Mayo Clinic Symptom Checker patient service as a benchmark. The results show that ChatGPT exhibits high accuracy, surpassing the previous GPT-3.5-turbo model, and demonstrating its potential as a medical training tool in learning health systems to enhance care quality and address health disparities.
Objective This study evaluates ChatGPT's symptom-checking accuracy across a broad range of diseases using the Mayo Clinic Symptom Checker patient service as a benchmark.Methods We prompted ChatGPT with symptoms of 194 distinct diseases. By comparing its predictions with expectations, we calculated a relative comparative score (RCS) to gauge accuracy.Results ChatGPT's GPT-4 model achieved an average RCS of 78.8%, outperforming the GPT-3.5-turbo by 10.5%. Some specialties scored above 90%.Discussion The test set, although extensive, was not exhaustive. Future studies should include a more comprehensive disease spectrum.Conclusion ChatGPT exhibits high accuracy in symptom checking for a broad range of diseases, showcasing its potential as a medical training tool in learning health systems to enhance care quality and address health disparities.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据