4.4 Article

Answering head and neck cancer questions: An assessment of ChatGPT responses

Journal

AMERICAN JOURNAL OF OTOLARYNGOLOGY
Volume 45, Issue 1, Pages -

Publisher

W B SAUNDERS CO-ELSEVIER INC
DOI: 10.1016/j.amjoto.2023.104085

Keywords

Head and neck cancer; Common questions; Chatgpt; Artificial intelligence; Patient education

Ask authors/readers for more resources

This study compared the performance of ChatGPT and Google websites in answering common head and neck cancer questions. The results showed that Google sources had significantly higher quality than ChatGPT responses, although both were similarly difficult to read. ChatGPT should not be considered reliable in its current form, and Google sources are preferred for patient educational materials.
Purpose: To examine and compare ChatGPT versus Google websites in answering common head and neck cancer questions.Materials and methods: Commonly asked questions about head and neck cancer were obtained and inputted into both ChatGPT-4 and Google search engine. For each question, the ChatGPT response and first website search result were compiled and examined. Content quality was assessed by independent reviewers using standardized grading criteria and the modified Ensuring Quality Information for Patients (EQIP) tool. Readability was determined using the Flesch reading ease scale.Results: In total, 49 questions related to head and neck cancer were included. Google sources were on average significantly higher quality than ChatGPT responses (4.2 vs 3.6, p = 0.005). According to the EQIP tool, Google and ChatGPT had on average similar response rates per criterion (24.4 vs 20.5, p = 0.09) while Google had a significantly higher average score per question than ChatGPT (13.8 vs 11.7, p < 0.001) According to the Flesch reading ease scale, ChatGPT and Google sources were both considered similarly difficult to read (33.1 vs 37.0, p = 0.180) and at a college level (14.3 vs 14.2, p = 0.820.)Conclusion: ChatGPT responses were as challenging to read as Google sources, but poorer quality due to decreased reliability and accuracy in answering questions. Though promising, ChatGPT in its current form should not be considered dependable. Google sources are a preferred resource for patient educational materials.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.4
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available