TY - JOUR
T1 - ChatGPT fails challenging the recent ESCMID brain abscess guideline
AU - Dyckhoff-Shen, Susanne
AU - Koedel, Uwe
AU - Brouwer, Matthijs C.
AU - Bodilsen, Jacob
AU - Klein, Matthias
N1 - Publisher Copyright: © The Author(s) 2024.
PY - 2024/4
Y1 - 2024/4
N2 - Background: With artificial intelligence (AI) on the rise, it remains unclear if AI is able to professionally evaluate medical research and give scientifically valid recommendations. Aim: This study aimed to assess the accuracy of ChatGPT’s responses to ten key questions on brain abscess diagnostics and treatment in comparison to the guideline recently published by the European Society for Clinical Microbiology and Infectious Diseases (ESCMID). Methods: All ten PECO (Population, Exposure, Comparator, Outcome) questions which had been developed during the guideline process were presented directly to ChatGPT. Next, ChatGPT was additionally fed with data from studies selected for each PECO question by the ESCMID committee. AI’s responses were subsequently compared with the recommendations of the ESCMID guideline. Results: For 17 out of 20 challenges, ChatGPT was able to give recommendations on the management of patients with brain abscess, including grade of evidence and strength of recommendation. Without data prompting, 70% of questions were answered very similar to the guideline recommendation. In the answers that differed from the guideline recommendations, no patient hazard was present. Data input slightly improved the clarity of ChatGPT’s recommendations, but, however, led to less correct answers including two recommendations that directly contradicted the guideline, being associated with the possibility of a hazard to the patient. Conclusion: ChatGPT seems to be able to rapidly gather information on brain abscesses and give recommendations on key questions about their management in most cases. Nevertheless, single responses could possibly harm the patients. Thus, the expertise of an expert committee remains inevitable.
AB - Background: With artificial intelligence (AI) on the rise, it remains unclear if AI is able to professionally evaluate medical research and give scientifically valid recommendations. Aim: This study aimed to assess the accuracy of ChatGPT’s responses to ten key questions on brain abscess diagnostics and treatment in comparison to the guideline recently published by the European Society for Clinical Microbiology and Infectious Diseases (ESCMID). Methods: All ten PECO (Population, Exposure, Comparator, Outcome) questions which had been developed during the guideline process were presented directly to ChatGPT. Next, ChatGPT was additionally fed with data from studies selected for each PECO question by the ESCMID committee. AI’s responses were subsequently compared with the recommendations of the ESCMID guideline. Results: For 17 out of 20 challenges, ChatGPT was able to give recommendations on the management of patients with brain abscess, including grade of evidence and strength of recommendation. Without data prompting, 70% of questions were answered very similar to the guideline recommendation. In the answers that differed from the guideline recommendations, no patient hazard was present. Data input slightly improved the clarity of ChatGPT’s recommendations, but, however, led to less correct answers including two recommendations that directly contradicted the guideline, being associated with the possibility of a hazard to the patient. Conclusion: ChatGPT seems to be able to rapidly gather information on brain abscesses and give recommendations on key questions about their management in most cases. Nevertheless, single responses could possibly harm the patients. Thus, the expertise of an expert committee remains inevitable.
KW - AI
KW - Brain abscess
KW - ChatGPT
KW - Guideline
UR - http://www.scopus.com/inward/record.url?scp=85183340027&partnerID=8YFLogxK
U2 - 10.1007/s00415-023-12168-1
DO - 10.1007/s00415-023-12168-1
M3 - Article
C2 - 38279999
SN - 0340-5354
VL - 271
SP - 2086
EP - 2101
JO - Journal of neurology
JF - Journal of neurology
IS - 4
ER -