Bioengineering (Jan 2024)

Exploring Diagnostic Precision and Triage Proficiency: A Comparative Study of GPT-4 and Bard in Addressing Common Ophthalmic Complaints

  • Roya Zandi,
  • Joseph D. Fahey,
  • Michael Drakopoulos,
  • John M. Bryan,
  • Siyuan Dong,
  • Paul J. Bryar,
  • Ann E. Bidwell,
  • R. Chris Bowen,
  • Jeremy A. Lavine,
  • Rukhsana G. Mirza

DOI
https://doi.org/10.3390/bioengineering11020120
Journal volume & issue
Vol. 11, no. 2
p. 120

Abstract

Read online

In the modern era, patients often resort to the internet for answers to their health-related concerns, and clinics face challenges to providing timely response to patient concerns. This has led to a need to investigate the capabilities of AI chatbots for ophthalmic diagnosis and triage. In this in silico study, 80 simulated patient complaints in ophthalmology with varying urgency levels and clinical descriptors were entered into both ChatGPT and Bard in a systematic 3-step submission process asking chatbots to triage, diagnose, and evaluate urgency. Three ophthalmologists graded chatbot responses. Chatbots were significantly better at ophthalmic triage than diagnosis (90.0% appropriate triage vs. 48.8% correct leading diagnosis; p p = 0.008), grader satisfaction for patient use (81.3% vs. 55.0%; p p = 0.010). More descriptors improved the accuracy of diagnosis for both GPT-4 and Bard. These results indicate that chatbots may not need to recognize the correct diagnosis to provide appropriate ophthalmic triage, and there is a potential utility of these tools in aiding patients or triage staff; however, they are not a replacement for professional ophthalmic evaluation or advice.

Keywords