Investigative Ophthalmology & Visual Science Cover Image for Volume 65, Issue 7
June 2024
Volume 65, Issue 7
Open Access
ARVO Annual Meeting Abstract  |   June 2024
Evaluation of Learning Model Advice Regarding Common Ocular Symptoms
Author Affiliations & Notes
  • Sanjana Harihar
    University of Pittsburgh, Pittsburgh, Pennsylvania, United States
  • Anindya Samanta
    Department of Ophthalmology, Baylor College of Medicine, Houston, Texas, United States
  • Amer Alsoudi
    Department of Ophthalmology, Baylor College of Medicine, Houston, Texas, United States
  • Kieran Lyons
    To The Top Education Group, Pennsylvania, United States
  • Chris Allen
    Department of Emergency Medicine, Allegheny Health Network, Pittsburgh, Pennsylvania, United States
  • Austin Williams
    Emergent Medical Associates, Hollywood Presbyterian Medical Center Medical Library, Los Angeles, California, United States
  • Elham Sadeghi
    Department of Ophthalmology, UPMC, Pittsburgh, Pennsylvania, United States
  • Jay Chhablani
    Department of Ophthalmology, UPMC, Pittsburgh, Pennsylvania, United States
  • Footnotes
    Commercial Relationships   Sanjana Harihar None; Anindya Samanta None; Amer Alsoudi None; Kieran Lyons None; Chris Allen None; Austin Williams None; Elham Sadeghi None; Jay Chhablani Novartis, Allergan, OD-OS, Erasca, Bausch + Lomb, Code C (Consultant/Contractor)
  • Footnotes
    Support  none
Investigative Ophthalmology & Visual Science June 2024, Vol.65, 341. doi:
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Sanjana Harihar, Anindya Samanta, Amer Alsoudi, Kieran Lyons, Chris Allen, Austin Williams, Elham Sadeghi, Jay Chhablani; Evaluation of Learning Model Advice Regarding Common Ocular Symptoms. Invest. Ophthalmol. Vis. Sci. 2024;65(7):341.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Purpose : Patients are seeking out medical advice on publicly available learning models. Medical advice from these learning models has yet to be fully evaluated. In this study we aim to evaluate the medical advice of four learning models and assess physician and non-physician perception of the advice.

Methods : 11 common eye symptoms from American Academy of Ophthalmology online website, "For Public & Patient" section, were put into four publicly available learning models (Chat GPT-3.5, Chat GPT 4.0, Bing, and Bard) in August 2023. Follow-up questions included etiology therapy, and follow-up with the physician. The answers were anonymized and graded by 6 evaluators (2 ophthalmologists, 2 emergency medicine physicians, and 2 non-medical persons with graduate degrees) on a 5-point Likert scale in three categories: accuracy, helpfulness, and specificity. The results were analyzed to determine the quality of response from each model.

Results : Average ratings for each learning model by each grader are shown in Figure 1. Overall Chat GPT 3.5 was the highest rated for all 6 graders and had the highest average. Bard had the lowest rated average, Chat GPT 3.5 was selected by evaluators as having the best response for 6/11 of the questions. Chat GPT 4.0, Bing and Bard were selected by the evaluators for each having the best response to 2 of the questions each.
Spearman’s correlation between the two non-medical persons was 0.51; between the two emergency medicine physicians was 0.49; and between two ophthalmologist was 0.63 (Figure 2). ANOVA of the responses between non-ophthalmologists was significant (p<0.005) but between physicians was not significant ( p>0.05).

Conclusions : Overall, Chat GPT 3.5 tends to provide the best advice at present. While within each group there was moderate to strong correlation on how each grader perceived the advice, non-medical persons evaluated the advice differently than physicians.

This abstract was presented at the 2024 ARVO Annual Meeting, held in Seattle, WA, May 5-9, 2024.

 

The overall average rating for each learning model by each gender. L1 and L2 were two non-medical persons with graduate degrees. E1 and E2 were two emergency medicine physicans. O1 and O2 were two ophthalmologists.

The overall average rating for each learning model by each gender. L1 and L2 were two non-medical persons with graduate degrees. E1 and E2 were two emergency medicine physicans. O1 and O2 were two ophthalmologists.

 

The relationship between the grader's response in each group, as shown by Spearman's correlation coefficient (SCC) within each group. L1 and L2 were two non-medical persons with graduate degress. E1 and E2 were two emergency medicine phycians. O1 and O2 were two ophthalmologists.

The relationship between the grader's response in each group, as shown by Spearman's correlation coefficient (SCC) within each group. L1 and L2 were two non-medical persons with graduate degress. E1 and E2 were two emergency medicine phycians. O1 and O2 were two ophthalmologists.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×