Investigative Ophthalmology & Visual Science Cover Image for Volume 65, Issue 7
June 2024
Volume 65, Issue 7
Open Access
ARVO Annual Meeting Abstract  |   June 2024
Analyzing the Role of AI in Medical Education: An Evaluation on the Performance of ChatGPT on Ophthalmology Trainee Examination Questions
Author Affiliations & Notes
  • Carolina J. Ramos
    Rutgers New Jersey Medical School, Newark, New Jersey, United States
  • Kimberly Cartagena
    Rutgers New Jersey Medical School, Newark, New Jersey, United States
  • Kailyn Amanda Ramirez
    Rutgers Robert Wood Johnson Medical School, Piscataway, New Jersey, United States
  • Alexandra Sanchez
    Rutgers New Jersey Medical School, Newark, New Jersey, United States
  • Maria Vega Garces
    Rutgers New Jersey Medical School, Newark, New Jersey, United States
  • Marko Oydanich
    Rutgers New Jersey Medical School, Newark, New Jersey, United States
  • Albert S Khouri
    Rutgers New Jersey Medical School, Newark, New Jersey, United States
  • Footnotes
    Commercial Relationships   Carolina Ramos None; Kimberly Cartagena None; Kailyn Ramirez None; Alexandra Sanchez None; Maria Vega Garces None; Marko Oydanich None; Albert Khouri None
  • Footnotes
    Support  None
Investigative Ophthalmology & Visual Science June 2024, Vol.65, 359. doi:
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Carolina J. Ramos, Kimberly Cartagena, Kailyn Amanda Ramirez, Alexandra Sanchez, Maria Vega Garces, Marko Oydanich, Albert S Khouri; Analyzing the Role of AI in Medical Education: An Evaluation on the Performance of ChatGPT on Ophthalmology Trainee Examination Questions. Invest. Ophthalmol. Vis. Sci. 2024;65(7):359.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Purpose : Chat GPT is a large-scale language model trained on various datasets to learn, analyze, and generate human-like answers to user’s questions. As artificial intelligence (AI) progresses and its use expands into medical education, it is increasingly important to assess the validity of its output. When considering its applicability in medical education, more information is required to understand whether its analyses are producing accurate and coherent responses. This study aims to evaluate the performance and accuracy of ChatGPT on practice questions used to prepare for the Ophthalmic Knowledge Assessment Program (OKAP).

Methods : Ophthalmology questions were obtained from a widely utilized study resource, Ophthoquestions. 13 sections, each with a differing ophthalmic subtopic, were sampled and 10 questions were collected from each section. Questions containing images or tables wereexcluded. 98 out of 130 questions and their respective answer choices were input into ChatGPT-3.5. ChatGPT responses were evaluated via the properties of natural coherence (Table 1). Incorrect responses were categorized as either logical fallacy, informational fallacy, or explicit fallacy (Table 2). ChatGPT accuracy was analyzed using Microsoft Excel and chi-square tests to determine statistical significance of categorical variables.

Results : ChatGPT answered 52 out of 98 questions correctly (53%). Logical reasoning, internal information and external information were identified in 82.7%, 84.7%, and 78.6% of the responses, respectively.Of the incorrect answers, informational was the most frequent fallacy (43.5%), followed by logical fallacy (32.6%) and external fallacy (23.9%). The use of logical reasoning (p=0.02) and internal information (p=0.01) was found to be statistically significant when stratified by correct and incorrect responses.

Conclusions : ChatGPT may be a potential study aid in resident education and serve as an additional resource for board preparation in ophthalmology. Given the recent advancements in AI, future studies should assess whether ChatGPT can positively influence resident performance when implemented as a learning tool.

This abstract was presented at the 2024 ARVO Annual Meeting, held in Seattle, WA, May 5-9, 2024.

 

 

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×