Investigative Ophthalmology & Visual Science Cover Image for Volume 65, Issue 7
June 2024
Volume 65, Issue 7
Open Access
ARVO Annual Meeting Abstract  |   June 2024
Large Language Models: A New Frontier in Pediatric Cataract Patient Education
Author Affiliations & Notes
  • Andrew David Brown
    University of Arkansas for Medical Sciences College of Medicine, Little Rock, Arkansas, United States
    Ophthalmology, University of Arkansas for Medical Sciences, Little Rock, Arkansas, United States
  • Qais Dihan
    Rosalind Franklin University of Medicine and Science Chicago Medical School, North Chicago, Illinois, United States
    Ophthalmology, University of Arkansas for Medical Sciences, Little Rock, Arkansas, United States
  • Muhammad Z Chauhan
    Ophthalmology, University of Arkansas for Medical Sciences, Little Rock, Arkansas, United States
  • Taher K Eleiwa
    Ophthalmology, Benha University, Benha, Qalyubia, Egypt
  • Amr K Hassan
    Ophthalmology, South Valley University, Qena, Qena, Egypt
  • Ahmed B Sallam
    Ophthalmology, University of Arkansas for Medical Sciences, Little Rock, Arkansas, United States
    Ophthalmology, Ain Shams University Faculty of Medicine, Cairo, Egypt
  • Abdelrahman M Elhusseiny
    Ophthalmology, University of Arkansas for Medical Sciences, Little Rock, Arkansas, United States
    Ophthalmology, Boston Children's Hospital, Boston, Massachusetts, United States
  • Footnotes
    Commercial Relationships   Andrew Brown None; Qais Dihan None; Muhammad Chauhan None; Taher Eleiwa None; Amr Hassan None; Ahmed Sallam None; Abdelrahman Elhusseiny None
  • Footnotes
    Support  None
Investigative Ophthalmology & Visual Science June 2024, Vol.65, 347. doi:
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Andrew David Brown, Qais Dihan, Muhammad Z Chauhan, Taher K Eleiwa, Amr K Hassan, Ahmed B Sallam, Abdelrahman M Elhusseiny; Large Language Models: A New Frontier in Pediatric Cataract Patient Education. Invest. Ophthalmol. Vis. Sci. 2024;65(7):347.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Purpose : Patient education materials (PEMs) available online for pediatric cataract have been shown to be composed at a reading level which exceeds the average American's comprehension skills. Large language models (LLMs) can rewrite text materials, and generate novel text. Our cross-sectional comparative study evaluated the ability of 3 LLMs (ChatGPT-3.5, ChatGPT-4, and Google Bard) to generate novel PEMs, and improve the readability of existing PEMs on pediatric cataract.

Methods : We compared LLMs’ responses to 3 prompts. Prompt A (control) requested they write an "educational handout on pediatric cataract that was easily understandable by the average American." Prompt B contained a modifier statement that requested the handout be written "at a 6th grade reading level (as recommended by the American Medical Association), using the SMOG (Simple Measure of Gobbledygook) readability formula." Prompt C rewrote existing PEMs on pediatric cataract "to a 6th grade reading level using the SMOG readability formula." Reponses were then compared on their readability (SMOG, Flesch-Kincaid Grade Level (FKGL)), quality (DISCERN tool, Patient Education Materials Assessment Tool (PEMAT)), and accuracy (Likert misinformation scale).

Results : All generated PEMs were of high quality (median DISCERN score ≥4, PEMAT understandability ≥70%) and accuracy (Likert score=1). Generated PEMs were not actionable (PEMAT actionability <70%). PEMs generated in response to Prompt B were significantly more readable than Prompt A responses (p<0.001). ChatGPT-4 generated PEMs more readable (lower SMOG and FGKL scores; 5.6 ± 0.5 and 4.3 ± 0.7, respectively) than the other two LLMs (p<0.001). Although all LLMs improved readability of existing PEMs (Prompt C), only ChatGPT-4 consistently rewrote them to or below the specified 6th grade reading level (SMOG: 5.1 ± 0.3, FKGL 3.8 ± 0.6).

Conclusions : This study underscores the value of LLMs, particularly ChatGPT-4, as strong supplementary tools for generating high-quality, readable, and accurate PEMs on pediatric cataract. Through effectively improving the readability of existing PEMs, they demonstrate potential to make information on pediatric cataract more accessible to a greater number of patients.

This abstract was presented at the 2024 ARVO Annual Meeting, held in Seattle, WA, May 5-9, 2024.

 

A) Prompt A vs B: Generated patient education materials’ readability
B) Differences in Prompt B readability scores

A) Prompt A vs B: Generated patient education materials’ readability
B) Differences in Prompt B readability scores

 

A) Readability: Online materials vs Prompt C rewrites
B) Prompt C rewrites: head-to-head comparison

A) Readability: Online materials vs Prompt C rewrites
B) Prompt C rewrites: head-to-head comparison

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×