June 2023
Volume 64, Issue 8
Open Access
ARVO Annual Meeting Abstract  |   June 2023
Multimodal Transformer Model to Detect Glaucoma from OCT and Retinal Nerve Fiber Layer (RNFL) Thickness
Author Affiliations & Notes
  • Justin Huynh
    Viterbi Family Department of Ophthalmology, University of California, San Diego, Hamilton Glaucoma Center, Shiley Eye Institute, La Jolla, California, United States
  • Ruben Gonzalez
    Viterbi Family Department of Ophthalmology, University of California, San Diego, Hamilton Glaucoma Center, Shiley Eye Institute, La Jolla, California, United States
  • Evan Walker
    Viterbi Family Department of Ophthalmology, University of California, San Diego, Hamilton Glaucoma Center, Shiley Eye Institute, La Jolla, California, United States
  • Benton Gabriel Chuter
    Viterbi Family Department of Ophthalmology, University of California, San Diego, Hamilton Glaucoma Center, Shiley Eye Institute, La Jolla, California, United States
  • Akram Belghith
    Viterbi Family Department of Ophthalmology, University of California, San Diego, Hamilton Glaucoma Center, Shiley Eye Institute, La Jolla, California, United States
  • Christopher Bowd
    Viterbi Family Department of Ophthalmology, University of California, San Diego, Hamilton Glaucoma Center, Shiley Eye Institute, La Jolla, California, United States
  • Michael Henry Goldbaum
    Viterbi Family Department of Ophthalmology, University of California, San Diego, Hamilton Glaucoma Center, Shiley Eye Institute, La Jolla, California, United States
  • Sally L. Baxter
    Viterbi Family Department of Ophthalmology, University of California, San Diego, Hamilton Glaucoma Center, Shiley Eye Institute, La Jolla, California, United States
    Health Department of Biomedical Informatics, University of California San Diego, La Jolla, California, United States
  • Robert N Weinreb
    Viterbi Family Department of Ophthalmology, University of California, San Diego, Hamilton Glaucoma Center, Shiley Eye Institute, La Jolla, California, United States
  • Ali Tafreshi
    Viterbi Family Department of Ophthalmology, University of California, San Diego, Hamilton Glaucoma Center, Shiley Eye Institute, La Jolla, California, United States
  • Linda M Zangwill
    Viterbi Family Department of Ophthalmology, University of California, San Diego, Hamilton Glaucoma Center, Shiley Eye Institute, La Jolla, California, United States
  • Mark Christopher
    Viterbi Family Department of Ophthalmology, University of California, San Diego, Hamilton Glaucoma Center, Shiley Eye Institute, La Jolla, California, United States
  • Footnotes
    Commercial Relationships   Justin Huynh None; Ruben Gonzalez None; Evan Walker None; Benton Chuter None; Akram Belghith None; Christopher Bowd None; Michael Goldbaum None; Sally Baxter voxelcloud.io, Code C (Consultant/Contractor), Optomed, Topcon, Code F (Financial Support), iVista Medical Education, Code I (Personal Financial Interest); Robert Weinreb Abbvie, Aerie Pharmaceuticals, Allergan, Equinox, Eyenovia, Nicox, Topcon, Code C (Consultant/Contractor), Heidelberg Engineering, Carl Zeiss Meditec, Konan Medical, Optovue, Centervue, Bausch&Lomb, Topcon, Code F (Financial Support), Toromedes, Carl Zeiss Meditec, Code P (Patent); Ali Tafreshi None; Linda Zangwill Abbvie Inc. Digitial Diagnostics, Code C (Consultant/Contractor), National Eye Institute, Carl Zeiss Meditec Inc., Heidelberg Engineering GmbH, Optovue Inc., Topcon Medical Systems Inc., Code F (Financial Support), Zeiss Meditec, Code P (Patent); Mark Christopher AISight Health, Code P (Patent)
  • Footnotes
    Support  This work is supported by National Institutes of Health/National Eye Institute Grants (R01EY034148, R01EY029058, R01EY11008, R01EY19869, R01EY027510, R01EY026574, EY018926, P30EY022589, K99EY030942, R01EY034146, T35EY033704, DP5OD029610); and an unrestricted grant from Research to Prevent Blindness (New York, NY). The sponsor or funding organization had no role in the design or conduct of this research.
Investigative Ophthalmology & Visual Science June 2023, Vol.64, 362. doi:
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Justin Huynh, Ruben Gonzalez, Evan Walker, Benton Gabriel Chuter, Akram Belghith, Christopher Bowd, Michael Henry Goldbaum, Sally L. Baxter, Robert N Weinreb, Ali Tafreshi, Linda M Zangwill, Mark Christopher; Multimodal Transformer Model to Detect Glaucoma from OCT and Retinal Nerve Fiber Layer (RNFL) Thickness. Invest. Ophthalmol. Vis. Sci. 2023;64(8):362.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Purpose :
Most prior works using deep learning (DL) models to detect glaucoma were applied to individual imaging modalities, such as fundus photographs, optical coherence tomography (OCT) or visual field (VF). However, a clinician often analyzes multiple sources of data when making a diagnosis. A multimodal model that incorporates multiple types of data may improve glaucoma detection performance. The purpose of this study was to compare a single modality convolutional neural network (CNN), a single modality transformer model, and a multimodal transformer model in detecting glaucoma from raw OCT data and RNFL thickness measurements.

Methods :
A set of 22,464 OCT Spectralis optic nerve head (ONH) circle scans and corresponding RNFL segmentations collected as part of the Diagnostic Innovations in Glaucoma Study (DIGS) and African Descent and Glaucoma Evaluation Study (ADAGES) were analyzed. Data were split on the patient level into training (571 patients, 969 eyes, 20,528 images) and testing (48 patients, 83 eyes, 1936 images) sets. Images were labeled as glaucomatous or healthy based on presence of glaucomatous visual field defects (GVFD). For the single modality models, ResNet50 and Vision Transformer (ViT) architectures were trained solely on OCT RNFL circle scans to detect glaucoma. For the multimodal model, a modified ViT (Figure 1) was trained on unsegmented RNFL circle scans and compared to a vector of 768 thickness measurements derived from the segmented RNFL circle scans to detect glaucoma.

Results :
The ResNet50 trained on RNFL scans had an accuracy of 0.85, area under receiver operating characteristic (AUROC) of 0.91, sensitivity 0.84, and specificity 0.84 (Figure 2). The ViT trained on RNFL scans had an accuracy of 0.85, AUROC of 0.92, sensitivity 0.84, and specificity 0.84. The multimodal ViT trained on RNFL scans and RNFL thickness values had an accuracy of 0.86, AUROC of 0.93, sensitivity 0.84, and specificity 0.84. The transformer based models, ViT and multimodal ViT performed marginally better than the CNN-based model ResNet50. The multimodal ViT performed modestly better than the single modality models.

Conclusions :
A multimodal DL model trained on OCT and RNFL thickness modestly outperforms single modality models trained only on OCT. Multimodal data may provide more accurate deep learning models, and is worth investigating further.

This abstract was presented at the 2023 ARVO Annual Meeting, held in New Orleans, LA, April 23-27, 2023.

 


Model architecture.


Model architecture.

 


Results on glaucoma detection.


Results on glaucoma detection.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×