Investigative Ophthalmology & Visual Science Cover Image for Volume 65, Issue 7
June 2024
Volume 65, Issue 7
Open Access
ARVO Annual Meeting Abstract  |   June 2024
A Unified Framework for Visual Field Test Estimation and Forecasting using Convolution and Attention Networks
Author Affiliations & Notes
  • Ashkan Abbasi
    Department of Ophthalmology, Casey Eye Institute, Oregon Health & Science University, Portland, Oregon, United States
  • Sowjanya Gowrisankaran
    Department of Ophthalmology, Casey Eye Institute, Oregon Health & Science University, Portland, Oregon, United States
  • Bhavna Josephine Antony
    Institute of Innovation, Science and Sustainability, Information Technology, Federation University Australia, Ballarat, Victoria, Australia
  • Xubo Song
    Department of Medical Informatics and Clinical Epidemiology, Oregon Health & Science University, Portland, Oregon, United States
  • Gadi Wollstein
    Department of Ophthalmology, NYU Langone Health, New York University, New York, New York, United States
    Department of Biomedical Engineering, New York University Tandon School of Engineering, Brooklyn, New York, United States
  • Joel S Schuman
    Wills Eye Hospital, Philadelphia, Pennsylvania, United States
    Thomas Jefferson University Sidney Kimmel Medical College, Philadelphia, Pennsylvania, United States
  • Hiroshi Ishikawa
    Department of Ophthalmology, Casey Eye Institute, Oregon Health & Science University, Portland, Oregon, United States
    Department of Medical Informatics and Clinical Epidemiology, Oregon Health & Science University, Portland, Oregon, United States
  • Footnotes
    Commercial Relationships   Ashkan Abbasi None; Sowjanya Gowrisankaran None; Bhavna Antony None; Xubo Song None; Gadi Wollstein None; Joel Schuman AEYE Health,Alcon Laboratories, Inc., Boehringer Ingelheim, Carl Zeiss Meditec, Ocugenix, Ocular Therapeutix, Opticient,Perfuse Inc., Regeneron Pharmaceuticals, Inc., SLACK;, Code C (Consultant/Contractor), AEYE Health, CarlZeiss Meditec, Ocugenix, Ocular Therapeutix, Opticient, Code I (Personal Financial Interest), Carl Zeiss Meditec, Ocugenix, Code P (Patent); Hiroshi Ishikawa Gobiquity, Code O (Owner)
  • Footnotes
    Support  This work was partially supported by NIH grants number R01EY030929, R01EY013178, and P30 EY010572 core grant, the Malcolm M. Marquis, MD Endowed Fund for Innovation, and an unrestricted grant from Research to Prevent Blindness (New York, NY) to Casey Eye Institute, Oregon Health & Science University.
Investigative Ophthalmology & Visual Science June 2024, Vol.65, 2361. doi:
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Ashkan Abbasi, Sowjanya Gowrisankaran, Bhavna Josephine Antony, Xubo Song, Gadi Wollstein, Joel S Schuman, Hiroshi Ishikawa; A Unified Framework for Visual Field Test Estimation and Forecasting using Convolution and Attention Networks. Invest. Ophthalmol. Vis. Sci. 2024;65(7):2361.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Purpose : Different methods have been reported for visual field (VF) estimation from optical coherence tomography (OCT) or forecasting future VF using prior VFs. These methods are modality-specific and hard to compare with each other. Our goal is to test the efficacy of our unified framework to estimate and forecast VF using different input data (2D or 3D OCT, and VF).

Methods : From our longitudinal glaucoma cohort, we collected 8,390 pairs of 2 consecutive VFs (Humphrey, 24-2 SITA Standard, Zeiss, Dublin, CA) and their corresponding 3D OCT images (Cirrus HD-OCT, 200x200 ONH Scan, Zeiss). The average number of days between sessions was 342 days (range 90-2400). The dataset was split to perform 10-fold cross-validation without patient overlap. We utilized a hybridized convolution and transformer network (CoTrNet) architecture (Table 1) composed of inverted residual convolution and transformer (relative self-attention and a fully connected layer) blocks to capture local and global patterns. In VF estimation, either a 2D (e.g. en face image, layer thickness map, etc.) or a down-sampled 3D OCT image can be used to estimate the corresponding VF (52 out of 54 values, excluding 2 blind spots). In VF forecasting, the input is made up of the current VF and the time difference (between the current and future VFs).

Results : In VF estimation with enface images, 2D ResNet and CoTrNet achieved the global mean absolute error (MAE) of 3.91 ± 0.24, and 3.52 ± 0.26 dBs, respectively. However, the overall performance was improved by using 3D OCT images. In Figure 1, MAE and its pointwise heatmap are reported for the compared methods. It shows that 3D ResNet’s error was 3.39 ± 0.21 while CoTrNet’s error was 3.10 ± 0.15. In VF forecasting, CoTrNet (MAE=2.10 ± 0.11) significantly outperformed the identity function (2.70 ± 0.15), recurrent neural network (RNN) (2.54 ± 0.21), and CascadeNet-5 (2.27 ± 0.13) methods. Our analysis showed that VF forecasting performance stayed stable until the time interval hit 4 years.

Conclusions : Our unified framework supported the use of 2D and 3D OCT, as well as VF, as inputs. The proposed model outperformed problem- and modality-specific methods for both VF estimation and forecasting by harnessing the power of local and global processing via the integration of convolutions and transformers in the network.

This abstract was presented at the 2024 ARVO Annual Meeting, held in Seattle, WA, May 5-9, 2024.

 

 

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×