June 2021
Volume 62, Issue 8
Open Access
ARVO Annual Meeting Abstract  |   June 2021
Improved training efficiency for deep learning models using disease severity comparison labels
Author Affiliations & Notes
  • Adam Hanif
    Ophthalmology, Oregon Health & Science University, Portland, Oregon, United States
  • Ilkay Yildiz
    Electrical and Computer Engineering, Northeastern University, Boston, Massachusetts, United States
  • Peng Tian
    Electrical and Computer Engineering, Northeastern University, Boston, Massachusetts, United States
  • Beyza Kalkanli
    Electrical and Computer Engineering, Northeastern University, Boston, Massachusetts, United States
  • Deniz Erdogmus
    Electrical and Computer Engineering, Northeastern University, Boston, Massachusetts, United States
  • Stratis Ioannidis
    Electrical and Computer Engineering, Northeastern University, Boston, Massachusetts, United States
  • Jennifer Dy
    Electrical and Computer Engineering, Northeastern University, Boston, Massachusetts, United States
  • Jayashree Kalpathy-Cramer
    Department of Radiology, Athinoula A Martinos Center for Biomedical Imaging Clinical Computational Neuroimaging Group, Charlestown, Massachusetts, United States
  • Susan Ostmo
    Ophthalmology, Oregon Health & Science University, Portland, Oregon, United States
  • Karyn Jonas
    Ophthalmology, University of Illinois at Chicago College of Medicine, Chicago, Illinois, United States
  • R.V. Paul Chan
    Ophthalmology, University of Illinois at Chicago College of Medicine, Chicago, Illinois, United States
  • Michael F Chiang
    Ophthalmology, Oregon Health & Science University, Portland, Oregon, United States
  • J. Peter Campbell
    Ophthalmology, Oregon Health & Science University, Portland, Oregon, United States
  • Footnotes
    Commercial Relationships   Adam Hanif, None; Ilkay Yildiz, None; Peng Tian, None; Beyza Kalkanli, None; Deniz Erdogmus, None; Stratis Ioannidis, None; Jennifer Dy, None; Jayashree Kalpathy-Cramer, None; Susan Ostmo, None; Karyn Jonas, None; R.V. Paul Chan, None; Michael Chiang, Genentech (F), InTeleretina (I), NIH (F), Novartis (C), NSF (F); J. Peter Campbell, Genentech (F)
  • Footnotes
    Support  This study was supported by grants R01EY19474, R01EY031331, and P30EY10572 from the National Institutes of Health (Bethesda, MD); by unrestricted departmental funding, and a Career Development Award (JPC) from Research to Prevent Blindness (New York, NY).
Investigative Ophthalmology & Visual Science June 2021, Vol.62, 2108. doi:
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Adam Hanif, Ilkay Yildiz, Peng Tian, Beyza Kalkanli, Deniz Erdogmus, Stratis Ioannidis, Jennifer Dy, Jayashree Kalpathy-Cramer, Susan Ostmo, Karyn Jonas, R.V. Paul Chan, Michael F Chiang, J. Peter Campbell; Improved training efficiency for deep learning models using disease severity comparison labels. Invest. Ophthalmol. Vis. Sci. 2021;62(8):2108.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Purpose : Neural network performance relies on large, high-quality training sets. In medical image recognition tasks, small datasets and high inter-labeler variance frequently limit models’ diagnostic accuracy. In this study, we compare the efficiency of training neural networks to predict disease severity using “comparison” labels versus the traditional method of using diagnostic “class” labels from a retinopathy of prematurity retinal image dataset.

Methods : 100 fundus images were each assigned “class” labels indicating plus disease severity per the majority vote of 3 experts between either “Plus”, “Pre-plus” or “No Plus”. Additionally, all combinations of image pairs within the set were assigned “comparison” labels reflecting relative disease severity obtained from 5 experts (4950 labels total). Deep learning models were first trained with “class” labels from up to 60 randomly sampled images, and validated on a set of 20 images with “class” labels. Then, this process was repeated using “comparison” labels. All models were then evaluated on a test set of 5561 pre-labeled fundus images in two binary classification experiments: “Normal vs. Abnormal” and “Plus vs. Non-plus”. For each model, predictive performance was measured by area under the receiver operating curves (AUC).

Results : For a given number of images, models trained on “comparison” labels consistently outperformed those trained on “class” labels. For the same number of labels, the performance of “class” and “comparison” labels was similar, but models trained on class labels exhibited wider confidence intervals by up to 0.2% in “Normal vs. Abnormal” experiments and 0.4% in “Plus vs. Non-plus” experiments (Figure 1).

Conclusions : "Comparison" labels are more informative per image than "class" labels. Further, the inherent subjectivity of "class" labels generates higher variability in model performance. This offers a solution for training highly accurate image classification models with fewer data.

This is a 2021 ARVO Annual Meeting abstract.

 

AUC of models trained with either “comparison” or “class” labels from sets of up to 60 corresponding images (A, B) or individual labels (C, D). Models' accuracy in disease severity prediction was assessed through binary image classification experiments: Normal vs. Abnormal (A, C) and Plus vs. Non-plus (B, D). Confidence intervals on reported metrics are indicated by the shaded region around the mean curve.

AUC of models trained with either “comparison” or “class” labels from sets of up to 60 corresponding images (A, B) or individual labels (C, D). Models' accuracy in disease severity prediction was assessed through binary image classification experiments: Normal vs. Abnormal (A, C) and Plus vs. Non-plus (B, D). Confidence intervals on reported metrics are indicated by the shaded region around the mean curve.

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×