Investigative Ophthalmology & Visual Science Cover Image for Volume 65, Issue 7
June 2024
Volume 65, Issue 7
Open Access
ARVO Annual Meeting Abstract  |   June 2024
Annotation of Surgical Tool Depth in Vitreoretinal Surgical Videos: Agreement and Performance Between Vitreoretinal Surgeons vs. Non-Surgeon Graders
Author Affiliations & Notes
  • Jason Dagoon
    Center of Translational Vision Research, University of California Irvine, Irvine, California, United States
    University of California Irvine Gavin Herbert Eye Institute, Irvine, California, United States
  • Pierre F. Baldi
    Information and Computer Sciences, University of California Irvine, Irvine, California, United States
    Center of Translational Vision Research, University of California Irvine, Irvine, California, United States
  • Sherif Abdelkarim
    Information and Computer Sciences, University of California Irvine, Irvine, California, United States
    Center of Translational Vision Research, University of California Irvine, Irvine, California, United States
  • Junze Liu
    Information and Computer Sciences, University of California Irvine, Irvine, California, United States
    Center of Translational Vision Research, University of California Irvine, Irvine, California, United States
  • Mohammad Esfahani Riazi
    University of California Irvine Gavin Herbert Eye Institute, Irvine, California, United States
    Center of Translational Vision Research, University of California Irvine, Irvine, California, United States
  • Marisabel Andrade
    University of California Irvine Gavin Herbert Eye Institute, Irvine, California, United States
    Center of Translational Vision Research, University of California Irvine, Irvine, California, United States
  • Amr Azzam
    University of California Irvine Gavin Herbert Eye Institute, Irvine, California, United States
    Center of Translational Vision Research, University of California Irvine, Irvine, California, United States
  • Parsa Riazi Esfahani
    University of California Irvine Gavin Herbert Eye Institute, Irvine, California, United States
  • Steven Chang
    Center of Translational Vision Research, University of California Irvine, Irvine, California, United States
  • Andrew Browne
    Center of Translational Vision Research, University of California Irvine, Irvine, California, United States
    University of California Irvine Gavin Herbert Eye Institute, Irvine, California, United States
  • Footnotes
    Commercial Relationships   Jason Dagoon None; Pierre Baldi None; Sherif Abdelkarim None; Junze Liu None; Mohammad Riazi None; Marisabel Andrade None; Amr Azzam None; Parsa Riazi Esfahani None; Steven Chang None; Andrew Browne Jcyte (2021-22), Alimera (2022), JeniVision (2022), Code C (Consultant/Contractor), United States patent US20200336638, United States patent US20200163737, United States patent US10295526, Code P (Patent)
  • Footnotes
    Support  Gavin Herbert Eye Institute 20/20 Society Pilot Research, BrightFocus Foundation, NIH/NEI 1K08EY034912 - 01, The Retina Society Research and International Retina Research Foundation, Unrestricted grant to UC Irvine department of ophthalmology from Research to Prevent Blindness
Investigative Ophthalmology & Visual Science June 2024, Vol.65, 3762. doi:
  • Views
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Jason Dagoon, Pierre F. Baldi, Sherif Abdelkarim, Junze Liu, Mohammad Esfahani Riazi, Marisabel Andrade, Amr Azzam, Parsa Riazi Esfahani, Steven Chang, Andrew Browne; Annotation of Surgical Tool Depth in Vitreoretinal Surgical Videos: Agreement and Performance Between Vitreoretinal Surgeons vs. Non-Surgeon Graders. Invest. Ophthalmol. Vis. Sci. 2024;65(7):3762.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Purpose : AI implementation in vitreoretinal surgery may enhance surgical training, provide safety mechanisms, and enable objective research from surgical videos. However, there is a lack of standards for the appropriate surgical annotator and annotation methods. This study examines the variability among surgical video graders of varying clinical experience by assessing the agreement and performance between vitreoretinal surgeons and non-surgeons in the annotation of surgical tool depth.

Methods : Three vitreoretinal surgeons and three lay scientists independently annotated an identical set of surgical videos comprising 66,920 frames. To annotate tool depth, each grader categorized the distance of the surgical tool tip from the retina into four zones: far, intermediate, near, and contact. Frames were arranged chronologically, with video playback. Computer Vision Annotation Tool (CVAT) was used for recording annotations. A deep learning model was trained using frames where the surgeons agreed (57,050). Performance was evaluated by comparing percent accuracy against the predictions of the trained model.

Results : Non-surgeons annotated a significantly higher percentage of frames than surgeon graders. However, surgeons showcased greater agreement than any two non-surgeons (57,050 vs. 30,134 frame consensus). Surgeons and non-surgeons agreed on only 9,020 frames. Compared with the trained model for instrument depth prediction, no grader achieved accuracy greater than 60%. Surgeons demonstrated more consistency and slightly higher accuracy in their performance compared to non-surgeons.

Conclusions : Our data suggests that non-surgeon annotators may benefit from additional training for more accurate and uniform annotations in evaluating tool depth. Graders performed sub optimally against the trained model, suggesting the existing method of using CVAT for surgical tool depth annotation in surgical videos is insufficient. There is a need for custom annotation software and workflow modification for refined annotation strategies.

This abstract was presented at the 2024 ARVO Annual Meeting, held in Seattle, WA, May 5-9, 2024.

 

Figure: Metrics for graders performing depth annotation task and their performance compared to a trained AI model

Figure: Metrics for graders performing depth annotation task and their performance compared to a trained AI model

×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×