Clinical Ophthalmology (Mar 2024)

Evaluation of Methods for Detection and Semantic Segmentation of the Anterior Capsulotomy in Cataract Surgery Video

  • Zeng Z,
  • Giap BD,
  • Kahana E,
  • Lustre J,
  • Mahmoud O,
  • Mian SI,
  • Tannen B,
  • Nallasamy N

Journal volume & issue
Vol. Volume 18
pp. 647 – 657

Abstract

Read online

Zixue Zeng,1,* Binh Duong Giap,2,* Ethan Kahana,3 Jefferson Lustre,4 Ossama Mahmoud,5 Shahzad I Mian,2 Bradford Tannen,2 Nambi Nallasamy2,6 1School of Public Health, University of Michigan, Ann Arbor, MI, USA; 2Kellogg Eye Center, Department of Ophthalmology and Visual Sciences, University of Michigan, Ann Arbor, MI, USA; 3Department of Computer Science, University of Michigan, Ann Arbor, MI, USA; 4School of Medicine, University of Michigan, Ann Arbor, MI, USA; 5School of Medicine, Wayne State University, Detroit, MI, USA; 6Department of Computational Medicine and Bioinformatics, University of Michigan, Ann Arbor, MI, USA*These authors contributed equally to this workCorrespondence: Nambi Nallasamy, Kellogg Eye Center, University of Michigan, 1000 Wall Street, Ann Arbor, MI, 48105, USA, Tel +1 (734) 763-5506, Fax +1 (734) 936-2340, Email [email protected]: The capsulorhexis is one of the most important and challenging maneuvers in cataract surgery. Automated analysis of the anterior capsulotomy could aid surgical training through the provision of objective feedback and guidance to trainees.Purpose: To develop and evaluate a deep learning-based system for the automated identification and semantic segmentation of the anterior capsulotomy in cataract surgery video.Methods: In this study, we established a BigCat-Capsulotomy dataset comprising 1556 video frames extracted from 190 recorded cataract surgery videos for developing and validating the capsulotomy recognition system. The proposed system involves three primary stages: video preprocessing, capsulotomy video frame classification, and capsulotomy segmentation. To thoroughly evaluate its efficacy, we examined the performance of a total of eight deep learning-based classification models and eleven segmentation models, assessing both accuracy and time consumption. Furthermore, we delved into the factors influencing system performance by deploying it across various surgical phases.Results: The ResNet-152 model employed in the classification step of the proposed capsulotomy recognition system attained strong performance with an overall Dice coefficient of 92.21%. Similarly, the UNet model with the DenseNet-169 backbone emerged as the most effective segmentation model among those investigated, achieving an overall Dice coefficient of 92.12%. Moreover, the time consumption of the system was low at 103.37 milliseconds per frame, facilitating its application in real-time scenarios. Phase-wise analysis indicated that the Phacoemulsification phase (nuclear disassembly) was the most challenging to segment (Dice coefficient of 86.02%).Conclusion: The experimental results showed that the proposed system is highly effective in intraoperative capsulotomy recognition during cataract surgery and demonstrates both high accuracy and real-time capabilities. This system holds significant potential for applications in surgical performance analysis, education, and intraoperative guidance systems.Keywords: cataract surgery, capsulotomy, deep learning, image segmentation, image classification

Keywords