
Code-Free Models Effective at Classifying Retinal Pathologies
Models that are code-free were able to perform as accurately as expert-designed bespoke models when classifying retinal pathologies.
A study published in
AI has been developed for ophthalmology by using imaging data in the past, with AI capable of
This study aimed to compare the performance of multiclass CFDL models compared with bespoke models when it came to diagnosing retinal pathologies based on optical coherence tomography (OCT) images.
The comparative analysis used an OCT video dataset to test the capabilities of DL models created by 2 groups when designed to diagnose. The first model was created by an ophthalmology trainee who did not have coding experience, and the second was created by an expert in AI. Both models were tasked with determining if an eye was normal/healthy retina or had 1 of macular hole (MH), epiretinal membrane (ERM), wet age-related macular degeneration (AMD), and diabetic macular edema (DME) based on an OCT image. The models were trained using videos collected from the Maisonneuve-Rosemont Hospital in Montreal. All OCTs were taken by an experienced OCT technician.
The CFDL was developed in Google Cloud AutoML Video Intelligence and Google Cloud AutoML Vision. All videos and images were uploaded into the Cloud bucket for evaluation. A postdoctoral researcher with experience in DL was the developer of the bespoke model.
There were 1173 videos in the overall dataset used to train and validate the 2 models. Overall, the CFDL video model had an area under the precision-recall curve (AUPRC) of 0.984, which was 94.1% precision, 94.1% recall, and 94.1% calculated accuracy at the 0.5 confidence threshold cutoff. The AUPRC was lowest for ERM at 0.954 and highest for MH and AMD at 1.0. Normal retina videos were also classified correctly 96% of the time, with the lowest accuracy coming from classifying ERM at 90%.
When compared with the best model developed by the expert, the CFDL model had a higher AUPRC value (0.984 vs 0.9097). The AI model was better in precision and sensitivity but only by 0.0009 in both categories; accuracy was higher in the AI model by a more significant amount (0.941 vs 0.9667).
The image model for CFDL was higher than the video model, as it had an AUPRC of 0.990 with a sensitivity and precision of 96.6%; the overall accuracy was 95.8%. The performance was the same across all metrics between the CFDL image model and the best bespoke algorithm.
There were some limitations to this study. The models were trained on a single dataset, which does not take into account datasets of different sizes or complexities. External validation was not included as a part of this study. The OCT images were not focused on the main pathological findings but were instead centered on the fovea. This could have affected the performance of the model.
The researchers concluded that AI developed by those who were not experts in AI tools were just as effective as those developed by experts. CFDL being an effective model for diagnosing could make AI more accessible for ophthalmologists and in other fields of medicine.
References
- Touma S, Hammou BA, Antaki F, Boucher MC, Duval R. Comparing code-free deep learning models to expert-designed models for detecting retinal diseases from optical coherence tomography. Int J Retina Vitreous. 2024;10:37. doi:10.1186/s40942-024-00555-3
- Bonavitacola J. AI effective in screening for referable diabetic retinopathy. AJMC. September 27, 2023. Accessed April 29, 2024. https://www.ajmc.com/view/ai-effective-in-screening-for-referable-diabetic-retinopathy
Newsletter
Stay ahead of policy, cost, and value—subscribe to AJMC for expert insights at the intersection of clinical care and health economics.