Figures
Abstract
Objective
To test the applicability of deep learning models for detecting and staging rhegmatogenous retinal detachment (RRD) based on morphological features using two- and three-dimensional optical coherence tomography (OCT) scans.
Design
Retrospective study using deep learning-based image classification analysis of 2D and 3D OCT scans combined with clinical baseline data.
Methods
A total of 252 eyes with RRD and 770 control eyes were included. All OCT scans and clinical baseline data were reviewed and graded. Binary and multiclass classification approaches were applied.
Main outcome measures
Area under the curve (AUC) and precision-recall area under the curve (PR AUC) for detection, stage classification and duration estimation of RRD.
Results
We employed both statistical and deep learning-based approaches using 2D and 3D OCT data. We evaluated an automated 3D OCT classification model in a multiclass analysis to distinguish RRD scans by macula status from a non-RRD group with macula-on cases (PR AUC = 0.66 ± 0.12, AUC = 0.96 ± 0.01) vs. macula-off cases (PR AUC = 0.86 ± 0.07, 0.98 ± 0.01) against non-RRD cases (PR AUC = 1.00, AUC = 1.00) Furthermore, the 3D model was able to classify the duration of macula-off status (< 3 days) with a PR AUC of 0.68 ± 0.2 and a AUC of 0.97 ± 0.2 when compared to a mixed group including longer macular-off, macular-on and non RRD cases. Lastly, manually graded RRD Stages were correlated with best corrected visual acuity (BCVA), as well as macula-off Duration and classified via a 2D model. A 2D model used for RRD stage classification achieved its best performance for stage 4, with a PR AUC of 0.56 ± 0.11 and an AUC of 0.94 ± 0.02.
Citation: Beuse A, Lopes IV, Spitzer MS, Druchkiv V, Grohmann C, Skevas C (2025) Deep learning detection of retinal detachment: Optical coherence tomography staging and estimation of duration of macular detachment. PLoS One 20(9): e0329951. https://doi.org/10.1371/journal.pone.0329951
Editor: Koichi Nishitsuka, Saitama Medical Center: Saitama Ika Daigaku Sogo Iryo Center, JAPAN
Received: March 31, 2025; Accepted: July 23, 2025; Published: September 2, 2025
Copyright: © 2025 Beuse et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: The machine learning algorithm, along with example DICOM files, are available in the public Github repository associated with this publication. This GitHub repository can be accessed here: https://link.uke-augenklinik.de/rds. According to local restrictions, the whole dataset cannot be published. In case of further questions or collaboration inquiries with the datasets please contact the corresponding author or the local ethics committee at ethik@aekhh.de.
Funding: The author(s) acknowledge financial support from the Open Access Publication Fund of UKE - Universitätsklinikum Hamburg-Eppendorf.
Competing interests: The authors have declared that no competing interests exist.
Abbreviations: 2D, two-dimensional; 3D, three-dimensional; AUC, area under the curve; BCVA, best corrected visual acuity; CI, confidence interval; MLA, machine learning algorithm; ORC, outer retinal corrugations; PR, precision recall; px, pixel; ResNet, Residual Network; RPE, retinal pigment epithelium; ROC, receiver operating characteristic; RRD, rhegmatogenous retinal detachment; SD, standard deviation
Introduction
Rhegmatogenous retinal detachment (RRD) is a serious and potentially sight-threatening ophthalmological emergency. The visual prognosis of RRD depends on multiple factors, most notably the macular involvement at presentation (macula-on vs. macula-off), the duration of central vision loss and the timing of surgical intervention [1–3]. It is well established that patients with an attached fovea have a more favorable visual prognosis, typically achieving a higher visual acuity than those with foveal involvement, regardless of symptom duration [1].
If left undiagnosed or improperly managed, an RRD can progress to foveal detachment, turning it into a time-critical surgical emergency. The acceptable time until surgical intervention in macula-off RRD remains controversial [2]. While earlier studies suggested that macular detachment leads to irreversible functional damage, potentially reducing the urgency compared to macula-on RRD [4], more recent studies demonstrated that photoreceptor cell death begins within the first 12 hours after the detachment and peaks after 2–3 days. Although the underlying cellular and molecular mechanisms are not yet fully understood, these findings suggest that even macula-off cases may benefit from a prompt surgical intervention [3,5]. A recent meta-analysis further supports this, showing that patients treated within three days achieved better visual outcomes than those repaired between days 4 and 7 [6].
Although the optimal timing for surgery in cases of fovea-involving RRD remains undefined, the duration of the foveal detachment is increasingly recognized as a key factor in determining the urgency of surgical intervention. However, accurately assessing the duration of central vision loss can, for several reasons, be challenging in clinical practice [7,8].
Optical coherence tomography (OCT), routinely performed in cases of RRD at the time of presentation, may provide valuable information that support clinicians in estimating the duration of foveal attachment. Martins Melo et al. proposed an innovative staging system based on the sequential outer retinal changes observed on swept-source OCT (SS-OCT) of patients with RRD [8]. The authors identified six distinct stages, which were further supported by observations from prior histological studies.
The use of machine learning algorithms (MLAs) in ophthalmic imaging has been extensively explored across a broad spectrum of conditions, ranging from common ones like diabetic retinopathy and age-related macular degeneration to rarer pathologies such as central retinal artery or vein occlusion [9–12]. In the context of retinal detachment, initial research has demonstrated the feasibility of MLAs, particularly using fundus photography and ultra-widefield imaging [13,14].
Methods
Data
This single-center retrospective study included 252 eyes from consecutive patients diagnosed with RRD, as well as 770 eyes without relevant pathomorphological changes, who presented to the University Medical Center Hamburg-Eppendorf, Germany, between January 2018 and January 2022. Only primary RRD cases were included. Exclusion criteria comprised any previous retinal defects or detachment, laser therapy or other prior retinal interventions, regardless of the time of occurrence. After preprocessing, 251 RRD scans were included in the MLA analysis.
The study was approved by the local ethics committee (Ethics Committee Hamburg, ID: 101389-BO) and conducted in accordance with the Declaration of Helsinki. All patient data were anonymized prior to analysis and no access to identifiable information was granted after data collection. The complete dataset was accessed and locked for analysis on February 23, 2025.
A retrospective, anonymized dataset from the University Medical Center Hamburg-Eppendorf (Hamburg, Germany) was used for model training, validation and testing. To ensure balanced representation across all categories, the dataset was organized using stratified 5-fold cross-validation based on RRD stages and the non-RRD control group. This allowed for robust evaluation of stage classification performance, while also enabling independent assessment of macular involvement and detachment duration.
The staging system was based on anatomical changes in the outer retina and the photoreceptor layer, following the six-stage classification proposed by Martins Melo et al (8). The stages were defined as follows: Stage 1: Separation of the neurosensory retina from the retinal pigment epithelium (RPE); Stage 2: Thickening of the photoreceptor layer; Stage 3: Formation of outer retinal corrugations (ORC), further divided in Stage 3a: low frequency ORCs and Stage 3b: high frequency ORCs; Stage 4: Progressive loss of ORC definition with concurrent thickening of the photoreceptor layer; Stage 5: Partly (“moth-eaten”) or complete loss of photoreceptors. For the purpose of MLA analysis s stages 4 and 5 were combined.
OCT-Scans obtained using our TOPCON device were carefully assessed by a senior retina specialist. For quality control, we defined an image quality level (IQV) of 60 as the minimum threshold. Scans with an IQV below 60 were excluded, as lower-quality images did not allow a reliable evaluation of the subtle morphological changes necessary to accurately distinguish between RRD stages. Clinical baseline characteristics and an estimation of the onset of foveal detachment, based on patient history as documented in the electronic medical records, were assessed.
The OCT dataset comprised a diverse patient population of both female and male subjects, with ages ranging from 18 to 93 years. Three-dimensional (3D) volume OCT scans, used for analysis of macular status and detachment duration, and corresponding two-dimensional (2D) B-scans, used for stage analysis, were collected using a Spectral-Domain OCT device (Triton OCT, Topcon Corporation, Tokyo, Japan) during clinical visits. Baseline characteristics of both the RRD and non-RRD control groups are summarized in Table 1.
The study included 252 OCT scans from patients with RRD, with a median age of 64 years, predominantly male (66.7%) and a slight majority involving the right eye (54.8%). Most RRD cases were in stages 3b (38%) and 3a (20%). The macula was detached in 75% of cases. The reported onset of central vision loss was used as a surrogate for the duration of foveal detachment. Inclusion criteria required the absence of prior ocular surgery. Only high-quality OCT scans were included, manually selected and then reviewed and pre-classified by an experienced retinal specialist. For quality control, we defined an image quality level (IQV) of 60 as the minimum threshold. The non-RRD control group consisted of 770 OCT scans from eyes without RRD or other relevant pathomorphological changes. For this purpose, our institutional database was screened for patients over 18 years of age with OCT scans without known retinal disease. All selected scans were manually reviewed to confirm the absence of retinal detachment, macular pathology or other structural anomalies.
The non-RRD control group was age-matched to the RRD group, with an age range of 18–92 years (median age 49) and a gender distribution of 47% male and 54% female.
The OCT scans used in this study were acquired with the Topcon DRI Triton Plus device, using a horizontal 3D wide volume scan mode (12 mm x 9 mm) with a resolution of 512 x 992 pixels. Each scan typically consisted of 512 columns, 992 rows and 256 slices. The fixation point was centered on the fovea; if central fixation was not possible, an external fixation light was used for the fellow eye. The average slice thickness was 0.035 mm, with an axial resolution of 0.0026 mm.
In the 2D deep learning model, all scans were interpolated to 496 x 496 pixels. For the 3D model, the volumetric data were resampled to 256 x 256 x 64 voxels, preserving the relative spatial structure across axes.
Model and statistics
We used Python (version 3.9.7, Python Software Foundation, Beaverton, USA) to manage the data infrastructure and perform statistical analyses. The deep learning framework for this project was developed using PyTorch, running on a Quadro RTX 8000 GPU (NVIDIA, Santa Clara, CA, USA) [15].
For statistical analysis of baseline characteristics, we used R (Version 4.4.2.). Incidence of macula-off status and the corresponding confidence intervals were calculated for each OCT stage. These incidences were then compared pairwise using Fisher’s exact test, and the p-values were adjusted using the Holm method. The duration of macula-off status and BCVA were summarized using the median and quartiles. Pairwise comparisons across OCT stages were conducted using the Mann-Whitney U test.
We applied transfer learning by utilizing a pre-trained residual network (ResNet) convolutional neural network (CNN) model for both two- and three-dimensional analyses. For 2D OCT slide classification, a ResNet-18 model comprising 18 layers was used, while a 3D ResNet-18 was implemented for volumetric analysis [16,17]. These models employ shortcut connections to mitigate issues such as vanishing gradients and training divergence.
The tested model architecture begins with a convolutional layer followed by max pooling. This is followed by four convolutional blocks, each containing two residual units with two convolutional layers per unit. Our model then uses an averaging layer for pooling with spatial dimension shrinkage to produce a single fully connected layer. Unlike the 2D model, the 3D model incorporates 3D convolutional layers, enabling the automatic capture and processing of volumetric data with further spatial and temporal dependencies.
For optimal configuration, an Adam algorithm (adaptive moment estimation) was employed with a learning rate of 5x5-5 for weight update optimization [18].
Separate PyTorch data loaders were implemented for training, validation and test sets For the multiclass classification, the algorithm was configured to produce up to six output classes. Each approach had a pre-trained ImageNet configuration before being fine-tuned on the study data [19]. CrossEntropyLoss was used as criterion. For the 2D model, an early stopping method was applied. During training and validation, gradients were processed, and model weights were updated and saved accordingly. We excluded updating the gradient for the test or evaluation phase, for which we selected the best weights from previous steps. The test set consisted of data unseen during training or validation. The hyperparameters used in model training are listed in Table 2. All results were bootstrapped, and 95% confidence intervals calculated and reported.
Multiclass classification comparing non-RRD scans with macula-on and macula-off RRD scans. The PR AUC indicate moderate discriminative performance for macula-on (PR AUC of 0.66) and macula-off (PR AUC of 0.86) cases. ROC analysis shows high classification performance across all groups, with AUC values > 0,95, reflecting high sensitivity and specificity.
Preprocessing
For our 2D approach, data augmentation was applied using the Albumentations library within the PyTorch framework to enhance training variability [20]. The following transformations were utilized during training: random shifts, scaling, rotation and resizing through interpolation. In contrast, validation and test sets were resized without any further augmentation (20). All processed images were normalized and subsequently converted into PyTorch tensors using a channel-first structure. For the 3D tasks, DICOM volumes were preprocessed as stacked tensors via normalization, and interpolated. The 3D-Approach utilized a trilinear calculation. For the three configurations, an early stopping mechanism was implemented using the macro-averaged one-vs-rest average precision metric. Training was halted when the predefined patient threshold of 10 epochs was reached without improvement. The model checkpoint with the best validation score was retained to prevent overfitting and reduce computational effort. The final fully-connected layer is sized to the task’s label count (2, 3 or 6).
Cross validation
To ensure robust model evaluation, the dataset was split and stratified by stage to be then randomly divided into five relatively evenly balanced folds. We employed a two-level cross-validation loop. The outer loop treats one of the five folds as an untouched test set, while the inner loop cycles through the remaining four folds each time using one for validation and the other three for training. For each iteration, one-fold was used as the test set, while the remaining four served for training and validation purposes. Each run through has been used independently for the three different calculations.
Some controls allowed usage of multiple eyes. The dataset was distributed into training, validation and test sets, respectively, with an approximate ratio of 80:20. Class imbalance was addressed by maintaining label proportions across all folds using stratified five-fold-cross-validation. Model selection and performance reporting were based on macro-averaged metrics, especially Average Precision and ROC AUC, applied for both the training/validation folds and the independent test hold out set, respectively). The number of cases and split ratios for training, validation and testing are stated in Table 3. For the first iteration a total of 648 scans were used for training, 176 for validation and 197 for testing.
Results
Data characteristics
This single-center, retrospective study included 252 eyes with RRD and a total of 1,021 OCT scans, with 251 RRD scans mobilized for MLA analysis after preprocessing. The RRD cohort ranged in age from 18 to 93 years, with a mean age of 63.41 years. The majority of participants consists of 67% male and 33% were female. The right eye was affected in 55% of cases and the left eye in 45.2%. Macula-off cases accounted for 75% of the cases, while 25% were classified as macula-on. Regarding the OCT staging, the most frequent stages were 3b (38%) and 3a (20%), with fewer cases in stages 4 (7.5%) and 5 (3%). Best-corrected visual acuity (BCVA) at presentation ranged from 0.00 to 2.70 LogMAR (mean: 1.19) and improved postoperatively to 0.00–2.30 LogMAR (mean: 0.61).
Model performance
Correct classification using the 3D model to distinguish between macula-on, macula-off and non-RRD cases was achieved with a strong performance metrics. For macula-off cases, the model reached a PR AUC of 0.86 ± 0.07 and an AUC of 0.98 ± 0.01. For macula-on cases, it achieved a PR AUC of 0.66 ± 0.12 and an AUC of 0.96 ± 0.01 for macula-on cases. (Fig 1). In the direct automatic 3D classification of macula-off RRD cases with a symptom duration less than 3 days versus those with a duration of more than 3 days, the model demonstrated a (PR AUC of 0.68 ± 0.2, indicating moderate classification performance with variability across thresholds (Fig 2). The corresponding AUC was 0.97 ± 0.20, reflecting strong overall performance in terms of sensitivity and specificity.
PR AUC values of 0.66 for macula-on and 0.86 for macula-off indicate moderate performance with some variability. ROC curves show high classification performance, with AUC values exceeding 0.95 for all classes demonstrating strong sensitivity and specificity.
The PR AUC of 0.68 ± 0.2 indicates moderate performance with some variability. The ROC curve shows strong classification performance, with an AUC of 0.97 ± 0.2, reflecting high sensitivity and specificity.
While the PR curve showed reduced precision at lower recall levels, the AUC curve maintained a consistently upward trajectory, suggesting robust model performance across all decision thresholds.
For the 2D RRD stage classification the AUC illustrate classification performance across six classes, with AUC values ranging from 0.90 to 1.00 (Fig 3). For non-RRD, the PR AUC and AUC is 1.00, resembling strong classification. Stage 1 has an AUC of 0.93, while for Stage 2, 3a, 3b, and 4 + 5, the AUC values remain high (0.90–0.94). In the early stages, performance is more challenging, with Stage 1 reaching an PR AUC of 0.28 ± 0.13 and improves slightly with an PR AUC of 0.22 ± 0.08, while Stage 3b shows the highest PR AUC among the RRD stages, achieving 0.56 ± 0.11. Stage 4 + 5 follows with PR AUC 0.35 ± 0.17, reflecting reasonable classification but some overlap with other stages.
The models shows variability in performance across the six classes. The non-RRD class achieved the highest PR AUC (1.00), while Stage 2 showed the lowest performance (0.56 ± 0.11). The highest PR AUC among the RRD stages was observed for stage 3b (0.56 ± 0.11). AUC values ranged from 0.90 to 1.00 across all classes, reflecting overall high sensitivity and specificity.
Statistics
The analysis of macula-off duration and BCVA (LogMar) in relation to OCT grading reveals that higher OCT grades are significantly associated with increased macula-off duration. In particular, Stages 4 and 5 exhibit the longest median duration (Figs 5 and 7). Statistical comparisons between OCT stages show significant differences, particularly between intermediate stages (e.g., 3b) and advanced stages (e.g., 4).
Two-dimensional GRAD-CAM feature map illustrate the estimated regions used by the model to classify RRD stages. On the left side is the true stage depicted and, on the right, the correctly estimated stage. Panels A-C show original OCT scans labeled as Stage 1, Stage 3 and Stage 4, respectively. Panels D-F display the corresponding correctly predicted scans for the same stages.
The boxplot shows BCVA values, measured in LogMAR, at the first clinical assessment across different manually graded OCT stages of RRD. A trend toward poorer visual acuity (higher LogMAR values) is observed in more advanced stages (3a, 3b, 4, and 5).
The boxplot shows the distribution of BCVA (LogMAR) values, at the time of the last clinical follow-up across the different manually graded OCT stages. Visual outcomes remain worse in higher stages, indicating that initial OCT grading may correlate with long-term visual prognosis.
The boxplot illustrates the duration of macular detachment (in days) across different manually graded OCT stages of RRD. A trend toward longer macula-off duration is observed in more advanced stages, particularly Stage 4 and Stage 5.
Similarly, BCVA (LogMar) at the initial assessment worsens with increasing OCT stage. Early stages (1 and 2) present with lower median LogMar values, while advanced stages (4 and 5) show the higher LogMAR values.
This trend persists at the final assessment (Fig 6), where later stages continue to demonstrate significantly worse visual outcomes. These differences were statistically confirmed using Holm-adjusted p-values, highlighting a robust association between OCT-based RRD staging, macula-off duration and visual function at both timepoints.
Feature map
To enable visual interpretation of the model´s decision-making process, we implemented a 2D GRAD-CAM (Gradient-weighted Class Activation Mapping) module in PyTorch for the OCT stage classification task [21]. For illustration, we randomly selected correctly classified test images from multiple folds and visualized the attention maps after applying the argmax function to the model’s output layer.
Discussion
We present significant findings that validate previous results regarding the relationship between OCT-based grading, macula-off duration and visual acuity [6,8,22]. Figs 5 and 6 illustrate the distribution of BCVA (LogMAR) at baseline and follow-up, respectively. In both a clear decline in visual acuity is observed with increasing OCT stage. Early-stage detachments (OCT stages 1 and 2) are associated with lower LogMAR values, indicative of better visual function, whereasmore advanced stages (4 and 5) show significantly higher LogMAR values, corresponding to severe visual impairment. These differences are statistically significant after correction using the Holm method, supporting prior findings that higher OCT grades correlate strongly with poorer visual outcomes both at presentation and at follow-up.
Fig 7 further illustrates the relationship between macula-off duration and OCT stage. A progressive increase in duration is evident from early (stages 1 and 2) to intermediate (3a, 3b) and more advanced stages (4,5). While median duration remains relatively low in early OCT stages, a marked increase is observed in later stages, suggesting that rather higher-grade detachments are associated with prolonged macula-off duration. These explorative results emphasize the importance of early diagnosis and intervention to prevent further progression and preserve visual function. They also provide further support for the potential use of OCT-based grading as an objective and data-driven tool for assessing disease chronicity.
Building on these associations, we evaluated the performance of MLA for automated classification of macula-on from macula-off RRDs versus non-RRD cases using a 3D model. The PR AUC for this multi-class task was 0.66 ± 0.12 and 0.86 ± 0.07 for macula-on and macula-off cases respectively, indicating substantial classification performance with some variability. AUC scores further underscored the model´s strong performance, with values of AUC of 0.96 ± 0.01 and 0.98 ± 0.01, for macula-on and macula-off, respectively, highlighting high sensitivity and specificity.
These results demonstrate the model´s robust diagnostic capabilities, while also reflecting areas where trade-offs between precision and recall may affect performance.
To further investigate the potential clinical utility of the approach, we tested the MLAs ability to distinguishing macula-off cases with an estimated duration of less than 3 days from a heterogeneous comparison group consisting of control cases, macula-on cases and macula-off cases older than 3 days. Here, the PR AUC was 0.68 ± 0.2, suggesting moderate discriminative performance with notable variability. However, the AUC reached 0.97 ± 0.02, reflecting high overall sensitivity and specificity. These findings highlight the model’s overall discriminative ability while also reflecting the challenges in precisely distinguishing between these clinical subgroups.
Lastly, we evaluated the model´s ability to directly classify OCT-based RRD stages. In this multi-class task, the PR AUC showed considerable variability across classes. The non-RRD group achieved the highest accuracy (PR AUC of 1.00), demonstrating excellent model reliability for healthy scans. The best PR AUC among the RRD stages is observed in Stage 3b (0.56 ± 0.11), followed by Stages 4 + 5 with 0.35 ± 0.17. The PR AUC reveals substantial challenges in early-stage classification with relatively low PR AUCs highlight limitations in the model’s current discriminative performance. This study represents a retrospective and exploratory analysis, and the findings should therefore be interpreted with caution. Larger datasets are essential to improve classification performance and to enable a more robust evaluation of the model’s clinical applicability.
These findings underscore the potential of MLA’s not only to differentiate RRD cases from normal controls, but also to accurately identify macular involvement and disease onset. Such applications could support clinical decision-making by assisting in diagnosis, enabling triage for surgical intervention and optimizing clinical workflows. In particular, the integration of feature maps not only allows for visualization of possible regions of pathomorphological interest, but also aids in assessing the reliability of model predictions (Fig 4).
This study further identifies robust relationships between OCT-based RRD grading, macula-off duration and visual acuity. Our analysis demonstrates clear and statistically significant differences between OCT stages and their respective clinical implications (Figs 5 and 6). After adjusting for multiple comparisons using the Holm method, we observed a pronounced decline in visual acuity (logMAR) from early (stages 1 and 2) to advanced stages (stages 4 and 5). These results highlight the crucial importance of early detection and timely surgical intervention to preserve visual function.
Additionally, Fig 7 illustrates a progressive increase in macula-off duration with higher OCT grades. This further reinforces the importance of prompt diagnosis and treatment to prevent extended macular detachment and associated visual loss. This data-driven approach supports the clinical importance of OCT-based RRD staging in managing and prognosticating patient outcomes. Integrating feature maps enhances visualization by pinpointing regions of pathomorphological interest. This can help to validate model predictions and provide a deeper understanding of RRD progression.
MLAs have been extensively investigated in Ophthalmology for a range of prevalent conditions, including diabetic retinopathy, glaucoma, retinal venous occlusion, as well as for less common entities such as central retinal artery occlusion [9–12]. In the context of retinal detachment, prior studies have primarily focused on detection using fundus photography or ultra-widefield imaging [13,14]. Our study expands upon this foundation by applying pre-trained MLA’s for classifying macular status in RRD, distinguishing RRD stages and identifying macula-off cases with short symptom duration (Figs 1-3.). The models achieved high sensitivity and specificity, particularly in differentiating between RRD stages and detecting early macular involvement.
These findings highlight the practical potential of MLAs in supporting timely diagnosis, facilitating surgical triage, and streamlining clinical workflows, for ultimately improving patient care. Importantly, in regions with limited access to ophthalmologic care, such as rural or underserved areas, such models could provide valuable pre-screening support. Such tools may help prioritize patients based on disease severity and urgency, ensuring that those in need of prompt surgical intervention are identified and referred more efficiently.
While our findings are promising, several limitations must be addressed. A key limitation of this study is the relatively small and homogeneous sample size. Future research should aim to incorporate a more diverse population across age groups, ethnic backgrounds, and clinical presentations to enhance the generalizability and applicability of the results. Furthermore, the dataset was limited to a single OCT device and central, non-widefield scans, which may have restricted the detection of peripheral RRDs, particularly macula-on cases not centrally located. This highlights a structural limitation of relying solely on standard OCT volume scans.
From a methodological perspective, future studies should consider the integration of data from various OCT devices and other ophthalmologic imaging modalities such as fundus photography or ultra-widefield imaging. Multimodal approaches could improve classification performance and enable a more holistic assessment of retinal detachment. While our current binary and multi-class models provide a foundation, expanding to multi-label classification could allow simultaneous recognition of multiple pathological features, leading to a more nuanced clinical insight. Another limitation lies in the restricted spectrum of RRD etiologies included in this study. Future investigations should encompass a wider range of retinal detachments, including tractional or exudative types, to evaluate MLA applicability across diverse clinical scenarios.
Despite these limitations, our study contributes significantly to the current understanding of MLA use in retinal detachment. First, we validate prior findings showing a strong correlation between OCT-based RRD staging and clinical implications, such as macula-off duration and visual acuity outcomes. Second, we demonstrate that MLAs can not only differentiate between various RRD stages, but also detect macula-off cases with symptom duration under three days. These compatibilities highlight the potential of MLAs to assist in early diagnosis, support triage decisions for surgical intervention and ultimately streamline clinical workflows.
References
- 1. Williamson TH, Shunmugam M, Rodrigues I, Dogramaci M, Lee E. Characteristics of rhegmatogenous retinal detachment and their relationship to visual outcome. Eye (Lond). 2013;27(9):1063–9. pmid:23788207
- 2. Yorston D, Donachie PHJ, Laidlaw DA, Steel DH, Sparrow JM, Aylward GW, et al. Factors affecting visual recovery after successful repair of macula-off retinal detachments: findings from a large prospective UK cohort study. Eye (Lond). 2021;35(5):1431–9. pmid:32581389
- 3. Elghawy O, Duong R, Nigussie A, Bogaard JD, Patrie J, Shildkrot Y. Effect of surgical timing in 23-g pars plana vitrectomy for primary repair of macula-off rhegmatogenous retinal detachment, a retrospective study. BMC Ophthalmol. 2022;22(1):136. pmid:35337296
- 4. Kang HK, Luff AJ. Management of retinal detachment: a guide for non-ophthalmologists. BMJ. 2008;336(7655):1235–40. pmid:18511798
- 5. Frings A, Markau N, Katz T, Stemplewitz B, Skevas C, Druchkiv V, et al. Visual recovery after retinal detachment with macula-off: is surgery within the first 72 h better than after?. Br J Ophthalmol. 2016;100(11):1466–9. pmid:26868705
- 6. van Bussel EM, van der Valk R, Bijlsma WR, La Heij EC. Impact of duration of macula-off retinal detachment on visual outcome: a systematic review and meta-analysis of literature. Retina. 2014;34(10):1917–25.
- 7. Xu D, Yee C, Traustason K, Flaxel CJ. Optimal timing for surgical repair of macula-involving rhegmatogenous retinal detachment. Invest Ophthalmol Visual Sci. 2021;62(8):3092.
- 8. Martins Melo I, Bansal A, Naidu S, Oquendo PL, Hamli H, Lee WW, et al. Morphologic stages of rhegmatogenous retinal detachment assessed using swept-source OCT. Ophthalmol Retina. 2023;7(5):398–405. pmid:36464211
- 9. Khan Z, Khan FG, Khan A, Rehman ZU, Shah S, Qummar S, et al. Diabetic retinopathy detection using VGG-NIN a deep learning architecture. IEEE Access. 2021;9:61408–16.
- 10. Rasel RK, Wu F, Chiariglione M, Choi SS, Doble N, Gao XR. Assessing the efficacy of 2D and 3D CNN algorithms in OCT-based glaucoma detection. Sci Rep. 2024;14(1):11758. pmid:38783015
- 11. Beuse A, Wenzel DA, Spitzer MS, Bartz-Schmidt KU, Schultheiss M, Poli S. Automated detection of central retinal artery occlusion using OCT imaging via explainable deep learning. Ophthalmol Sci. 2025;5(2):100630.
- 12. Chen Q, Yu W-H, Lin S, Liu B-S, Wang Y, Wei Q-J, et al. Artificial intelligence can assist with diagnosing retinal vein occlusion. Int J Ophthalmol. 2021;14(12):1895–902. pmid:34926205
- 13. Li Z, Guo C, Nie D, Lin D, Zhu Y, Chen C, et al. Deep learning for detecting retinal detachment and discerning macular status using ultra-widefield fundus images. Commun Biol. 2020;3(1):15. pmid:31925315
- 14. Christ M, Habra O, Monnin K, Vallotton K, Sznitman R, Wolf S. Deep learning-based automated detection of retinal breaks and detachments on fundus photography. Transl Vis Sci Technol. 2024;13(4):1.
- 15. Paszke A, Gross S, Massa F, Lerer A, Bradbury J, Chanan G, et al. Pytorch: an imperative style, high-performance deep learning library. In: Advances in neural information processing systems. 2019;32.
- 16. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit, 2016. 770–8.
- 17. Tran D, Wang H, Torresani L, Ray J, Lecun Y, Paluri M. A closer look at spatiotemporal convolutions for action recognition. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2018. 6450–9.
- 18. Kingma DP, Ba JL. Adam: A method for stochastic optimization. In: 3rd Int Conf Learn Represent ICLR 2015 - Conf Track Proc, 2015. 1–15.
- 19.
Deng J, Dong W, Socher R, Li LJ, Li K, Fei-Fei L. ImageNet: a large-scale hierarchical image database. In: 2009 IEEE Conf Comput Vis Pattern Recognition, CVPR 2009. 2009;248–55.
- 20. Buslaev A, Iglovikov VI, Khvedchenya E, Parinov A, Druzhinin M, Kalinin AA. Albumentations: fast and flexible image augmentations. Information. 2020;11(2):125.
- 21. Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D. Grad-CAM: visual explanations from deep networks via gradient-based localization. In: Proceedings of the IEEE International Conference on Computer Vision, 2017. 618–26.
- 22. Lee CS, Shaver K, Yun SH, Kim D, Wen S, Ghorayeb G. Comparison of the visual outcome between macula-on and macula-off rhegmatogenous retinal detachment based on the duration of macular detachment. BMJ Open Ophthalmol. 2021;6(1):e000615. pmid:33782654