Figures
Abstract
Introduction
The Brazilian Multilabel Ophthalmological Dataset (BRSET) addresses the scarcity of publicly available ophthalmological datasets in Latin America. BRSET comprises 16,266 color fundus retinal photos from 8,524 Brazilian patients, aiming to enhance data representativeness, serving as a research and teaching tool. It contains sociodemographic information, enabling investigations into differential model performance across demographic groups.
Methods
Data from three São Paulo outpatient centers yielded demographic and medical information from electronic records, including nationality, age, sex, clinical history, insulin use, and duration of diabetes diagnosis. A retinal specialist labeled images for anatomical features (optic disc, blood vessels, macula), quality control (focus, illumination, image field, artifacts), and pathologies (e.g., diabetic retinopathy). Diabetic retinopathy was graded using International Clinic Diabetic Retinopathy and Scottish Diabetic Retinopathy Grading. Validation used a ConvNext model trained during 50 epochs using a weighted cross entropy loss to avoid overfitting, with 70% training (20% validation), and 30% testing subsets. Performance metrics included area under the receiver operating curve (AUC) and Macro F1-score. Saliency maps were calculated for interpretability.
Results
BRSET comprises 65.1% Canon CR2 and 34.9% Nikon NF5050 images. 61.8% of the patients are female, and the average age is 57.6 (± 18.26) years. Diabetic retinopathy affected 15.8% of patients, across a spectrum of disease severity. Anatomically, 20.2% showed abnormal optic discs, 4.9% abnormal blood vessels, and 28.8% abnormal macula. A ConvNext V2 model was trained and evaluated BRSET in four prediction tasks: “binary diabetic retinopathy diagnosis (Normal vs Diabetic Retinopathy)” (AUC: 97, F1: 89); “3 class diabetic retinopathy diagnosis (Normal, Proliferative, Non-Proliferative)” (AUC: 97, F1: 82); “diabetes diagnosis” (AUC: 91, F1: 83); “sex classification” (AUC: 87, F1: 70).
Discussion
BRSET is the first multilabel ophthalmological dataset in Brazil and Latin America. It provides an opportunity for investigating model biases by evaluating performance across demographic groups. The model performance of three prediction tasks demonstrates the value of the dataset for external validation and for teaching medical computer vision to learners in Latin America using locally relevant data sources.
Author summary
In low-resource settings, access to open medical datasets is crucial for research. Regions such as Latin America often face underrepresentation, resulting in health biases and inequities. To face the scarcity of diverse ophthalmological datasets in these areas, especially in Brazil and Latin America, we introduce the Brazilian Multilabel Ophthalmological Dataset (BRSET) as a means to alleviate biases in medical AI research. Comprising 16,266 color fundus retinal photos from 8,524 Brazilian patients, BRSET integrates sociodemographic information, empowering researchers to investigate biases across demographic groups and diseases. BRSET was extracted from São Paulo outpatient centers, and includes demographics, clinical history, and retinal images labeled for anatomical features, quality control, and pathologies like diabetic retinopathy. Validation was performed in a set of selected prediction tasks, such as diabetes diagnosis, sex classification, and diabetic retinopathy diagnosis. BRSET’s inclusion of sociodemographic data and experiment metrics underscores its potential efficacy across diverse classification objectives and patient groups, providing crucial insights for medical AI in underrepresented regions.
Citation: Nakayama LF, Restrepo D, Matos J, Ribeiro LZ, Malerbi FK, Celi LA, et al. (2024) BRSET: A Brazilian Multilabel Ophthalmological Dataset of Retina Fundus Photos. PLOS Digit Health 3(7): e0000454. https://doi.org/10.1371/journal.pdig.0000454
Editor: Miguel Ángel Armengol de la Hoz, Fundación Progreso y Salud: Junta de Andalucia Consejeria de Salud y Familias Fundacion Progreso y Salud, SPAIN
Received: January 23, 2024; Accepted: June 3, 2024; Published: July 11, 2024
Copyright: © 2024 Nakayama et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All the codes used in this paper for the dataset setup, data analysis, and experiments are found in a GitHub repository at https://github.com/luisnakayama/BRSET. The BRSET is available in https://physionet.org/content/brazilian-ophthalmological/1.0.0/.
Funding: The author(s) received no specific funding for this work.
Competing interests: The authors have declared that no competing interests exist.
Introduction
In ophthalmological practice, imaging assists in the diagnosis and follow-up of ocular conditions, including retinal fundus photos, ocular anterior segment photos, corneal topography, visual field tests, and optical coherence tomography [1,2]. Artificial intelligence (AI) algorithms can potentially improve medical care by facilitating access to screening, diagnosis, and monitoring in resource-limited settings and assist with the decision-making process [1–3]. In ophthalmology, AI holds promise for ocular diseases such as diabetic retinopathy, age-related macular degeneration, glaucoma, and retinopathy of prematurity [1,2,4–8]. While AI represents a breakthrough technology, unfair outcomes can arise from every step of the AI lifecycle, with models developed using non-representative and local data leading to unequal algorithms [9–12]. Among the sources of biases that can be induced in deep learning, those that stand out the most are those introduced by the data, the model, and the decision-making. A way to tackle biases introduced by data, is providing a representative open source dataset at both the sample size, as well as in knowledge through not only clinical variables but also demographic and sociocultural variables [13].
The open science movement in healthcare has not gained traction in Latin America [14]. In ophthalmology, the majority of the available datasets come from high-income countries, as can be seen in Table 1. In addition, datasets lack demographic and crucial clinical information such as comorbidities [15]. In low and middle-income countries (LMIC), the number of ophthalmologists relative to the population is not adequate [16]. Autonomous systems may increase ophthalmological coverage and reduce preventable blindness; however, datasets that do not adequately represent those who are disproportionately impacted by the disease lead to biased and harmful algorithms. For instance, conditions like diabetic retinopathy and infectious retinopathies have a disproportionately higher prevalence in patients from LMICs, yet these individuals are often underrepresented in ophthalmological datasets [15,17].
The BRSET stands as a pioneering initiative within Latin America’s ophthalmological landscape, offering a dataset that distinguishes itself by its multifaceted approach. While previous datasets have existed within the region, BRSET elevates the standard by encompassing a diverse spectrum of diagnoses and sociodemographic categories. In essence, it marks a significant advancement as the inaugural repository to comprehensively integrate multiple diagnoses alongside sociodemographic parameters.
Materials and methods
This study was approved by the Sao Paulo Federal University (UNIFESP) IRB (CAAE 33842220.7.0000.5505), and informed consent was waived due to efforts to ensure data anonymity, low risk of reidentification, and interruption of data sharing in any reidentification risk. Before accessing the dataset, the users must be credentialed, accept the Health Data use agreement, and complete training on data research. In this dataset, identifiable patient information was manually reviewed and removed from all images.
Data sources
We included data from three outpatient Brazilian ophthalmological centers in São Paulo evaluated from 2010 to 2020 and from the Sao Paulo Federal University ophthalmology sector.
Data collection
Images present in this dataset were collected using different retinal fundus cameras, including Nikon NF505 (Nikon, Tokyo, Japan), Canon CR-2 (Canon Inc, Melville, NY, USA) retinal camera, and Phelcom Eyer (Phelcom Technologies, MA, USA). Retinal photos were taken by previously trained non-medical professionals in pharmacological mydriasis.
Dataset preparation
The file identification was removed from all fundus photos, as well as sensitive data (e.g., patient name, exam date). Every image was reviewed to ensure the absence of protected health information in images. The images were exported directly from retinal cameras in JPEG format, and no preprocessing techniques were performed. The image viewpoint can be macula-centered or optic disc-centered. The dataset does not include fluorescein angiogram photos, non-retinal images, or duplicated images.
Metadata
Each retinal image is labeled with the retinal camera device, image center position, patient nationality, age in years, sex, comorbidities, insulin use, and duration of diabetes diagnosis. The demographics and medical features were collected from the electronic medical records.
Labeling
A retinal specialist ophthalmologist labeled all the images based on criteria that were established by the research group [31]. The following characteristics were labeled:
- Anatomic classification: The retinal optic disc (vertical cup-disc ratio of ≥0.65 [32]), retinal vessels (tortuosity and width), and macula (abnormal findings) were categorized as either normal or abnormal.
- Quality control parameters: Parameters including image focus, illumination, image field, and artifacts, were assessed and classified as satisfactory or unsatisfactory. The criteria are defined in Table 2.
- Pathological classifications: The images were classified according to the pathological classification list: diabetic retinopathy, diabetic macular edema, scar (toxoplasmosis), nevus, age-related macular degeneration (AMD), vascular occlusion, hypertensive retinopathy, drusens, nondiabetic retinal hemorrhage, retinal detachment, myopic fundus, increased cup disc ratio, other.
- Diabetic retinopathy classification: Diabetic retinopathy was classified using the International Clinic Diabetic Retinopathy (ICDR) grading and Scottish Diabetic Retinopathy Grading (SDRG), as can be seen in Table 3.
Data records
This dataset may be used to build computer vision models that predict demographic characteristics and multi-label disease classification. BRSET consists of 16,266 images from 8,524 Brazilian patients, and a metadata file. Columns are detailed in S1 File.
Data storage
The dataset images and labels are stored on the PhysioNet repository entitled “A Brazilian Multilabel Ophthalmological Dataset (BRSET)”. Link: https://physionet.org/content/brazilian-ophthalmological/1.0.0/.
Descriptive analytics and technical validation
Descriptive analysis.
The BRSET database contains 10,592 (65.1%) images taken from a Canon CR2 retinal camera, and 5,674 (34.9%) are from the Nikon NF5050 retinal camera. The sex distributions in the dataset are 6,214 (38.2%) male patients, and 10,052 (61.8%) female patients. The average age is 57.6 years (standard deviation of 18.26 years).
The database includes 2,579 (15.8%) patients with a diagnosis of diabetes mellitus. Among those patients 1,922 (74.5%) do not have retinopathy, 107 (4.1%) have mild non-proliferative retinopathy, 181 (7%) have moderate non-proliferative retinopathy, 127 (4.9%) have severe non-proliferative retinopathy, and 242 (9.4%) have proliferative retinopathy. A sample of the images can be seen in Fig 1.
Among the anatomical criteria, 3,281 (20.2%) of the images have an abnormal optic disc, 807 (4.9%) have abnormal blood vessels, and 4,685 (28.8%) have an abnormal macula. The distribution of normal exams and pathological findings is described in Table 4.
Data quality assessment
Image quality.
In terms of quality criteria, 542 (3.3%) images were classified as having inadequate focus, 84 (0.5%) as having inadequate lighting, 1401 (8.6%) as having an inadequate field, and 57 (0.3%) as having artifacts.
Quality metadata.
On the other hand, regarding the quality characteristics of the metadata, duplicate data and missing data were assessed. The dataset does not have duplicate data, and only 3 columns contain missing data: patient age column (33.47%), diabetes duration (88.26%), and insulin usage (89.46%). Notably, the latter two are exclusively recorded for diabetic patients.
Experiments.
To explore the dataset, we provide use cases and benchmarks for future research. The codes required to run these experiments are available in a publicly available GitHub repository (see Code availability section).
In this section, we delineate the methodologies employed to validate the dataset proposed in our study. Our primary focus will be harnessing the application of ConvNext V2 as a deep learning method on the BRSET dataset for retina fundus photo analysis [35]. This investigation encompasses binary diabetic retinopathy classification, three-class diabetic retinopathy state classification (normal, non-proliferative, and proliferative retinopathy), diabetes mellitus prediction, and sex prediction.
As diabetic retinopathy is the leading cause of preventable blindness among adults and the most explored disease in publicly available datasets [15,17,36], we included diabetic retinopathy prediction and clinical diagnosis of diabetes mellitus through retina fundus photos.
To assess the ability to predict patient demographic characteristics through retinal fundus photos, extending beyond pathological findings, we included the prediction of patient sex, which is reported in previous studies [37,38].
Model selection.
We opted for ConvNext V2, diverging from the conventional use of ResNet-50 used in prior research on the BRSET dataset [39–41]. The reason for using ConvNext V2 instead of ResNet-50, is that ConvNext V2 is an evolution of the ResNet-50 architecture, and introduces several enhancements over its predecessors, such as layer normalization, expanded kernel sizes, or regularization techniques [35,42]. The architecture is characterized by its depth-wise convolutional layers, which facilitate efficient feature extraction and scalability.
Data preprocessing.
The BRSET dataset was stratified based on labels and divided into training (70%), validation (within training set: 20%), and testing sets (30%) using all the images available in the dataset. This stratification ensures a representative distribution of classes across each subset.
The images in the dataset present variations in the height, ranging from 874 to 2304 pixels; and the width, ranging from 951 to 2984 pixels. To ensure consistency with the model’s input all the images were first resized to 256x256 from the original size, and randomly cropped to ensure a 224x224 shape during training. Also, all the images were normalized by dividing them into 225 to ensure all pixel values were between 0 and 1.
Training strategy.
The model was trained using the full training set during 50 epochs incorporating an early stopping mechanism, keyed to F1 score improvements on the validation set, with a patience parameter set to 7 epochs. This implies that if no performance increase is observed over seven consecutive epochs, training ceases, reverting to the last optimal model state.
Loss function and evaluation metrics.
All the models were trained for a classification task using a cross-entropy loss. Given the class imbalance inherent in the dataset, a weighted cross-entropy loss [43] function was employed using the formula in the Eq (1) [44], where c represents the number of classes, n represents the batch size, wc represents the weights of class c, and y and y hat represent the ground truth and predicted value respectively.
Given that our predictive task is a classification task, the output of the model was defined using the number of classes as output neurons and using softmax activation function to return a set of probabilities for each class. The probability of each class was defined by the Eq (2) [45]; where z, is the logit corresponding to the class c of the i-th element in the batch. and k corresponds to the number of possible classes.
Combining Eqs (1) and (2), we can see that the loss function of the models is given by Eq (3).
The models were trained using class weights to tackle class imbalance during training. The weights were calculated using Eq (4), where w are the weights of class C, N is the total samples in the training dataset, K is the number of classes, and Nc is the samples of class in the training dataset.
Model performance was assessed using the Area Under the Receiver Operating Characteristic Curve (AUC-ROC), a performance metric used to evaluate the quality of predictions, extended to multi-class classification through macro-average.
The F1 score metric was also used to mitigate bias due to class imbalance. The F1 score can be seen in Eq (7) and is a harmonic mean of precision (5) and recall (6) [46]. For the multiclass problems, the macro average was applied to the F1 score to avoid biased results.
Optimization and interpretability.
Adam optimizer [47] was utilized with a learning rate of 1e-5 for diabetes prediction, binary diabetic retinopathy, and 3-class diabetic retinopathy; and 5e-6 was used for sex prediction, over a maximum of 50 epochs or until early stopping was triggered. To enhance model clinical interpretability, we generated saliency maps for select images across classes [48]. Saliency maps highlight regions within an image most influential to the model’s decision, computed by taking the gradient of the output with respect to the input image. Saliency maps are calculated using Eq 8 [48], where S(x) denotes the saliency map, Y is the model output, and X is the input image. This provides insights into model focus areas and potential biases.
Results
The efficacy of ConvNext V2 was benchmarked against state-of-the-art (SOTA) models documented in previous studies utilizing the same dataset. This comparison extends to saliency maps, offering a visual and quantitative analysis of model attention in diagnosing diabetic retinopathy and determining sex, underscoring the model’s practical applicability in real-world settings. (Table 5)
The tasks are binary diabetic retinopathy classification, three-class diabetic retinopathy state classification (normal, non-proliferative, and proliferative), diabetes mellitus prediction, and sex prediction.
The task of binary diabetic retinopathy classification achieved an AUC of 0.97 and a Macro F1-score of 0.89 with ConvNext V2, matching the AUC but significantly surpassing the F1-score of ResNet 50. This improvement in the F1-score, a measure sensitive to class imbalance, underscores the effectiveness of ConvNext V2 in balancing sensitivity and specificity, a crucial aspect in clinical diagnostics.
The three-class diabetic retinopathy diagnosis further emphasizes the model’s robustness, with ConvNext V2 achieving an AUC of 0.97 and a Macro F1-score of 0.82, establishing comparable results with the state-of-the-art (SOTA) results for the BRSET dataset.
The classification of diabetes mellitus, a more challenging task due to the subtler manifestation of the disease in retina images, saw a commendable performance with an AUC of 0.87 and a Macro F1-score of 0.70. This underscores the ConvNext V2’s capability to extract nuanced features indicative of early or less pronounced disease states, which are critical for early intervention and management.
In the sex classification task, the model achieved an AUC of 0.91 and a Macro F1-score of 0.83, demonstrating the model’s ability to discern patterns beyond pathological features, potentially leveraging anatomical and physiological differences captured in retina images.
Saliency maps have emerged as a pivotal tool in interpreting deep learning models, particularly in medical imaging, by highlighting regions within an image that significantly influence the model’s decision. The maps in Fig 2 and Fig 3 reveal that, for tasks such as sex prediction, ConvNext V2 tends to focus on more global features of the retina, whereas, for diabetic retinopathy classification, the model’s attention is drawn to more localized features indicative of the disease, such as microaneurysms, hemorrhages, and retinal exudates.
For sex prediction, the model assesses more global retinal features.
For diabetic retinopathy, localized regions indicative of pathology are highlighted.
Discussion
The Brazilian Multilabel Ophthalmological Dataset (BRSET) represents a significant stride towards mitigating biases in healthcare research. The inclusion of a comprehensive dataset encompassing of Brazilian patients, coupled with extensive metadata, provides a robust foundation for enhancing data representativeness in medical AI applications. This dataset not only facilitates the exploration of models by evaluating performance across diverse demographic groups but also serves as a valuable resource for teaching medical computer vision in regions where relevant local data is scarce.
The utility of the BRSET extends beyond mere data representation; it serves as a critical tool in addressing biases prevalent in healthcare AI, particularly in underrepresented populations. The integration of sociodemographic information enables a nuanced investigation into differential model performance, offering insights into potential biases and disparities. This aspect is crucial for Latin America and similar regions, where healthcare disparities are pronounced, and data representation in medical research is limited [14].
The employment of advanced deep learning techniques, such as ConvNext V2, further underscores the dataset’s utility [35] and has yielded notable advancements in the field of retina fundus photo analysis. The results showcase the superiority of the architecture over the previously established benchmarks, primarily those set by ResNet 50 in the classification tasks of binary diabetic retinopathy and three-class diabetic retinopathy states [39,41].
Notably, the ConvNext V2 model’s ability to discern subtle patterns in retinal images, as evidenced by its commendable performance in diabetes mellitus classification and sex prediction, highlights the dataset’s versatility, with results comparable to previous studies [37,38].
Interpretability methods, such as saliency maps, play a pivotal role in demystifying the decision-making process of AI models for clinical practitioners. In our study, saliency maps revealed the model’s focus areas, distinguishing between localized pathological features for diabetic retinopathy and more global retinal features for sex prediction, corroborating the human clinical classification. This level of interpretability is vital for clinical acceptance, allowing healthcare professionals to understand and trust AI-assisted diagnostics.
Usage notes
BRSET is available on PhysioNet as a database that requires user credentialing prior to access. Users must be registered on PhysioNet, have proper human subject research training, and sign a data use agreement that forbids re-identification of patients and sharing it with those who are not credentialed.
Limitations
Our dataset only includes one nationality and represents general ophthalmological outpatients. As a result, the disease distribution is imbalanced, with a high percentage of normal and mild cases (Table 4). Those limitations may restrict the generalizability of findings to broader Latin American populations. This limitation underscores the necessity for future research to expand the dataset, incorporating data from various Latin American countries to enhance its representativeness and utility.
A single expert evaluation risks introducing bias and undermining the credibility of labels. To enhance the robustness of the BRSET, our future steps target employing two ophthalmologists for labeling, supplemented by adjudication from either a retina specialist or a glaucoma specialist.
This project sought to address the issue of underrepresentation in ophthalmological data from Brazil and Latin America. However, biases in AI are complex and multifaceted. While improving data representativeness is a crucial step, it is only one aspect among many that needs consideration.
Our technical validation results are biased towards the distribution of age, sex, race, among many others in our dataset. Future work should analyze and address different types of social biases included in our dataset and during model development as well as the use of models, methods such as policy loss function, and model deployment. We highly encourage the community to expand our analysis over our dataset and models to detect and reduce other sources of biases.
Strengths and conclusions
The BRSET is the first multilabel ophthalmological dataset from Brazil and Latin America. BRSET aims to improve data representativeness and create the framework for ophthalmological dataset development. To the best of our knowledge, BRSET is the only publicly available retinal image dataset that also contains sociodemographic data such as sex and age in Latin America, which allows us to investigate algorithmic bias across different demographic groups.
The superior performance of ConvNext V2, demonstrated through enhanced AUC and Macro F1-scores across multiple classification tasks on the BRSET dataset, underscores its suitability for clinical applications. The model not only sets new benchmarks but also provides a deeper understanding of disease markers through interpretability tools like saliency maps, bridging the gap between AI models and clinical applicability. These insights can guide clinicians in decision-making and foster trust in AI-assisted diagnostics, marking a significant leap forward in the integration of deep learning technologies in healthcare.
Supporting information
S1 File. Data Dictionary–description of BRSET columns.
https://doi.org/10.1371/journal.pdig.0000454.s001
(DOCX)
References
- 1. Kras A, Celi LA, Miller JB. Accelerating ophthalmic artificial intelligence research: the role of an open access data repository. Curr Opin Ophthalmol. 2020;31: 337–350. pmid:32740059
- 2. Ting DSW, Pasquale LR, Peng L, Campbell JP, Lee AY, Raman R, et al. Artificial intelligence and deep learning in ophthalmology. Br J Ophthalmol. 2019;103: 167–175. pmid:30361278
- 3. Rudnisky CJ, Tennant MTS, Weis E, Ting A, Hinz BJ, Greve MDJ. Web-based grading of compressed stereoscopic digital photography versus standard slide film photography for the diagnosis of diabetic retinopathy. Ophthalmology. 2007;114: 1748–1754. pmid:17368543
- 4. Md A, Abramoff , Lavin PT, Birch M, Shah N, Folk JC. Pivotal trial of an autonomous AI-based diagnostic system for detection of diabetic retinopathy in primary care offices. Yearbook of Paediatric Endocrinology. 2019.
- 5. Bhaskaranand M, Ramachandra C, Bhat S, Cuadros J, Nittala MG, Sadda SR, et al. The value of automated diabetic retinopathy screening with the EyeArt system: A study of more than 100,000 consecutive encounters from people with diabetes. Diabetes Technol Ther. 2019;21: 635–643. pmid:31335200
- 6. Nakayama LF, Mitchell WG, Ribeiro LZ, Dychiao RG, Phanphruk W, Celi LA, et al. Fairness and generalisability in deep learning of retinopathy of prematurity screening algorithms: a literature review. BMJ Open Ophthalmol. 2023;8. pmid:37558406
- 7. Zhang G, Fu DJ, Liefers B, Faes L, Glinton S, Wagner S, et al. Clinically relevant deep learning for detection and quantification of geographic atrophy from optical coherence tomography: a model development and external validation study. Lancet Digit Health. 2021;3: e665–e675. pmid:34509423
- 8. Balaskas K, Glinton S, Keenan TDL, Faes L, Liefers B, Zhang G, et al. Prediction of visual function from automatically quantified optical coherence tomography biomarkers in patients with geographic atrophy using machine learning. Sci Rep. 2022;12: 15565. pmid:36114218
- 9. Nakayama LF, Matos J, Quion J, Novaes F, Mitchell WG, Mwavu R, et al. Unmasking biases and navigating pitfalls in the ophthalmic Artificial Intelligence lifecycle: A review. 2023.
- 10. Charpignon M-L, Byers J, Cabral S, Celi LA, Fernandes C, Gallifant J, et al. Critical Bias in Critical Care Devices. Crit Care Clin. 2023;39: 795–813. pmid:37704341
- 11. Gichoya JW, Thomas K, Celi LA, Safdar N, Banerjee I, Banja JD, et al. AI pitfalls and what not to do: Mitigating bias in AI. Br J Radiol. 2023;96: 20230023. pmid:37698583
- 12. Youssef A, Pencina M, Thakur A, Zhu T, Clifton D, Shah NH. External validation of AI models in health should be replaced with recurring local validation. Nat Med. 2023;29: 2686–2687. pmid:37853136
- 13.
Suresh H, Guttag J. A Framework for Understanding Sources of Harm throughout the Machine Learning Life Cycle. Proceedings of the 1st ACM Conference on Equity and Access in Algorithms, Mechanisms, and Optimization. New York, NY, USA: Association for Computing Machinery; 2021. pp. 1–9. https://doi.org/10.1145/3465416.3483305
- 14. Restrepo D, Quion J, Vásquez-Venegas C, Villanueva C, Anthony Celi L, Nakayama LF. A scoping review of the landscape of health-related open datasets in Latin America. PLOS Digit Health. 2023;2: e0000368. pmid:37878549
- 15. Khan SM, Liu X, Nath S, Korot E, Faes L, Wagner SK, et al. A global review of publicly available datasets for ophthalmological imaging: barriers to access, usability, and generalisability. The Lancet Digital Health. 2020. pmid:33735069
- 16. Resnikoff S, Lansingh VC, Washburn L, Felch W, Gauthier T-M, Taylor HR, et al. Estimated number of ophthalmologists worldwide (International Council of Ophthalmology update): will we meet the needs? Br J Ophthalmol. 2020;104: 588–592. pmid:31266774
- 17. Sun H, Saeedi P, Karuranga S, Pinkepank M, Ogurtsova K, Duncan BB, et al. IDF Diabetes Atlas: Global, regional and country-level diabetes prevalence estimates for 2021 and projections for 2045. Diabetes Res Clin Pract. 2022;183: 109119. pmid:34879977
- 18. Cuadros J, Bresnick G. EyePACS: an adaptable telemedicine system for diabetic retinopathy screening. J Diabetes Sci Technol. 2009;3: 509–516. pmid:20144289
- 19. Takahashi H, Tampo H, Arai Y, Inoue Y, Kawashima H. Applying artificial intelligence to disease staging: Deep learning for improved staging of diabetic retinopathy. PLoS One. 2017;12: e0179790. pmid:28640840
- 20. APTOS 2019 Blindness Detection. [cited 18 Feb 2024]. Available: https://www.kaggle.com/c/aptos2019-blindness-detection
- 21. Islam MT, Imran SA, Arefeen A, Hasan M, Shahnaz C. Source and Camera Independent Ophthalmic Disease Recognition from Fundus Image Using Neural Network. 2019 IEEE International Conference on Signal Processing, Information, Communication & Systems (SPICSCON). 2019.
- 22. Pachade S, Porwal P, Thulkar D, Kokare M, Deshmukh G, Sahasrabuddhe V, et al. Retinal Fundus Multi-Disease Image Dataset (RFMiD): A Dataset for Multi-Disease Detection Research. Brown Univ Dig Addict Theory Appl. 2021;6: 14.
- 23. Decencière E, Zhang X, Cazuguel G, Lay B, Cochener B, Trone C, et al. Feedback on a publicly distributed image database: The Messidor database. Image Anal Stereol. 2014;33: 231.
- 24. Pires R, Jelinek HF, Wainer J, Valle E, Rocha A. Advancing bag-of-visual-words representations for lesion classification in retinal images. PLoS One. 2014;9: e96814. pmid:24886780
- 25. Data sets. In: Rotterdam Ophthalmic Data Repository [Internet]. [cited 17 Feb 2024]. Available: http://www.rodrep.com/data-sets.html
- 26. Porwal P, Pachade S, Kamble R, Kokare M, Deshmukh G, Sahasrabuddhe V, et al. Indian Diabetic Retinopathy Image Dataset (IDRiD): A Database for Diabetic Retinopathy Screening Research. Brown Univ Dig Addict Theory Appl. 2018;3: 25.
- 27. Kauppi T, Kalesnykiene V, Kamarainen J-K, Lensu L, Sorri I, Uusitalo H, et al. DIARETDB0: Evaluation database and methodology for diabetic retinopathy algorithms. Machine Vision and Pattern Recognition Research Group, Lappeenranta University of Technology, Finland.
- 28. Nguyen AH. DIARETDB1—Standard Diabetic Retinopathy Database. 2021. Available: https://www.kaggle.com/datasets/nguyenhung1903/diaretdb1-standard-diabetic-retinopathy-database
- 29. Decencière E, Cazuguel G, Zhang X, Thibault G, Klein J-C, Meyer F, et al. TeleOphta: Machine learning and image processing methods for teleophthalmology. IRBM. 2013;34: 196–203.
- 30. Giancardo L, Meriaudeau F, Karnowski TP, Li Y, Garg S, Tobin KW Jr, et al. Exudate-based diabetic macular edema detection in fundus images using publicly available datasets. Med Image Anal. 2012;16: 216–226. pmid:21865074
- 31. Nakayama LF, Gonçalves MB, Ribeiro LZ, Malerbi FK, Regatieri CVS. Diabetic Retinopathy Labeling Protocol for the Brazilian Multilabel Ophthalmological Dataset. 2023.
- 32. Tan R, Teo KYC, Husain R, Tan NC, Lee QX, Hamzah H, et al. Evaluating the outcome of screening for glaucoma using colour fundus photography-based referral criteria in a teleophthalmology screening programme for diabetic retinopathy. Br J Ophthalmol. 2023. pmid:37852739
- 33. Wilkinson CP, Ferris FL 3rd, Klein RE, Lee PP, Agardh CD, Davis M, et al. Proposed international clinical diabetic retinopathy and diabetic macular edema disease severity scales. Ophthalmology. 2003;110: 1677–1682. pmid:13129861
- 34. Zachariah S, Wykes W, Yorston D. Grading diabetic retinopathy (DR) using the Scottish grading protocol. Community Eye Health. 2015;28: 72–73. Available: https://www.ncbi.nlm.nih.gov/pubmed/27418727 pmid:27418727
- 35.
Woo S, Debnath S, Hu R, Chen X, Liu Z, Kweon IS, et al. ConvNeXt V2: Co-designing and Scaling ConvNets with Masked Autoencoders. 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE; 2023. https://doi.org/10.1109/cvpr52729.2023.01548
- 36. Cheung N, Wong TY. Diabetic retinopathy and systemic vascular complications. Prog Retin Eye Res. 2008;27: 161–176. pmid:18249026
- 37. Korot E, Pontikos N, Liu X, Wagner SK, Faes L, Huemer J, et al. Predicting sex from retinal fundus photographs using automated deep learning. Sci Rep. 2021;11: 10286. pmid:33986429
- 38. Poplin R, Varadarajan AV, Blumer K, Liu Y, McConnell MV, Corrado GS, et al. Prediction of cardiovascular risk factors from retinal fundus photographs via deep learning. Nat Biomed Eng. 2018;2: 158–164. pmid:31015713
- 39.
Wu C, Yang X, Gilkes EG, Cui H, Choi J, Sun N, et al. De-identification and Obfuscation of Gender Attributes from Retinal Scans. Clinical Image-Based Procedures, Fairness of AI in Medical Imaging, and Ethical and Philosophical Issues in Medical Imaging. Springer Nature Switzerland; 2023. pp. 91–101. https://doi.org/10.1007/978-3-031-45249-9_9
- 40. Nakayama LF, Choi J, Cui H, Gilkes EG, Wu C, Yang X, et al. Pixel Snow and Differential Privacy in Retinal fundus photos de-identification. Invest Ophthalmol Vis Sci. 2023;64: 2399–2399. Available: https://iovs.arvojournals.org/article.aspx?articleid=2787142
- 41.
Gould DSW, Yang J, Clifton DA. Deep Learning for Multi-Label Disease Classification of Retinal Images: Insights from Brazilian Data for AI Development in Lower-Middle Income Countries. medRxiv. 2024. p. 2024.02.12.24302676. https://doi.org/10.1101/2024.02.12.24302676
- 42.
He K, Zhang X, Ren S, Sun J. Deep Residual Learning for Image Recognition. arXiv [cs.CV]. 2015. Available: http://arxiv.org/abs/1512.03385
- 43.
Zhang Z, Sabuncu MR. Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels. arXiv [cs.LG]. 2018. Available: http://arxiv.org/abs/1805.07836
- 44. Shannon CE. A mathematical theory of communication. The Bell System Technical Journal. 1948;27: 379–423.
- 45.
Bridle JS. Training stochastic model recognition algorithms as networks can lead to maximum mutual information estimation of parameters. Proceedings of the 2nd International Conference on Neural Information Processing Systems. Cambridge, MA, USA: MIT Press; 1989. pp. 211–217.
- 46. Van Rijsbergen CJ. Information Retrieval. Butterworths; 1979. Available: https://play.google.com/store/books/details?id=t-pTAAAAMAAJ
- 47.
Kingma DP, Ba J. Adam: A Method for Stochastic Optimization. arXiv [cs.LG]. 2014. Available: http://arxiv.org/abs/1412.6980
- 48.
Simonyan K, Vedaldi A, Zisserman A. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps. arXiv [cs.CV]. 2013. Available: http://arxiv.org/abs/1312.6034