Skip to main content
Advertisement
  • Loading metrics

Detection of trachoma using machine learning approaches

  • Damien Socia,

    Roles Investigation, Methodology, Software, Validation, Visualization, Writing – review & editing

    Affiliation Division of Surgical Research, Department of Surgery, Larner College of Medicine, University of Vermont, Burlington, Vermont, United States of America

  • Christopher J. Brady,

    Roles Conceptualization, Methodology, Resources, Supervision, Writing – original draft, Writing – review & editing

    Affiliations Division of Surgical Research, Department of Surgery, Larner College of Medicine, University of Vermont, Burlington, Vermont, United States of America, Division of Ophthalmology, Department of Surgery, Larner College of Medicine, University of Vermont, Burlington, Vermont, United States of America

  • Sheila K. West,

    Roles Conceptualization, Methodology, Writing – review & editing

    Affiliation Dana Center for Preventive Ophthalmology, Wilmer Eye Institute, Baltimore, Maryland, United States of America

  • R. Chase Cockrell

    Roles Conceptualization, Formal analysis, Methodology, Project administration, Supervision, Writing – original draft, Writing – review & editing

    robert.cockrell@med.uvm.edu

    Affiliation Division of Surgical Research, Department of Surgery, Larner College of Medicine, University of Vermont, Burlington, Vermont, United States of America

Abstract

Background

Though significant progress in disease elimination has been made over the past decades, trachoma is the leading infectious cause of blindness globally. Further efforts in trachoma elimination are paradoxically being limited by the relative rarity of the disease, which makes clinical training for monitoring surveys difficult. In this work, we evaluate the plausibility of an Artificial Intelligence model to augment or replace human image graders in the evaluation/diagnosis of trachomatous inflammation—follicular (TF).

Methods

We utilized a dataset consisting of 2300 images with a 5% positivity rate for TF. We developed classifiers by implementing two state-of-the-art Convolutional Neural Network architectures, ResNet101 and VGG16, and applying a suite of data augmentation/oversampling techniques to the positive images. We then augmented our data set with additional images from independent research groups and evaluated performance.

Results

Models performed well in minimizing the number of false negatives, given the constraint of the low numbers of images in which TF was present. The best performing models achieved a sensitivity of 95% and positive predictive value of 50–70% while reducing the number images requiring skilled grading by 66–75%. Basic oversampling and data augmentation techniques were most successful at improving model performance, while techniques that are grounded in clinical experience, such as highlighting follicles, were less successful.

Discussion

The developed models perform well and significantly reduce the burden on graders by minimizing the number of false negative identifications. Further improvements in model skill will benefit from data sets with more TF as well as a range in image quality and image capture techniques used. While these models approach/meet the community-accepted standard for skilled field graders (i.e., Cohen’s Kappa >0.7), they are insufficient to be deployed independently/clinically at this time; rather, they can be utilized to significantly reduce the burden on skilled image graders.

Author summary

Trachoma is an infectious disease, experienced primarily in the developing world, and is a leading cause of global blindness. As recent efforts to address the disease have led to a significant reduction in disease prevalence, it has become difficult to train health workers to detect trachoma, due to its rarity; this if often referred to as the “last mile” problem. To address this issue, we have implemented a convolutional neural network to detect the presence of TF in images of everted eyelids. The trained network has comparable performance to trained, but non-expert, human image graders. Further, we found that misclassified images were typically characterized by poor image quality (e.g., blurry, eyelid not in image, etc.), which could be addressed by a standardization of the image acquisition protocol.

Introduction

Despite intensive worldwide control efforts, trachoma remains the most important infectious cause of vision loss [1] and one of the overall leading causes of global blindness,[2,3] with nearly 125 million people at risk of vision loss [4]. The World Health Organization (WHO) set a goal for global elimination of trachoma as a public health problem by 2020[5], and while this has contributed to a significant reduction in prevalence of the disease, additional strategies will be necessary for global elimination. This is due to the “last mile” problem [6][79], which, in essence, describes situations in which the additional application of conventional resources leads to diminishing returns. In this case, the decreasing prevalence of TF makes it more difficult to both train new field graders to detect the disease, and for existing graders to maintain their skills—without ongoing exposure to the clinical sign of TF, graders’ clinical accuracy atrophies [810].

The conventional means of disease control has been aggressive surveillance for clinical signs of active disease (key sign: trachomatous inflammation—follicular, TF) and whole district treatment with azithromycin if the prevalence of TF in 1–9-year-olds is 5% or higher. Crucially, when the disease prevalence approaches 5% or less, clinical accuracy tolerances should theoretically be much tighter than when prevalence is high because the decision about continuing treatment (or restarting programs in the face of disease re-emergence) has such important and expensive ramifications. Travel to endemic areas has been used for training and ongoing standardization of graders, but there are increasingly few places with an adequate prevalence of TF to justify the cost and logistical complexity of travel. With the global COVID-19 pandemic, travel has become even more problematic and has further escalated the need for new solutions.

Ophthalmic photography has been used for standardized clinical diagnosis for research purposes in many conditions including trachoma,[1113] but the utility of imaging at scale in control programs that operate almost exclusively in rural areas of resource-poor countries remains unknown. Additionally, expert-grading of images in a conventional reading-center model is labor-intensive and expensive, and unlike the upfront investment for extensive training and potential travel to certify a skilled grader, these costs would generally be expected to ongoing, per patient, for the life of the program. While we are not aware of a commercial fee schedule for remote grading of trachoma images, in the United States, grading of eye images has a 2022 Medicare allowable charge between $20–28 per patient [14]. While more cost-effective than conventional in-person examination for some eye conditions in the US, this cost is still likely orders of magnitude higher than could be scalable for trachoma control programs [1518]. As such, there are two distinct, but inter-related efforts, with the goal of facilitating the identification of districts in which population-based interventions are needed to address endemic trachoma: crowdsourcing [1921], used successfully in our prior work, and artificial intelligence [22]. Previous work conducted by our team using a smartphone-based image capture system (ICAPS) demonstrated that eyelid photography of sufficient quality for non-expert grading can be acquired in a district-level trachoma survey, transmitted from the field via the internet, and graded remotely in a Virtual Reading Center by both an expert grader or by nonexpert crowdsourcing workers at lower cost per image than clinical grading [6,23].

Other groups [22] have utilized artificial neural networks to grade trachoma images and shown that, though this is a promising scalable technique, it would need further development to be utilized as a clinical tool. In this work, we examine the feasibility of augmenting and/or replacing the expert grader/crowdsourcing workflow with an AI. In this study, we utilize state-of-the-art Convolutional Neural Network (CNN) architectures to evaluate and grade the same set of images that were used in the ICAPS study for comparison. We validate our model against previously published data and demonstrate feasibility, but that further standardization of our image capture workflow is needed to render the Artificial Intelligence (AI) model clinically relevant.

Methods

Ethics statement

This work was deemed research that does not involve human subjects by the Institutional Review Board at the University of Vermont, and thus exempt from Institutional Review Board review.

Dataset

The dataset of 2,614 everted upper eyelid images used for this study was collected during a 2019 district level survey in Chamwino, Tanzania as part of the Image Capture and Processing System (ICAPS) development study [6]. For this study we analyzed the set of 2,299 ICAPS images that received the same TF grade from a single international trachoma expert and a single field grader, randomly divided into a training/validation (n = 44 TF) and test set (n = 12 TF). This photo-field concordant grade was considered the “ground truth” for analysis of AI grading. To validate our work, we considered an additional dataset [22] consisting of 1546 total images with n = 421 TF for training and validation, and n = 106 TF for the test set.

Convolutional neural network classifier

The Convolutional Neural Network (CNN) [24,25], is a type of neural network that is commonly used in image processing [26,27] and is motivated by the biological architecture of the visual cortex. The neural network architectures used in this work were ResNet101 [28] and VGG16 [29], with the number of output classes for each model reduced to two (positive or negative for TF). Class weighting and batch accumulation [30] were utilized for efficient model training. Class weighting adds a weight to the positive class when calculating the loss, causing incorrect predictions for TF-positive images to have a larger impact on model training than incorrect predictions of negative images. Batch accumulation is a method to increase batch size on memory constrained machines, by accumulating the loss function from several batches prior to backpropagation. We used the binary cross-entropy loss function [31], as this is widely-recognized to lead to the most efficient training and most accurate results.

Data augmentation

We note that out of the approximately 2000 images in the primary dataset, only ~2.5% of the images were positive for TF. Due to the unbalanced nature of the dataset, we examined a variety of oversample ratios to optimize the performance of the chosen networks. Ultimately, we found that oversample ratios greater than 50% (meaning that additional copies of TF-positive images were inserted into the training set such that it resulted in a 2:1 ratio of TF-negative:TF-positive images) did not increase the performance of the network.

Additionally, we utilized several data augmentation techniques on the oversampled images (allowing the copies to be subtly different than their originals): horizontal flipping [32], rotation[33], perspective shift [34,35], and color jitter [36,37]. The horizontal flipping reflects the image across its central axis (changing an image of a left-eye to an image of the right-eye and vice versa). For rotation, the orientation of the image was adjusted by ±15 degrees, inspired by the potential camera-alignments of the data collectors. Perspective shift projects the two-dimensional image onto a three-dimensional plane, simulating image capture from different angles. Color jitter transforms randomly change the hue, saturation, and brightness of the image. During the training, each image transform was applied to the TF-positive images with a probability of 0.5; this probability was applied individually to each transform, allowing for various combinations of data augmentation techniques to be applied to each image. Data augmentation techniques were only applied to the training set. Subsequent to the data augmentation, color-spaces in all images were normalized to match the color distribution of the training set, and all images were resized such that they had an area of 224 x 224 pixels.

Follicle enhancement

Because the clinical definition of TF is based on the number of follicles ≥ 0.5mm in diameter, a method was devised to improve the contrast between the follicles and the rest of the tarsal plate in an attempt to enhance model performance using actual clinical metrics. Designing AI tool which function using the same logic as clinical decision making is thought to improve model explainability, which can be important for the acceptability of AI systems [3840]. First, the image is transformed from an RGB color space [41] to an HSV color space [42]. HSV was chosen by inspection of Trachoma positive images in various color spaces, as seen in Fig 1; we selected to use the space which most accentuates the follicles to a human observer, which was primarily in the S channel of the HSV space. The follicles were further enhanced through the application the Contrast Limited Adaptive Histogram Equalization [43] algorithm to the selected HSV space as seen in Fig 2. This method was then applied to all the images and the model was retrained.

thumbnail
Fig 1.

Color Space Illustration: Here we present the four color-spaces that were used when developing the initial neural network models.Each color space, represented as column in this image, is defined by a tripled, represented by the rows in this image.

https://doi.org/10.1371/journal.pntd.0010943.g001

thumbnail
Fig 2.

Follicle Enhancement: In an attempt to encourage the network to learn the clinical features that identify trachomatous inflammation (i.e., ≥ 5 follicles with ≥ 0.5 mm diameter), we enhanced the contrast of the follicles using the Contrast Limited Adaptive Histogram Equalization algorithm. The results of this on the S channel are shown in the third row and the final image is then shown in the fourth row.

https://doi.org/10.1371/journal.pntd.0010943.g002

Skilled grader overread

To mimic a conventional reading center with a tiered grading structure, all images with a predicted class of positive for TF were then graded (“overread”) by a skilled human grader different from the expert who completed the initial grading. The final grades were then compared with the consensus grade using sensitivity, specificity, and the kappa statistic. In elimination programs, a kappa of ≥ 0.7 is considered sufficient agreement to validate a field grader [44], however we note that in the seminal Global Trachoma Mapping Project (GTMP) protocol, the disease prevalence was recommended to be between 30% and 70% TF. Likewise, the most recent Tropical Data training documentation recommends a minimum prevalence of 10% for an intergrader assessment [45]. The GTMP and Tropical Data authors recognized that kappa is dependent on prevalence, such that in very low (or very high) prevalence environments the kappa threshold is harder to meet and thus the skill level of a grader or any new diagnostic tool in such an environment must be higher than what was historically considered acceptable in moderate prevalence environments [43].

Validation and model updates

In order to externally validate our model, we utilized an independent dataset, described in [22]. We note that this dataset contains three classes: without active trachoma, TF, and trachomatous inflammation—intense (TI), the latter two of which can coexist. We utilized only the data that either did (n = 527) or did not (n = 1019) fulfil the criteria for the two signs of the WHO simplified grading system. Images that were positive for both TF and TI were included, but images that were solely positive for TI were excluded. After validation, we synthesized our data with the independent dataset and retrained the model using the network architecture and hyper-parameters from our best-performing model.

Performance metrics

In any type of clinical classifier or binary diagnostic tool, balance must be sought between false positive and false negative results, the relative importance of which will be dictated by the specifics of the clinical scenario. Because this tool is designed for use in low TF prevalence environments, false positives have a disproportionate effect on the output of most public health interest: district level prevalence. Since the motivation of this stage is to reduce, rather than eliminate, the skilled image-evaluation burden (defined as the proportion of images requiring a skilled/human grade), if almost all normal images can be eliminated, expert graders can still be used to overread positives to improve the specificity of a positive prediction. As such, the primary metric we use to gauge model performance is recall or sensitivity, defined as the number of true positive predictions divided by the number of true positives and false negatives.

Results

As described in Methods, we applied a number of data augmentation/oversampling techniques and transforms in various combinations on a subset of the data. In Table 1, we present the recall from 10 stochastic replicates of network training, for the top 3 performing networks; our best performing model utilized the Resnet101 architecture with the horizontal flip, perspective shift, and rotation transforms. To tune the classification threshold (i.e., the boundary between the two classes), we examined the precision-recall curve, shown in Fig 3. We set the threshold at 0.2, which gives a recall of 89%. We note that this threshold is low, which introduces the danger of false positives, but minimizes false negatives. This is illustrated in Fig 4, a confusion matrix representing the best-performing model. For this confusion matrix, class 0 represents the TF-negative class and class 1 represents the TF-positive class; the ground truth is shown on the x-axis and model predictions are shown on the y-axis. As discussed above, there were a significant number of false positives, 146.

thumbnail
Table 1. Recall values for combinations of network architecture and data augmentation techniques described in Methods.

https://doi.org/10.1371/journal.pntd.0010943.t001

thumbnail
Fig 3. Precision-Recall Curve for Best Performing ICAPS Model: The precision-recall curve for the best performing model trained on ICAPS data with a threshold of 0.2.

Precision, displayed on the y-axis, is defined as the total number of images that are correctly identified as positive for TF divided by the total number of images that are correctly identified as positive for TF plus the total number of images that are incorrectly identified as positive for TF. Recall, shown on the x-axis, is defined as the total number of images that are correctly identified as positive for TF divided by the total number of images that are correctly identified as positive for TF plus the total number of images that are incorrectly identified as negative for TF.

https://doi.org/10.1371/journal.pntd.0010943.g003

thumbnail
Fig 4. Confusion Matrix for Best Performing ICAPS Model: A comparison of predicted vs ground truth classes for the best performing model trained on ICAPS data with a threshold of 0.2.

The number of images that were correctly predicted as negative for TF are displayed in the top left; images that were incorrectly classified as positive for TF are displayed in the top right; images that are incorrectly classified as negative for TF are displayed in the bottom left; images that are correctly classified as positive for TF are shown in the bottom left.

https://doi.org/10.1371/journal.pntd.0010943.g004

When the AI classifier was used as “first pass” with skilled human overread of the 157 positive images, the kappa agreement increased from 0.086 of the classifier prediction alone to 0.659. Specificity increased to 99.6% from 67.5%, though sensitivity decreased from 91.7% to 58.3%. The prevalence of TF in the test set per the ground truth grade was 2.60% (95% confidence interval (CI): 1.35%-4.50%) as compared to 1.74% (95% CI: 0.75%-3.34%) using overreads and 34.0% (95% CI: 29.7%-38.6%) with the AI classifier alone. Of note, the grading task was completed by the skilled grader in approximately 20 minutes and constituted a 66% reduction in grading burden.

To externally validate the ICAPS-trained model, we utilized an independent dataset containing 1546 images from Ethiopia and Niger [22], illustrated in Fig 5, which generated a precision of 0.46 and a recall of 0.79. This decrease in performance is not unexpected given the paucity of TF-positive cases in our dataset. We then combined the datasets, noting that data from [22] contained images of a third class, TI, which were excluded. This resulted in a training set with 466 images that were positive for TF and 3262 images that were negative for TF, with data augmentation performed as described above; 117 images in the test set were positive for TF. This addition significantly increased the model’s performance, illustrated in Fig 6, generating a precision of 0.56 and a recall of 0.93. We show the precision-recall curve for this model in Fig 7. After skilled overread of the 193 images that were classified as positive for TF by the CNN, kappa agreement increased to 0.787 from 0.634. Specificity increased to 97.6% from 87.1% while sensitivity decreased to 78.6% from 93.1%. The prevalence of TF in the combined set per the ground truth grade was much higher than in the ICAPS data set at 15.2% (95% CI: 12.7%-17.9%) as compared to 14.0% (95% CI: 11.6%-16.7%) using overreads and 25.1% (95% CI: 22.0–28.3%) with the AI classifier alone. In this instance, the grading task was completed by the skilled grader in approximately 30 minutes and once again constituted a 75% reduction in grading burden. Detailed training/validation/test data splits are available at https://github.com/An-Cockrell/Trachoma.

thumbnail
Fig 5. Confusion Matrix for Best Performing ICAPS Model Validated Against External Dataset.

A comparison of predicted vs ground truth classes for the best performing model trained on data from ICAPS and tested with Kim et al [22] with a threshold of 0.2. The number of images that were correctly predicted as negative for TF are displayed in the top left; images that were incorrectly classified as positive for TF are displayed in the top right; images that are incorrectly classified as negative for TF are displayed in the bottom left; images that are correctly classified as positive for TF are shown in the bottom left.

https://doi.org/10.1371/journal.pntd.0010943.g005

thumbnail
Fig 6. Confusion Matrix for Best Performing Combined Model.

A comparison of predicted vs ground truth classes for the best performing model trained on two independent datasets with a threshold of 0.2. The number of images that were correctly predicted as negative for TF are displayed in the top left; images that were incorrectly classified as positive for TF are displayed in the top right; images that are incorrectly classified as negative for TF are displayed in the bottom left; images that are correctly classified as positive for TF are shown in the bottom left.

https://doi.org/10.1371/journal.pntd.0010943.g006

thumbnail
Fig 7. Precision-Recall Curve for Best Performing Combined Model.

The precision-recall curve for the best performing model trained on two independent datasets with a threshold of 0.2. Precision, displayed on the y-axis, is defined as the total number of images that are correctly identified as positive for TF divided by the total number of images that are correctly identified as positive for TF plus the total number of images that are incorrectly identified as positive for TF. Recall, shown on the x-axis, is defined as the total number of images that are correctly identified as positive for TF divided by the total number of images that are correctly identified as positive for TF plus the total number of images that are incorrectly identified as negative for TF.

https://doi.org/10.1371/journal.pntd.0010943.g007

Additionally, we trained the networks using the follicle enhancement technique described in methods. Notably, these models performed worse when considering the goal of minimizing false negatives, though they were more successful at reducing false positives. We compare the relative precision/recall metrics for these models in Table 2.

thumbnail
Table 2. Precision, Recall, and Kappa scores for the best-performing model architecture trained and tested on various combinations of independent datasets.

https://doi.org/10.1371/journal.pntd.0010943.t002

Discussion

These results indicate that, given a sufficient training set, it should be possible to train an AI to detect TF with a similar accuracy to a trained expert clinician. In the current report, we relied on the AI classifier to provide a “first-pass” for a skilled grader and found an approximately 70% reduction in the grading burden with results comparable to conventional live grading, though we anticipate acceptable standalone results with future refinements as was demonstrated by retraining with an expanded external dataset. An automated tool to accurately identify TF is desirable for several reasons, including cost and efficiency of disease prevention, however, in our opinion, its greatest value would lie in the ability of the model to store knowledge/skill (the ability to diagnose TF) that is becoming increasingly hard to acquire and maintain given the decreasing prevalence of the disease. Additionally, we would like to emphasize three notable findings in Table 2: 1) the best results were achieved when using only the Kim et al. [22] dataset from Niger and Ethiopia that are characterized by a balanced number of TF-positive:TF-negative cases and high image quality with little variation in image capture technique between images; however, the high number of TF cases may not be representative of districts where TF is very low and thus, kappa may be relatively inflated. Steps must be taken to ensure systems are validated in realistic environments to mitigate spectrum bias/effect [46]; 2) models trained on a single dataset and tested on an independent dataset in general do not perform as well (see rows 2 and 5 in Table 2); 3) our combined dataset performs very well, but its performance is slightly degraded by the greater variation in image capture technique across studies.

In contrast with previous efforts to construct a Machine-Learning (ML) pipeline for the identification of TF, we did not develop a cropping procedure; however we note that there is a key qualitative difference in our images compared to theirs: the image collector in [22] wore examination gloves while our image collectors were ungloved. Upon visual inspection in alternate color space, we found the tissue underneath the thumbnail to be very similar to the epithelial tissue on the tarsal plate, which we expect would introduce difficulties into an automated cropping procedure. Despite this, we recognize the importance of maximizing the amount of useful information fed to the neural network through image standardization. As an alternative to automated cropping, we would suggest a more standardized method of image collection (e.g., distance to camera, centering of tarsal plate, etc.), and this would be easily achievable with a screen overlay on the AR-system utilized in the initial ICAPS study [47]. We note the significantly improved performance of the AI model upon training with the alternate dataset as evidence for this assertion. In addition to the use of white latex gloves to help differentiate examiner thumbnails from the subject’s tarsal plate, images utilized in [22] were typically in excellent focus and had the tarsal plate relatively centered in the image. Thus, the inclusion of this data can be considered to be predictive of what would happen if we could augment our dataset with higher-quality images. Given the wide range of features that can cause an image to be ungradable or incorrectly graded (e.g., blur, glare, presence of fingernails), we suggest that efforts be made to standardize image capture; these could include things like modifications to cell-phone camera screen, such as the inclusion of some targeting reticle/boundary in which the eyelid should be centered in the image. Future efforts to train AI classifiers to identify TF on poor quality images may be needed for final app deployment, but we believe efforts to standardize imaging should be prioritized.

We note that when transforming the images such that the clinical features used by expert graders are highlighted (the follicular enhancement transform described above), the model performance degraded. We speculate that this is due to the AI model not “learning” the clinical definition of TF, i.e., the presence of five or more follicles greater than 0.5 mm in diameter on the tarsal plate of a single eyelid. This is especially relevant in light of the recent paper from [40], describing, among other things, the dichotomous criteria ML models and physicians use to make diagnoses. While an AI model that meets or exceeds physician/expert classification performance is obviously desirable, it is more desirable to develop an AI model that meets or exceeds physician/expert classification performance while using the same methodology (identifying and counting follicles) as the expert grader, as this build confidence in the model and reduces the “black-box” sentiment of AI [38]. In future work, we will be examining pre-training methods which we believe will guide the AI model towards this goal, inspired by the work presented in [39], though without explicit image segmentation techniques.

Ultimately, we anticipate that these models (and other models of this type) will initially be used to augment, rather than replace, the capabilities of trained and/or expert image graders. While the number of false-negative diagnoses for TF was well minimized in this study, with suggestion that it will be further minimized with the addition of new data, the number of false positive diagnoses would still lead to the initiation of unnecessary pharmacologic therapy without skilled overreading. Use of these models as pre-screening tools can significantly reduce the grading burden on expert graders and allow for more rapid and wide-reaching evaluations of TF prevalence in the developing world.

References

  1. 1. Bourne RR, Stevens GA, White RA, Smith JL, Flaxman SR, Price H, et al. Causes of vision loss worldwide, 1990–2010: a systematic analysis. Lancet Glob Health. 2013;1(6):e339–49. Epub 2014/08/12. pmid:25104599.
  2. 2. Lietman TM, Oldenburg CE, Keenan JD. Trachoma: Time to Talk Eradication. Ophthalmology. 2020;127(1):11–3. Epub 2019/12/23. pmid:31864470.
  3. 3. Williams LB, Prakalapakorn SG, Ansari Z, Goldhardt R. Impact and Trends in Global Ophthalmology. Curr Ophthalmol Rep. 2020;Jun 22:1–8 [online ahead of print]. Epub 2020/08/25. pmid:32837802; PubMed Central PMCID: PMC7306491.
  4. 4. World Health Organization. WHO Alliance for the Global Elimination of Trachoma: progress report on elimination of trachoma, 2021–Alliance de l’OMS pour l’élimination mondiale du trachome: rapport de situation sur l’élimination du trachome, 2021. Weekly Epidemiological Record = Relevé épidémiologique hebdomadaire. 2022;97(31):353–64.
  5. 5. W. H. O. Alliance for the Global Elimination of Trachoma. Meeting, W. H. O. Programme for the Prevention of Blindness and Deafness. Report of the third meeting of the WHO Alliance for the Global Elimination of Trachoma, Ouarzazate, Morocco, 19–20 October 1998 Geneva: World Health Organization; 1999 [2/14/2021]. Available from: https://apps.who.int/iris/handle/10665/65933.
  6. 6. Naufal F, Brady CJ, Wolle MA, Saheb Kashaf M, Mkocha H, Bradley C, et al. Evaluation of photography using head-mounted display technology (ICAPS) for district Trachoma surveys. PLoS Negl Trop Dis. 2021;15(11):e0009928. pmid:34748543
  7. 7. World Health Organization. Reaching the Last Mile Forum: Keynote Address 2019 [2/14/2021]. Available from: https://www.who.int/director-general/speeches/detail/reaching-the-last-mile-forum.
  8. 8. World Health Organization. Report of the 23rd Meeting of the WHO Alliance for the Global Elimination of Trachoma by 2020, Virtual meeting, 30 November–1 December 2020 [in press]. Geneva: World Health Organization, 2021.
  9. 9. World Health Organization. Network of WHO collaborating centres for trachoma: second meeting report, Decatour, GA, USA, 26 June 2016 Geneva: World Health Organization; 2017 [2/14/2021]. WHO/HTM/NTD/2016.8]. Available from: https://apps.who.int/iris/handle/10665/258687.
  10. 10. Solomon AW, Le Mesurier RT, Williams WJ. A diagnostic instrument to help field graders evaluate active trachoma. Ophthalmic Epidemiol. 2018;25(5–6):399–402. Epub 2018/08/02. pmid:30067432; PubMed Central PMCID: PMC6850902.
  11. 11. Nesemann JM, Seider MI, Snyder BM, Maamari RN, Fletcher DA, Haile BA, et al. Comparison of Smartphone Photography, Single-Lens Reflex Photography, and Field-Grading for Trachoma. The American journal of tropical medicine and hygiene. 2020;103(6):2488–91. Epub 2020/10/07. pmid:33021196; PubMed Central PMCID: PMC7695070.
  12. 12. Snyder BM, Sié A, Tapsoba C, Dah C, Ouermi L, Zakane SA, et al. Smartphone photography as a possible method of post-validation trachoma surveillance in resource-limited settings. Int Health. 2019;11(6):613–5. Epub 2019/07/23. pmid:31329890.
  13. 13. West SK, Taylor HR. Reliability of photographs for grading trachoma in field studies. Br J Ophthalmol. 1990;74(1):12–3. Epub 1990/01/01. pmid:2306438; PubMed Central PMCID: PMC1041969.
  14. 14. Center for Medicare & Medicaid Services. Search the Physician Fee Schedule. Available from: https://www.cms.gov/medicare/physician-fee-schedule/search.
  15. 15. Muqri H, Shrivastava A, Muhtadi R, Chuck RS, Mian UK. The Cost-Effectiveness of a Telemedicine Screening Program for Diabetic Retinopathy in New York City. Clinical Ophthalmology (Auckland, NZ). 2022;16:1505. pmid:35607437
  16. 16. Nguyen HV, Tan GSW, Tapp RJ, Mital S, Ting DSW, Wong HT, et al. Cost-effectiveness of a national telemedicine diabetic retinopathy screening program in Singapore. Ophthalmology. 2016;123(12):2571–80. pmid:27726962
  17. 17. Avidor D, Loewenstein A, Waisbourd M, Nutman A. Cost-effectiveness of diabetic retinopathy screening programs using telemedicine: a systematic review. Cost Effectiveness and Resource Allocation. 2020;18(1):1–9. pmid:32280309
  18. 18. Ullah W, Pathan SK, Panchal A, Anandan S, Saleem K, Sattar Y, et al. Cost-effectiveness and diagnostic accuracy of telemedicine in macular disease and diabetic retinopathy: A systematic review and meta-analysis. Medicine. 2020;99(25). pmid:32569163
  19. 19. Wang X, Mudie LI, Baskaran M, Cheng CY, Alward WL, Friedman DS, et al. Crowdsourcing to Evaluate Fundus Photographs for the Presence of Glaucoma. Journal of glaucoma. 2017;26(6):505–10. Epub 2017/03/21. pmid:28319525; PubMed Central PMCID: PMC5453824.
  20. 20. Brady CJ, Mudie LI, Wang X, Guallar E, Friedman DS. Improving Consensus Scoring of Crowdsourced Data Using the Rasch Model: Development and Refinement of a Diagnostic Instrument. Journal of medical Internet research. 2017;19(6):e222. Epub 2017/06/22. pmid:28634154; PubMed Central PMCID: PMC5497070.
  21. 21. Brady CJ, Villanti AC, Pearson JL, Kirchner TR, Gupta OP, Shah CP. Rapid grading of fundus photographs for diabetic retinopathy using crowdsourcing. Journal of medical Internet research. 2014;16(10):e233. Epub 2014/10/31. pmid:25356929; PubMed Central PMCID: PMC4259907.
  22. 22. Kim MC, Okada K, Ryner AM, Amza A, Tadesse Z, Cotter SY, et al. Sensitivity and specificity of computer vision classification of eyelid photographs for programmatic trachoma assessment. PLoS One. 2019;14(2):e0210463. pmid:30742639
  23. 23. Brady CJ, Naufal F, Wolle MA, Mkocha H, West SK. Crowdsourcing Can Match Field Grading Validity for Follicular Trachoma. Invest Ophthalmol Vis Sci. 2021;62(8):1788–.
  24. 24. O’Shea K, Nash R. An introduction to convolutional neural networks. arXiv preprint arXiv:151108458. 2015.
  25. 25. Li Z, Liu F, Yang W, Peng S, Zhou J. A survey of convolutional neural networks: analysis, applications, and prospects. IEEE Transactions on Neural Networks and Learning Systems. 2021.
  26. 26. Yamashita R, Nishio M, Do RKG, Togashi K. Convolutional neural networks: an overview and application in radiology. Insights into imaging. 2018;9(4):611–29. pmid:29934920
  27. 27. Schumaker G, Becker A, An G, Badylak S, Johnson S, Jiang P, et al. Optical Biopsy Using a Neural Network to Predict Gene Expression From Photos of Wounds. Journal of Surgical Research. 2022;270:547–54. pmid:34826690
  28. 28. He K, Zhang X, Ren S, Sun J, editors. Deep residual learning for image recognition. Proceedings of the IEEE conference on computer vision and pattern recognition; 2016.
  29. 29. Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:14091556. 2014.
  30. 30. Devarakonda A, Naumov M, Garland M. Adabatch: Adaptive batch sizes for training deep neural networks. arXiv preprint arXiv:171202029. 2017.
  31. 31. Liu L, Qi H, editors. Learning effective binary descriptors via cross entropy. 2017 IEEE winter conference on applications of computer vision (WACV); 2017: IEEE.
  32. 32. Shijie J, Ping W, Peiyi J, Siping H, editors. Research on data augmentation for image classification based on convolution neural networks. 2017 Chinese automation congress (CAC); 2017: IEEE.
  33. 33. Quiroga F, Ronchetti F, Lanzarini L, Bariviera AF, editors. Revisiting data augmentation for rotational invariance in convolutional neural networks. International Conference on Modelling and Simulation in Management Sciences; 2018: Springer.
  34. 34. Wang K, Fang B, Qian J, Yang S, Zhou X, Zhou J. Perspective transformation data augmentation for object detection. IEEE Access. 2019;8:4935–43.
  35. 35. Mikołajczyk A, Grochowski M, editors. Data augmentation for improving deep learning in image classification problem. 2018 international interdisciplinary PhD workshop (IIPhDW); 2018: IEEE.
  36. 36. Lasseck M, editor Acoustic bird detection with deep convolutional neural networks. Proceedings of the Detection and Classification of Acoustic Scenes and Events 2018 Workshop (DCASE2018); 2018.
  37. 37. Qiao Y, Su D, Kong H, Sukkarieh S, Lomax S, Clark C, editors. Data augmentation for deep learning based cattle segmentation in precision livestock farming. 2020 IEEE 16th International Conference on Automation Science and Engineering (CASE); 2020: IEEE.
  38. 38. Zhang Y, Weng Y, Lund J. Applications of Explainable Artificial Intelligence in Diagnosis and Surgery. Diagnostics. 2022;12(2):237. pmid:35204328
  39. 39. Ramesh A, Dhariwal P, Nichol A, Chu C, Chen M. Hierarchical Text-Conditional Image Generation with CLIP Latents. arXiv preprint arXiv:220406125. 2022.
  40. 40. Varoquaux G, Cheplygina V. Machine learning for medical imaging: methodological failures and recommendations for the future. npj Digital Medicine. 2022;5(1):1–8.
  41. 41. Süsstrunk S, Buckley R, Swen S, editors. Standard RGB color spaces. Color and imaging conference; 1999: Society for Imaging Science and Technology.
  42. 42. Shuhua L, Gaizhi G, editors. The application of improved HSV color space model in image processing. 2010 2nd International Conference on Future Computer and Communication; 2010: IEEE.
  43. 43. Zuiderveld K. Contrast limited adaptive histogram equalization. Graphics gems. 1994:474–85.
  44. 44. Solomon AW, Pavluck AL, Courtright P, Aboe A, Adamu L, Alemayehu W, et al. The Global Trachoma Mapping Project: methodology of a 34-country population-based study. Ophthalmic Epidemiol. 2015;22(3):214–25. pmid:26158580
  45. 45. Courtright P, MacArthur C, Macleod C, Dejene M, Gass K, Lewallen S. Tropical data: training system for trachoma prevalence surveys. London: International Coalition for Trachoma Control. 2016.
  46. 46. Wu J-H, Liu TA, Hsu W-T, Ho JH-C, Lee C-C. Performance and limitation of machine learning algorithms for diabetic retinopathy screening: meta-analysis. J Med Internet Res. 2021;23(7):e23863. pmid:34407500
  47. 47. Naufal F, West S, Mkocha H, Bradley C, Kabona G, Ngondi J, et al., editors. A Novel Hands-Free Augmented-Reality System to Document the Prevalence of Active Trachoma. CUG?H 2021 Virtual Conference: CUGH.