Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Influence of artificial intelligence on ophthalmologists’ judgments in glaucoma

  • Kenji Kashiwagi ,

    Roles Conceptualization, Data curation, Funding acquisition, Investigation, Project administration, Supervision, Validation, Writing – original draft

    kenjik@yamanashi.ac.jp

    Affiliation Department of Ophthalmology, University of Yamanashi Faculty of Medicine, Chuo, Japan

  • Masahiro Toyoura,

    Roles Methodology, Software

    Affiliation Department of Computer Science and Engineering, University of Yamanashi, Kofu, Japan,

  • Xiaoyang Mao,

    Roles Methodology, Supervision

    Affiliation Department of Computer Science and Engineering, University of Yamanashi, Kofu, Japan,

  • Kazuhide Kawase,

    Roles Data curation, Project administration, Supervision

    Affiliations Yasuma Eye Clinic, Nagoya, Japan,, Department of Ophthalmology Protective Care for Sensory Disorders, Nagoya University Graduate School of Medicine, Nagoya, Japan

  • Masaki Tanito,

    Roles Data curation, Project administration

    Affiliation Department of Ophthalmology, Shimane University Faculty of Medicine, Izumo, Japan

  • Toru Nakazawa,

    Roles Data curation, Project administration

    Affiliation Department of Ophthalmology, Tohoku University Graduate School of Medicine, Sendai, Japan

  • Atsuya Miki,

    Roles Data curation, Project administration

    Affiliation Department of Myopia Control Research, Aichi Medical University, Nagoya, Japan

  • Kazuhiko Mori,

    Roles Data curation, Project administration

    Affiliations Baptist Eye Institute, Kyoto, Japan,, Department of Ophthalmology, Kyoto Prefectural University of Medicine, Kyoto, Japan

  • Takeshi Yoshitomi

    Roles Data curation, Project administration

    Affiliation Department of Orthoptics, Fukuoka International University of Health and Welfare, Fukuoka, Japan

Abstract

Purpose

To examine the influence of artificial intelligence (AI) on physicians’ judgments regarding the presence and severity of glaucoma on fundus photographs in an online simulation system.

Methods

Forty-five trainee and expert ophthalmologists independently evaluated 120 fundus photographs, including 30 photographs each from patients with no glaucoma, mild glaucoma, moderate glaucoma, and severe glaucoma. A second trial was conducted at least one week after the initial trial in which photograph presentation order was randomized. During the second trial, 30% of the glaucoma judgments made by the AI system were intentionally incorrect. The evaluators were asked about their thoughts on AI in ophthalmology via a 3-item questionnaire.

Results

The percentage of correct responses for all images significantly improved (P < 0.001) from 48.4 ± 24.8% in the initial trial to 59.6 ± 20.3% in the second trial. The improvement in the correct response rate was significantly greater for trainees (14.2 ± 19.0%) than for experts (8.6 ± 11.4%) (P = 0.04). The correct response rate was 63.9 ± 20.6% when the AI response was correct, significantly greater than the 47.9 ± 26.6% when the AI response was incorrect (P < 0.0001). For trainees, the correct response rate was significantly greater when the AI’s response was correct than when it was incorrect. However, for experts, the effect was less pronounced. The decision time was significantly longer when the AI response was incorrect than when it was correct (P = 0.003).

Conclusion

In fundus photography-based glaucoma detection, the results of AI systems can influence physicians’ judgments, particularly those of physicians with less experience.

Introduction

In recent years, artificial intelligence (AI)-based imaging algorithms have demonstrated high accuracy, equivalent to or even superior to that of ophthalmologists, in the diagnosis of many ophthalmic diseases [17]. The development of better-performing AI algorithms is expected to improve the level of medical care, but several challenges remain in the use of AI in clinical practice. Neither AI systems nor physicians have 100% diagnostic accuracy, and therefore some misdiagnoses will still be made. For image-based diagnoses made by a physician, diagnostic guidelines can be referenced so that other physicians and healthcare workers can confirm the diagnosis. However, when AI is involved, the basis for the diagnosis may not be clear; this lack of transparency has become a factor in physicians’ distrust of AI [8]. Neural networks and deep learning-based AI algorithms lack clarity and transparency, but to ensure physician satisfaction, it is crucial that these algorithms be made accountable for the results of their analyses. In many diagnostic imaging procedures, the accuracy of the judgments of the AI system is similar to or even better than that of a human physician; therefore, a physician with high diagnostic skills is still needed to verify whether the AI has given an incorrect response. Furthermore, since AI is often reported to have high diagnostic accuracy, it is possible that the diagnostic results of AI may influence physician judgment, resulting in a bias in the final diagnosis.

Early detection of glaucoma is extremely important, as subjective symptoms are often rare, and the damage from the disease is irreversible. Although fundus photographs are useful for diagnosing glaucoma, Soh et al. reported that the incidence of undetected glaucoma has not improved in recent years [9]. This may be because accurate determination of glaucoma largely depends on the diagnostic ability of the physician and because the lack of a diagnostic environment hinders proper diagnosis by ophthalmologists. Many recent studies have reported the high accuracy of AI in diagnosing glaucoma via fundus photographs; therefore, AI is expected to be useful in the management of glaucoma patients. However, its increasing use may lead to two major problems. First, if an AI system misdiagnoses a patient, the ophthalmologist may overlook the patient, which can lead to delayed diagnosis and disease progression. Second, the ophthalmologists’ ability to diagnose glaucoma may decrease if they begin relying more heavily on the AI system in their diagnosis. This reduced diagnostic capacity may have a negative impact on providing appropriate glaucoma management.

A previous study revealed that the diagnostic accuracy of general practitioners was influenced by a fictitious AI diagnostic support system in the diagnosis of skin lesions. When the AI provided the correct diagnosis, the physician’s diagnostic ability improved. However, when the AI provided the wrong information, only a small percentage of the physicians were able to disagree with the AI [10]. Since the impact of AI on ophthalmologists’ judgment in the diagnosis of glaucoma has not been determined, we used an online simulation system referred to as the Fictitious AI-based Diagnosis System to examine the influence of AI on ophthalmologists’ ability to detect and assess the severity of glaucoma on fundus photographs.

Subjects and methods

This study was approved by the Ethics Committee of the University of Yamanashi School of Medicine (Approval Code 2635). The study was performed in accordance with the Declaration of Helsinki. All the images used in this study have been anonymized, and the ethics committee has waived the requirement for informed consent. Investigators accessed the data for research purposes on 09/03/2023.

Fundus photograph subjects

Consecutive patients aged 40–70 years who attended the Department of Ophthalmology, University of Yamanashi Hospital, between January and June 2021 were considered for the study. The inclusion criteria were as follows: phakia coupled with best corrected visual acuity of 18/20 or better, refractive error within +3.0 D to −6.0 D, and clear fundus color photographs, while control eyes had to present with no ocular diseases other than mild cataracts. The exclusion criteria for the fundus images were as follows: unclear focus, smudged peripheral areas, exclusion of the papillae and the posterior pole, inclusion of eyelashes and other artifacts, and poor image quality as judged by the main author (KK).

Image acquisition

24-2 testing was performed with a Humphrey visual field analyzer (HFA) (Carl-Zeiss Medics, Tokyo, Japan) during the entry period, employing visual field tests with solid vision failure, false negative, and false positive rates of less than 20%. Fundus color photography was performed with a fundus camera (CR-2 AF, CANON, Tokyo, Japan); information such as personal identifiers, angle of view information, and date and time of acquisition was removed from the photographs prior to analysis.

The images were divided into four severity groups, namely, nonglaucomatous, mildly glaucomatous, moderately glaucomatous and severely glaucomatous eyes, with 30 images assigned to each group. Glaucoma severity was defined as mild if the mean deviation according to the HFA 24-2 test was better than −6.0 dB, moderate if it was between −6.0 dB and −12.0 dB, and severe if it was worse than −12.0 dB.

Image evaluators

A total of 49 ophthalmologists who understood the purpose of the study and agreed to cooperate in the research initially participated as evaluators: 26 had at least five years of clinical ophthalmology experience and had achieved an ophthalmology specialization (ophthalmology experts), and 23 were in the process of obtaining an ophthalmology specialization (within two years, ophthalmology trainees). Four evaluators were excluded from the study because of protocol violations; therefore, a total of 45 evaluators were included in the final analysis.

Evaluation

Initial assessment.

Both the initial and second trials were performed with an online decision system developed for this study by the Department of Computer Science and Engineering, Faculty of Engineering, University of Yamanashi. A representative example of the presentation screen and fundus images is shown in Fig 1. Thirty images from each of the four previously described severity groups were presented in a random order on a computer monitor; the evaluator was given 30 seconds to assess each fundus photo to determine whether glaucoma was present and, if so, whether the glaucoma was mild, moderate, or severe. The evaluator was instructed to assess the images with breaks to allow sufficient time to concentrate on the assessment.

thumbnail
Fig 1. Representative screenshot of the internet-based judgment system for the first trials.

https://doi.org/10.1371/journal.pone.0321368.g001

Reassessment.

A second trial involving the same instructions was performed at least one week after the initial trial, but the order of the presentation of the images was randomly rearranged. Supplementary information in the form of the AI judgment, termed the mimic AI system, was presented on the monitor, in contrast to the initial image (Fig 2).

thumbnail
Fig 2. Representative screenshot of the internet-based judgment system used for the second trials.

https://doi.org/10.1371/journal.pone.0321368.g002

AI judgment results in the online diagnostic system

Among the AI judgment results presented during the second trial, 30% were intentionally incorrect. The results of the mimic AI system shown onscreen are summarized in Table 1.

thumbnail
Table 1. Criteria for presenting incorrect judgment outcomes.

https://doi.org/10.1371/journal.pone.0321368.t001

Questionnaire survey of evaluators

After the second trial, a questionnaire consisting of 3 items answered via a 4-point Likert scale (none, little, mild, strong) was administered to the evaluators. The questionnaire asked the following questions: How much influence did the AI response have on their second judgment? Do they feel threatened by the use of AI in ophthalmology? Do they have expectations for the use of AI in ophthalmology?

Statistical analysis

A paired t test (or Wilcoxon rank-sum test for unpaired data) was used to analyze the judgments and response times between the two groups. The Steel–Dwass test (multiple analyses of nonparametric comparisons) was used to assess significant differences among the nonglaucoma, mild impairment, moderate impairment, and severe impairment groups. Repeated-measures analysis of variance (ANOVA) was performed via JMP17 software (JMP Statistical Discovery LLC, USA), and the significance level for all tests was set at P < 0.05.

Results

Rates of correct responses in the initial and second trials

Fig 3 shows the rate of correct responses for the first and second trials. For all the evaluators, the correct response rate (CRR) for all the images was 48.4 ± 24.8% in the first trial, and significant improvements were seen in the second trial, with a CRR of 59.6 ± 20.3% (P < 0.001). In both trials, the CRR was highest in diagnosing nonglaucoma, followed by severe glaucoma. The CRR in the second trial was significantly greater than that in the first trial in all patient groups except for the nonglaucoma group (P < 0.001).

thumbnail
Fig 3. Comparison of the correct response rates between the first and second trials.

(Paired t test, *P<0.0001 between two trials) (bar = SD).

https://doi.org/10.1371/journal.pone.0321368.g003

Comparison of the correct response rates between experts and trainees

Both experts and trainees showed significant improvements in the CRR between the first and second trials in all severity groups except the nonglaucoma group (Figs 4a and b).

thumbnail
Fig 4. Comparison of expert and trainee correct response rates in the first and second trials.

The rates of correct responses in the first and second trials were compared between experts (a) and residents (b). (Paired t test, *P<0.001 between two trials) (bar = SD).

https://doi.org/10.1371/journal.pone.0321368.g004

The improvement in the CRR between the first and second trials was significantly greater for trainees (14.2 ± 19.0%) than for experts (8.6% ± 11.4%, P = 0.04), and the trainees tended to show greater performance for all severity groups. A significant improvement (P = 0.03) was observed for the moderate glaucoma severity group (Fig 5).

thumbnail
Fig 5. Comparison of accuracy improvements between trainees and experts (Steel–Dwass test, *P = 0.03) (bar = SD).

https://doi.org/10.1371/journal.pone.0321368.g005

Impact of the supplementary presentation of AI results on the CRRs of the evaluators

For all the evaluators, the CRR for all severity groups was 63.9 ± 20.6% when the mimic AI system gave a correct answer, significantly greater than the CRR of 47.9 ± 26.6% when the mimic AI system gave an incorrect answer (P < 0.0001). As shown in Fig 6, the CRRs for all groups were significantly greater when the AI mimic system gave a correct response than when it gave an incorrect response. For trainees, the CRR for all severity groups was 66.5 ± 18.5% when the AI mimic system gave a correct answer, significantly greater than the 41.5 ± 18.5% CRR when the AI mimic system gave an incorrect answer (P < 0.0001). Across all groups, the CRR was significantly greater when the AI mimic system indicated a correct answer than when it indicated an incorrect answer. The experts tended to have higher CRRs when the AI mimic system indicated a correct answer than when it indicated an incorrect answer for all severity groups, but the differences were not statistically significant (Table 2). The rate of improvement for images with correct responses was 24.9 ± 26.7% for trainees and 9.7 ± 21.7% for experts, and this difference was significant (P < 0.0001). The improvement rate was significantly better for trainees than for experts in all severity groups, and the most significant difference was found in the severe glaucoma group (Fig 6).

thumbnail
Fig 6. Effects of the mimic AI system results on the correct response rate.

(Steel–Dwass test, *P = 0.03, **P = 0.003, §P < 0.0001) (bar = SD).

https://doi.org/10.1371/journal.pone.0321368.g006

thumbnail
Table 2. Comparison of the effects of the AI judgment results on the CRR between experts and trainees.

https://doi.org/10.1371/journal.pone.0321368.t002

Comparison of CRRs when the AI mimic system showed an incorrect result

During the second trial, the CRRs obtain when incorrect AI mimic system responses were shown were compared to the CRRs obtained during the first trial. No significant difference was found between the first and second CRRs for all evaluators (Fig 7).

thumbnail
Fig 7. Correct response rate comparison between the first and second trials with incorrect mimic AI system results (Steel–Dwass test) (bar = SD).

https://doi.org/10.1371/journal.pone.0321368.g007

Analysis of response times

Overall, the evaluators had a significantly shorter response time in the second trial (9.0 ± 2.5 sec) than in the first trial (10.8 ± 4.3 sec, P = 0.0005). For trainees, the response time in the second trial (11.4 ± 5.1 sec) was significantly shorter than that in the first trial (8.4 ± 2.7 sec, P < 0.0001). Additionally, the response times among the experts for the first and second trials were 10.4 ± 3.7 seconds and 9.4 ± 2.2 seconds, respectively (P = 0.009). The reduction in response time for the second trial relative to the first trial was 2.4 ± 3.9 sec for trainees and 1.0 ± 2.7 sec for experts; this difference was statistically significant (P < 0.0001) (Fig 8).

thumbnail
Fig 8. Comparison of response times between experts and trainees (bar = SD).

https://doi.org/10.1371/journal.pone.0321368.g008

Relationship between the number of correct responses and response time

The response times of all the evaluators for correct and incorrect responses were also compared. In the first trial, the response time for correct responses was 10.0 ± 3.7 seconds, and the response time for incorrect responses was 11.7 ± 4.7 seconds; however, this difference was not significant (P=0.063). In the second trial, the difference in response times for correct responses (8.2 ± 2.0 seconds) and incorrect responses (9.7 ± 2.7 seconds) was significant (P = 0.003).

Comparison of response times between correct and incorrect ai answers in the second trial

The response times when the AI mimic system presented a correct response and when the AI mimic system presented an incorrect response were compared. For all the evaluators, the response time was 9.31 ± 2.9 s when the AI mimic response was incorrect and 8.5 ± 2.3 s when the AI mimic system response was correct; this difference was statistically significant (P = 0.001). When comparing experts and trainees, the response times were longer for both groups when the AI mimic system indicated an incorrect response; this tendency was more pronounced for experts. The difference in response times among the experts was 1.64 seconds, which was slightly longer than the 1.32 seconds observed among the trainees (Fig 9).

thumbnail
Fig 9. Comparison of response times between correct and incorrect AI responses.

(Paired t test, *P < 0.001, **P = 0.009 between two trials) (bar = SD).

https://doi.org/10.1371/journal.pone.0321368.g009

Results of the questionnaire administered after the second trial (Table 3)

The most common choice of both the trainees and experts was that the response of the mimic AI system had a mild influence on their decisions. The trainees tended to consider themselves more influenced than the experts, but the difference was not significant. A greater proportion of trainees felt threatened by the future use of AI in ophthalmic practice. On the other hand, the experts were more likely to have higher expectations of AI. Overall, the trainees were more concerned about the use of AI in medicine than the experts were.

thumbnail
Table 3. Comparison of questionnaire results between experts and trainees.

https://doi.org/10.1371/journal.pone.0321368.t003

Discussion

In this study, the impact of AI diagnosis on ophthalmologists’ judgments—which, to date, had not yet been explored—was investigated via a simulation system. The results showed that the AI system’s judgments influenced ophthalmologists’ diagnoses and that this influence was particularly strong for ophthalmologists with fewer years of experience. The accuracy of ophthalmologists’ judgments regarding the presence of glaucoma was relatively high, but the accuracy of judging the severity of glaucomatous optic neuropathy was not sufficiently high. Furthermore, ophthalmologists’ acceptance of the use of AI in glaucoma care was relatively positive.

In recent years, the diagnostic capabilities of AI have improved rapidly, with many reports indicating that these capabilities are equal to or even better than those of a physician [4,6,1114]. Because of its promising results, AI has begun to be used in clinical settings; therefore, it is essential that physicians are also aware of its negative impact on their diagnostic ability.

There have been several reports in other fields on the impact of AI on physicians’ diagnostic abilities [15,16]. Nagendram et al. reported that explainable AI information influences physicians prescriptions [17], whereas Jacob et al. reported that incorrect AI judgments may adversely impact clinicians’ treatment selections [18]. Micocci et al. conducted a study using a fictitious AI diagnostic support system and reported that the accuracy of skin lesion diagnoses by general practitioners improved when the AI gave a correct answer; however, they did not disagree with the AI when presented with an incorrect answer [10]. In the field of ophthalmology, there is little research on how AI-based diagnostic outcomes affect ophthalmologists. Previous studies have demonstrated that AI systems can assess glaucoma with a high degree of accuracy [2,5,6,19]. In this study, we investigated whether the AI system influenced physicians’ clinical judgments. Evaluators were informed that the system is capable of making highly accurate assessments regarding both the presence and severity of glaucoma, consistent with findings from previous studies5. While the AI system used in this study was trained on expert evaluations by ophthalmologists, we aimed to determine whether providing this information would impact the evaluators’ decision-making process.

In this study, we showed not only that ophthalmology trainees with insufficient diagnostic abilities were more likely to rely on the judgments provided by AI but also that experts could also be influenced to the decisions of such systems. The response time for trainees was shorter than that for experts regardless of whether the AI judgment was correct or incorrect, which further suggests that trainees are more likely to rely on AI judgment. Stewart et al. reported that physicians may depend entirely on AI systems and not make any personal judgments [20]. Therefore, ophthalmologists, especially those with less experience, must continue training until they become experts in handling AI systems properly.

The introduction of AI into glaucoma practice has demonstrated several advantages [5,6,14,19]. For example, AI may improve diagnostic ability and prevent misdiagnosis. In the present study, the evaluators showed relatively high performance in distinguishing between fundus images with and without glaucoma, but their ability to evaluate the severity of glaucoma was relatively poor. In future AI studies, the assessment of disease severity in addition to the presence or absence of disease could be useful for physicians. Moreover, AI may be useful for preventing missed glaucoma diagnoses when glaucoma is complicated by other ophthalmic diseases; AI is also expected to help reduce medical costs. With all the possible benefits of this technology, AI is expected to be actively introduced into glaucoma management in the future. However, careful attention should be given to the potential impact of AI misjudgment on the practices of physicians, especially those who do not yet have high levels of experience.

Physician’s knowledge of and experience with AI are still insufficient. Although there were differences between the experts and trainees in this study in terms of feeling threatened by and their expectations of AI, the differences were not significant. Few participants from both the expert and trainee groups indicated that they were hardly affected by the AI judgment, while most reported being affected to some extent. Therefore, there is concern that the introduction of AI may cause confusion in clinical practice. One of the concerns of physicians about the introduction of AI in clinical practice is that the basis for AI decisions is unclear—the so-called black box problem [8,21]. Neural networks and deep learning-based AI algorithms lack clarity and transparency, which can make clinicians hesitant and uncertain when making prognostic and diagnostic decisions. Many physicians are reluctant to apply AI to clinical medicine, as the majority of current AIs only report judgment results without providing sufficient diagnostic evidence to convince physicians of these results. Hua et al. reviewed factors influencing the acceptability of AI in medical practice among physicians and emphasized that human-centered practice is essential [22]. These points all need to be considered before AI can be used in glaucoma practice.

Given that the number of evaluators in this study was relatively small, further studies should be conducted with a larger number of evaluators in the future. In the present study, the impact of the AI system was compared separately for experts and trainees, but it is possible that within each group, the diagnostic skills of the evaluators were not homogenous, which may have affected the results. In addition, the degree to which the evaluators deferred to the accuracy of the diagnosis of glaucoma by the AI system was not analyzed; therefore, the evaluation of the degree of influence may not have been sufficient. The fundus photographs were judged on the same computer in each trial by each evaluator, but the monitor’s resolution and other factors, including experimental circumstances, were not exactly the same among the evaluators; the possibility that these factors may have affected the judgment results cannot be completely ruled out. Although there was a sufficient interval between the first and second trials and the order of image presentation was changed randomly in the second trial, the possibility that a learning effect from the first trial affected the second trial cannot be completely eliminated. While visual field impairment is commonly used as a criterion for determining the severity of glaucoma, there is no clear consensus on its definition. In this study, the severity classification for training the AI system was based on the mean deviation (MD) value of the visual field, as described in the Methods section. Evaluators were instructed to adhere to this criterion when making their assessments.

Conclusion

This study revealed that AI responses may influence ophthalmologists’ responses, especially among ophthalmologists with insufficient experience, but ophthalmologists need to fully understand that the AI system is not perfect. Improving physician knowledge about AI and providing a human-centered AI system for more adequate glaucoma management are important. Notably, the interaction between humans and AI in clinical settings should be balanced and subject to scrutiny as previously reported [23].

Acknowledgments

We would like to express our deepest gratitude to the following individuals for their participation in this study: Ichiya Sano MD, Masato Matsuo MD, Akiko Harano MD, Sho Ichioka MD, Yuji Takayanagi MD (Shimane University), Shinichi Usui MD, Rumi Kawashima MD, Hiroshi Kubota MD, Yuki Maeno MD, Yasuyuki Yamada MD (Osaka University), Moe Ono MD, Airi Takahashi MD, Yumi Shigemoto MD, Natsuki Okabe MD, Yuka Kasai MD, Satoshi Kogure MD, Kazuyoshi Kitamura MD, Yoshiko Fukuda MD, Shinya Minakuchi MD, Yuka Hasebe MD, Mio Matsubara MD, Masako Sakamoto MD, Misa Kimura MD, Kengo Fukunaga MD, Takanori Fukuyama MD, Shuhei Hosoda MD (University of Yamanashi), Masato Ishikawa MD, Yu Yokoyama MD, Kazuko Omodaka MD, Satoru Tsuda MD, Takeshi Yabana MD, Hiroki Takizawa MD, Yurina Yamada MD, Kanna Sasaki MD, Takuro Matsumoto MD, Chiaki Yamaguchi MD (Tohoku University), Masahiro Maeda MD (Nagoya University), Yoko Ikeda, MD, Morio Ueno MD, Hiroki Mieno MD, Yuko Maruyama MD, Yo Okada MD, Renako Kozaki MD, Kosuke Minami MD, Hiroto Yuge MD, Hiromi Nagano MD (Kyoto Prefectural University of Medicine).

References

  1. 1. Gulshan V, Peng L, Coram M, Stumpe MC, Wu D, Narayanaswamy A, et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA. 2016;316(22):2402–10. pmid:27898976
  2. 2. Kim SJ, Cho KJ, Oh S. Development of machine learning models for diagnosis of glaucoma. PLoS One. 2017;12(5):e0177726. pmid:28542342
  3. 3. Rahimy E. Deep learning applications in ophthalmology. Curr Opin Ophthalmol. 2018;29(3):254–60. pmid:29528860
  4. 4. Chen X, Xu Y, Wong DWK, Wong TY, Liu J. Glaucoma detection based on deep convolutional neural network. Conf Proc IEEE Eng Med Biol Soc. 2015;2015:715–8.
  5. 5. Huang X, Islam MR, Akter S, Ahmed F, Kazami E, Serhan HA, et al. Artificial intelligence in glaucoma: opportunities, challenges, and future directions. Biomed Eng Online. 2023;22(1):126. pmid:38102597
  6. 6. Phan S, Satoh S, Yoda Y, Kashiwagi K, Oshika T, Japan Ocular Imaging Registry Research Group. Evaluation of deep convolutional neural networks for glaucoma detection. Jpn J Ophthalmol. 2019;63(3):276–83. pmid:30798379
  7. 7. Soleimani M, Cheraqpour K, Sadeghi R, Pezeshgi S, Koganti R, et al. Artificial intelligence and infectious keratitis: where are we now? Life (Basel). 2023;13.
  8. 8. Wadden JJ. Defining the undefinable: the black box problem in healthcare artificial intelligence. J Med Ethics. 2021.
  9. 9. Soh Z, Yu M, Betzler BK, Majithia S, Thakur S, Tham YC, et al. The global extent of undetected glaucoma in adults: a systematic review and meta-analysis. Ophthalmology. 2021;128(10):1393–404. pmid:33865875
  10. 10. Micocci M, Borsci S, Thakerar V, Walne S, Manshadi Y, Edridge F, et al. Attitudes towards trusting artificial intelligence insights and factors to prevent the passive adherence of GPs: a pilot study. J Clin Med. 2021;10(14):3101. pmid:34300267
  11. 11. Ting DSW, Cheung CY-L, Lim G, Tan GSW, Quang ND, Gan A, et al. Development and validation of a deep learning system for diabetic retinopathy and related eye diseases using retinal images from multiethnic populations with diabetes. JAMA. 2017;318(22):2211–23. pmid:29234807
  12. 12. Hashimoto Y, Asaoka R, Kiwaki T, Sugiura H, Asano S, Murata H, et al. Deep learning model to predict visual field in central 10° from optical coherence tomography measurement in glaucoma. Br J Ophthalmol. 2021;105(4):507–13. pmid:32593978
  13. 13. Poplin R, Varadarajan AV, Blumer K, Liu Y, McConnell MV, Corrado GS, et al. Prediction of cardiovascular risk factors from retinal fundus photographs via deep learning. Nat Biomed Eng. 2018;2(3):158–64. pmid:31015713
  14. 14. Aggarwal R, Sounderajah V, Martin G, Ting DSW, Karthikesalingam A, King D, et al. Diagnostic accuracy of deep learning in medical imaging: a systematic review and meta-analysis. NPJ Digit Med. 2021;4(1):65. pmid:33828217
  15. 15. Goddard K, Roudsari A, Wyatt JC. Automation bias - a hidden issue for clinical decision support system use. Stud Health Technol Inform. 2011;164:17–22. pmid:21335682
  16. 16. Vicente L, Matute H. Humans inherit artificial intelligence biases. Sci Rep. 2023;13(1):15737.
  17. 17. Nagendram S, Singh A, Harish Babu G, Joshi R, Pande SD, Ahammad SKH, et al. Stochastic gradient descent optimisation for convolutional neural network for medical image segmentation. Open Life Sci. 2023;18(1):20220665. pmid:37589001
  18. 18. Jacobs M, Pradier MF, McCoy TH Jr, Perlis RH, Doshi-Velez F, Gajos KZ. How machine-learning recommendations influence clinician treatment selections: the example of the antidepressant selection. Transl Psychiatry. 2021;11(1):108. pmid:33542191
  19. 19. Xiao X, Xue L, Ye L, Li H, He Y. Health care cost and benefits of artificial intelligence-assisted population-based glaucoma screening for the elderly in remote areas of China: a cost-offset analysis. BMC Public Health. 2021;21(1):1065. pmid:34088286
  20. 20. Stewart J, Freeman S, Eroglu E, Dumitrascu N, Lu J, Goudie A, et al. Attitudes towards artificial intelligence in emergency medicine. Emerg Med Australas. 2024;36(2):252–65. pmid:38044755
  21. 21. Holzinger A, Langs G, Denk H, Zatloukal K, Müller H. Causability and explainability of artificial intelligence in medicine. Wiley Interdiscip Rev Data Min Knowl Discov. 2019;9(4):e1312. pmid:32089788
  22. 22. Hua D, Petrina N, Young N, Cho J-G, Poon SK. Understanding the factors influencing acceptability of AI in medical imaging domains among healthcare professionals: A scoping review. Artif Intell Med. 2024;147:102698. pmid:38184343
  23. 23. Maehara H, Ueno Y, Yamaguchi T, Kitaguchi Y, Miyazaki D, Nejima R, et al. The importance of clinical experience in AI-assisted corneal diagnosis: verification using intentional AI misleading. Sci Rep. 2025;15(1):1462. pmid:39789113