Figures
Abstract
Purpose
This study evaluates the impact of artificial intelligence (AI) assistance on the diagnostic performance of radiologists with varying levels of experience in interpreting mammograms in a Malaysian tertiary referral center, particularly in women with dense breasts.
Methods
A retrospective study including 434 digital mammograms interpreted by two general radiologists (12 and 6 years of experience) and two trainees (2 years of experience). Diagnostic performance was assessed with and without AI assistance (Lunit INSIGHT MMG), using sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), and area under the receiver operating characteristic curve (AUC). Inter-reader agreement was measured using kappa statistics.
Results
AI assistance significantly improved the diagnostic performance of all reader groups across all metrics (p < 0.05). The senior radiologist consistently achieved the highest sensitivity (86.5% without AI, 88.0% with AI) and specificity (60.5% without AI, 59.2% with AI). The junior radiologist demonstrated the highest PPV (56.9% without AI, 74.6% with AI) and NPV (90.3% without AI, 92.2% with AI). The trainees showed the lowest performance, but AI significantly enhanced their accuracy. AI assistance was particularly beneficial in interpreting mammograms of women with dense breasts.
Conclusion
AI assistance significantly enhances the diagnostic accuracy and consistency of radiologists in mammogram interpretation, with notable benefits for less experienced readers. These findings support the integration of AI into clinical practice, particularly in resource-limited settings where access to specialized breast radiologists is constrained.
Citation: Ramli Hamid MT, Ab Mumin N, Abdul Hamid S, Mohd Ariffin N, Mat Nor K, Saib E, et al. (2025) Comparative analysis of diagnostic performance in mammography: A reader study on the impact of AI assistance. PLoS One 20(5): e0322925. https://doi.org/10.1371/journal.pone.0322925
Editor: Yuki Arita, Memorial Sloan Kettering Cancer Center, UNITED STATES OF AMERICA
Received: October 31, 2024; Accepted: March 31, 2025; Published: May 7, 2025
Copyright: © 2025 Ramli Hamid et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: The dataset underlying this study has been made publicly available in Figshare in accordance with PLOS data-sharing requirements. The data can be accessed at DOI: 10.6084/m9.figshare.28529132. This dataset includes de-identified summary statistics relevant to the study while ensuring compliance with institutional policies and patient confidentiality regulations.
Funding: This research was supported by the Universiti Teknologi MARA Research Grant Scheme: 600/TNCPI-5/3/DDF (MEDIC)(003/2021), 600/TNCPI-5/3/DDJ(HUITM)(003/2021), and 600/RMC-LESTARI SDG-T5/3. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Introduction
Breast cancer remains a leading cause of morbidity and mortality among women worldwide, emphasizing the critical need for early detection and accurate diagnosis to improve treatment outcomes and survival rates [1]. Mammography is the primary imaging modality for breast cancer screening, but its diagnostic performance is inherently influenced by the radiologist’s experience, interpretative skills, and workload pressures [2]. Given these limitations, artificial intelligence (AI) has emerged as a promising tool in medical imaging, offering the potential to augment radiologists’ capabilities, improve diagnostic accuracy, reduce human error, and optimize workflow efficiency [3,4].
Despite the increasing body of evidence supporting AI’s role in mammography, its impact on diagnostic performance across different levels of clinical expertise, particularly in high-volume, tertiary referral settings, remains an area requiring further exploration [5,6]. In resource-constrained environments like Malaysia—a middle-income country with a limited number of breast radiologists and the absence of routine second-reader programs for screening mammography—these challenges are even more pronounced. AI could serve as a crucial second reader and decision-support tool, enhancing diagnostic accuracy and efficiency where specialized expertise is scarce [7,8].
The implications of AI in mammography extend beyond Malaysia to other low- and middle-income countries facing similar challenges in breast cancer screening. In many regions with a shortage of trained breast radiologists, AI-driven decision-support systems could help bridge the gap by improving diagnostic consistency, particularly among less experienced readers. Understanding how AI performs in the Malaysian context provides valuable insights that may be generalizable to other countries with comparable healthcare infrastructures, supporting its broader adoption in global breast cancer screening programs
This study, conducted at a tertiary referral center in Malaysia, aims to address this gap by evaluating the diagnostic performance of trainee radiologists and general radiologists in interpreting mammograms, both with and without AI assistance. The integration of AI in mammography offers the potential to improve diagnostic accuracy and potentially democratize expertise by supporting less experienced readers, thereby democratizing expertise in breast imaging. However, concerns remain regarding potential over-reliance on AI, its effect on interpretative skills, and the evolving role of radiologists in breast cancer diagnosis [9,10].
This reader study assesses the impact of AI assistance on the diagnostic performance of readers with varying levels of expertise, providing valuable insights into the practical benefits and limitations of AI in mammography. By examining diagnostic outcomes with AI assistance, this study contributes to the ongoing discourse on optimizing AI implementation in radiology. It also aims to inform policies and guidelines for AI application in mammography, ensuring that such technologies are utilized to maximize patient care, improve diagnostic accuracy, and enhance healthcare accessibility across diverse clinical settings.
Methodology
Study design
This was a retrospective cross-sectional study approved by the Institutional Review Board
(Medical Research Ethics ID 2022530–11258).) involving mammography cases from January 2021 to May 2022 done in Hospital Al-Sultan Abdullah UiTM. We excluded cases with post operative changes, personal history of breast cancer, incomplete or missing data, and patients lost from follow-up. Clinical data were extracted from electronic medical records (EMR). Data on patients age, ethnicity, breast density, American College of Radiology-Breast Imaging and Reporting and Data Systems classification (BIRADS) assessment and final histopathological (HPE) diagnosis were recorded.
Equipment and technique
Mammography examination were performed using a 3D digital mammography system (Selenia Dimension, Hologic, Bedford, Massachusetts, USA). All sets consist of craniocaudal (CC) and mediolateral oblique (MLO) views. The processed image data were sent from the acquisition workstation to the reading workstation.
AI software & algorithm
Lunit INSIGHT MMG (version 1.1.7.2, Lunit, South Korea) was selected for this study based on its FDA approval and prior validation studies, which demonstrated its effectiveness in enhancing breast cancer detection [11,12]. Previous research has shown that Lunit INSIGHT MMG improves diagnostic performance, particularly in increasing sensitivity and aiding radiologists in detecting subtle malignancies [13–16].
The AI system assigns a malignancy probability score ranging from 0% to 100% to each detected lesion. A threshold of 10% was used in this study, where scores below 10% were deemed clinically insignificant based on prior validation studies [12,13]. This means that findings scoring less than 10% were classified as normal or clinically insignificant findings, with no clinical action required. The system prioritizes suspicious findings with higher malignancy scores, assisting radiologists in their decision-making process.
Data collection and storage
Data were kept in the hospital picture archiving and communication system (PACS) (Infinitt, South Korea). Data were accessed for research purposes from 1st June 2023–28th February 2024. All image datasets were anonymized and subsequently stored in digital imaging and communication in medicine (DICOM) format in a password-protected external hard disk. Images with incomplete datasets, poor diagnostic quality, and post-operative cases were excluded.
Image assessment protocol
AI technology was not used in routine clinical practice at the time of image acquisition. The AI-assisted interpretations in this study were conducted solely in a research setting, ensuring that the readers had no prior exposure to AI during routine reporting. All readers underwent training on the application of the AI software prior to the research.
Four readers (2 trainee radiologists (ES – R1) and (NAM – R4) and two general radiologists (NMA-R3 and KMN – R2) with 6 and 12 years’ experience, independently reviewed same set of mammography images in a randomized order, in two separate sessions: first without AI assistance and then with AI assistance following a six-week washout period to minimize recall bias. Fig 1 illustrates the study design workflow.
Mammograms were first analysed by radiologists without AI assistance, followed by a six-week washout period before re-evaluation with AI support.
The readers assigned BI-RADS classifications and density assessments based on ACR BI-RADS 5th edition (2013) lexicon which includes density assessment, calcifications, mass, asymmetric density and architectural distortion. All readers are blinded to the radiology report, HPE findings (if performed), and the patients’ outcome. A final BIRADS assessment category score were given for each case.
Statistical analysis
All the data were collected in an Excel spreadsheet, and statistical analysis was performed using IBM SPSS software v.23. A p-value of < 0.05 will be taken as statistically significant. Comparisons between reader groups with and without AI assistance were conducted using Cohen kappa tests (κ). Inter-reader correlation (ICC) coefficient was calculated to ascertain agreement between readers in BI-RADS category assessment and breast density.
Diagnostic performance metrics, including sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), and area under the receiver operating characteristic (ROC) curve (AUC), were calculated for each reader group, both with and without AI assistance. The gold standard for final diagnosis was the HPE results if a biopsy was taken. If no biopsy was performed, the cases were followed up for up to 2-year, and considered benign when remains stable.
Sensitivity was when BIRADS 4 or 5 category given by radiologist and HPE results were malignant. Specificity was BIRADS 1,2, or 3 category given by radiologist on imaging and HPE or follow up imaging were benign. Positive predictive value (PPV) were malignant result in HPE and BIRADS category assigned were 4 or 5. Negative predictive value (NPV) were benign result in HPE or on follow-up, and BIRADS category assigned were 1, 2, or 3.
Ethical approval
This study was approved by the Ethics Committee of University Malaya Medical Centre and Hospital Al-Sultan Abdullah UiTM (Medical Research Ethics ID 2022530–11258) and supported by University Teknologi MARA Research Grant Scheme: 600/TNCPI-5/3/DDF (MEDIC)(003/2021), 600/TNCPI-5/3/DDJ(HUITM)(003/2021) and 600/RMC-LESTARI SDG-T5/3.
Results
A total of 434 mammography were analysed, consisting of 339 (78.1%) Malays, 71 (16.4%) Chinese, and 24 (5.5%) Indians. The patients’ ages ranged from 30 to 80, with a mean age of 52.1 (SD: 0.513).
HPE findings
Tissue biopsies were performed on 134 (30.8%) breasts whilst 300 (69%) were followed up. All patients had at least 2 years of follow-up after mammogram and/or biopsy. Of the lesions biopsied, 58 (43.2%) were benign and 76 (56.7%) were malignant. 9 (6.7%) B3, premalignant lesions, all of which were papilloma, were included in the benign category for this study.
Table 1 outlines the detailed HPE findings of the lesions. The most common benign lesion were fibroadenoma and fibrocystic change, and the most common malignant lesion was invasive ductal carcinoma. Notably, all reported benign and normal findings showed no interval cancers during a minimum of one and a half year follow-up period.
Mammogram findings
There were 288 (66.4%) women with dense breast (BIRADS density C and D). Supplementary breast ultrasound was performed in all density C and D patients. Of the 134 biopsied performed, 98 (73.1%) were in dense breast. 45 (81.8%) HPE proven benign and 53 (69.7%) malignant lesions were observed in dense breasts.
Table 2 outlines the lesion presentation on mammogram and HPE. Masses, calcifications and architectural distortion were observed in a higher proportion of malignant compared to benign cases.
There was moderate to substantial agreement in density reading between the readers and AI software with K value 0.549–0.627, p = < 0.001.
A significant negative correlation were seen between breast density and mass detection on mammograms on Spearman correlation indicating that as breast density increases, mass detection on mammograms tends to decrease, or vice versa. On the other hand, significant positive correlation were seen between breast density and detection of calcifications on mammograms.
Final BI-RADS assessment category
Intraclass Correlation Coefficient (ICC) showed a very good to excellent agreement observed between all the four readers, for the BI-RADS final assessment category, with a slightly higher agreement observed when using AI software. Without AI software, the average measures ICC was 0.910 with a 95% CI of 0.894 to 0.924, F(433, 1299), p < 0.001. With AI software, the average measures ICC was 0.931 with a 95% CI of 0.920 to 0.941, F (433, 1299), p < 0.001. The significant p-values from the ANOVA tests indicate that the differences in measurements between the readers with and without AI software are statistically significant. Table 3 presents the frequency distribution of BI-RADS categories assigned by each reader, both with and without AI assistance. Fig 2 provides a visual representation of BI-RADS category distribution, facilitating comparison between AI-assisted and non-AI interpretations.
The graph highlights differences in classification patterns, illustrating trends in BI-RADS assignments based on AI involvement.
Diagnostic performance overall
AI assistance improved the diagnostic performance of all reader groups (Table 4, Fig 3). This is evident from higher sensitivity, specificity, PPV, and AUC for all reader groups with AI (R1AI, R2AI, R3AI, R4AI) compared to without AI (R1, R2, R3, R4). General radiologist (R2), with 12 years of experience, achieved the highest sensitivity and specificity, both with and without AI assistance, suggesting R2 was the most accurate in distinguishing between malignant and benign cases. General radiologist (R3), with 6 years of experience, had the highest PPV and NPV, indicating better accuracy in confirming true positive and true negative cases, both with and without AI assistance. Trainee radiologists (R1 and R4) exhibited the lowest performance, both with and without AI assistance, likely due to less experience. However, AI significantly improved their sensitivity and PPV. Fig 3 (radar plot), provides a visual comparison of overall diagnostic performance, illustrating the increase in key metrics with AI assistance. An example of AI-assisted detection in a patient with a fatty breast composition is illustrated in Fig 4. The AI system successfully identified high-density lesions, with findings confirmed by radiologists both with and without AI assistance.
The graph illustrates key performance metrics, including Sensitivity, Specificity, PPV, NPV, and AUC. The blue line represents AI-assisted performance, while the red dashed line represents performance without AI assistance.
Craniocaudal (A) and mediolateral oblique (B) views show two high-density lesions in the left breast. AI-assisted analysis (C) highlights the suspicious regions with a high abnormality score, prompting further evaluation. Both AI-assisted and non-AI interpretations categorized this case as suspicious.
Diagnostic performance in dense breast
AI assistance enhanced performance across all reader groups for mammograms of women with dense breasts (Table 5, Fig 5). General radiologist (R2) continued to show the highest sensitivity and specificity, indicating superior accuracy in classifying malignant and benign cases even in dense breasts. General radiologist (R3) showed higher PPV without AI assistance, but higher NPV with AI, suggesting AI’s role in reducing false negatives in dense breast assessments. Trainee radiologists (R1 and R4) performed the lowest overall, but AI improved their diagnostic accuracy, particularly in terms of NPV and AUC. The radar plot (Fig 5) visually compares the diagnostic performance in dense breast, demonstrating how AI assistance enhances detection but with varying impacts depending on reader experience. Fig 6 demonstrates AI-assisted interpretation in a patient with a dense breast composition. AI highlighted multiple suspicious regions, supporting radiologists in categorizing this case as suspicious, consistent with non-AI assessments.
The figure illustrates how AI affects Sensitivity, Specificity, PPV, NPV, and AUC when interpreting dense breast cases. The blue line represents AI-assisted performance, while the red dashed line represents performance without AI assistance.
Craniocaudal (A) and mediolateral oblique (B) views reveal multiple high-density lesions in the right breast. AI-assisted analysis (C) assigns a high abnormality score, supporting the identification of suspicious areas. Both AI-assisted and non-AI interpretations categorized this case as suspicious.
The sensitivity and specificity values were lower for all reader groups in women with dense breasts compared to overall performance, suggesting greater challenges in interpreting mammograms in this subgroup. AI assistance led to an improvement in diagnostic performance across all metrics, though the extent of improvement varied among readers.
The trend of AI-assisted improvement is clearly visualized in the radar plots (Figs 3 and 5), where AI-assisted interpretation resulted in higher sensitivity, specificity, and AUC across both overall and dense breast subgroups. These figures illustrate the differences in AI’s impact on performance across different breast densities and reader experience levels.
Discussion
AI has demonstrated significant potential in improving breast cancer screening, particularly by aiding less experienced radiologists [2,17–20]. However, its effect varies based on radiologist expertise, dataset quality, and mammographic findings [3,4,6]. Automation bias and the potential over-reliance on AI are critical considerations in AI-assisted radiology. While AI can enhance diagnostic accuracy, excessive dependence on AI outputs may reduce radiologists’ critical thinking and decision-making skills, potentially impacting patient care [19]. The observed slight decrease in AUC for some readers suggests that AI may not always enhance performance and could introduce unintended challenges in interpretation [19].
One critical factor influencing AI’s performance in mammography interpretation is image quality. High-quality images with good contrast resolution enable AI to detect subtle abnormalities more effectively, whereas poor-quality images—especially in cases of dense breast tissue—can reduce AI’s diagnostic accuracy [5,7]. Mammographic noise, artifacts, and varying exposure levels may contribute to false-positive or false-negative results [8]. In particular, dense breast tissue can obscure lesions, making it difficult for both AI and human readers to achieve high sensitivity [9,10]. The performance of AI systems should, therefore, be continuously evaluated across different imaging conditions to ensure optimal reliability [17].
Furthermore, the variability in AI impact across different readers highlights the importance of experience level when integrating AI into radiology workflows [3,6]. While AI has been particularly beneficial for trainee radiologists by improving their sensitivity and overall performance, its effect on experienced radiologists has been more variable [4,18]. This suggests that AI should serve as an adjunct rather than a replacement for human interpretation, ensuring that radiologists remain actively engaged in decision-making rather than becoming over-reliant on AI outputs [19]. This variability in AI performance may explain the observed decrease in AUC values for R1 and R2, as shown in Tables 4 and 5, where AI’s impact was not uniformly positive.
The study findings also emphasize the potential role of AI in resource-limited settings, such as Malaysia and other low- and middle-income countries with a shortage of dedicated breast radiologists [1,7,9]. In such environments, AI can function as a second reader, helping to bridge the gap in expertise and reducing diagnostic discrepancies between general radiologists and breast imaging specialists [20]. The insights gained from this study may be generalizable to other countries with similar healthcare challenges, reinforcing AI’s role in addressing global disparities in breast cancer detection [10,17].
The performance variations observed among different readers in our study also highlight the potential limitations of AI assistance. Some radiologists, particularly those with more experience, may interpret AI outputs differently or be more cautious in adjusting their diagnoses. Additionally, the observed decrease in AUC for some readers suggests that AI’s effectiveness depends on user interaction and experience level. This underscores the need for further training and calibration of AI tools to optimize their integration into clinical practice.
Limitations and future research directions
This study has several limitations. As a single-center retrospective study, the findings may not be fully generalizable to broader populations. Additionally, the lack of direct comparison with other AI models limits the ability to conclude whether Lunit INSIGHT MMG performs better than alternative AI systems. Another concern is the potential for automation bias, particularly if radiologists become overly reliant on AI predictions rather than using them as a supportive tool. Finally, there is a need for prospective, multicenter validation studies to assess AI’s clinical impact across diverse patient populations and real-world screening settings. Future research should explore AI’s role in multicenter trials, evaluate its cost-effectiveness, and examine its integration into routine radiological workflows to determine its long-term viability in breast cancer screening programs.
Conclusion
his study highlights the significant role of artificial intelligence in enhancing breast cancer screening, particularly in resource-limited settings like Malaysia. AI has demonstrated the potential to improve diagnostic accuracy, reduce inter-reader variability, and support less experienced radiologists, contributing to a more standardized and equitable healthcare system.
Beyond Malaysia, these findings are relevant to other low- and middle-income countries facing similar challenges. AI can help bridge the expertise gap and optimize resources, facilitating earlier cancer detection and improved patient outcomes. However, careful implementation is necessary to prevent automation bias and ensure AI complements rather than replaces human expertise.
While this study provides valuable insights, further research, including multi-center trials and cost-effectiveness studies, is needed to establish AI’s long-term clinical impact. Strategic adoption and ongoing refinement will be crucial to maximizing AI’s potential in breast cancer detection and patient care.
References
- 1. Barrios CH. Global challenges in breast cancer detection and treatment. Breast. 2022;62 Suppl 1(Suppl 1):S3–6. pmid:35219542
- 2. Clerkin N, Ski CF, Brennan PC, Strudwick R. Identification of factors associated with diagnostic performance variation in reporting of mammograms: a review. Radiography (Lond). 2023;29(2):340–6. pmid:36731351
- 3. Arzamasov K, Vasilev Y, Vladzymyrskyy A, Omelyanskaya O, Shulkin I, Kozikhina D, et al. An international non-inferiority study for the benchmarking of AI for routine radiology cases: chest X-ray, fluorography and mammography. Healthcare (Basel). 2023;11(12):1684. pmid:37372802
- 4. Hardy M, Harvey H. Artificial intelligence in diagnostic imaging: impact on the radiography profession. Br J Radiol. 2020;93(1108):20190840. pmid:31821024
- 5. van Leeuwen KG, de Rooij M, Schalekamp S, van Ginneken B, Rutten MJCM. How does artificial intelligence in radiology improve efficiency and health outcomes? Pediatr Radiol. 2022;52(11):2087–93. pmid:34117522
- 6. Najjar R. Redefining radiology: a review of artificial intelligence integration in medical imaging. Diagnostics (Basel). 2023;13(17):2760. pmid:37685300
- 7. Schaffter T, Buist DSM, Lee CI, Nikulin Y, Ribli D, Guan Y, et al. Evaluation of combined artificial intelligence and radiologist assessment to interpret screening mammograms. JAMA Netw Open. 2020;3(3):e200265. pmid:32119094
- 8. Taylor CR, Monga N, Johnson C, Hawley JR, Patel M. Artificial intelligence applications in breast imaging: current status and future directions. Diagnostics (Basel). 2023;13(12):2041. pmid:37370936
- 9. Lokaj B, Pugliese M-T, Kinkel K, Lovis C, Schmid J. Barriers and facilitators of artificial intelligence conception and implementation for breast imaging diagnosis in clinical practice: a scoping review. Eur Radiol. 2024;34(3):2096–109. pmid:37658895
- 10. Koh D-M, Papanikolaou N, Bick U, Illing R, Kahn CE Jr, Kalpathi-Cramer J, et al. Artificial intelligence and machine learning in cancer imaging. Commun Med (Lond). 2022;2:133. pmid:36310650
- 11. Salim M, Wåhlin E, Dembrower K, Azavedo E, Foukakis T, Liu Y, et al. External evaluation of 3 commercial artificial intelligence algorithms for independent assessment of screening mammograms. JAMA Oncol. 2020;6(10):1581–8. pmid:32852536
- 12. Kim H-E, Kim HH, Han B-K, Kim KH, Han K, Nam H, et al. Changes in cancer detection and false-positive recall in mammography using artificial intelligence: a retrospective, multireader study. Lancet Digit Health. 2020;2(3):e138–48. pmid:33334578
- 13. Kwon M-R, Chang Y, Ham S-Y, Cho Y, Kim EY, Kang J, et al. Screening mammography performance according to breast density: a comparison between radiologists versus standalone intelligence detection. Breast Cancer Res. 2024;26(1):68. pmid:38649889
- 14. Kizildag Yirgin I, Koyluoglu YO, Seker ME, Ozkan Gurdal S, Ozaydin AN, Ozcinar B, et al. Diagnostic performance of AI for cancers registered in a mammography screening program: a retrospective analysis. Technol Cancer Res Treat. 2022;21:15330338221075172. pmid:35060413
- 15. Chang Y-W, An JK, Choi N, Ko KH, Kim KH, Han K, et al. Corrigendum: artificial intelligence for breast cancer screening in mammography (AI-STREAM): a prospective multicenter study design in korea using AI-based CADe/x. J Breast Cancer. 2022;25(2):147. pmid:35506581
- 16. Ahn JS, Shin S, Yang S-A, Park EK, Kim KH, Cho SI, et al. Artificial intelligence in breast cancer diagnosis and personalized medicine. J Breast Cancer. 2023;26(5):405–35. pmid:37926067
- 17. van Winkel SL, Rodríguez-Ruiz A, Appelman L, Gubern-Mérida A, Karssemeijer N, Teuwen J, et al. Impact of artificial intelligence support on accuracy and reading time in breast tomosynthesis image interpretation: a multi-reader multi-case study. Eur Radiol. 2021;31(11):8682–91. pmid:33948701
- 18. Conant EF, Toledano AY, Periaswamy S, Fotin SV, Go J, Boatsman JE, et al. Improving accuracy and efficiency with concurrent use of artificial intelligence for digital breast tomosynthesis. Radiol Artif Intell. 2019;1(4):e180096. pmid:32076660
- 19. Dratsch T, Chen X, Rezazade Mehrizi M, Kloeckner R, Mähringer-Kunz A, Püsken M, et al. Automation bias in mammography: the impact of artificial intelligence BI-RADS suggestions on reader performance. Radiology. 2023;307(4):e222176. pmid:37129490
- 20. Leibig C, Brehmer M, Bunk S, Byng D, Pinker K, Umutlu L. Combining the strengths of radiologists and AI for breast cancer screening: a retrospective analysis. Lancet Digit Health. 2022;4(7):e507–19.