Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Research on the establishment of a four-classification model for breast mass ultrasound images based on transfer learning

  • Jinlao Li ,

    Contributed equally to this work with: Jinlao Li, Bo Zhao

    Roles Conceptualization, Data curation, Formal analysis, Methodology, Project administration, Software, Validation, Visualization, Writing – original draft, Writing – review & editing

    309827646@qq.com (JL); 2871629986@qq.com (SYZ)

    Affiliation Diagnostic Ultrasound Center, The First Hospital of Jilin University, Changchun, China

  • Bo Zhao ,

    Contributed equally to this work with: Jinlao Li, Bo Zhao

    Roles Investigation, Methodology, Supervision, Writing – review & editing

    Affiliation Breast Surgery Department, The First Hospital of Jilin University, Changchun, China

  • YongQi Chen,

    Roles Formal analysis, Investigation, Software, Validation

    Affiliation Diagnostic Ultrasound Center, The First Hospital of Jilin University, Changchun, China

  • Guang Wang,

    Roles Formal analysis, Investigation, Resources

    Affiliation Breast Surgery Department, The First Hospital of Jilin University, Changchun, China

  • XueLiang Tian,

    Roles Investigation, Project administration, Resources, Software

    Affiliation Breast Surgery Department, The First Hospital of Jilin University, Changchun, China

  • ShuYu Zhou

    Roles Conceptualization, Formal analysis, Methodology, Supervision, Writing – review & editing

    309827646@qq.com (JL); 2871629986@qq.com (SYZ)

    Affiliation Pharmaceutical College, University Of beihua University, Yanji, China

Abstract

Breast mass classification via ultrasound is critical for early diagnosis but remains challenging due to overlapping morphological features. This study evaluates the efficacy of optimized transfer learning (TL) based on deep convolutional neural networks (DCNNs) in distinguishing four breast mass categories: invasive ductal carcinoma (IDC), fibroadenoma (FA), mucinous carcinoma (MC), and inflammatory mass (IM). Our approach comprises four key steps: (1) applying extensive data augmentation to a retrospective dataset of 346 ultrasound images from 294 patients (November 2021-March 2023) classified by pathological confirmation; (2) implementing transfer learning with five pretrained DCNN models (ResNet18, ResNet50, DenseNet121, MobileNetV2, GoogLeNet); (3) optimizing model hyperparameters for four-class classification; (4) comparing performance metrics (accuracy, precision, recall, F1-score, AUC-ROC) against two senior radiologists (8- and 10-years’ experience) on a held-out test set. Results: The DenseNet121 and GoogLeNet models demonstrated superior overall accuracy (0.912 and 0.926, respectively), significantly outperforming the radiologists’ consensus (0.691) (p < 0.01). In class-specific analysis, the optimized models achieved notably higher accuracy for IDC, MC, and IM. Statistical testing confirmed the significant performance improvement of the top models over human experts. Optimized TL with DCNNs, particularly DenseNet121 and GoogLeNet, enables highly accurate four-class breast mass classification in ultrasound images, surpassing expert radiologist performance with statistical significance. This approach holds promise for clinical decision support. Code is publicly available at https://github.com/jinlao777/BCC.

1. Introduction

In recent years, breast diseases have become one of the most common medical issues among women in China. According to the report from the American Cancer Society [1], The incidence and mortality rates of breast cancer remain persistently high each year, posing a serious threat to women’s health. Research has found that early screening and diagnosis are crucial for the treatment and prognosis of breast cancer. Common imaging methods include mammography, ultrasound, and magnetic resonance imaging [2]. Among them, breast ultrasound examination has attracted much attention due to its advantages of non-ionizing radiation, rapid imaging, portability, and low cost. In clinical practice, radiologists classify ultrasound images as benign or malignant based solely on experience and provide preliminary diagnostic results. This diagnostic process is subjective and highly dependent on the experience of the radiologist. Therefore, since the 1980s, computer-aided diagnosis (CAD) systems have been used to assist radiologists in improving the efficiency of medical image diagnosis, and have achieved good results [3].

Machine learning is a core area of computer algorithms that can automatically improve through experience, making it a key application area of artificial intelligence. Common machine learning algorithms include random forests, support vector machines, and neural networks. Deep learning is a branch of machine learning built on neural network algorithms and is therefore also known as deep convolutional neural networks (DCNNs) [4]. Deep learning is a widely recognized method capable of processing raw image data and automatically learning high-level abstract features from it. This gives it the potential to perform tasks in medical image analysis, including nodule classification and organ segmentation [5], This greatly reduces the time required for manual feature extraction in traditional machine learning processes, and it is also capable of handling large-scale data samples. Furthermore, it is a machine learning method based on multi-layer neural network structures, which processes information and makes predictions by automatically learning and adjusting the connection strengths between different layers [6].

Currently, deep learning-related methods have become the mainstream technology in the field of computer-aided diagnosis, widely applied to various tasks, including disease classification, Region of Interest (ROI) segmentation, object detection, and image registration [7], The applications of deep learning are also widespread [5], including but not limited to the breast [8], liver [9], carotid artery [10], thyroid [11], and lungs [12].

However, research on deep learning in the field of breast ultrasound image classification often focuses on the classification of benign and malignant nodules. There is limited research and application regarding the rich information present in breast mass ultrasound images for further detailed classification [1317]. Therefore, expanding the scope of research on breast ultrasound image classification to focus more attention on detailed classification of breast masses is of significant importance for improving diagnostic accuracy and clinical applicability. This study deliberately selected four categories that present significant challenges and hold considerable importance in clinical diagnosis. Research has found that fibroadenomas can increase the subsequent risk of breast cancer by approximately 2 to 3 times [18]. Furthermore, although fibroadenomas typically exhibit sonographic features of benign masses on ultrasound, many present with atypical characteristics in clinical practice. These atypical features include irregular shape, heterogeneous internal echogenicity, indistinct or irregular margins, angulations, or microlobulations, and may also exhibit rich blood flow signals, creating a significant overlap in ultrasound appearance with breast carcinoma.

Mucinous carcinoma of the breast, conversely, is a unique pathological type of breast cancer that occurs predominantly in postmenopausal or elderly patients [19]. Its histological presentation is distinctive, often showing low cellular atypia. On ultrasound images, the masses frequently demonstrate a relatively regular shape and well-defined borders, with complex internal echogenicity. These characteristics often lead to misdiagnosis as fibroadenomas or other benign lesions, thereby increasing the risk of a missed diagnosis [20].

Inflammatory breast masses, as a common benign breast condition, have a relatively low incidence. Particularly for non-lactational inflammatory masses, the exact etiology remains unclear. Their clinical manifestations and physical findings can resemble infectious mastitis or inflammatory carcinoma. In most cases, a multidisciplinary approach is required to narrow the differential diagnosis [21]. Sonographically, some of their features bear a high resemblance to breast cancer and may be confused with it. This not only poses challenges for early diagnosis but also may cause patients to miss the optimal treatment window, thereby presenting a significant challenge to hospital diagnostic capabilities [22].

Invasive ductal carcinoma (IDC) is the most common type of breast cancer clinically, typically developing when ductal cancer cells breach the basement membrane and invade the stroma [23]. Its microscopic features include cancer cells with considerable variation in size and shape, marked pleomorphism, frequent mitotic figures, arranged in cord-like or nest-like structures, sometimes accompanied by glandular formations. Characteristic ultrasound findings include a small, hard mass with marked posterior acoustic shadowing; a spiculated or crab-leg-like margin is its most distinctive sonographic feature [24]. However, these subtle features may be insufficiently pronounced or overlooked on ultrasound, particularly in some early-stage cases, potentially leading to inadequate attention during the diagnostic process.Here, we present the first systematic evaluation of optimized transfer learning for four-category classification of breast ultrasound masses (invasive ductal carcinoma, fibroadenoma, mucinous carcinoma, and inflammatory masses), demonstrating superior diagnostic accuracy to senior radiologists through rigorous benchmarking of five DCNN architectures.

Our principal contributions are threefold:(1)We propose a comprehensive transfer learning framework that systematically optimizes and benchmarks five contemporary DCNN architectures (ResNet18, ResNet50, DenseNet121, MobileNetV2, GoogLeNet) for the challenging task of four-category breast mass classification. (2) We establish a rigorous clinical benchmark by comparing model performance directly against two senior radiologists with 8- and 10-years of experience, providing a realistic assessment of clinical applicability. (3)We demonstrate that our optimized DenseNet121 model not only achieves state-of-the-art classification accuracy but also significantly outperforms human experts across all four mass categories, offering a potent tool for clinical decision support. (4)To ensure full reproducibility and foster further research, we will make our complete code implementation publicly available on GitHub.

The remainder of this paper is organized as follows. Section 2 details the methodology, including data collection procedures, preprocessing techniques, deep learning model architectures, and statistical analysis. Section 3 presents the experimental results, comprising patient characteristics and ultrasound image features, along with performance comparisons between transfer learning models and radiologists. Section 4 provides a comprehensive discussion of the findings, clinical implications, study limitations, and future research directions.The paper concludes with Section 5, which summarizes the principal findings and their clinical significance.

2. Materials and methods

This retrospective study was approved by the Ethics Committee of the First Bethune Hospital of Jilin University, and written informed consent was waived (Approval No. 2023-KS-191).

2.1 Data collection

This study retrieved data from the hospital’s United-Imaging system for 1,184 patients who visited the Breast Surgery Department of the First Bethune Hospital of Jilin University between January 2020 and March 2023. Ultimately, 346 ultrasound images from 294 patients who underwent breast ultrasound examinations and had confirmed pathological results met the inclusion criteria. Among them, there were 75 cases of fibroadenoma, 67 cases of inflammatory mass, 99 cases of invasive carcinoma, and 53 cases of mucinous carcinoma. Most lesions were solitary, with a small portion being multiple. To prevent data leakage and ensure a fair evaluation, all eligible images were randomly split into training and test sets at a ratio of 8:2 after de-identification. Additionally, two senior ultrasound radiologists classified the images in the training set (Radiologist 1: 8 years of ultrasound diagnostic experience, Radiologist 2: 10 years of ultrasound diagnostic experience). Ultrasound images were acquired using the Canon Xario200 ultrasound diagnostic system, Canon Aplio300 ultrasound diagnostic system, Toshiba 790 ultrasound diagnostic system, and Philips HD9 ultrasound diagnostic system. Basic clinical data, including age, gender, pathological results, and ultrasound diagnostic reports, were obtained from the hospital’s electronic medical record system. Given the retrospective design of this study and the fact that all data were fully anonymized, the ethics committee approved a waiver of written informed consent. The patient information obtained during the study does not affect subsequent treatment nor compromise patient privacy.Fig 1 illustrates the flowchart of the study process.

Inclusion criteria: Definite pathological diagnosis results; ultrasound examination performed before biopsy or surgery; a single B-mode ultrasound image capable of fully displaying the boundary of the mass. Exclusion criteria: Uncertain pathological results; poor image quality; history of radiotherapy or chemotherapy; lack of ultrasound examination images.

2.2 Data preprocessing

The obtained breast mass ultrasound images were preprocessed by manually extracting the rectangular region containing the breast mass from the entire original ultrasound image to eliminate redundant information in the image. Finally, ultrasound images containing only other human tissue structures such as skin, subcutaneous fat, glandular tissue, chest muscles, intercostal muscles, etc., were obtained. In order to address issues such as insufficient feature learning, poor network performance, instability, and overfitting that may arise from small sample datasets during the training process, preprocessing and enhancement steps were applied to the image data, including image rotation, Gaussian blur, scaling, contrast, and brightness adjustments.All programs were executed in Python version Python 3.9.0.

2.3 Training and interpretation of deep learning models

Transfer learning, as a machine learning method, aims to apply knowledge or experience learned from one task to another related task [25]. The study selected transfer learning using pre-trained deep convolutional neural network models. During the model training phase, pre-training was conducted using the ImageNet dataset [26], Five deep learning models, namely ResNet18, ResNet50 [27], DenseNet121 [28], MobileNetV2 [29], and GoogLeNet [30], were used as base networks. The parameters of the convolutional layers were retained, and fine-tuning was performed on the classification layer parameters using the training set to classify breast mass ultrasound images into four types.

This study chose the ADAM optimizer as the optimization algorithm for the model and utilized it to update gradients and network parameters. Cross-entropy loss function was employed to measure the difference between model output and true labels. Regarding hyperparameter settings, the batch size (BS) was set to 32, the number of epochs was set to 50, and the initial learning rate (LR) was set to 0.001. Additionally, a learning rate decay strategy was adopted, halving the learning rate every 5 epochs to aid better convergence of the model. The Softmax function was used in the output layer of the model as a classification function to accomplish multi-class classification tasks. Fig 2 Overview of feature based transfer learning framework.

thumbnail
Fig 2. Overview of feature based transfer learning framework.

https://doi.org/10.1371/journal.pone.0340111.g002

To better interpret the model’s prediction results, we employed class activation mapping (CAM) methods to generate heatmaps, visualizing the most indicative regions of interest in breast mass images as identified by the model [31,32]. All heat maps were generated using the OpenCV package (version 4.5.5).

2.4 Statistical analysis

The diagnostic performance of the five deep learning models was evaluated on the four-class breast mass classification task using standard metrics: accuracy, precision, recall (sensitivity), specificity, and F1-score. The models’ predictions were further visualized using confusion matrices and receiver operating characteristic (ROC) curves, with the area under the curve (AUC) calculated for each class. For benchmark comparison, the test set was independently classified by two senior ultrasound radiologists with 8 and 10 years of experience, respectively. McNemar’s test was employed to assess the statistical significance of the performance difference between each model and the radiologists’ consensus, with a p-value threshold of 0.01. These computations were carried out using standard scientific libraries in Python. The metrics are defined as follows:

Accuracy measures the overall proportion of correctly classified instances and is calculated as:

(1)

Precision quantifies the proportion of correctly predicted positive instances among all instances predicted as positive:

(2)

Recall (also known as Sensitivity) indicates the proportion of actual positive instances that are correctly identified:

(3)

Specificity assesses the model’s ability to correctly identify negative instances:

(4)

F1-score represents the harmonic mean of precision and recall, providing a single metric that balances both concerns:

(5)

3. Results

3.1 Patient and ultrasound image characteristics

This study included 346 ultrasound images of breast masses from 294 patients with breast masses, including 99 cases of invasive ductal carcinoma, 75 cases of fibroadenoma, 67 cases of inflammatory masses, and 53 cases of mucinous carcinoma, as shown in Table 1. All ultrasound images were randomly allocated to the training set and testing set, with 278 breast mass images used to train the models and optimize their parameters, and the remaining 68 breast mass images used as an independent test set to validate the performance of the trained optimal models.

3.2 Comparison of diagnostic performance between the five transfer learning models and radiologists

Among the five pre-trained deep convolutional neural network models used in this study, DenseNet121 performed excellently in classifying the four categories. Additionally, the accuracy of the DenseNet121 model and senior radiologists for invasive ductal carcinoma, fibroadenoma, mucinous carcinoma, and inflammatory mass were 0.9853, 0.9706, 0.9559, 0.9118, and 0.8382, 0.8529, 0.8382, 0.8529, respectively. Fig 3 shows the classification confusion matrices of different models and senior ultrasound radiologists, detailing the numbers of true positive, false positive, true negative, and false negative results. Specific diagnostic performance metrics and results are listed in Table 2.

thumbnail
Table 2. Classification performance of transfer learning models and radiologists.

https://doi.org/10.1371/journal.pone.0340111.t002

thumbnail
Fig 3. Confusion matrices of transfer learning models and radiologists on the test set.

https://doi.org/10.1371/journal.pone.0340111.g003

Table 3 summarizes the diagnostic performance of the deep learning models in comparison to radiologists. The DenseNet121 and GoogLeNet architectures significantly outperformed the radiologists’ consensus (OA: 0.691), achieving overall accuracies of 0.912 and 0.926, respectively (McNemar’s test, p < 0.01). While all models attained higher overall accuracy than the radiologists, the performance advantage of ResNet18, ResNet50, and MobileNet-V2 was not statistically significant (p > 0.05).

thumbnail
Table 3. Performance comparison of transfer learning models versus radiologists in breast mass classification.

https://doi.org/10.1371/journal.pone.0340111.t003

Fig 4 shows the ROC curves for different models and senior ultrasound radiologists on the test set for classifying different masses. The DenseNet121 model had the highest AUC value in classifying invasive ductal carcinoma (0.98, 95% CI: 0.917–1.000). The GoogLeNet model performed best in classifying fibroadenoma, inflammatory masses, and mucinous carcinoma, with AUC values of 0.97 (95% CI: 0.932–1.000), 0.93 (95% CI: 0.833–1.0000), and 0.95 (95% CI: 0.850–1.000), respectively. Senior radiologists had AUC values of 0.79 (95% CI: 0.683–0.792), 0.81 (95% CI: 0.702–0.919), 0.80 (95% CI: 0.676–0.912), and 0.77 (95% CI: 0.629–0.904) for classifying invasive ductal carcinoma, fibroadenoma, inflammatory masses, and mucinous carcinoma, respectively.

thumbnail
Fig 4. Receiver operating characteristic curves of transfer learning models and radiologists on the test set.

https://doi.org/10.1371/journal.pone.0340111.g004

Fig 5 shows the training curves of the five pre-trained deep learning models. The ascending curves represent the accuracy on the training and validation sets, while the descending curves represent the loss on the training and test sets, indicating the degree of fit between the predictions and the true pathological labels. As the number of training epochs increases, the accuracy on both the training and test sets gradually rises and stabilizes, while the loss function curves for both the training and test sets continuously decrease. The similarity between the two loss curves indicates that the model does not exhibit significant overfitting.

In Fig 6, images A and C show the ultrasound images of fibroadenoma and inflammatory mass in the test set, respectively, while images B and D show the corresponding images with heatmaps overlaid using the pre-trained DenseNet121 model. In Fig 7, images A and C show the ultrasound images of invasive ductal carcinoma and mucinous carcinoma in the test set, respectively, while images B and D show the corresponding images with heatmaps overlaid using the pre-trained DenseNet121 model. The model correctly predicted each of their true categories.

thumbnail
Fig 6. Figures A and C show B-mode ultrasound images of a fibroadenoma and an inflammatory masses, respectively, from the test set, and Figures C and D show the corresponding images after overlaying the thermograms using the pre-trained DenseNet121 model.

https://doi.org/10.1371/journal.pone.0340111.g006

thumbnail
Fig 7. Figures A and C show the B-mode ultrasound images of invasive ductal carcinoma and mucinous carcinoma in the test set, respectively, and Figures C and D show the respective images after overlaying the thermograms using the pre-trained DenseNet121 model.

https://doi.org/10.1371/journal.pone.0340111.g007

4. Discussion

This study demonstrated the efficacy of optimized transfer learning with pre-trained deep convolutional neural networks (DCNNs) for the four-category classification of breast tumors in ultrasound images. The employed models—ResNet18, ResNet50, DenseNet121, GoogLeNet, and MobileNetV2—all achieved satisfactory classification outcomes, particularly the DenseNet121 model, which produced commendable results on the test set. The areas under the receiver operating characteristic curve (AUC) for invasive ductal carcinoma, fibroadenoma, mucinous carcinoma, and inflammatory breast tumors were 0.976 (95% CI: 0.917, 1.000), 0.947 (95% CI: 0.870, 1.000), 0.940 (95% CI: 0.845, 1.000), and 0.899 (95% CI: 0.802, 0.974), respectively, surpassing the performance of two experienced sonographers. Notably, McNemar’s test revealed statistically significant differences between radiologists and both GoogleNet (p < 0.001) and DenseNet121 (p = 0.001), indicating that these models performed significantly better than human experts. In contrast, ResNet18 (p = 0.070), ResNet50 (p = 0.663), and MobileNet-V2 (p = 0.110) showed no statistically significant difference from radiologists, suggesting comparable performance levels.

From Table 2, we observe that although all five models achieved satisfactory prediction results on the test set, there were slight differences in classification among the different categories. Overall, DenseNet121 and GoogLeNet performed exceptionally well across various categories. In clinical practice, for the ultrasound diagnosis of breast tumors, we aim to find the best method to minimize the misdiagnosis rate of benign tumors and the missed diagnosis rate of malignant tumors.To this end, we analyzed diagnostic errors and identified specific patterns. A key finding was that inflammatory masses were most frequently misclassified as malignant tumors, particularly mucinous carcinoma, which could lead to unnecessary biopsies. Conversely, the majority of missed malignancies were also mucinous carcinomas, as their sonographic features overlap with benign conditions. Considering the overall performance, DenseNet121 appears to be a more reliable choice for clinical application in the classification and prediction of breast tumor ultrasound images.

Figs 5 and 6 shows the heatmaps generated by the pre-trained DenseNet121 model using the CAM class activation mapping method. These heatmaps illustrate the decision basis for the model’s classification of breast tumors. In these images, the model correctly predicted the true category of each image. Notably, the model’s attention is not limited to the tumor region but also includes parts of the tumor margins. This indicates that these regions also contribute value to the model’s prediction process. It also suggests that the model comprehensively understands the image and makes accurate classification predictions based on the different features of each region.

Despite the achievements of this study, several limitations remain. First, the retrospective design and relatively small dataset may pose risks of overfitting and affect the model’s generalizability. To mitigate overfitting, we implemented a rigorous random assignment of images during data preprocessing and employed extensive data augmentation techniques. More importantly, the use of transfer learning with pre-trained models, which leverages features learned from large-scale natural image datasets (e.g., ImageNet), significantly enhances feature extraction capability and robustness when fine-tuned on smaller medical datasets. The consistently superior performance of our models, particularly DenseNet121, strongly suggests that the models have effectively learned generalizable pathological features rather than merely memorizing dataset-specific noise. However, further validation is required to assess model performance on external datasets acquired from different institutions and imaging equipment.

Second, since the examinations were conducted by multiple radiologists, some variations in image quality and acquisition parameters are inevitable, which, while reflecting real-world clinical practice, may introduce confounding factors. Third, the selection of pre-trained deep convolutional neural network models, though encompassing diverse architectures (e.g., residual networks, densely connected networks), is not exhaustive, and other state-of-the-art architectures could be explored.

Lastly, the model’s classification performance for rare or atypical cases requires further validation. The clinical impact of misclassification, particularly false negatives and false positives, must be carefully considered. A false negative (e.g., misclassifying a malignant mass as benign) could lead to delayed diagnosis and treatment, with serious consequences for patient outcomes. Conversely, a false positive (e.g., classifying a benign mass as malignant) may cause unnecessary patient anxiety, follow-up examinations, or invasive biopsies, increasing healthcare costs and patient burden. Therefore, any clinical deployment must prioritize high sensitivity, especially for malignant categories.

For future clinical implementation, the proposed model is best positioned not as a replacement, but as a decision support tool integrated into the radiology workflow. It could function as a pre-screening system to flag suspicious cases for prioritized radiologist review, thereby improving diagnostic efficiency. Alternatively, it could serve as a “second reader” to highlight potential discrepancies and reduce perceptual errors. To realize this vision, prospective, multi-center studies are essential to validate the model’s efficacy in a live clinical setting and to understand its impact on radiologists’ diagnostic accuracy and workflow. Furthermore, generalization capability must be enhanced by training on larger, more diverse, multi-institutional datasets and by refining the model architecture to improve its sensitivity to atypical and rare lesions.

5. Conclusion

In conclusion, this study successfully explored the use of pre-trained deep convolutional neural networks for optimized transfer learning to achieve four-class prediction of breast masses in ultrasound images. This method could become an effective tool in clinical practice for early screening of different types of breast masses, supporting radiologists in better formulating treatment strategies, increasing the chances of early detection of breast cancer, and positively impacting the health of female patients.

References

  1. 1. Saka AH, Giaquinto AN, McCullough LE, Tossas KY, Star J, Jemal A, et al. Cancer statistics for African American and Black people, 2025. CA Cancer J Clin. 2025;75(2):111–40. pmid:39976243
  2. 2. Luo L, Wang X, Lin Y, Ma X, Tan A, Chan R, et al. Deep learning in breast cancer imaging: a decade of progress and future directions. IEEE Rev Biomed Eng. 2025;18:130–51. pmid:38265911
  3. 3. Ibrahim A, Mohamed HK, Maher A, Zhang B. A survey on human cancer categorization based on deep learning. Front Artif Intell. 2022;5:884749. pmid:35832207
  4. 4. Zhang XY, Wei Q, Wu GG, Tang Q, Pan XF, Chen GQ, et al. Artificial intelligence - based ultrasound elastography for disease evaluation - a narrative review. Front Oncol. (2023) 13:1197447.
  5. 5. Liu S, Wang Y, Yang X, Lei B, Liu L, Li SX, et al. Deep Learning in Medical Ultrasound Analysis: A Review. Engineering. 2019;5(2):261–75.
  6. 6. Dong S, Wang P, Abbas K. A survey on deep learning and its applications. Computer Science Review. 2021;40:100379.
  7. 7. Jalloul R, Chethan HK, Alkhatib R. A Review of Machine Learning Techniques for the Classification and Detection of Breast Cancer from Medical Images. Diagnostics (Basel). 2023;13(14):2460. pmid:37510204
  8. 8. Luo L, Wang X, Lin Y, Ma X, Tan A, Chan R, et al. Deep Learning in Breast Cancer Imaging: A Decade of Progress and Future Directions. IEEE Rev Biomed Eng. 2025;18:130–51. pmid:38265911
  9. 9. Chen J, Zhang W, Bao J, Wang K, Zhao Q, Zhu Y, et al. Implications of ultrasound-based deep learning model for preoperatively differentiating combined hepatocellular-cholangiocarcinoma from hepatocellular carcinoma and intrahepatic cholangiocarcinoma. Abdom Radiol (NY). 2024;49(1):93–102. pmid:37999743
  10. 10. Ottakath N, Al-Maadeed S, Zughaier SM, Elharrouss O, Mohammed HH, Chowdhury MEH, et al. Ultrasound-Based Image Analysis for Predicting Carotid Artery Stenosis Risk: A Comprehensive Review of the Problem, Techniques, Datasets, and Future Directions. Diagnostics (Basel). 2023;13(15):2614. pmid:37568976
  11. 11. Yu Y, Ouyang W, Huang Y, Huang H, Wang Z, Jia X, et al. Artificial intelligence-based multi-modal multi-tasks analysis reveals tumor molecular heterogeneity, predicts preoperative lymph node metastasis and prognosis in papillary thyroid carcinoma: a retrospective study. Int J Surg. 2025;111(1):839–56. pmid:38990290
  12. 12. Attention_Enhanced_InceptionNeXt-Based_Hybrid_Deep_Learning_Model_for_Lung_Cancer_Detection.
  13. 13. Han S, Kang H-K, Jeong J-Y, Park M-H, Kim W, Bang W-C, et al. A deep learning framework for supporting the classification of breast lesions in ultrasound images. Phys Med Biol. 2017;62(19):7714–28. pmid:28753132
  14. 14. Fujioka T, Kubota K, Mori M, Kikuchi Y, Katsuta L, Kasahara M, et al. Distinction between benign and malignant breast masses at breast ultrasound using deep learning method with convolutional neural network. Jpn J Radiol. 2019;37(6):466–72. pmid:30888570
  15. 15. Tanaka H, Chiu S-W, Watanabe T, Kaoku S, Yamaguchi T. Computer-aided diagnosis system for breast ultrasound images using deep learning. Phys Med Biol. 2019;64(23):235013. pmid:31645021
  16. 16. Stoffel E, Becker AS, Wurnig MC, Marcon M, Ghafoor S, Berger N, et al. Distinction between phyllodes tumor and fibroadenoma in breast ultrasound using deep learning image analysis. Eur J Radiol Open. 2018;5:165–70. pmid:30258856
  17. 17. Yu T-F, He W, Gan C-G, Zhao M-C, Zhu Q, Zhang W, et al. Deep learning applied to two-dimensional color Doppler flow imaging ultrasound images significantly improves diagnostic performance in the classification of breast masses: a multicenter study. Chin Med J (Engl). 2021;134(4):415–24. pmid:33617184
  18. 18. Dupont WD, Page DL, Parl FF, Vnencak-Jones CL, Plummer WD Jr, Rados MS, et al. Long-term risk of breast cancer in women with fibroadenoma. N Engl J Med. 1994;331(1):10–5. pmid:8202095
  19. 19. Kuzmiak CM, Calhoun BC. Pure mucinous carcinoma of the breast: radiologic-pathologic correlation. J Breast Imaging. 2023;5(2):180–7. pmid:38416927
  20. 20. Sonography. Linh - Mucinous breast carcinoma A case series and literature review. 2023.
  21. 21. Alper F, Abbasguliyev H, Ozmen S, Yalcin A, Yilmaz Cankaya B, Akcay MN. Clinical, histopathological, imaging, and treatment perspectives of inflammatory granulomatous mastitis: review of the literature. Eurasian J Med. 2023;54(Supp1):S172–8.
  22. 22. Barco-Manrique JF, Jaramillo-Botero N, Ochoa-Vera ME, Bolivar-Aguilera I, Vera-Campos SN. Mastitis granulomatosa: un reto radiológico. Revista chilena de radiología. 2021;27:127–31.
  23. 23. Makki J. Diversity of breast carcinoma: histological subtypes and clinical relevance. Clin Med Insights Pathol. 2015;8:23–31. pmid:26740749
  24. 24. Nguyen QD, He J. Invasive Ductal Carcinoma NST and special subtypes: radiology-pathology correlation. Curr Breast Cancer Rep. 2021;13(4):347–64.
  25. 25. Pan SJ, Yang Q. A survey on transfer learning. IEEE Trans Knowl Data Eng. 2010;22(10):1345–59.
  26. 26. V SS, S RIR. Segmentation of nano-particles from SEM images using transfer learning and modified U-Net. ijacsa. 2025;16(1).
  27. 27. He K, Zhang X, Ren S, Sun J. Deep Residual Learning for Image Recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2016. 770–8. https://doi.org/10.1109/cvpr.2016.90
  28. 28. Huang G, Liu Z, van der Maaten L, Weinberger KQ. Densely connected convolutional networks. 30th IEEE/CVF conference on computer vision and pattern recognition. 2017.
  29. 29. Sandler M, Howard A, Zhu M, Zhmoginov A, Chen L-C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2018. 4510–20. https://doi.org/10.1109/cvpr.2018.00474
  30. 30. Szegedy C, Wei Liu, Yangqing Jia, Sermanet P, Reed S, Anguelov D, et al. Going deeper with convoluntions. In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2015. 1–9. https://doi.org/10.1109/cvpr.2015.7298594
  31. 31. Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D. Grad-CAM: Visual explanations from deep networks via gradient-based localization. Int J Comput Vis. 2019;128(2):336–59.
  32. 32. Cheon D, Nam K, Oh S. Design and robust control of a precise torque controllable steering module for steer-by-wire systems. IEEE Trans Ind Electron. 2022;69(12):13245–54.