Correction
17 Nov 2023: The PLOS ONE Staff (2023) Correction: Comparison between single and serial computed tomography images in classification of acute appendicitis, acute right-sided diverticulitis, and normal appendix using EfficientNet. PLOS ONE 18(11): e0294794. https://doi.org/10.1371/journal.pone.0294794 View correction
Figures
Abstract
This study aimed to develop a convolutional neural network (CNN) using the EfficientNet algorithm for the automated classification of acute appendicitis, acute diverticulitis, and normal appendix and to evaluate its diagnostic performance. We retrospectively enrolled 715 patients who underwent contrast-enhanced abdominopelvic computed tomography (CT). Of these, 246 patients had acute appendicitis, 254 had acute diverticulitis, and 215 had normal appendix. Training, validation, and test data were obtained from 4,078 CT images (1,959 acute appendicitis, 823 acute diverticulitis, and 1,296 normal appendix cases) using both single and serial (RGB [red, green, blue]) image methods. We augmented the training dataset to avoid training disturbances caused by unbalanced CT datasets. For classification of the normal appendix, the RGB serial image method showed a slightly higher sensitivity (89.66 vs. 87.89%; p = 0.244), accuracy (93.62% vs. 92.35%), and specificity (95.47% vs. 94.43%) than did the single image method. For the classification of acute diverticulitis, the RGB serial image method also yielded a slightly higher sensitivity (83.35 vs. 80.44%; p = 0.019), accuracy (93.48% vs. 92.15%), and specificity (96.04% vs. 95.12%) than the single image method. Moreover, the mean areas under the receiver operating characteristic curve (AUCs) were significantly higher for acute appendicitis (0.951 vs. 0.937; p < 0.0001), acute diverticulitis (0.972 vs. 0.963; p = 0.0025), and normal appendix (0.979 vs. 0.972; p = 0.0101) with the RGB serial image method than those obtained by the single method for each condition. Thus, acute appendicitis, acute diverticulitis, and normal appendix could be accurately distinguished on CT images by our model, particularly when using the RGB serial image method.
Citation: Park SH, Kim YJ, Kim KG, Chung J-W, Kim HC, Choi IY, et al. (2023) Comparison between single and serial computed tomography images in classification of acute appendicitis, acute right-sided diverticulitis, and normal appendix using EfficientNet. PLoS ONE 18(5): e0281498. https://doi.org/10.1371/journal.pone.0281498
Editor: Kathiravan Srinivasan, Vellore Institute of Technology: VIT University, INDIA
Received: July 17, 2022; Accepted: January 24, 2023; Published: May 24, 2023
Copyright: © 2023 Park et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All relevant data are within the paper and its Supporting Information files.
Funding: This research was supported by the Ministry of Science and ICT (MSIT), Korea, under the Information Technology Research Center (ITRC) support program (IITP−2021−2017−0−01630) super-vised by the Institute for Information and communications Technology Promotion (IITP), the GRRC program of the Gyeonggi province (No. GRRC-Gachon2020 (B01), and the research was supported by G−ABC FRD2019−11−02(3).
Competing interests: The authors have declared that no competing interests exist.
Abbreviations: CNN, convolutional neural network; CT, computed tomography; RGB, red, green, blue; ROI, region-of-interest
Introduction
Acute lower abdominal pain is a common symptom among patients visiting hospitals [1, 2]. Abdominopelvic computed tomography (CT) is commonly used to evaluate acute right lower abdominal pain in adults with suspected acute appendicitis or acute right-sided diverticulitis [1, 3, 4]. Although radiologists can easily distinguish between these conditions on CT images, the rapidly increasing volume of CT examinations in the emergency room can place a burden on radiologists [5]. Distinguishing between these diseases and a normal appendix is also sometimes difficult for a physician (e.g., an internal medicine specialist or a surgeon). Although clinical presentations of the two diseases are similar, the treatments differ between surgical and medical treatments [4, 6–8], with acute appendicitis requiring surgery and acute diverticulitis requiring medical treatment in uncomplicated cases [9, 10]. Thus, correct classification of each condition in patients with acute abdominal pain is crucial for rapid and accurate decision-making regarding appropriate treatment [8].
A convolutional neural network (CNN) is a deep learning-based neural network and can be used to analyze radiological image data. A CNN was recently used to diagnose acute appendicitis automatically in a single-center, single-protocol cohort of 319 pre-contrast abdominopelvic CT images [11]. However, to the best of our knowledge, CNNs using contrast-enhanced abdominopelvic CT images to classify acute appendicitis, acute diverticulitis, and normal appendix have not been developed to date. Although CNNs using single-image analysis can outperform conventional machine-learning methods [12–14], they fail to leverage the depth information from CT images (e.g., the tubular structure of the appendix on serial axial CT images as opposed to the round structure of the appendix on a single cross-sectional axial CT image). Consecutive CT image slices with red, green, and blue (RGB) channel superposition provide information that can be used to improve the automated diagnostic performance, and, in cases of serial CT imaging, implement deep learning methods based on CNNs [15, 16].
The EfficientNet algorithm consists of compounded coefficients for scaling the depth, width, and resolution dimensions of a CNN [17]. EfficientNet, which presents a new approach for performing transfer learning in classification, reduces the analysis time and computing resources required, and consequently, its use in various deep learning studies has increased recently [18–21]. This algorithm is composed of 8 models, from B0 to B7, with increasing numbers referring to more parameters and higher accuracy [17, 22]. This influences the neural architecture used to explore the baseline. EfficientNet-B0 has a better trade-off on accuracy and calculation time (i.e., floating-point operations per second) [17, 23]. We have previously used the B0 model. which covers 4 million trainable parameters [24].
In the present study, we used a large dataset of CT images to develop and validate a CNN, using the EfficientNet algorithm and the corresponding training method, for the automated classification of acute appendicitis, acute diverticulitis, and normal appendix. We also compared the diagnostic performance of the CNN when applied to single CT images versus when applied to serial CT images with the RGB superposition method.
Methods
Study population
Approval for this retrospective study was obtained from the institutional review board (GDIRB2020-096) of the Gil medical center. The requirement for obtaining written informed patient consent was waived because of the retrospective nature of the study. The study was conducted according to the principles of the Declaration of Helsinki.
We searched the CT database and electronic medical records of patients who visited a hospital complaining of acute lower abdominal pain and who underwent contrast-enhanced abdominopelvic CT between January 2017 and August 2019 and were diagnosed with acute appendicitis or right-sided diverticulitis. For the acute appendicitis group, the inclusion criteria were as follows: (a) patients (≥ 20-years-old) with acute lower abdominal pain who were diagnosed with acute appendicitis that was detected via CT (i.e., we searched CT reports to find patients with acute appendicitis and reviewed their clinical symptoms) and surgically confirmed, and (b) who had undergone preoperative abdominopelvic CT within 1 week of surgery. The exclusion criteria were as follows: (a) cecal cancer with secondary appendicitis or appendiceal cancer with appendicitis and (b) severe motion artifacts on CT images. For the acute right-sided diverticulitis group, the inclusion criteria were as follows: (a) patients (≥ 20-years-old) with acute lower abdominal pain and acute right-sided diverticulitis detected via CT, who were (b) medically treated and improved during follow-up (> 1 month). The exclusion criteria were (a) bowel perforation due to acute diverticulitis and (b) severe motion artifacts on CT images. For the normal appendix group, we identified all patients aged ≥ 20 years and searched CT reports between January 2017 and December 2020 to find those with a normal appendix by using the key words: “normal appendix” (n = 410). The exclusion criteria were (a) CT images showing acute diverticulitis or colitis and (b) severe motion artifacts on CT images. Among these 410 cases, we randomly selected 215 patients, to ensure a similar number of CT slides (n = 1296) as for the acute appendicitis (n = 1959) and acute diverticulitis (n = 823) cases (Fig 1). Selected patients were retrospectively enrolled in the study, and their data were collected in our imaging laboratory.
CT imaging
The CT scanners used in this study included SOMATOM Edge, SOMATOM Definition AS, SOMATOM Definition Flash, and SOMATOM Force (Siemens Healthcare, Erlangen, Germany). The scan parameters varied: tube voltages of 80, 100, or 120 kVp and reference tube currents of 170–298 mAs. No low-dose CT protocol was used for these datasets. The CT images were reconstructed on the axial plane with a thickness of 2–5 mm.
Validation of the deep learning model
We used five-fold cross-validation and a test dataset to train the deep learning model. For each fold, the dataset samples were randomly assigned as follows: 60–70% to a training dataset to train and optimize the model; 15–20% to a validation dataset to evaluate the best configuration and tune the model parameters; and 15–25% to a test dataset to evaluate its classification performance. In addition, the training dataset was analyzed after the augmentation of acute appendicitis, diverticulitis, and normal appendix data to prevent training disturbances caused by data imbalances. More details on model training are provided in the Implementation and training of the deep learning model subsection of the methods section.
Preprocessing with region-of-interest extraction and RGB method superposition
We manually cropped box-shaped regions-of-interest (ROIs) at the same position across all axial planes from the start to the end positions of the visible targets for acute appendicitis, acute diverticulitis, and normal appendix. The cropped images were then resized to 224 × 224 pixels based on an image size conversion reference [25]. Next, we obtained three consecutive images from the CT slice sequence (sorted along the z-position) and represented them in the RGB method to acquire a new color image. Fig 2 illustrates the process for obtaining a CT image with the RGB superposition method.
The RGB superposition method enhances the differences in the connectivity and shape of the target in serial CT images. The connected region of a continuous slice is almost gray in color, whereas the unconnected region(s) show(s) different shapes (i.e., primary colors). A three-dimensional effect can be obtained from two-dimensional images based on the representation of the connectivity information pertaining to the organization that can be confirmed by the three-dimensional volume using the RGB color model.
Implementation and training of the deep learning model
We used the IBM Power System AC922 server with four Tesla V100s, NVLink GPUs, and 16 GB memory (NVIDIA Corp., Sta. Clara, CA, USA) to run the deep learning model based on the EfficientNet-B0 architecture (Fig 3A). The code was implemented in Python 3.7.11 and TensorFlow frameworks (Version 2.2.0) were run on the Ubuntu 18.04.3 operating system. The code for image pre-processing was implemented on OpenCV-Python 3.4.10, Pydicom 2.2.2 and read-roi 1.6.0.
(a) EfficientNet-B0 architecture (b) Example of a prediction outcome label.
For training, we set the batch size to 64, number of epochs to 200, and used Adam as the optimizer [26]. We adapted early stopping to avoid overfitting during training; thus, the training was stopped when the validation loss did not improve after over 10 epochs. We described the loss function using categorical cross-entropy [27]. In addition, we set the learning rate to 0.0001 (1e-4), and for cases with more than 30 epochs, the learning rate was set to 0.00001 (1e−5) to ensure that this rate does not increase significantly as learning progresses. Application of the deep learning model yielded the prediction probability for the three statuses (probability for acute appendicitis, acute diverticulitis, or normal appendix). Among the three, the status with the highest probability was taken as the prediction label (Fig 3B).
Statistical analysis
The diagnostic accuracy, sensitivity, and specificity of CNN classification of acute appendicitis, acute diverticulitis, and normal appendix using the single image and serial image RGB methods were calculated and determined by considering 95% confidence intervals. The CNN classification performances of the two methods were compared in terms of the area under the curve (AUC) values, obtained from receiver operating characteristic curves, and sensitivity, using the McNemar test. P < 0.05 was considered as statistically significant. The statistical analyses were conducted using the MedCalc software (version 18.11.3) and IBM SPSS Statistics for Windows (version 21.0; IBM Corp., Armonk, NY, USA).
Results
We analyzed the training, test, and validation datasets using the data obtained from 715 consecutive patients (347 women, 368 men; age, 44.3 ± 18.4 years). Of these, 246 had acute appendicitis, 254 had acute diverticulitis, and 215 had normal appendix. The details of the test dataset are presented in Table 1. The test dataset contained 4,078 CT image slices (acute appendicitis, 1,959; acute diverticulitis, 823; normal appendix, 1,296), which were derived from the scans of 715 patients. More details on the model case and the data obtained using five-fold cross-validation are presented in Table 2 and S1 Appendix. As can be seen, the amount of training data (number of CT slides) among the three groups after data augmentation are relatively similar.
The diagnostic performance of the two CNN classification approaches are listed in Table 3. For the classification of acute appendicitis, the RGB method demonstrated slightly higher sensitivity (87.85% vs. 85.60%; p = 0.047), accuracy (87.94% vs. 86.07%), and specificity (88.01% vs. 96.04%) than the single image method. For the classification of acute diverticulitis, the RGB method yielded slightly higher sensitivity (83.35 vs. 80.44%; p = 0.019), accuracy (93.48% vs. 92.15%), and specificity (96.04% vs. 95.12%) than the single image method. For the classification of normal appendix, the RGB method showed slightly higher sensitivity (89.66 vs. 87.89%; p = 0.244), accuracy (93.62% vs. 92.35%), and specificity (95.47% vs. 94.43%) than the single method. For classification of the three statuses, the RBG method revealed a higher overall accuracy than the single method (87.51% vs. 85.29%). The RGB method showed a significantly higher mean AUC for acute appendicitis (0.951 vs. 0.937; p < 0.0001), acute diverticulitis (0.972 vs. 0.963; p = 0.0025), and normal appendix (0.979 vs. 0.972; p = 0.0101) than the single image method (Fig 4).
Area under the receiver operating characteristic curve for classifying acute appendicitis (blue), acute diverticulitis (orange), and normal appendix (green) using (a) the single image and (b) serial image red, green, blue (RGB) methods.
Detailed comparisons between the single image and RBG methods in the classification of the three statuses are presented in Table 4 and Figs 5–7. Among the 1,959 CT images of acute appendicitis, 146 and 123 images were misclassified as acute diverticulitis and 136 and 115 images were predicted to be normal appendices by the single image and RBG methods, respectively. In Among the 823 CT images of acute diverticulitis, 142 and 126 images were predicted to be acute appendicitis and 19 and 11 images were predicted to be normal appendices using these two methods, respectively. Among the 1,296 CT images of normal appendices, 144 and 128 images were misclassified as representing acute appendicitis using the two methods, respectively.
Examples of acute appendicitis cases correctly classified by the convoluted neural network by (CNN) using the (a) single image and (b) red, green, blue (RGB) methods, with and without class activation maps (CAMs).
Examples of acute diverticulitis cases correctly classified by the convoluted neural network (CNN) using the (a) single image and (b) red, green, blue (RGB) methods, with and without class activation maps (CAMs).
Examples of normal appendix cases correctly classified by the convoluted neural network (CNN) using the (a) single image and (b) red, green, blue (RGB) methods, with and without class activation maps (CAMs).
Discussion
We developed and applied a CNN model using the EfficientNet-B0 architecture to classify acute appendicitis, acute right-sided diverticulitis, and normal appendix from single and serial RGB contrast-enhanced abdominopelvic CT images. All the AUC values obtained using the RGB method were higher than those obtained using the single image method.
The sensitivity (87.89, 89.66%; single, RGB) and specificity (94.43, 95.47%; single, RGB) values for normal appendix were the highest among the three diseases. Appropriate classification of a normal appendix using a CNN could help inexperienced physicians identify a normal appendix from CT images in an emergency room. To improve the performance of our CNN, we applied data augmentation on the training dataset and used the EfficientNet-B0 architecture. Although we enrolled a similar number of patients in each of the three groups, the amount of CT data in these groups varied significantly. For example, the appendix dataset involved a larger number of CT images than the acute diverticulitis dataset, because the appendix has a tubular structure, as can be seen from the several images on the axial CT scan, whereas diverticulitis has a small, round shape. For this reason, we applied data augmentation on the training dataset (particularly on the diverticulitis dataset) to minimize data imbalance to ensure that the training effectiveness is not hampered.
Using a region of low entropy and a fully convolutional network, Walid et al. [11] designed a CNN that could detect acute appendicitis on non-contrast CT images with a specificity of 68.8–92.6%. In the classification of acute appendicitis, our CNN was comparably more specific (86.50–88.01%). Unlike the model developed by Walid et al. [11], ours was developed using contrast-enhanced CT scans from a larger cohort. In practice, contrast-enhanced CT is more commonly used than non-enhanced CT in patients with suspected acute appendicitis. Our results showed that the developed model can be applied to clinical conditions when physicians need to classify acute appendicitis, normal appendix, and diverticulitis. Rajpurkar et al. [28] recently developed a CNN using video pre-training to classify acute appendicitis on contrast-enhanced CT scans. They used eight CT slices from eight sequential CT images, whereas we used three serial CT images with the RGB superposition method, including the information from continuous CT images (i.e., previous, current [lesion], and next image). This allowed us to establish the three-dimensional volume of the lesion using the CNN. Our CNN maximizes the diagnostic performance of image classification by leveraging deep learning and maintains a low error rate while fine-tuning the hyperparameters from small available datasets [29]. Moreover, it can adequately handle accuracy saturation with increasing network depth [29].
The EfficientNet algorithm models can scale the depth, width, and resolution dimensions of a CNN [17]. Transfer learning using the ImageNet dataset can save time and computing resources [17]. The EfficientNet algorithm provides superior efficiency and higher accuracy than other CNNs, including GoogleNet, AlexNet, and MobileNetV2 [17]. In this study, we trained our CNN model using EfficientNet-B0 architecture to classify acute appendicitis, acute diverticulitis, and normal appendix. The model was determined by considering the image resolution of the original image (before image resizing), while the training time affected the number of parameters [30].
We hypothesized that the use of three serial CT images in the RGB method would improve the CNN performance for classifying the three output labels (acute appendicitis, acute diverticulitis, and normal appendix) by enhancing the spatial features of the target lesion. Few studies have reported CNNs using three CT slices with the RGB method [15, 16, 31] or multiple CT slices from videos [28, 32, 33] to enhance spatiotemporal features. The sensitivity of the model for classifying acute diverticulitis was slightly higher when using the RGB method (83.35%) than when using the single image method (80.44%), as was the specificity (96.04% and 95.12%, respectively). The sensitivity and specificity for classifying acute appendicitis were also slightly higher when using the RGB method (87.85% and 88.01%, respectively) than when using the single image method (85.60% and 86.50, respectively). The AUCs obtained with the RGB method (0.951 for acute appendicitis; 0.972 for acute diverticulitis; and 0.979 for normal appendix) were significantly higher than those obtained with the single image method (0.937, 0.963, and 0.972, respectively; P < 0.01). Therefore, spatial information from multiple slices may contribute to the classification accuracy when using contrast-enhanced abdominopelvic CT scans.
There are some limitations of our study that must be addressed. First, because this was a retrospective study, a selection bias may have occurred. Thus, a prospective study should be conducted to validate our classification results. Second, the number of CT images diagnosed as acute diverticulitis was relatively smaller than those diagnosed as acute appendicitis or normal appendix. Therefore, we analyzed the training dataset after augmenting each CT dataset to minimize the training disturbance caused by unbalanced datasets. Third, we did not include cases of acute diverticulitis with complications (e.g., perforation or abscess) as such cases were beyond the scope of our study but will be addressed in our future studies. Fourth, we did not develop a localization tool to detect normal appendix, appendicitis, or acute diverticulitis. We therefore believe that identifying the location of such conditions using artificial intelligence will be helpful [11]. The development of such detection tools will be attempted in a future study. Finally, we did not assess the CNN performance using coronal reformatted CT images. Notably, combining axial and coronal CT images may enhance CNN performance, and we plan to investigate this combination in the future.
In conclusion, we developed an automated CNN-based model employing EfficientNet for classifying acute appendicitis, acute diverticulitis, and normal appendix on contrast-enhanced CT scans. All the aforementioned conditions were better classified by the CNN when serial images were used with the RGB method than when a single image method was used. This presumably reflects the detailed and rich spatial information provided by the former, which leverages information from multiple CT images.
Supporting information
S1 Appendix. Five-fold cross-validation.
(a,b) Single methods (c,d) RGB methods for classification of acute appendicitis, acute diverticulitis, and normal appendix using EfficientNet.
https://doi.org/10.1371/journal.pone.0281498.s002
(DOCX)
References
- 1. Radiology ACo. ACR Appropriateness Criteria® Right lower quadrant pain-Suspected appendicitis. [cited 2020 21 February][Internet]. Available from: https://acsearch.acr.org/docs/69357/Narrative/.
- 2. Jacobs DO. Diverticulitis. New England Journal of Medicine. 2007;357(20): 2057–2066.
- 3. Stoker J, van Randen A, Laméris W, Boermeester MA. Imaging patients with acute abdominal pain. Radiology. 2009;253(1): 31. pmid:19789254
- 4. Song JH, Kim YW, Lee S, Do HH, Seo JS, Lee JH, et al. Clinical difference between acute appendicitis and acute right-sided colonic diverticulitis. Emergency Medicine International. 2020;2020. pmid:32953179
- 5. Wang DC, Parry CR, Feldman M, Tomlinson G, Sarrazin J, Glanc P. Acute abdomen in the emergency department: is CT a time-limiting factor? American Journal of Roentgenology. 2015;205(6): 1222–1229. pmid:26587929
- 6. Kahveci S, Tokmak TT, Yıldırım A, Mavili E. Acute right-sided colonic diverticulitis mimicking appendicitis: a report of two cases. J Clin Ultrasound. 2013;41(4): 238–241. pmid:22855407
- 7. Anderton M, Griffiths B, Ferguson G. Giant sigmoid diverticulitis mimicking acute appendicitis. Ann R Coll Surg Engl. 2011;93(6): e89–90. pmid:21929895
- 8. Kruis W, Morgenstern J, Schanz S. Appendicitis/diverticulitis: diagnostics and conservative treatment. Dig Dis. 2013;31(1): 69–75. pmid:23797126
- 9. Stollman N, Smalley W, Hirano I, Adams MA, Dorn SD, Dudley-Brown SL, et al. American Gastroenterological Association Institute guideline on the management of acute diverticulitis. Gastroenterology. 2015;149(7): 1944–1949. pmid:26453777
- 10. Ergenç M, Uprak TK. Appendiceal Diverticulitis Presenting as Acute Appendicitis and Diagnosed After Appendectomy. Cureus. 2022;14(3): e23050. pmid:35464547
- 11. Al WA, Yun ID, Lee KJ. Reinforcement Learning-based Automatic Diagnosis of Acute Appendicitis in Abdominal CT. arXiv preprint arXiv:190900617. 2019.
- 12. Rawat W, Wang Z. Deep convolutional neural networks for image classification: A comprehensive review. Neural computation. 2017;29(9): 2352–2449. pmid:28599112
- 13. Yamashita R, Nishio M, Do RKG, Togashi K. Convolutional neural networks: an overview and application in radiology. Insights into Imaging. 2018;9(4): 611–629. pmid:29934920
- 14. Soffer S, Ben-Cohen A, Shimon O, Amitai MM, Greenspan H, Klang E. Convolutional Neural Networks for Radiologic Images: A Radiologist’s Guide. Radiology. 2019;290(3): 590–606. pmid:30694159
- 15. Ye W, Gu W, Guo X, Yi P, Meng Y, Han F, et al. Detection of pulmonary ground-glass opacity based on deep learning computer artificial intelligence. Biomedical engineering online. 2019;18(1): 6. pmid:30670024
- 16. Gao M, Jiang J, Zou G, John V, Liu Z. RGB-D-based object recognition using multimodal convolutional neural networks: A survey. IEEE Access. 2019;7: 43110–43136.
- 17.
Tan M, Le Q. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In: Kamalika C, Ruslan S, editors. Proceedings of the 36th International Conference on Machine Learning. PMLR; 2019. p. 6105–6114. Available from: https://proceedings.mlr.press/v97/tan19a.html
- 18. Marques G, Agarwal D, de la Torre Díez I. Automated medical diagnosis of COVID-19 through EfficientNet convolutional neural network. Appl Soft Comput. 2020;96: 106691–106691. pmid:33519327
- 19. Duong LT, Nguyen PT, Di Sipio C, Di Ruscio D. Automated fruit recognition using EfficientNet and MixNet. Computers and Electronics in Agriculture. 2020;171: 105326.
- 20. Alhichri H, Alswayed AS, Bazi Y, Ammour N, Alajlan NA. Classification of Remote Sensing Images Using EfficientNet-B3 CNN Model With Attention. IEEE Access. 2021;9: 14078–14094.
- 21.
Mantha T, Reddy BE, editors. A Transfer Learning method for Brain Tumor Classification using EfficientNet-B3 model. 2021 IEEE International Conference on Computation System and Information Technology for Sustainable Solutions (CSITSS); 2021 16–18 Dec. 2021.
- 22.
Sandler M, Howard A, Zhu M, Zhmoginov A, Chen L-C, editors. Mobilenetv2: Inverted residuals and linear bottlenecks. Proceedings of the IEEE conference on computer vision and pattern recognition; 2018.
- 23.
Tan M, Le Q. EfficientNetV2: Smaller Models and Faster Training. In: Marina M, Tong Z, editors. Proceedings of the 38th International Conference on Machine Learning. PMLR; 2021. p. 10096–10106. Available from: https://proceedings.mlr.press/v139/tan21a.html
- 24. Fu Y. Image classification via fine-tuning with Efficientnet. Keras Disponıvel em https://kerasio/examples/vision/image_classification_efficientnet_fine_tuning/Acessoem. 2020;11.
- 25.
Ferreira CA, Melo T, Sousa P, Meyer MI, Shakibapour E, Costa P, et al., editors. Classification of breast cancer histology images through transfer learning using a pre-trained inception resnet v2. International conference image analysis and recognition; 2018: Springer.
- 26. Zhang Z, Sabuncu M. Generalized cross entropy loss for training deep neural networks with noisy labels. Advances in neural information processing systems. 2018;31.
- 27. Kingma DP, Ba J. Adam: A method for stochastic optimization. arXiv preprint arXiv:14126980. 2014.
- 28. Rajpurkar P, Park A, Irvin J, Chute C, Bereket M, Mastrodicasa D, et al. AppendiXNet: Deep Learning for Diagnosis of Appendicitis from A Small Dataset of CT Exams Using Video Pretraining. Scientific reports. 2020;10(1): 3958. pmid:32127625
- 29.
He K, Zhang X, Ren S, Sun J, editors. Deep residual learning for image recognition. Proceedings of the IEEE conference on computer vision and pattern recognition; 2016.
- 30. Munien C, Viriri S. Classification of Hematoxylin and Eosin-Stained Breast Cancer Histology Microscopy Images Using Transfer Learning with EfficientNets. Comput Intell Neurosci. 2021;2021: 5580914. pmid:33897774
- 31. Lee JH, Kim YJ, Kim YW, Park S, Choi YI, Kim YJ, et al. Spotting malignancies from gastric endoscopic images using deep learning. Surgical endoscopy. 2019;33(11): 3790–3797. pmid:30719560
- 32. Le QV, Zou WY, Yeung SY, Ng AY, editors. Learning hierarchical invariant spatio-temporal features for action recognition with independent subspace analysis. CVPR 2011; 2011 20–25 June 2011.
- 33.
Wang X, Gupta A, editors. Unsupervised learning of visual representations using videos. Proceedings of the IEEE International Conference on Computer Vision; 2015.