Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

CO-IRv2: Optimized InceptionResNetV2 for COVID-19 detection from chest CT images

  • M. Rubaiyat Hossain Mondal ,

    Roles Conceptualization, Project administration, Supervision, Writing – original draft, Writing – review & editing

    rubaiyat97@iict.buet.ac.bd

    Affiliation Institute of Information and Communication Technology, Bangladesh University of Engineering and Technology, Dhaka, Bangladesh

  • Subrato Bharati,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Writing – original draft, Writing – review & editing

    Affiliation Institute of Information and Communication Technology, Bangladesh University of Engineering and Technology, Dhaka, Bangladesh

  • Prajoy Podder

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Validation, Writing – review & editing

    Affiliation Institute of Information and Communication Technology, Bangladesh University of Engineering and Technology, Dhaka, Bangladesh

Abstract

This paper focuses on the application of deep learning (DL) in the diagnosis of coronavirus disease (COVID-19). The novelty of this work is in the introduction of optimized InceptionResNetV2 for COVID-19 (CO-IRv2) method. A part of the CO-IRv2 scheme is derived from the concepts of InceptionNet and ResNet with hyperparameter tuning, while the remaining part is a new architecture consisting of a global average pooling layer, batch normalization, dense layers, and dropout layers. The proposed CO-IRv2 is applied to a new dataset of 2481 computed tomography (CT) images formed by collecting two independent datasets. Data resizing and normalization are performed, and the evaluation is run up to 25 epochs. Various performance metrics, including precision, recall, accuracy, F1-score, area under the receiver operating characteristics (AUC) curve are used as performance metrics. The effectiveness of three optimizers known as Adam, Nadam and RMSProp are evaluated in classifying suspected COVID-19 patients and normal people. Results show that for CO-IRv2 and for CT images, the obtained accuracies of Adam, Nadam and RMSProp optimizers are 94.97%, 96.18% and 96.18%, respectively. Furthermore, it is shown here that for the case of CT images, CO-IRv2 with Nadam optimizer has better performance than existing DL algorithms in the diagnosis of COVID-19 patients. Finally, CO-IRv2 is applied to an X-ray dataset of 1662 images resulting in a classification accuracy of 99.40%.

1. Introduction

In late 2019, China first reported novel coronavirus disease (COVID-19) caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) [1]. In March 2020, the World Health Organization (WHO) declared COVID-19 as a pandemic. Some common symptoms related to COVID-19 include cough, fever, dyspnea, and sore throat [24]. The respiratory droplets and the interaction between human-to-human are the leading causes of the spread of this infection [5]. The incubation time of COVID-19 is significantly longer, causing difficulty in the control of this pandemic. Moreover, some COVID-19 patients are asymptomatic, which causes the spreading of the infection [6]. Due to these causes, the early recognition, diagnosis, and control of infectious diseases like COVID-19 are considered hard.

Efforts have been made to find an efficient method for diagnosing COVID-19. RT-PCR test is an authorized method for detecting COVID-19 using a nasopharyngeal or oral swab [7]. The RT-PCR test is time-consuming and complex. Besides, RT-PCR test has some limitations, which result in insufficient sensitivity and a higher false-negative rate [8, 9]. Several patients who are incorrectly detected as negative are likely to interact with a massive number of people. Hence, a superior diagnosis technique with a low false-negative rate is required to control the risk of COVID-19. Nevertheless, radiology imaging, i.e., X-rays [10], ultrasound, magnetic resonance imaging (MRI), and computed tomography (CT) image scanning are also applied to diagnose various diseases, including COVID-19 [1018]. Sometimes underlying pathology may hide viral infection, such as in patients with pulmonary fibrosis. The diagnosis of COVID-19 is considerably hard. However, sometimes X-rays can provide a false-negative rate. In contrast, chest CT scan images provide detail view useful for the rapid diagnosis of COVD-19. This also helps in the timely isolation of infected patients and thus managing COVID-19 [2, 1921].

Concisely, the early and accurate diagnosis and treatment of COVID-19 suspects can play a significant role in treating patients. This is important for the COVID-19 control, prognosis of patients, and the security of public health. Chest CT can detect small areas of ground-glass opacity (GGO) for its high resolution. In the early phases of COVID-19-pneumonia, pulmonary outcomes in lung CT images may be present with peripheral, subpleural, and minor GGOs [22], which may take much longer than large-involved and isolated GGOs or/and patterns of integration. Moreover, radiologists’ visual fatigue will increase the possible risks of missing a diagnosis for several minor lesions. Hence, it is essential to develop an artificial intelligence (AI) system for the computer-aided diagnosis of COVID-19.

As a subset of AI, deep learning (DL) has been applied to automatically detect respiratory diseases and illnesses [2328]. However, lesions need to be distinguished for most DL-based approaches for the diagnosis of CT segments. Harmful lesions of COVID-19 cost much effort for radiologists, while COVID-19 spreads rapidly and is not adaptable for radiologists. One of the most accessible labels to diagnose COVID-19 is the patient-level label, where patients are labeled negative or positive. An appropriate DL-based scheme can be crucial for the diagnosis of COVID-19. Hence, this work explores a DL model for the automatic diagnosis of COVID-19.

The main contributions of this paper are summarized as follows:

  • A new approach termed as Optimized InceptionResNetV2 for COVID-19 (CO-IRv2) is proposed in this work. Hyperparameter tuning is conducted for the optimization of CO-IRv2.
  • A dataset of 2481 CT images is formed by combining two different databases.
  • The proposed CO-IRv2 is applied to the new dataset of 2481 CT images and an existing dataset of X-ray images. The effectiveness of several optimizers in CO-IRv2 is evaluated in classifying normal people and suspected COVID-19 patients.

This paper is structured as follows. The background of these studies is introduced in Section 2. Moreover, in Section 3, dataset descriptions are briefly presented. Furthermore, methodology and experimental discussions are provided in Section 4 and Section 5, respectively. Finally, the concluding remarks are provided in Section 6.

2. Literature review

A number of research work proposed various DL models for detecting or diagnosis of COVID-19. These studies are summarized in Table 1 and are described in the following.

thumbnail
Table 1. Summary of literature reviews for CT images using DL.

https://doi.org/10.1371/journal.pone.0259179.t001

Wu et al. [29] proposed a framework of coronavirus screening based on ResNet50, a modified form of CNN. The dataset was collected from two hospitals where the images of COVID-19 and other cases of pneumonia were 368 and 127 images, respectively. All images were resized to a size of 256×256 by using image augmentation. They achieved a specificity of 61.5%, a recall of 81.1%, an accuracy of 76% and an area under the receiver operating characteristics curve (AUC) of 81.9% after preprocessing. The authors of [29] created a validation, training, and testing set using just slices of the maximal lung area in three perspectives. Additional CT slices in various perspectives were also provided as input to the suggested model to increase its performance. Other clinical data, such as body mass index and severity of COVID-19 infection, were not available for subgroup analysis. Nonetheless, this study’s age and gender subgroup analyses enabled an initial assessment of the model. Despite the importance of this work, more powerful AI algorithms need to be created to facilitate and accurately diagnose COVID-19. The model in [29] could not minimize misdiagnosis and thus affected clinical outcomes.

The researchers of [30] classified different types of diseases like Influenza-A viral pneumonia and COVID-19 pneumonia from CT images using CNN variants such as ResNet18. They used 618 CT images from three different datasets provided by different hospitals. Their proposed model provided 80.8% precision, 81.5% recall, 86.7% accuracy and 81.1% F1-score [30]. As reported in [30], the symptoms of COVID-19 could overlap with those of other pneumonia, including eosinophilic pneumonia, organizing pneumonia and influenza-A viral pneumonia (IAVP). The study in [30] compared the CT manifestations of IAVP and COVID-19. One limitation of the work was the use of limited number of samples. Moreover, the work [30] did not address the complicated clinical condition and further multicenter clinical trials.

The dataset of the work of modified ResNet50v2 [31] included 48,260 CT scanning images of 282 normal people and 15,589 images of 95 COVID-19 patients. The initial stage in the image processing technique was to evaluate the view of the lung and discard any CT images that were not properly visible in the lung. This significantly lowered processing time and false detections. The authors then created a novel architecture to categorize the ResNet50v2 model over a range of image resolutions, ensuring that the model did not lose data from small objects. The system evaluated the patient’s condition using a predefined threshold. The authors tested their system in different methods using ResNet50v2, Xception, and modified ResNet50v2 [31].

In [32], a self-built model named CTnet-10 was developed with an accuracy of 82.1% for the diagnosis of COVID-19. Additionally, the authors experimented with other models, including VGG-16, VGG-19, ResNet-50, DenseNet169, and InceptionV3. With an accuracy of 94.52%, the VGG-19 outperformed all other DL models considered in the study. The work [32] only used RMSProp optimizer for few CT images. U-Net CNN architecture was proposed to detect COVID-19 from 540 images [33], where the U-Net was applied for lung segmentation. The results of the segmentation were provided as the 3D-CNN input for the prediction and possibility of COVID-19. Their models provided 91.1% specificity, 90.7% recall, and 95.9% AUC [33]. One limitation was that the lung segmentation did not include temporal information and was trained using imprecise ground-truth masks. Another limitation was that the research data were derived from a single hospital, and no cross-center validations were conducted.

Moreover, the work [34] proposed a 2D CNN model for scanning viral pneumonia and COVID-19 from 1065 CT images. The authors of [34] also modified the inception DL algorithm to establish the model, followed by external and internal validation, where external validation showed 83% specificity with 79.3% accuracy. On the other hand, internal validation obtained 87% recall, 88% specificity, and 89.5% accuracy [34]. Several factors, including low signal to noise and complex data integration affected the efficacy of the DL used in [34]. Because of the relatively large amount of CT scan parts, especially those irrelevant for diagnosing pneumonia, classification was a difficult task.

The authors of [35] considered CT images and built a U-Net++ segmentation model to detect COVID-19. They invented the VB-Net method for extracting lung regions and infected lungs. They provided accurate medical research quantification data, including a quantitative assessment of progression and disease prediction. The study obtained successful performance by including the human loop method into the development of a segmentation network based on the VB-Net. The drawbacks of this study were its retrospective nature, selection bias (absence of severe COVID-19 patients), small sample size, and evaluation bias in the radiologist-defined CT score. Another study considered 2D and 3D U-Net models for COVID-19 detection [36]. In addition, the authors of [37] suggested a 2D DL architecture with a U-Net backbone to detect lung areas. The authors conducted two-division tasks: the first was a segmented abnormality in the chest CT scan specific to COVID-19 infection, and the second was a segmented CT lung space. Together with the volumetric assessment, the two segmentation tasks enabled a chest CT scan prediction to give statistical information on COVID-19 anomalies. The authors got 84.7% of mean intersection over union (M-IoU) and an F1 score of 97.32%. For semantic segmentation, their research utilized lung CT image sets from Kaggle and GitHub.

The work of [38] automatically extracted radiographic features indicating developing pneumonia from a radiograph, most notably the GGO. They developed a model of Deep Pneumonia to assist doctors in recognizing and diagnosing COVID-19 pneumonia. Their strategy was divided into three phases. First, they isolated the damaged lung areas and used lung segmentation filter banks to eliminate any noise. The second step utilized the deep temporal regression network (DTRN) to extract top characteristics from CT images and to provide picture-level predictions. Image-level predictions were aggregated in the last stage to generate a patient-level diagnosis. Their model was based on 88 COVID-19 patients and 100 patients with bacterial pneumonia from 777 image datasets. They achieved an AUC of 0.95 for DRE-Net, which is much higher than the AUCs reported for other models such as ResNet, VGG16, and DenseNet. The study’s drawback was the use of a small number of CT scan images to detect COVID-19.

The authors of [39] developed a DL ensemble model for identifying COVID-19 CT images. A total of 2933 COVID-19 lung CT pictures were gathered from public sources, past publications, and major media stories. The images were preprocessed in order to get 2500 images of exceptional quality. A hospital obtained 2500 CT scans of a lung tumor and 2500 normal lung images. Transfer learning was used to establish the model parameters, and three deep convolutional network models were pre-trained: GoogleNet, ResNet, and AlexNet. All the images were extracted using these models. Softmax was used to classify the layers in a completely linked manner. A relative majority vote was used for classification of the images.

In one study, a total of 15 pre-trained CNN architectures were used: ResNet-50, ResNext50, SeResnet 50, DenseNet121, EfficientNets(B0-B5), Xception, NasNetMobile, NasNetLarge, InceptionV3, and Inception ResNetv2 [40]. Utilizing the optimum mix of deep transfer learning output, the work then created a group method based on a majority vote to further improve the classification performance. The authors of [40] analyzed a publicly available dataset of CT scans that included 349 CT scans categorized as COVID-19 positive, and 397 CT scans having COVID-19 negative samples [40].

Furthermore, the authors of [41] established a DL method named CovidCTNet to diagnose COVID-19 infection from CT images. This work applied U-Net model for developing BCDU-Net architecture. This algorithm could distinguish CAP, COVID-19, and control lungs from CT images. They used 89145 images with 32230 CT images of COVID-19 confirmed patients, 25699 images of CAP, and 32216 images of other disorders or healthy lungs. For the evaluation, the holdout method was used with 90% of data samples used for training, and 10% for testing. Their proposed method provided a recall of 87.5%, an AUC of 95%, a specificity of 94%, and an accuracy of 91.66% [41]. A novel DL method CRNet was proposed for COVID-19 detection from 746 CT images, where the images from three open-access datasets were divided into 60% training, 25% testing, and 15% validation [42]. The method achieved a value of 94% AUC, 85% F1-score, and 86% accuracy [42].

In [43], a DL scheme containing a multi-layer perceptron, two decoders, and an encoder was applied to 1044 images [43]. The number of COVID-19, healthy individuals, lung cancer patients and different types of pathology were 449, 100, 98 and 397, respectively. The data was split into 80% training, 10% testing and 10% validation. This system obtained the highest AUC, recall, specificity and accuracy of 93%, 94%, 79% and 86%, respectively [43]. Aside from the numerous advantages of using CT images to identify and isolate early COVID-19 patients, DL methods based on CT images can also aid physicians in combating this disease [44]. This is because DL can be used not only to segment and classify images in the field of healthcare, but also to predict treatment outcomes [44]. Another effective technique is to utilize unsupervised learning algorithms to identify the picture and identify lesions when only a limited dataset is accessible [45]. These marginally better tactics could help fight against the coronavirus COVID-19, where just a few databases are usually available, and clinicians cannot provide a big volume of labeled data.

Zhu et al. [46] proposed DL methods where they used multi-task learning and single-task learning. They used 408 real CT images from two different sources. They achieved 85.91% accuracy for their proposed best model [46]. However, the focus of that study was to predict whether the patient would develop severe symptoms based on mild symptoms. That would result in a classification imbalance because only a small number of patients would progress to severe conditions. First, the imbalanced categorization of the dataset was biased, with 86 severe instances compared to 322 non-severe instances. That complicated the process of developing good classification models [46]. Moreover, another DL model, Elastic Net [47] segmented 1224 images for training and testing, where the total images were 1381, including 1200 non-COVID-19 images and 181 COVID-19 images. The combination of 3D and 2D CNN was used for segmentation purposes. Next, the dataset was evaluated by Elastic Net. This model obtained an AUC of 88.2%, while the testing data had 641 patients. The highest AUC, recall, accuracy and specificity were 94%, 79%, 90% and 91%, respectively [47].

Narin et al. obtained 96.10% accuracy when they applied ResNet50 model to 3141 X-ray images where 341 are of COVID-19 patients, and the remaining are of normal people [48]. One study [49] considered a dataset of 224 COVID-19 X-ray images and obtained a classification accuracy of 98.75% using pre-trained DL methods. Another study used ResNet50 along with a support vector machine to 381 X-ray images resulting in an accuracy of 98.66% [50]. A DL model achieved 98.08% accuracy when applied to detect COVID-19 patients from an X-ray dataset where 127 cases are for COVID-19 patients [16].

The above studies reveal that the lack of large COVID-19 datasets is challenging to validate different DL models. Some of the existing models described above were only tested for CT images, so their effectiveness in the case of X-ray images was not reported. Some others only focused on X-ray images without considering CT images at all. Furthermore, most of these works split image samples using the holdout method without considering the cross-validation method. Hence, this paper generates a new dataset of 2481 CT images by combining two different databases. This paper proposes a new DL algorithm CO-IRv2, and its performance is evaluated for CT images and X-ray images. Furthermore, this paper considers the holdout method as well as cross-validation methods for splitting the training and testing images.

3. Description of database

Our experimental dataset was retrieved from two different open access sources [51, 52]. The dataset of [51] had 829 CT images. On the other hand, the dataset of [52] had 2482 CT images where normal and COVID-19 patients were 1230 and 1252, respectively. After collecting the CT images, we formed a database where we retrieved the 829 images from the dataset reported in [51], and the rest of the images were retrieved from the dataset available in [52]. The total CT images for our experiments were 2481, where the images of normal and COVID-19 patients were 1229 and 1252, respectively. The resultant dataset is now made available in [53]. Fig 1 shows samples of our generated dataset [53], Fig 1(A) is for non-COVID and Fig 1(B) is for COVID-19 cases. Next, we split our database into training and testing. The details of the training and testing samples are shown in Table 2. This work also considers a dataset of X-ray images [54] containing 1662 images, of which 1583 images are of normal people and 79 are of COVID-19 patients.

thumbnail
Fig 1.

Sample of the (a) non-COVID-19 and (b) COVID-19 CT scan dataset [53].

https://doi.org/10.1371/journal.pone.0259179.g001

thumbnail
Table 2. The number of CT images per class applied in training and testing stages.

https://doi.org/10.1371/journal.pone.0259179.t002

4. Methodology

The experimentation had several stages. Data resizing and normalization were performed to facilitate generalization and avoid overfitting. The dataset was divided into two parts: training and testing. The proposed CO-IRv2 model was trained using training data. The evaluation was run up to 25 epochs. The proposed model got the desired accuracy within 25 epochs. Next, we performed hyperparameters’ fine-tuning on the proposed model. The CO-IRv2 model avoided underfitting by using several dense layers described in Section 4.2. Moreover, CO-IRv2 avoided overfitting of the model by using regularization techniques: data augmentation described in Section 4.1, and dropout discussed in Section 4.2. Afterwards, the system was evaluated with respect to precision, recall, F1-score, confusion matrix, receiver operating characteristics curve (ROC) curves, and accuracy. A basic system of our proposed work is depicted in Fig 2.

4.1 Data Preprocessing

Several steps, including normalization and data augmentation, were taken to tune the hyperparameters of CO-IRv2. Normalization of the data is a critical step to guarantee numerical stability. Normalization accelerates the training of a model and increases the likelihood of steady gradient descent. The samples of the CT images were of different sizes. Hence, we resized the images to 224×224 pixels with RGB color. The pixel values of the input images were normalized to a value between 0 and 1. The data sets include grayscale photos that have been rescaled by multiplying 1/255 by the pixel values. Next, we applied data augmentation [51, 52] that increases the sample size as models need a wide range of data for successful training. However, in many cases, the number of CT images in a dataset is less, since collecting medical data is difficult. Data augmentation increases the number and the variability of images while maintaining class labels. In this work, the CT images in the dataset were extended by applying the following methods: (1) the images were rotated by angles of 180 degrees clockwise, (2) the images were scaled by 15%, (3) the images were horizontally rotated and (4) Gaussian noise with a mean zero and variance 0.25 were added. Besides applying flipping, resizing, and rotation, we applied random horizontal flips to improve the simplification of the architecture for each probable location of CT images. Finally, a number of images were produced by using a 180-degree rotation. In data augmentation, random horizontal flip helped in the identification of COVID-19 from chest symptoms and random resized crop helped in deeper recognition of pixels’ relationship by varying image intensity. The mentioned augmentation and development strategies were applied to promote the generalizations of the proposed model. It should be noted that all these methods were used for training samples. Finally, there was a bigger training set of 2481 images, 5 times more than the initial training images.

4.2 Proposed CO-IRv2

This section discusses the proposed CO-IRv2 algorithm. A part of the CO-IRv2 scheme is derived from the concepts of Inception and ResNet [55] with hyperparameter tuning, while the remaining part is a new architecture consisting of a global average pooling layer, batch normalization, dense and dropout layers.

Fig 3 shows the CO-IRv2 architecture where the upper part (within the rectangular box) is derived from InceptionNet and ResNet, while the lower part is a new arrangement of multiple layers. In the upper part, there are the stem, Inception [56], and reduction modules. The Inception model is substantially tunable, which means some changes like changing the number of filters and layers have to be performed without impacting the quality of the fully connected layers. By carefully tuning the sizes of the layers, the training speed can be optimized. For our proposed CO-IRv2 model, uniform selections are performed for the Inception modules for all grid sizes. In CO-IRv2, batch-normalizations are applied distinctly on top of the conventional layers, instead of above the combinations of InceptionNet and ResNet. CO-IRv2 is based on Inception architecture having residual connections instead of filter concatenation operation. The Inception module selects all three filter sizes at each layer instead of choosing one for each layer. The use of multiple filters enables the selection of the best features resulting in an excellent performance. The addition of ResNet and Inception allows the model to have excellent accuracy than standalone Inception and ResNet models. In CO-IRv2, the residual connectors are used to combine different convolution filters. The residual connectors cause a reduction in the computation time during the training phase. Specifically, the residual blocks are utilized to permit that the Inception modules can increase their quantity. Hence, the depth of the network can also be increased. The extremely deep networks of CNN is the training stage, where the computation time is reduced by the residual networks [55, 56]. When more than 1,000 filters are deployed in the network, the network reduces the residue to solve the training problem [56].

The stem, Inception and reduction modules consist of convolution layers, pooling layers and Rectified-Linear-Unit (ReLU) activation functions. Convolution layers are applied to generate features from images. A convolution kernel operates on the input images by sliding across them. This is done with a set of movement parameters known as a stride. The size of the output of convolution operation depends on the size of the kernel and that of the slide. Optimization is done on the bias and the weights of the kernels. The activation layers enhance the nonlinearity of the output obtained from the convolution layers. The activation function is a decision-making function that aids in the learning of complex patterns. The use of the right activation function can speed up the learning process. ReLU is an activation function having a fast calculation speed. Pooling layers are used to reduce the feature map. Pooling is a form of local operation that processes similar data in the vicinity of the receptive field. Pooling layers also reduce the upper layers in terms of the number of parameters and size in space. These pooling layers help the CO-IRv2 model to remain invariant to the effects of translation, distortion and transformation. The stem, Inception and reduction elements are further described with illustrations later in this section.

Next, we describe the lower part of the CO-IRv2 architecture, as shown in Fig 3. Firstly, there is a global average pooling layer where a pooling operation gp(.) generates feature map from input feature-map as follows.

(1)

The next operation is the batch normalization that addresses the issues of internal covariance shift inside feature maps. It smoothes the gradient flow and works as a regulatory component, so assisting in the network’s generalization. The normalized feature map for a mini-batch can be represented as follows. (2) where μ is the mean and represents the variance of a feature-map, and the term ε is considered to address the divide by zero issue. After that there are multiple dense layers that are associated with a leaky ReLU and a dropout layer. These dense layers contribute to improving classification accuracy. A minor negative side is allowed by a leaky ReLU. To keep a regulated negative component, leaky ReLU activation with a value of 0.2 is used for each dense layer. The dropout process sets the output of each hidden neuron to zero with a probability of half. The neurons that are dropped out in this fashion do not participate in forward or backpropagation. Consequently, the neural network samples a different design each time an input is supplied, yet all of these structures have the same weights. Dropout layers are used for regularization by randomly dropping out or ignoring some layer outputs or units. Dropout layers can reduce the effect of over-fitting. This is done by reducing the number of parameters of the model. The dropout layer increases the robustness of the CO-IRv2 framework. Fig 3 shows one dropout layer in CO-IRv2 configured at rate 0.5, while the other with a rate of 0.4. Softmax layers are used to transform the input values into probabilities so that the sum of the values becomes unity. The softmax function is most typically employed as an activation function in a neural network model. Softmax is used as the activation function for multi-class classification problems that need class membership on more than two class labels. In this way, softmax enables multi-class classification. Softmax is a mathematical function that converts an integer vector into a probability vector. This is done by ensuring the probability of each value is proportional to the relative scale of the vector.

Fig 4 describes the details of the stem in CO-IRv2. It can be seen that the initial blocks within the stem compute three 3×3 convolutional operations on the given input. Next, there are three inception blocks. The first one has two paths, one for 3×3 convolutional operations and the other for max pooling. The two paths are concatenated and passed to the next inception block. The second inception block also has two paths; one path has 1×1 and 3×3 convolutional operations, while the other has 1×1, 7×1, 1×7 and 3×3 convolutional operations. The output of the two paths is concatenated. Similar to the first block, the third inception block has two paths, one for 3×3 convolutional operations and the other for max pooling. The final output is the concatenation operation of the two paths within the third inception block.

Fig 5(A) and 5(B) depict the detailed architecture of Inception-A, and Reduction A., respectively in CO-IRv2. In the Inception-A module, there are three paths of 1×1 and 3×3 convolution operations. The illustrations of Inception-Res-Net B, Reduction B, Inception-Res-Net C modules are similar to those reported in [56]. The Inception-ResNet-B and Inception-ResNet-C have two paths of convolution operations with different convolution filter sizes. Each inception block is connected to a filter layer, a form of 1 × 1 convolution operation. This filter layer obtains transformation in dimensions for matching with the input. This compensates for the reduction in dimensions at the inception stage. Inception-A module shown in Fig 5(A) is created for 35×35 grid blocks of the traditional Inception-v4 network. Conversely, Inception-B module has 17×17 grid modules and Inception-C has 8×8 grid blocks. There are also Reduction-A and Reduction-B layers. The Reduction-A block has one path for max-pooling and two paths for convolution operation. The Reduction-B block consists of three convolution paths and one max-pooling path. The Reduction-A module shown in Fig 5(B) reduces 35×35 to 17×17 modules. Moreover, the Reduction-B module reduces 17×17 modules to 8×8 grid.

thumbnail
Fig 5.

Illustration of CO-IRv2 (a) Inception-A module (b) Reduction-A module.

https://doi.org/10.1371/journal.pone.0259179.g005

A number of optimizers namely Adam, Nadam and RMSprop are used for CO-IRv2. Adam is a method for calculating adaptive learning rates. It keeps an exponentially decaying average of past squared gradients as well as an exponentially decaying average of past gradients like momentum. Nadam is a combination of Adam and Nesterov accelerated gradient (NAG). The NAG term is incorporated in Adam by changing the momentum term achieved by first moving in the direction of the previous momentum vector and then moving in the direction of the current gradient. RMSprop is a gradient-based optimization strategy suitable for neural network training. Table 3 presents the parameters of optimizers such as Adam, Nadam and RMSProp with learning rates. CO-IRv2 has 56,444,642 parameters for building the model where the number of non-trainable parameters is 54,340,832 and the number of trainable parameters is 2,103,810.

thumbnail
Table 3. Optimizers with various parameters for training and testing of CO-IRv2.

https://doi.org/10.1371/journal.pone.0259179.t003

5. Experimental results and discussion

In this section, the experimental results of CO-IRv2 are first described for the case of CT images and then for X-ray images.

5.1 Results for CT images

This section discusses the experimentation of CO-IRv2 applied to CT images. Google Collaboratory was used for experimental purposes. This is a cloud service based on Jupyter notebook for distributing knowledge and applying DL. It deals with a fixed GPU with free-of-charge and fully optimized runtime for DL or ML. In the experiments, Tensorflow, Matplotlib, Sklearn and Numpy were also applied as libraries. In the following, the performance metrics are described and then the results are presented.

Firstly, several performance metrics, including confusion matrix, precision, recall, accuracy, F1-score, ROC curve [1113, 5759] are defined in the context of CO-IRv2 model. The result of a suspected patient is negative if the lung is not infected from coronavirus and is positive when the coronavirus has infected the lungs. The outcome of this test for all COVID-19 patients may not or may match the actual cases of the patients. As an element of the confusion matrix, the true positive (TP) denotes the positive COVID-19 patients correctly identified. Another term called false positive indicates incorrectly identified COVID-19 patients who may or may not have other lung diseases. At the same time, true negative (TN) means correctly detected COVID-19 negative patients. False-negative represents the incorrectly identified non-infected patients. Classification accuracy is the measure of the correctness of identifying a normal case as normal and an abnormal cases as abnormal. Accuracy, A, is given as.

(3)

Recall is the number of correctly classified patients to the number of suspected patients. Recall or sensitivity, R, is given by.

(4)

Specificity, S, refers to the prediction accuracy of normal cases expressed as (5)

Precision, P, is the ratio of accurately classified positive cases to the overall predicted positive cases given by.

(6)

F1-score, F1, is the harmonic mean of precision, P, and recall, R, calculated as follows.

(7)

Next, results are provided for the case where holdout method is used to separate training and testing images. Results are for the case of 25 epochs. Fig 6 shows the values of the confusion matrix of CO-IRv2 for Adam, Nadam and RMSProp. It can be seen from Fig 6 that the values achieved for TP, TN, FN, FP are 250, 222, 17, 8, respectively, for Adam optimizer. Moreover, for the case of Nadam optimizer, TP, TN, FN, FP have values of 246, 232, 7, 12, respectively. Furthermore, the values achieved for TP, TN, FN, FP are 236, 242, 17, 2, respectively, for RMSProp optimizer in CO-IRv2. The obtained accuracies of Adam, Nadam and RMSProp optimizers are 94.97%, 96.18% and 96.18%, respectively. This is shown in Table 4. For Nadam and RMSProp optimizers, our proposed model obtains the highest accuracy where the accuracy of normal and COVID-19 is 96.18% individually. Table 4 shows that Nadam provides a recall value of 97% normal and 95% COVID-19, whereas RMSProp has a recall value of 93% normal and 99% COVID-19.

thumbnail
Fig 6.

Confusion matrix of CO-IRv2 for (a) Adam, (b) Nadam and (c) RMSProp.

https://doi.org/10.1371/journal.pone.0259179.g006

thumbnail
Table 4. Classification of COVID-19 and normal patients for different optimizers using CO-IRv2.

https://doi.org/10.1371/journal.pone.0259179.t004

The efficiency of CO-IRv2 is also apparent from the AUC-ROC curves for CO-IRv2 and those of existing InceptionNetV3 model. The ROC curves in Fig 7 depict the performance of InceptionNetV3 for different optimizers, while Fig 8 presents that of CO-IRv2. Fig 7 shows that in the case of InceptionNetV3, the AUC values for Adam, Nadam and RMSProp are 91%, 93% and 93%, respectively. From Fig 8 it can be seen that for the case of CO-IRv2, the AUC values of Adam, Nadam and RMSProp CO-IRv2 are 93%, 95% and 93%, respectively. Hence, for each optimizer, CO-IRv2 has higher AUC values than InceptionNetV3. Furthermore, it can be seen that the Nadam optimizer is more efficient than other optimizers when CO-IRv2 is applied to diagnose the infection of COVID-19. Fig 9 compares different optimizers for CO-IRv2 model for each class. Fig 9 presents that for CO-IRv2, the classification accuracy is higher for RMSProp and Nadam compared to Adam optimizer.

thumbnail
Fig 7.

ROC plot of InceptionNetV3 for the optimizer of (a) Adam, (b) Nadam and (c) RMSProp.

https://doi.org/10.1371/journal.pone.0259179.g007

thumbnail
Fig 8.

ROC plot of CO-IRv2 for the optimizer of (a) Adam, (b) Nadam and (c) RMSProp.

https://doi.org/10.1371/journal.pone.0259179.g008

thumbnail
Fig 9. Comparison of different optimizers for CO-IRv2 model for each classes.

https://doi.org/10.1371/journal.pone.0259179.g009

Next, heatmap is used for evaluating the performance of CO-IRv2. It visualizes the operation of the proposed algorithm. Fig 10 visualizes ground truth and heatmap of COVID-19 CT images when CO-IRv2 is used. The ground truth achieved from the heatmap indicates the actual results of the detection of COVID-19 infection.

thumbnail
Fig 10. Ground truth and heatmap of COVID-19 CT images when CO-IRv2 is used.

https://doi.org/10.1371/journal.pone.0259179.g010

In the experiments, the performance of CO-IRv2 was also evaluated using the cross-validation method to separate the training and testing images in the dataset. For this, five-fold cross-validation was chosen. Since Nadam optimizer exhibited the best results for the holdout method for CT images, we used Nadam for cross-validation. The total images were separated into five groups marked as groups 1–5. For the first fold, the images from group 1 were used for testing, while the images from groups 2–5 were used for training. With this consideration, the classification accuracy and other metrics were calculated for the first fold. Next, group 4 was used for testing, while the images from groups 1–3 and group 5 were used for training. This arrangement enabled the calculations for the second fold. Similarly, the third fold, fourth fold, and fifth fold metrics were calculated considering group 3, group 4, and group 5 for testing, respectively, with the remaining ones used as training images. The metrics obtained for folds 1–5 were averaged to obtain the final results. The obtained values for precision, recall, specificity, F1-score and accuracy were 97.85%, 91.20%, 97.96%, 94.41% and 94.55%, respectively.

5.2 Comparison of CO-IRv2 with existing methods for CT images

This section provided comparative results of the proposed CO-IRv2 with the existing methods for the case of CT images. Table 5 shows comparative results of CO-IRv2 with the existing InceptionNetV3 model for Adam, Nadam and RMSProp optimizers. Table 5 also indicates that compared to InceptionNetV3, our proposed CO-IRv2 model achieves higher values of sensitivity, precision, recall, F1-score and accuracy. The execution time of CO-IRv2 is comparable to that of InceptionNetV3 for each of the optimizers. Table 6 compares the execution time for CO-IRv2 with several existing models. CO-IRv2, VGG-19, CTnet-10, and Inception V3 models were trained and evaluated using a Tesla K80 Graphical Processing Unit (GPU) supplied by Google Colab, while DenseNet-169 model was run using an eighth-generation Intel i5 CPU. On the other hand, MobileNet, VGG16, DenseNet121, DenseNet169, NasNet Large, Xception and InceptionResNetV2 were run using the Google Collaboratory with Tesla K80 GPU Card in conjunction with an Intel i7-core @3.6 GHz processor, 16 GB RAM and 64-bit Windows 10 operating system. From Table 6, it can be seen that CTnet-10 has a low execution time of 130.90 seconds, while running on Tesla K80 GPU. On the other hand, MobileNet has a low execution time of 374 seconds while running on a Tesla K80 GPU Intel i7-core @3.6 GHz processor and 16 GB RAM. The proposed CO-IRv2 has an execution time of 707 seconds.

thumbnail
Table 5. Comparison of CO-IRv2 with existing InceptionNetV3 model.

https://doi.org/10.1371/journal.pone.0259179.t005

thumbnail
Table 6. Computational time of various models for CT scans of COVID-19 patients.

https://doi.org/10.1371/journal.pone.0259179.t006

Table 7 briefly compares CO-IRv2 model with the state-of-art methods in terms of recall, precision, recall, F1-score and accuracy. Our proposed CO-IRv2 model achieves the highest accuracy and recall when compared with the existing models in Table 7. The accuracy obtained by CO-IRv2 outperforms all referred models taken into consideration. The accuracies described by Ref. [30, 42, 43, 46] are around 86% for CT images. The ElasticNet [47] and UNet [41] models have accuracy values of 90% and 91.66%, respectively. However, the accuracy of CO-IRv2 is 96.18%. Hence, CO-IRv2 has the potential to be considered as an efficient system for binary classification of lung CT images.

thumbnail
Table 7. Comparison of the proposed method with a number of existing literature.

https://doi.org/10.1371/journal.pone.0259179.t007

5.3 Results for X-ray images

This section provides brief results of CO-IRv2 applied to a dataset of X-ray images [54]. The dataset contains 1662 images of which 1583 images are of normal people and 79 are of COVID-19 patients. Using the holdout concept, 80% of the total images are used for training and the remaining 20% for testing. This means 1329 images are for the training, of which 1261 and 68 are for normal and COVID-19 patients, respectively. On the other hand, 333 images are used for testing where 11 are of COVID-19 patients, and the remaining 322 are of normal people. Table 8 shows the performance results for CO-IRv2 with different optimizers for the case of X-ray image dataset. The performance is evaluated using a Tesla K80 GPU supplied by Google Colab. It can be seen from Table 8 that the best results are obtained by Adam optimizer in terms of accuracy, precision, recall, F1-score and specificity. CO-IRv2 with Adam optimizer achieves classification accuracy of 99.40% and a recall value of 99.38% for an execution time of 430 seconds. The classification accuracy value of Adam is better than Nadam and RMSProp. However, Nadam and RMSProp optimizers have slightly lower execution times than Adam. The results obtained for CO-IRv2 are comparable to the results reported in the literature of X-ray images. For example, the studies in [16, 4850] consider different X-ray datasets and report accuracy values of 96.10%, 98.75%, 98.66% and 98.08%, respectively. On the other hand, CO-IRv2 with Adam optimizer can obtain an accuracy of 99.40% for the case of X-ray images.

thumbnail
Table 8. Comparison of CO-IRv2 for different optimizer using X-ray images.

https://doi.org/10.1371/journal.pone.0259179.t008

5.4 The implication of the results

It is shown in this paper that CO-IRv2 can be successfully applied to the images of CT scans and X-ray for COVID-19 screening. Processes involving Image testing and clinical testing are different. A clinical test can be unpleasant compared to imaging trials, which do not require a lengthy procedure. Clinical testing has slightly higher accuracy when doctors analyze the reports themselves, and in the field of image testing, a computer does this until detailed instructions are given. Image analysis is less expensive than clinical analysis. However, CT scan radiation can be a concern particularly for pregnant women and those with metal implants. CO-IRv2 approach classifies CT scans as COVID-19 negative or positive. We may extend our CO-IRv2 model further to detect positive COVID-19 CT scan images, depending on the extent of COVID-19 dissemination in the pulmonary area. Our method is well organized and can be used by doctors to screen large numbers of individuals. It is likely to be comparable with the accuracy and speed of the current RT-PCR method in COVID-19 diagnosis. The CO-IRv2 approach can categorize CT scan images of COVID-19 patients, allowing doctors to obtain data efficiently and timely.

6 Conclusions and future work

This paper introduces and describes a new scheme termed as CO-IRv2. This new CO-IRv2 scheme has two major blocks. One is the combination of the idea of InceptionNet and ResNet, while the other is a combination of a global average pooling layer, batch normalization, dense layers, and dropout layers. This CO-IRv2 is suitable for diagnosing and evaluating each class (normal vs. COVID-19) by applying different optimizers such as Adam, Nadam and RMSProp. The proposed model is applied to a new dataset of CT images which is formed here as the combination of two databases. A number of data preprocessing techniques, including data segmentation, augmentation, rescaling and data normalization are applied to the new dataset. For CO-IRv2, hyperparameters fine-tuning are applied for optimization. For the case of CT images, CO-IRv2 achieves the highest 96.18% accuracy, 97.23% recall, 95% AUC and 96.28% F1-score for Nadam optimizer. The proposed CO-IRv2 is also applied to an X-ray dataset of 1662 images. It is found that CO-IRv2 with Adam optimizer achieves classification accuracy of 99.40% and a recall value of 99.38%.

The proposed CO-IRv2 model can be executed in the medical field to classify chest X-ray and CT images to find out the COVID-19 patients. This can be utilized as a pre-assessment method for shortening the workload of physicians and doctors, diagnosing and treating diseases at initial stages. This method can be used to determine the condition of the patients before applying treatment quickly. Although CO-IRv2 can predict COVID-19 patients, the effectiveness of the method depends on the datasets. If a dataset contains incomplete labels or the images have significant distortions and noise, the disease prediction may be inaccurate. Like other DL methods, CO-IRv2 requires large-sized reliable data for training so that the prediction can be reliable. There is a limited number of large and balanced datasets. Moreover, there is no benchmark dataset for COVID-19, so it is challenging to compare CO-IRv2 with others in a precise way. Nevertheless, in the future, the effectiveness of CO-IRv2 should be evaluated for large datasets of COVID-19.

Acknowledgments

This work has been carried out at the Institute of Information and Communication Technology (IICT) of Bangladesh University of Engineering and Technology (BUET), Bangladesh. Hence, the authors would like to thank BUET for providing the support.

References

  1. 1. Velavan TP, Meyer CG. The COVID‐19 epidemic. Tropical medicine & international health. 2020;25(3):278. pmid:32052514
  2. 2. Huang C, Wang Y, Li X, Ren L, Zhao J, Hu Y, et al. Clinical features of patients infected with 2019 novel coronavirus in Wuhan, China. The lancet. 2020;395(10223):497–506.
  3. 3. Mondal MRH, Bharati S, Podder P, Podder P. Data analytics for novel coronavirus disease. Informatics in Medicine Unlocked. 2020;20:100374. pmid:32835073
  4. 4. Podder P, Khamparia A, Mondal MRH, Rahman MA, Bharati S. Forecasting the Spread of COVID-19 and ICU Requirements. International Journal of Online and Biomedical Engineering (iJOE). 2021;17(5): 81–99.
  5. 5. Phan LT, Nguyen TV, Luong QC, Nguyen TV, Nguyen HT, Le HQ, et al. Importation and human-to-human transmission of a novel coronavirus in Vietnam. New England Journal of Medicine. 2020;382(9):872–874. pmid:31991079
  6. 6. Sakurai A, Sasaki T, Kato S, Hayashi M, Tsuzuki S-i, Ishihara T, et al. Natural history of asymptomatic SARS-CoV-2 infection. New England Journal of Medicine. 2020;383(9):885–886. pmid:32530584
  7. 7. Lan L, Xu D, Ye G, Xia C, Wang S, Li Y, et al. Positive RT-PCR test results in patients recovered from COVID-19. Jama. 2020;323(15):1502–3. pmid:32105304
  8. 8. Kucirka LM, Lauer SA, Laeyendecker O, Boon D, Lessler J. Variation in false-negative rate of reverse transcriptase polymerase chain reaction–based SARS-CoV-2 tests by time since exposure. Annals of internal medicine. 2020;173(4):262–7. pmid:32422057
  9. 9. Guyatt G, Rennie D, Meade M, Cook D. Users’ guides to the medical literature: a manual for evidence-based clinical practice: AMA press Chicago; 2002.
  10. 10. Bharati S, Podder P, Mondal MRH, Prasath VBS. CO-ResNet: Optimized ResNet model for COVID-19 diagnosis from X-ray images. International Journal of Hybrid Intelligent Systems. 2021;17:71–85.
  11. 11. Bharati S, Podder P, Mondal MRH. Hybrid deep learning for detecting lung diseases from X-ray images. Informatics in Medicine Unlocked. 2020; 20:100391. pmid:32835077
  12. 12. Bharati S, Podder P, Mondal MRH, Gandhi N. Optimized NASNet for Diagnosis of COVID-19 from Lung CT Images. 20th International Conference on Intelligent Systems Design and Applications (ISDA 2020): Springer; 2020; 1351:647–656
  13. 13. Khamparia A, Bharati S, Podder P, Gupta D, Khanna A, Phung TK, et al. Diagnosis of Breast Cancer Based on Modern Mammography using Hybrid Transfer Learning. Multidimensional Systems and Signal Processing. 2021;32(2):747–65. pmid:33456204
  14. 14. Bharati S, Podder P, Mondal MRH. Artificial Neural Network Based Breast Cancer Screening: A Comprehensive Review. International Journal of Computer Information Systems and Industrial Management Applications. 2020;12:125–37.
  15. 15. Li K, Wu J, Wu F, Guo D, Chen L, Fang Z, et al. The clinical and chest CT features associated with severe and critical COVID-19 pneumonia. Investigative radiology. 2020. pmid:32118615
  16. 16. Ozturk T, Talo M, Yildirim EA, Baloglu UB, Yildirim O, Acharya UR. Automated detection of COVID-19 cases using deep neural networks with X-ray images. Computers in biology and medicine. 2020;121:103792. pmid:32568675
  17. 17. Soldati G, Smargiassi A, Inchingolo R, Buonsenso D, Perrone T, Briganti DF, et al. Is there a role for lung ultrasound during the COVID‐19 pandemic? Journal of Ultrasound in Medicine. 2020.
  18. 18. Mondal MRH, Bharati S, Podder P. Diagnosis of COVID-19 Using Machine Learning and Deep Learning: A Review. Current Medical Imaging. 2021. pmid:34259149
  19. 19. Guan W-j, Ni Z-y, Hu Y, Liang W-h, Ou C-q, He J-x, et al. Clinical Characteristics of Coronavirus Disease 2019 in China. New England Journal of Medicine. 2020;382(18):1708–20. pmid:32109013
  20. 20. Lei J, Li J, Li X, Qi X. CT imaging of the 2019 novel coronavirus (2019-nCoV) pneumonia. Radiology. 2020;295(1):18. pmid:32003646
  21. 21. Chung M, Bernheim A, Mei X, Zhang N, Huang M, Zeng X, et al. CT imaging features of 2019 novel coronavirus (2019-nCoV). Radiology. 2020;295(1):202–7. pmid:32017661
  22. 22. Pan F, Ye T, Sun P, Gui S, Liang B, Li L, et al. Time course of lung changes on chest CT during recovery from 2019 novel coronavirus (COVID-19) pneumonia. Radiology. 2020.
  23. 23. Ardila D, Kiraly AP, Bharadwaj S, Choi B, Reicher JJ, Peng L, et al. End-to-end lung cancer screening with three-dimensional deep learning on low-dose chest computed tomography. Nature medicine. 2019;25(6):954–61. pmid:31110349
  24. 24. Cowton J, Kyriazakis I, Plötz T, Bacardit J. A combined deep learning gru-autoencoder for the early detection of respiratory disease in pigs using multiple environmental sensors. Sensors. 2018;18(8):2521. pmid:30072607
  25. 25. González G, Ash SY, Vegas-Sánchez-Ferrero G, Onieva Onieva J, Rahaghi FN, Ross JC, et al. Disease staging and prognosis in smokers using deep learning in chest computed tomography. American journal of respiratory and critical care medicine. 2018;197(2):193–203. pmid:28892454
  26. 26. Coudray N, Ocampo PS, Sakellaropoulos T, Narula N, Snuderl M, Fenyö D, et al. Classification and mutation prediction from non–small cell lung cancer histopathology images using deep learning. Nature medicine. 2018;24(10):1559–67. pmid:30224757
  27. 27. Suzuki K. Overview of deep learning in medical imaging. Radiological physics and technology. 2017;10(3):257–73. pmid:28689314
  28. 28. Bharati S, Podder P, Mondal MRH, Prasath VBS. Medical Imaging with Deep Learning for COVID-19 Diagnosis: A Comprehensive Review. International Journal of Computer Information Systems and Industrial Management Applications. 2021;13:91–112.
  29. 29. Wu X, Hui H, Niu M, Li L, Wang L, He B, et al. Deep learning-based multi-view fusion model for screening 2019 novel coronavirus pneumonia: a multicentre study. European Journal of Radiology. 2020:109041. pmid:32408222
  30. 30. Xu X, Jiang X, Ma C, Du P, Li X, Lv S, et al. A deep learning system to screen novel coronavirus disease 2019 pneumonia. Engineering. 2020. pmid:32837749
  31. 31. Rahimzadeh M, Attar A, Sakhaei SM. A fully automated deep learning-based network for detecting covid-19 from a new and large lung ct scan dataset. Biomedical Signal Processing and Control. 2021;68:102588. pmid:33821166
  32. 32. Shah V, Keniya R, Shridharani A, Punjabi M, Shah J, Mehendale N. Diagnosis of COVID-19 using CT scan images and deep learning techniques. Emergency radiology. 2021;28(3):497–505. pmid:33523309
  33. 33. Zheng C, Deng X, Fu Q, Zhou Q, Feng J, Ma H, et al. Deep learning-based detection for COVID-19 from chest CT using weak label. medRxiv. 2020.
  34. 34. Wang S, Kang B, Ma J, Zeng X, Xiao M, Guo J, et al. A deep learning algorithm using CT images to screen for Corona Virus Disease (COVID-19). European radiology. 2021:1–9. pmid:32767103
  35. 35. Shen C, Yu N, Cai S, Zhou J, Sheng J, Liu K, et al. Quantitative computed tomography analysis for stratifying the severity of Coronavirus Disease 2019. Journal of Pharmaceutical Analysis. 2020;10(2):123–9. pmid:32292624
  36. 36. Gozes O, Frid-Adar M, Greenspan H, Browning PD, Zhang H, Ji W, et al. Rapid ai development cycle for the coronavirus (covid-19) pandemic: Initial results for automated detection & patient monitoring using deep learning ct image analysis. arXiv preprint arXiv:200305037. 2020.
  37. 37. Kuchana M, Srivastava A, Das R, Mathew J, Mishra A, Khatter K. AI aiding in diagnosing, tracking recovery of COVID-19 using deep learning on Chest CT scans. Multimedia tools and applications. 2021;80(6):9161–75.
  38. 38. Song Y, Zheng S, Li L, Zhang X, Zhang X, Huang Z, et al. Deep learning enables accurate diagnosis of novel coronavirus (COVID-19) with CT images. IEEE/ACM Transactions on Computational Biology and Bioinformatics. 2021. pmid:33705321
  39. 39. Zhou T, Lu H, Yang Z, Qiu S, Huo B, Dong Y. The ensemble deep learning model for novel COVID-19 on CT images. Applied Soft Computing. 2021;98:106885. pmid:33192206
  40. 40. Shalbaf A, Vafaeezadeh M. Automated detection of COVID-19 using ensemble of transfer learning with deep convolutional neural network based on CT scans. International journal of computer assisted radiology and surgery. 2021;16(1):115–23. pmid:33191476
  41. 41. Javaheri T, Homayounfar M, Amoozgar Z, Reiazi R, Homayounieh F, Abbas E, et al. CovidCTNet: an open-source deep learning approach to diagnose covid-19 using small cohort of CT images. NPJ digital medicine. 2021;4(1):1–10. pmid:33398041
  42. 42. He X, Yang X, Zhang S, Zhao J, Zhang Y, Xing E, et al. Sample-efficient deep learning for covid-19 diagnosis based on ct scans. MedRxiv. 2020.
  43. 43. Amyar A, Modzelewski R, Li H, Ruan S. Multi-task deep learning based CT imaging analysis for COVID-19 pneumonia: Classification and segmentation. Computers in Biology and Medicine. 2020;126:104037. pmid:33065387
  44. 44. Amyar A, Ruan S, Gardin I, Herault R, Clement C, Decazes P, et al. Radiomics-net: Convolutional neural networks on FDG PET images for predicting cancer treatment response. Journal of Nuclear Medicine. 2018;59:324.
  45. 45. Amyar A, Modzelewski R, Vera P, Morard V, Ruan S. Weakly Supervised PET Tumor Detection Using Class Response. Journal of Nuclear Medicine. 2021; 62 (supplement 1): 1432.
  46. 46. Zhu X, Song B, Shi F, Chen Y, Hu R, Gan J, et al. Joint prediction and time estimation of COVID-19 developing severe symptoms using chest CT scan. Medical image analysis. 2021;67:101824. pmid:33091741
  47. 47. Guiot J, Vaidyanathan A, Deprez L, Zerka F, Danthine D, Frix A-N, et al. Development and validation of an automated radiomic CT signature for detecting COVID-19. Diagnostics. 2021;11(1):41.
  48. 48. Narin A, Kaya C, Pamuk Z. Automatic detection of coronavirus disease (covid-19) using x-ray images and deep convolutional neural networks. Pattern Analysis and Applications. 2021:1–4. pmid:33994847
  49. 49. Apostolopoulos ID, Mpesiana TA. Covid-19: automatic detection from x-ray images utilizing transfer learning with convolutional neural networks. Physical and Engineering Sciences in Medicine. 2020;43(2):635–40. pmid:32524445
  50. 50. Sethy PK, Behera SK. Detection of coronavirus disease (covid-19) based on deep features. Preprints. 2020;2020030300:2020.
  51. 51. COVID-19 CT segmentation datasets. Available from: http://medicalsegmentation.com/covid19/ [cited 2021 Jul 25].
  52. 52. Eduardo P. SARS-COV-2 Ct-Scan Dataset: kaggle; 2020 [cited 2020 18 December]. Available from: https://www.kaggle.com/plameneduardo/sarscov2-ctscan-dataset.
  53. 53. Bharati S, Podder P, Mondal MRH. COVID-19 CT scans for CO-IRv2 model. 2021.
  54. 54. Bharati S, Podder P, Mondal MRH. X-ray images three levels. figshare; 2021 [cited 2021 Jul 25]. Available from: https://figshare.com/articles/dataset/X-ray_images_three_levels/14755965/1
  55. 55. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition. 2016 (pp. 770–778).
  56. 56. Szegedy C, Ioffe S, Vanhoucke V, Alemi A. Inception-v4, inception-resnet and the impact of residual connections on learning. Thirty-first AAAI conference on artificial intelligence. 2017.
  57. 57. Raihan-Al-Masud M, Mondal MRH. Data-driven diagnosis of spinal abnormalities using feature selection and machine learning algorithms. Plos one. 2020;15(2):e0228422. pmid:32027680
  58. 58. Bharati S, Robel MRA, Rahman MA, Podder P, Gandhi N. Comparative Performance Exploration and Prediction of Fibrosis, Malign Lymph, Metastases, Normal Lymphogram Using Machine Learning Method. InInternational Conference on Innovations in Bio-Inspired Computing and Applications. 2019;1180:66–77.
  59. 59. Bharati S, Rahman MA, Podder P. Breast cancer prediction applying different classification algorithm with comparative analysis using WEKA. 2018 4th International Conference on Electrical Engineering and Information & Communication Technology (iCEEiCT); 2018; Dhaka, Bangladesh: IEEE.
  60. 60. Taresh MM, Zhu N, Ali TAA, Hameed AS, Mutar ML. Transfer Learning to Detect COVID-19 Automatically from X-Ray Images Using Convolutional Neural Networks. International Journal of Biomedical Imaging. 2020;2021.