Next Article in Journal
CD4+ T Cell Regulatory Network Underlies the Decrease in Th1 and the Increase in Anergic and Th17 Subsets in Severe COVID-19
Next Article in Special Issue
Scoring Enzootic Pneumonia-like Lesions in Slaughtered Pigs: Traditional vs. Artificial-Intelligence-Based Methods
Previous Article in Journal
Validating Immunomodulatory Responses of r-LdODC Protein and Its Derived HLA-DRB1 Restricted Epitopes against Visceral Leishmaniasis in BALB/c Mice
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Batch Normalized Convolution Approach for Improving COVID-19 Detection from Chest X-ray Images

by
Ibrahim Al-Shourbaji
1,
Pramod H. Kachare
2,
Laith Abualigah
3,4,5,6,7,*,
Mohammed E. Abdelhag
8,
Bushra Elnaim
9,
Ahmed M. Anter
10,11 and
Amir H. Gandomi
12,13,*
1
Department of Computer Science, University of Hertfordshire, Hatfield AL10 9AB, UK
2
Department of Electronics & Telecommunication Engineering, Ramrao Adik Institute of Technology, Nerul, Navi Mumbai 400706, Maharashtra, India
3
Computer Science Department, Prince Hussein Bin Abdullah Faculty for Information Technology, Al al-Bayt University, Mafraq 25113, Jordan
4
Hourani Center for Applied Scientific Research, Al-Ahliyya Amman University, Amman 19328, Jordan
5
Faculty of Information Technology, Middle East University, Amman 11831, Jordan
6
Applied Science Research Center, Applied Science Private University, Amman 11931, Jordan
7
School of Computer Sciences, Universiti Sains Malaysia, Pulau Pinang 11800, Malaysia
8
Department of Information Technology and Security, Jazan University, Jazan 45142, Saudi Arabia
9
Department of Computer Science, College of Science and Humanities in Al-Sulail, Prince Sattam bin Abdulaziz University, Riyadh 11671, Saudi Arabia
10
Egypt-Japan University of Science and Technology (E-JUST), Alexandria 21934, Egypt
11
Faculty of Computers and Artificial Intelligence, Beni-Suef University, Benisuef 62511, Egypt
12
Faculty of Engineering and IT, University of Technology Sydney, Ultimo, NSW 2007, Australia
13
University Research and Innovation Center (EKIK), Óbuda University, 1034 Budapest, Hungary
*
Authors to whom correspondence should be addressed.
Pathogens 2023, 12(1), 17; https://doi.org/10.3390/pathogens12010017
Submission received: 16 November 2022 / Revised: 12 December 2022 / Accepted: 19 December 2022 / Published: 22 December 2022

Abstract

:
Pre-trained machine learning models have recently been widely used to detect COVID-19 automatically from X-ray images. Although these models can selectively retrain their layers for the desired task, the output remains biased due to the massive number of pre-trained weights and parameters. This paper proposes a novel batch normalized convolutional neural network (BNCNN) model to identify COVID-19 cases from chest X-ray images in binary and multi-class frameworks with a dual aim to extract salient features that improve model performance over pre-trained image analysis networks while reducing computational complexity. The BNCNN model has three phases: Data pre-processing to normalize and resize X-ray images, Feature extraction to generate feature maps, and Classification to predict labels based on the feature maps. Feature extraction uses four repetitions of a block comprising a convolution layer to learn suitable kernel weights for the features map, a batch normalization layer to solve the internal covariance shift of feature maps, and a max-pooling layer to find the highest-level patterns by increasing the convolution span. The classifier section uses two repetitions of a block comprising a dense layer to learn complex feature maps, a batch normalization layer to standardize internal feature maps, and a dropout layer to avoid overfitting while aiding the model generalization. Comparative analysis shows that when applied to an open-access dataset, the proposed BNCNN model performs better than four other comparative pre-trained models for three-way and two-way class datasets. Moreover, the BNCNN requires fewer parameters than the pre-trained models, suggesting better deployment suitability on low-resource devices.

1. Introduction

The coronavirus (COVID-19) remains a global health problem that negatively impacts our lives and the global economy. The initial infection was reported on January 2020 of twenty-seven patients with Pneumonia and an epidemiological link to a live wild animal market [1]. The actual host remained unknown; however, the transmission can be through inhaling duplets produced by the cough or sneeze of an infected person within six feet. Likewise, studies have shown that the transmission can be through touching virus-covered surfaces and passing it to their mouth, nose, or possibly eyes [2,3]. To tackle this, there has to be a solution where the suspected patients receive faster confirmation of the presence of the disease. This can be provided by imaging modalities, such as chest radiographs (X-ray) or Computed Tomography (CT) scan images. The radiologists can give an opinion by analyzing and applying some image analysis methods to diagnose COVID-19 cases. It was confirmed that X-rays or CTs contain valuable information related to this disease. Therefore, a faster and more accurate diagnosis can be performed using radiologic images with Machine Learning (ML) techniques to determine COVID-19 cases [4].
Nowadays, ML techniques have become essential tools in medical diagnosis, and several studies have confirmed their capabilities for diagnosing COVID-19 [2,3,5,6]. Furthermore, these techniques assist physicians and health experts in diagnosing patients with pulmonary embolism, pulmonary circulation, or body temperature to improve the speed and accuracy in detecting COVID-19 cases [7,8,9,10]. Recently, the deep learning algorithm has gained special attention due to its potential to effectively extract relevant features from medical images as a part of its search process and classify them for disease diagnosis. These models can help radiologists and health experts to interpret, identify, and triage positive-infected COVID-19 cases.
Visual Geometry Group (VGG)-16, VGG-19, and other pre-trained models require large-scale datasets to optimize their model parameters in TL [11]. Although Transfer Learning (TL) models allow flexibility in retraining some or all the layers of such models for the desired task, the output of most layers remains biased due to a considerable number of pre-trained weights and parameters by these models [12]. Despite promising results reported for COVID-19 detection from chest X-ray images, most existing approaches have not provided parameter requirements in the pre-trained models to improve model classification capability. Therefore, further efforts are still needed to reduce computation complexity and the required parameters to train those models. In this paper, a novel model, namely BNCNN, is proposed. It uses batch normalization [13] to increase the model’s generalization and dropout layers [14] to improve training speed and avoid the model overfitting COVID-19 detection from X-ray images. The BNCNN can be considered a diagnostic system for COVID-19, which is necessary for critical conditions, as it can be helpful to the radiologists and their decisions in specifying COVID-19 cases from X-ray images. The main contributions of this paper can be summarised as follows:
A BNCNN model comprising a pre-processor, a feature extractor, and a classifier is proposed. The pre-processor resizes and normalizes input X-ray images to improve feature mapping and reduce classification loss. The feature extractor comprises four repetitions of the cascaded convolution layer to learn feature maps, a batch normalization layer to solve the internal covariance shift of feature maps, and a max-pooling layer to increase the convolution span for generating high-level patterns. The classifier comprises two repetitions of a dense layer to learn complex feature maps, a batch normalization layer to standardize internal feature maps, and a dropout layer to avoid overfitting while improving generalization. The proposed BNCNN model retains the pre-processor and features extractor while only the number of neurons in the output layer of the classifier is changed for binary and muti-class COVID-19 detection.
The performance of the BNCNN model is investigated against four pre-trained models, namely, VGG-16, VGG-19, Inception-V3, and ResNet-50, for COVID-19 detection from X-ray images.
The proposed model requires fewer parameters than other pre-trained models making it more suitable for implementation on low-resource devices.
This paper is organized as follows: Section 2 describes the COVID-19 dataset, feature extraction, and adaptation of the pre-trained image classification models, and Section 3 presents experimental results. Finally, Section 4 concludes the paper.

2. Proposed Model

The pre-trained VGG models inspire the proposed model intending to reduce computation complexity while increasing the classification accuracy of COVID-19 detection. Various tasks explored the state-of-the-art VGG models due to their excellent capability for feature extraction. These models can be well understood in two sections: Feature extraction and classifier. The first section embeds the raw input into low-dimensional vectors further accepted by the classifier for generating desired class labels. The proposed BNCNN is inspired by the repetition of block structures of feature extraction, as in the VGG model. The proposed BNCNN-based COVID-19 detection system has three primary phases: Data pre-processing, Feature extraction, and Classification. We explain each phase in the following subsections.

2.1. Data Pre-Processing

The proposed model uses chest X-ray images of COVID-19 patients and other subjects. Researchers from the University of Doha and the University of Dhaka collected this dataset, and it is publicly available with metadata on the ‘Kaggle’ [15]. The dataset contains three classes, including COVID-19, Normal and Viral Pneumonia images. Samples of X-ray images from each of the three classes in the dataset are provided in Figure 1. For experiments and evaluation, the dataset is partitioned into three mutually exclusive and exhaustive subsets for training (80%), validation (10%), and testing (10%). These subsets are summarised in Table 1.
The X-ray images in the dataset are uniformly pre-processed to facilitate the learning process. Each X-ray image is resized to 150 × 150 pixels, unlike 224 × 224 as in VGG models. It reduces the network’s input dimension and trainable and non-trainable weights of the BNCNN model. Each X-ray image is divided by 255, resulting in a normalized image in 0–1, which facilitates weight learning by avoiding vanishing and exploding gradients. Data augmentation strategies are used to simulate real-life scenarios and avoid the risks of overfitting. All subsets are augmented independently with a rotation ranging from −10 to 10 degrees, a zooming range of 0–10%, shearing of 0–10%, a horizontal stride of 0–10%, a vertical stride of 0–10%, and horizontally flip to improve generalization and increase diversity in the learning process by the models. The pixel values unavailable in the augmented image are replaced by the nearest pixel values. Examples of pre-processed chest X-ray images from each of the three classes are shown in Figure 1. The vertical flip is avoided because it will be easy for an ordinary user to identify vertical orientation in chest X-ray images. The class label for each image is encoded using one-hot encoding. It converts the label into a 3D vector with all zeros except for the one corresponding to the image class. It can be noted that the actual order of the classes in these three dimensions does not affect the classifier performance [16].

2.2. Feature Extraction

The feature extraction phase comprises the first twelve layers, while the remaining are for the classifier phase. Architectural details of the BNCNN model, the output dimension at each layer, and the number of trainable/non-trainable parameters are provided in Table 2. Inspired by the VGG models, the feature extraction phase of the proposed BNCNN uses four repetitions of a block of similar layers. Each block comprises a convolution layer to learn kernel weights suitable for the features map, followed by a batch normalization layer to solve the internal covariance shift of feature maps [13] and a max-pooling layer to find the highest-level patterns in the input images by increasing the span of the convolution calculation [16].
f m a x m , n = m a x I h ,   v ,   I h + 1 ,   v ,   I h ,   v + 1 ,   I h + 1 ,   v + 1
where h = m + s and v = n + s .
Herein, m and n are horizontal and vertical indices of the image, and s is stride. A detailed illustration can be found in [17]. Each convolution layer comprises 3 × 3 filters with stride and padding of 1 and a ReLU activation function. The number of filters in convolution layers increases from the input to the output layers. A batch normalization layer after every convolution layer increases the model’s generalization capability. It standardizes the output of the previous layer to have a mean of zero and a Standard Deviation of one. These layers keep track of input variable statistics during training and standardize input during testing. The standardized variables can be scaled at the transformed output to have the desired statistics updated during training and maintained for testing. Hence, these layers have an equal number of trainable and non-trainable parameters. A max-pooling layer reduces the dimension of the input feature map without any parameters. All max-pooling layers use a stride and maximum size of two, halving the dimension of the input feature map at each occurrence.

2.3. Classification

The classifier starts with the flattening layer to convert all output of the previous phase (i.e., feature extraction) into a vector. This phase replaces convolution and max-pooling layers with dense and dropout layers. It comprises two repetitions of a block of the dense layer for feature mapping, followed by a batch normalization layer to standardize internal feature maps. The dropout layer avoids overfitting while aiding the model generalization, and a softmax layer generates a probabilistic output for each class label.
The number of neurons in the dense layer decreases from 256 to 128 from input to output to remove redundant features. Furthermore, it reduces the model’s computational complexity as the number of trainable parameters for dense layers increases exponentially with the number of neurons. All dropout layers use a drop factor of 0.2, indicating that randomly selected 20% weights are updated in each iteration to increase the model’s generalization.
The architecture of the BNCNN remains unchanged for three-way and two-way classifications except for the softmax layer. In three-way classification, the softmax layer uses three output nodes with 387 trainable parameters, resulting in 2,786,435 trainable parameters. While in two-way classification, two output nodes with 258 trainable parameters are used, resulting in 2,786,306 trainable parameters.
The Adaptive Moment Estimation (Adam) optimizer iteratively updates network weights using the training data. This optimizer combines Stochastic Gradient Descent (SGD) and Root Mean Square Propagation (RMSP). The choice of Adam optimizer facilitates adjusting the learning rate for each weight in the network by computing the first and second moments of the gradient, adaptive learning rate, and history-based updates for faster convergence. Moreover, it shows the best accuracy compared to SGD and RMSP optimization algorithms [17].
For the nth training example, let y p r e d (n, i) be the ith scalar value corresponding to predicted class probability ( i y p r e d n ,   i = 1 ,   n ) and y (n, i) be the ith scalar value corresponding to the actual one-hot-encoded class label. The cross-entropy loss (L) is calculated as follows:
L = n = 1 N t r a i n i = 1 N c l a s s y n ,   i   .   log y p r e d n ,   i
where, N t r a i n is the total number of training examples and N c l a s s is the number of classes (i.e., 3 for three-way and 2 for two-way classification).
The number of non-trainable parameters remains unaltered in both classification types. The major contributors to the increasing number of trainable parameters, convolution, and dense layers are minimized in the proposed BNCNN layer. Figure 2 illustrates the overall process of the BNCNN model.
A callback is tailored to reduce the learning rate after every epoch if the validation accuracy stops improving. The callback’s primary goal is to monitor the validation accuracy and reduce the learning rate by a factor of 0.3 if no improvement is achieved for three consecutive epochs. After reducing the learning rate, it waits for at least five epochs before applying the reduction again. Another callback is tailored for early stopping the training by monitoring training and validation accuracies after each epoch. The callback stores the model weights if the current training and validation accuracies exceed the earlier epochs. The second callback avoids model overfitting without worrying about the exact number of epochs. The proposed BNCNN steps are represented in Algorithm 1.
Algorithm 1: BNCNN model
INPUT
IMG: Dataset of X-ray images [224 × 224 × 15,153]
LAB: Set of labels {‘COVID-19’, ‘Normal’, ‘Viral Pneumonia’} corresponding to X-ray images [1 × 15,153]
N c l a s s : number of classes used for training
OUTPUT
Y p r e d : a matrix of prediction probability of class labels [ N c l a s s × 683]
ALGORITHM:
If 3-way
N c l a s s   = 3
If 2-way
N c l a s s = 2
For I = 1 to length (LAB)
If LAB (1, i) == ‘COVID-19’ or LAB(1, i) == ‘Normal’
X(:, :, i), Y(1, i) = IMG(:, :, i), LAB(1, i)
End if
End_for
Else
N c l a s s = 2
X, Y = IMG, LAB
End_if
X = image_resize(X)/255
Y e n c = one-hot_encoder (Y, N c l a s s )
If model_name == ‘VGG-16’ or model_name == ‘VGG-16’
model = load_pre-trained_weigts(model_name)
model = freeze_feature_extration_layers(model)
Else_if model_name == ‘BNCNN’
model = construct_model()
End_if
X t r a i n , X v a l , X t e s t , Y t r a i n , Y v a l , Y t e s t = cross-validation (X, Y e n c , train = 0.8, val = 0.1, test = 0.1)
X t r a i n _ a u g , Y t r a i n _ a u g = data_augmentation( X t r a i n , Y t r a i n , rotation = [−20, 20], zoom = [0, 0.2],
shear = [0, 0.2], horizontal_flip = True, vertical_flip = False)
X v a l _ a u g , Y v a l _ a u g = data_augmentation( X t r a i n , Y t r a i n , rotation = [−20, 20], zoom = [0, 0.2],
Shear = [0, 0.2], horizontal_flip = True, vertical_flip = False)
model = train ( X t r a i n _ a u g , Y t r a i n _ a u g , X v a l _ a u g , Y v a l _ a u g , custom_callbacks)
Y p r e d = predict (model, X t e s t )
cm = calculate_confusion_matrix( Y p r e d , Y t e s t , N c l a s s )
acc, recall, precision, F1-score = calculate_evaluation_metrics(cm, N c l a s s )

3. Experiments and Results

This section provides the implementation details of the BNCNN model, followed by the results of the experiments.

3.1. Experimental Setup

The BNCNN model and other pre-trained models: VGG-16, [18], VGG-19, [18], Inception-V3, [19] and ResNet-50, [20] are evaluated on the dataset to classify chest X-ray images into 3-way classification: COVID-19, Normal and Viral Pneumonia and for two-way classification: COVID-19 and Normal X-ray images.
The BNCNN model is trained for 100 iterations using the Adam optimizer with an initial learning rate decay of 0.0001 to finish all the epochs and obtain the solutions without interruption. The cross-entropy loss function minimizes the distance between predicted and actual probability distributions.
The hyper-parameter settings of the Adam optimizer for the BNCNN model and other pre-trained models are provided in Table 3. These settings are assigned after we experimentally find that these are the best settings of parameters for training the models. All the models are implemented using Python and are executed using 12 GB NVIDIA Tesla P100 GPU and Intel Xenon CPU @ 2.00GHz with 13 GB RAM.

3.2. Evaluation Measures

For model evaluation, accuracy, sensitivity, Positive Predictive Value (PPV), and F1-score measures are used to assess the performance of the BNCNN and the other models. The equations for deriving the values of these metrics are provided in (3)–(6).
  Accuracy   Acc = TP + TN TP + TN + FN + FP
Sensitivity   Sen = TP TP + FN
Positive   Predictive   Value   PPV = TP TP + FP
F 1 score   F 1 = 2   Percision   × Recall Percision + Recall
True Positive (TP) denotes the cases where the predicted class label is the same as the class under consideration. True Negative (TN) refers to the cases where classes not under consideration are predicted as themselves. False Negative (FN) is a misclassified case where the class under consideration is predicted as any class other than itself. False Positive (FP) indicates the cases where the model wrongly identifies other classes as the class under consideration.

4. Results and Discussion

This section discusses the comparative performance of the BNCNN against the other pre-trained models.

4.1. Results of the Proposed BNCNN Model

The comparative performance results of the BNCNN model and the pre-trained models for three-way classification are provided in Table 4. Although the VGG-16 model performs slightly better than the proposed BNCNN model during training, the latter performs better during the validation and testing phases, as shown in Table 4. The ResNet-50 achieved the least performance results in training, validation, and testing datasets, followed by VGG-19 and Inception-V3. This is due to the large number of pre-trained network parameters causing excessive bias in the feature extraction phase. For three-way COVID-19 classification, the BNCNN model achieved an accuracy of 96.84% (95% CI: 91.26–97.45), a sensitivity of 93.06% (95% CI: 89.13–96.54), PPV of 97.40% (95% CI: 94.71–98.80), and F1 of 95.18% (95% CI: 88.45–0.97.13) in the testing phase.
For two-way classification, the performance results of the BNCNN model are comparatively higher than the VGG-16, VGG-19, Inception-V3, and ResNet-50, as shown in Table 5. The BNCNN model achieved 97.28% accuracy (95% CI: 95.25–98.73), 97.28% sensitivity (95% CI: 95.32–98.75), 97.23% PPV (95% CI: 95.19–98.68) and 97.25% F1 (95% CI: 95.33–98.13) in the training phase. In the validation phase, the BNCNN model achieved 98.55% accuracy (95% CI: 97.45–99.26), 98.52% sensitivity (95% CI: 96.84–99.22), 98.53% PPV (95% CI: 97.03–99.18), and 98.52% F1 (95% CI: 96.95–99.43), while achieving an accuracy, sensitivity, PPV, and F1 in the testing phase are 99.27% (95% CI: 98.35–99.63), 99.45% (95% CI: 98.37–99.73), 98.83% (95% CI: 97.64–99.36), and 99.14% (95% CI: 98.24–99.61), respectively. A similar degradation in the performance of the VGG-19 model can be observed in the three-way classification.
A confusion matrix is utilized to determine further the distribution of the predicted X-ray images in different classes. The confusion matrices for the three-way and two-way classification of the BNCNN model are shown in Figure 3. The BNCNN model is tested on the test dataset, including 384 COVID-19 X-ray images, 981 Normal images, and 153 Viral Pneumonia images for three-way classification. For two-way classification, 369 COVID-19 images and 1004 normal images are used.
Figure 3a shows that 360 of 384 COVID-19 images, 979 of 981 normal images, and 131 of 153 Viral Pneumonia images are correctly classified for the three-way classification. Figure 3b shows 361 out of 369 COVID-19 images, and 1002 out of 1004 Normal images are classified for two-way classification.

4.2. Convergence Analysis

The convergence analysis is performed to study the stability of the learning patterns by the BNCNN model over the number of epochs. Figure 4 plots the BNCNN model accuracy and loss on the training and validation datasets over the training number of epochs. Figure 4a is for three-way classification, while Figure 4b is for two-way classification. It can be observed from Figure 4 that the BNCNN model has shown the best fit and convergence for three-way and two-way classification.

4.3. Comparison with Existing Models

Several ML models have been developed to diagnose COVID-19 automatically from X-ray images. Ozturk et al. (2020) [21] suggested the DarkCovidNet model for automatic COVID-19 detection from X-ray images. The model used 17 convolution layers with different numbers of filters in each layer. The DarkCovidNet model reported high reliability with an accuracy of 98% for two-way classification (i.e., COVID-19 and Normal) and 87% for 3-way classification (i.e., COVID-19, Normal, and Viral Pneumonia). In another work, Khan et al. (2020) [22] reported a CoroNet model based on pre-trained Inception architecture for COVID-19 detection from X-ray images. The CoroNet reported an accuracy of 95% for three-way classification.
Apostolopoulos and Mpesiana (2020) [23] suggested a novel convolution neural network (CNN) architecture and examined VGG-19 for COVID-19 detection from X-ray images. The model reported an accuracy of 93.48% for three-way classification. Wang et al. (2020) [24] reported a COVID-Net model for 3-way classification with an accuracy of 92.4%. Sethy and Behera (2020) [25] combined TL based on ResNet-50 and the support vector machine to diagnose COVID-19 from X-ray images. Their combined model achieved 95.38% accuracy for three-way class classification.
Horry et al. (2020) [26] used a pre-trained VGG-19 model for COVID-19 detection using a dataset comprising 115 COVID-19, 60361 Normal, and 322 Pneumonia X-ray images. The results showed that the pre-trained model attained 81% accuracy for 3-way classification. In another work, Rahimzadeh and Attar (2020) [27] detected COVID-19 from X-ray images using several deep neural networks and reported an accuracy of 91.4% for 3-way classification. Song et al. (2021) [28] developed a computer-aided method to classify images into COVID-19, bacterial Pneumonia, and Normal cases from a dataset collected from two provinces in China. Experimental results showed that the reported model could accurately identify COVID-19 cases with an accuracy of 86%.
Hussain et al. (2021) [29] suggested a CoroDet model for COVID-19 detection from X-ray images. The results confirmed that the CoroDet model could effectively identify COVID-19 cases with an accuracy of 94.2% for three-way and 99.1% for two-way classification. Chen (2021) [30] employed a CNN model to detect COVID-19 cases from X-ray images, and the results showed an accuracy of 85% for 3-way classification. Vinod et al. (2021) [31] suggested DeepCovix-net to effectively diagnose COVID-19 from X-ray and CT medical images and reported an accuracy of 96.8% for 3-way class classification. Anter et al. (2021) [32] proposed a model for COVID-19 diagnosis from X-ray images called AFCM-LSMA. Their suggested model achieved an accuracy of 96% for two-way classification. Basha et al. (2021) [33] reported a neurotrophic model for COVID-19 diagnosis from chest X-ray images with an accuracy of 98.7% for two-way classification.
Table 6 depicts the accuracy of the achievements of the existing models. Most studies in Table 6 used different datasets to validate their proposed model’s efficiency. The dataset used in this work is collected from existing studies and is publicly available [34]. It is not fair to compare the performance of the BNCNN with the other models since the size and chrematistics of the datasets are different, but the performance of these models is still comparable. However, Table 6 depicts the achievements of 13 existing models against the proposed BNCNN model in terms of accuracy for three-way and two-way class classification. As per the results in Table 6, the proposed BNCNN model provides higher accuracy for three-way and two-way classification than the existing models.

4.4. Statistical Analysis

To further evaluate and show the significance of the results of the BNCNN, Friedman’s test is performed [35]. We partially trained models to analyze learning speed based on testing accuracy. Testing accuracy performance for the BNCNN and the other pre-trained models, corresponding to the respective model’s test accuracy, are evaluated after 100 epochs.
Friedman’s test is performed with the null hypotheses where the testing accuracy samples for BNCNN and other pre-trained models originated from the same distribution, i.e., all models under comparison have equal testing accuracy. The alternate hypothesis assumes at least one of the models predicts different testing accuracy than other COVID-19 detection models with a significance level (p < 0.05). The samples for testing accuracy are taken from partially trained models at every ten epochs. It should be noted that higher rankings indicate improved performance. The average ranks of all the models for three-way and two-way classification are shown in Figure 5. The p-value calculated using Friedman’s test for three-way classification is 0.0146 and for two-way classification is p = 0.0053, which is less than the value of p. The test indicates that testing accuracy is different for comparative models. From Figure 5, BNCNN is ranked first for both three-way and two-way classification.
Holm’s posthoc test is used to confirm the differences in the behavior of the BNCNN (controlled model) and the other comparative models, as provided in Table 7. It uses BNCNN as the controlled model because of the highest rank in Friedman’s test. The results in Table 7 show a significant difference between the BNCNN and other pre-trained models. This proves the efficiency of the BNCNN as an alternative model for COVID-19 detection.
In addition, the ability of the BNCNN and other models are assessed by comparing them in terms of Area under the Receiver Operating Curve (AUC). We used the DeLong test to assess differences between the AUC of the models (p ≤ 0.05 is considered statistically significant) [36]. Figure 6 shows the discriminative ability of all the used models.
The proposed BNCNN model showed better discriminative ability with an equal AUC of 0.92 (95% CI: 0.87–0.95) and 0.94 (95% CI: 0.89–0.96) for three-way and two-way classification, respectively. Discriminative analysis of other models compared to the proposed BNCNN is shown in Table 8. No competing model has a statistically similar AUC to the proposed BNCNN model. It can be observed that all models have smaller AUCs than the BNCNN model, indicating the better discriminative ability of the developed BNCNN model.

5. Discussion

This work aimed to introduce a novel model for COVID-19 detection from X-ray medical images as an intelligent platform, which can provide updates on the patient’s health conditions and then guide further treatment. The proposed BNCNN model performance and efficacy are investigated using several evaluation measurements and compared to VGG-16, VGG-19, Inception-V3, and ResNet-50 pre-trained models. It is observed from the performance measures comparison that the suggested model shows superior results compared to the tested pre-trained models. In addition, the statistical tests of significance proved the superiority of the proposed model compared to other pre-trained models, which reflects the reliability of the developed BNCNN model. This is because the batch normalization layers in the proposed BNCNN model extracted features much better, while the max pooling and dropout layers reduced the computational complexity in their structure.
The main limitation associated with our study is the dataset, which is restricted only to publically available chest X-ray images. X-ray images are not recommended as the first line-imaging test for diagnosing COVID-19 due to the low positive detection rate at the early stages, which may be related to the insensitivity of X-ray images to the density of Ground Glass Opacity (GGO) [37]. Conversely, CT showed significant advantages in monitoring disease progression and served as an effective clinical diagnostic tool for early screening and diagnosis of COVID-19 [38]. CT has proved to be a good choice for early detection, severity assessment, and timely therapeutic effects evaluation for COVID-19, with or without laboratory confirmation [39]. X-ray and CT are medical imaging techniques widely used to assess and diagnose COVID-19 pneumonia patients. However, CT shows greater sensitivity for early pneumonic change, disease progression, and alternative diagnosis than X-ray [40]. ML can analyze irregular symptoms and other ‘red flags ‘of the infected cases at the early stage by using advanced algorithms [41]. These methods show a promising way for optimizing healthcare and improving the results of diagnostic and therapeutic procedures. Therefore, the extension of the proposed BNCNN model can be improved by using a large CT dataset to build an intelligent, accurate, and cost-effective platform for COVID-19.

6. Conclusions and Future Work

The COVID-19 disease is becoming increasingly significant as infected cases rapidly increase. Many researchers have devoted their efforts to developing ML models for COVID-19 detection that would benefit radiologists and health experts. This paper proposed an improved model named BNCNN to detect COVID-19 from chest X-ray images. The BNCNN uses VGG-inspired repetitive block structure, and each block comprises a convolution, followed by batch normalization, and max-pooling layers to improve the model’s generalization and feature map reduction. For confirming the reliability of the proposed BNCNN and other pre-trained models: VGG-16, VGG-19, Inception-V3, and ResNet-50 pre-trained models, a dataset is employed from ‘Kaggle’. The results show the superiority of the proposed BNCNN model over the pre-trained models, which significantly outperforms other comparative pre-trained models. Hence, the proposed BNCNN can be used to recognize the COVID-19 virus accurately. In the future, the performance of the proposed BNCNN model on CT imaging needs to be investigated.

Author Contributions

Conceptualization, I.A.-S., P.H.K. and L.A.; methodology, I.A.-S. and P.H.K.; software, I.A.-S. and P.H.K.; validation, L.A., A.M.A. and A.H.G.; formal analysis, I.A.-S. and P.H.K.; investigation, I.A.-S. and P.H.K.; resources, I.A.-S. and P.H.K.; data acquisition, I.A.-S. and P.H.K.; writing—original draft preparation, I.A.-S. and P.H.K. writing—review and editing, I.A.-S., P.H.K., L.A., M.E.A., B.E., A.M.A. and A.H.G.; visualization, I.A.-S.; supervision, L.A. and A.H.G.; project administration, L.A. and A.H.G.; funding acquisition, A.H.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available in Kaggle at [15].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. World Health Organization (WHO). Health Topics. Available online: www.who.int/bulletin/volumes/98/7/20-254045/en/ (accessed on 25 October 2021).
  2. Hall, L.O.; Paul, R.; Goldgof, D.B.; Goldgof, G.M. Finding COVID-19 from chest x-rays using deep learning on a small dataset. arXiv 2020, arXiv:2004.02060. [Google Scholar]
  3. Fanelli, D.; Piazza, F. Analysis and forecast of COVID-19 spreading in China, Italy and France. Chaos Solitons Fractals 2020, 134, 109761. [Google Scholar] [CrossRef] [PubMed]
  4. Lee, E.Y.; Ng, M.Y.; Khong, P.L. COVID-19 Pneumonia: What has CT taught us? Lancet Infect. Dis. 2020, 20, 384–385. [Google Scholar] [CrossRef] [PubMed]
  5. Ardabili, S.F.; Mosavi, A.; Ghamisi, P.; Ferdinand, F.; Varkonyi-Koczy, A.R.; Reuter, U.; Atkinson, P.M. COVID-19 outbreak prediction with machine learning. Algorithms 2020, 13, 249. [Google Scholar] [CrossRef]
  6. Raza, K. Artificial intelligence against COVID-19: A meta-analysis of current research. In Big Data Analytics and Artificial Intelligence against COVID-19: Innovation Vision and Approach; Springer: Cham, Switzerland, 2020; Volume 78, pp. 165–176. [Google Scholar]
  7. Ameri, P.; Inciardi, R.M.; di Pasquale, M.; Agostoni, P.; Bellasi, A.; Camporotondo, R.; Danzi, G. Pulmonary embolism in patients with COVID-19: Characteristics and outcomes in the Cardio-COVID Italy multicenter study. Clin. Res. Cardiol. 2021, 110, 1–9. [Google Scholar] [CrossRef]
  8. Magazzino, C.; Mele, M.; Schneider, N. The relationship between air pollution and COVID-19-related deaths: An application to three French cities. Appl. Energy 2020, 279, 115835. [Google Scholar] [CrossRef]
  9. Sajadi, M.M.; Habibzadeh, P.; Vintzileos, A.; Shokouhi, S.; Miralles-Wilhelm, F.; Amoroso, A. Temperature and latitude analysis to predict potential spread and seasonality for COVID-19. JAMA Netw. 2020, 3, e2011834. [Google Scholar] [CrossRef]
  10. Demongeot, J.; Flet-Berliac, Y.; Seligmann, H. Temperature Decreases Spread Parameters of the New COVID-19 Case Dynamics. Biology 2020, 9, 94. [Google Scholar] [CrossRef]
  11. Denton, E.L.; Zaremba, W.; Bruna, J.; LeCun, Y.; Fergus, R. Exploiting linear structure within convolutional networks for efficient evaluation. In Proceedings of the 27th Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  12. Lin, J.D.; Wu, X.Y.; Chai, Y.; Yin, H.P. Structure optimization of convolutional neural networks: A survey. Acta Autom. Sin. 2020, 46, 24–37. [Google Scholar]
  13. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the 32nd International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  14. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  15. Rahman, T.; Khandakar, A.; Qiblawey, Y.; Tahir, A.; Kiranyaz, S.; Kashem, S.B.A.; Islam, M.T.; Maadeed, S.A.; Zughaier, S.M.; Khan, M.S.; et al. Exploring the Effect of Image Enhancement Techniques on COVID-19 Detection Using Chest X-ray Images. Comput. Biol. Med. 2021, 132, 104319. Available online: https://www.kaggle.com/tawsifurrahman/covid19-radiography-database (accessed on 1 March 2022). [CrossRef] [PubMed]
  16. Sharma, S.; Mehra, R. Implications of pooling strategies in convolutional neural networks: A deep insight. Found. Comput. Decis. Sci. 2019, 44, 303–330. [Google Scholar] [CrossRef] [Green Version]
  17. Ahsan, M.M.; Alam, T.E.; Trafalis, T.; Huebner, P. Deep MLP-CNN model using mixed-data to distinguish between COVID-19 and Non-COVID-19 patients. Symmetry 2020, 12, 1526. [Google Scholar] [CrossRef]
  18. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  19. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  20. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  21. Ozturk, T.; Talo, M.; Yildirim, E.A.; Baloglu, U.B.; Yildirim, O.; Acharya, U.R. Automated detection of COVID-19 cases using deep neural networks with X-ray images. Comput. Biol. Med. 2020, 121, 103792. [Google Scholar] [CrossRef]
  22. Khan, A.I.; Shah, J.L.; Bhat, M.M. CoroNet: A deep neural network for detection and diagnosis of COVID-19 from chest X-ray images. Comput. Methods Programs Biomed. 2020, 196, 105581. [Google Scholar] [CrossRef]
  23. Apostolopoulosa, I.D.; Mpesiana, T.A. COVID-19: Automatic detection from x-ray images utilizing transfer learning with convolutional neural networks. Phys. Eng. Sci. Med. 2020, 43, 635–640. [Google Scholar] [CrossRef] [Green Version]
  24. Wang, L.; Lin, Z.Q.; Wong, A. COVID-net: A tailored deep convolutional neural network design for detection of COVID-19 cases from chest x-ray images. Sci. Rep. 2020, 10, 19549. [Google Scholar] [CrossRef]
  25. Sethy, P.K.; Behera, S.K.; Ratha, P.K.; Biswas, P. Detection of coronavirus disease (COVID-19) based on deep features and support vector machine. Int. J. Math. Eng. Manag. Sci. 2020, 5, 643–651. [Google Scholar] [CrossRef]
  26. Horry, M.J.; Chakraborty, S.; Paul, M.; Ulhaq, A.; Pradhan, B.; Saha, M.; Shukla, N. X-ray image based COVID-19 detection using pre-trained deep learning models. engrxiv, 2020; preprint. [Google Scholar] [CrossRef]
  27. Rahimzadeh, M.; Attar, A. A modified deep convolutional neural network for detecting COVID-19 and Pneumonia from chest X-ray images based on the concatenation of Xception and ResNet50V2. Inform. Med. Unlocked 2020, 19, 100360. [Google Scholar] [CrossRef]
  28. Song, Y.; Zheng, S.; Li, L.; Zhang, X.; Zhang, X.; Huang, Z.; Yang, Y. Deep learning enables accurate diagnosis of novel coronavirus (COVID-19) with CT images. IEEE/ACM Trans. Comput. Biol. Bioinform. 2021, 18, 2775–2780. [Google Scholar] [CrossRef] [PubMed]
  29. Hussain, E.; Hasan, M.; Rahman, M.A.; Lee, I.; Tamanna, T.; Parvez, M.Z. CoroDet: A deep learning based classification for COVID-19 detection using chest X-ray images. Chaos Solitons Fractals 2021, 142, 110495. [Google Scholar] [CrossRef] [PubMed]
  30. Chen, J.I.Z. Design of Accurate Classification of COVID-19 Disease in X-Ray Images Using Deep Learning Approach. J. ISMAC 2021, 3, 132–148. [Google Scholar] [CrossRef]
  31. Vinod, D.N.; Jeyavadhanam, B.R.; Zungeru, A.M.; Prabaharan, S.R.S. Fully automated unified prognosis of COVID-19 chest X-ray/CT scan images using Deep Covix-Net model. Comput. Biol. Med. 2021, 136, 104729. [Google Scholar] [CrossRef] [PubMed]
  32. Anter, A.M.; Oliva, D.; Thakare, A.; Zhang, Z. AFCM-LSMA: New intelligent model based on Lévy slime moul, d algorithm and adaptive fuzzy C-means for identification of COVID-19 infection from chest X-ray images. Adv. Eng. Inform. 2021, 49, 101317. [Google Scholar] [CrossRef]
  33. Basha, S.H.; Anter, A.M.; Hassanien, A.E.; Abdalla, A. Hybrid intelligent model for classifying chest X-ray images of COVID-19 patients using genetic algorithm and neutrosophic logic. Soft Comput. 2021, 1–16. [Google Scholar] [CrossRef]
  34. Cohen, J.P.; Morrison, P.; Dao, L. COVID-19 image data collection. arXiv 2020, arXiv:2003.11597. [Google Scholar]
  35. Martin, L.; Leblanc, R.; Toan, N.K. Tables for the Friedman rank test. Can. J. Stat. 1993, 21, 39–43. [Google Scholar] [CrossRef]
  36. DeLong, E.R.; DeLong, D.M.; Clarke-Pearson, D.L. Comparing the areas under two or more correlated receiver operating characteristic curves: A nonparametric approach. Biometrics 1988, 44, 837–845. [Google Scholar] [CrossRef]
  37. Wong, H.Y.F.; Lam, H.Y.S.; Fong, A.H.T.; Leung, S.T.; Chin, T.W.Y.; Lo, C.S.Y.; Ng, M.Y. Frequency and distribution of chest radiographic findings in patients positive for COVID-19. Radiology 2020, 296, E72–E78. [Google Scholar] [CrossRef] [Green Version]
  38. Kato, S.; Ishiwata, Y.; Aoki, R.; Iwasawa, T.; Hagiwara, E.; Ogura, T.; Utsunomiya, D. Imaging of COVID-19: An update of current evidences. Diagn. Interv. Imaging 2021, 102, 493–500. [Google Scholar] [CrossRef]
  39. Li, J.; Long, X.; Wang, X.; Fang, F.; Lv, X.; Zhang, D.; Xiong, N. Radiology indispensable for tracking COVID-19. Diagn. Interv. Imaging 2021, 102, 69–75. [Google Scholar] [CrossRef] [PubMed]
  40. Li, L.; Qin, L.; Xu, Z.; Yin, Y.; Wang, X.; Kong, B.; Xia, J. Using artificial intelligence to detect COVID-19 and community-acquired Pneumonia based on pulmonary CT: Evaluation of the diagnostic accuracy. Radiology 2020, 296, E65–E71. [Google Scholar] [CrossRef] [PubMed]
  41. Ito, R.; Iwano, S.; Naganawa, S. A review on the use of artificial intelligence for medical imaging of the lungs of patients with coronavirus disease 2019. Diagn. Interv. Radiol. 2020, 26, 443. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Example of pre-processed X-ray images for (a) COVID-19, (b) Normal, and (c) Viral Pneumonia.
Figure 1. Example of pre-processed X-ray images for (a) COVID-19, (b) Normal, and (c) Viral Pneumonia.
Pathogens 12 00017 g001
Figure 2. Proposed BNCNN model for COVID-19 detection.
Figure 2. Proposed BNCNN model for COVID-19 detection.
Pathogens 12 00017 g002
Figure 3. Confusion matrices of the BNCNN model for (a) three-way and (b) two-way classification (red: FP, yellow: FN, green: TP, and blue: TN for COVID-19 class).
Figure 3. Confusion matrices of the BNCNN model for (a) three-way and (b) two-way classification (red: FP, yellow: FN, green: TP, and blue: TN for COVID-19 class).
Pathogens 12 00017 g003
Figure 4. Variation of accuracy and loss of the BNCNN model on training and validation datasets over 100 epochs, (a) three-way and (b) two-way classification.
Figure 4. Variation of accuracy and loss of the BNCNN model on training and validation datasets over 100 epochs, (a) three-way and (b) two-way classification.
Pathogens 12 00017 g004
Figure 5. Average ranking based on testing accuracy at (the Friedman test) for (a) three-way (p = 0.0146) and (b) two-way (p = 0.0053) classification.
Figure 5. Average ranking based on testing accuracy at (the Friedman test) for (a) three-way (p = 0.0146) and (b) two-way (p = 0.0053) classification.
Pathogens 12 00017 g005
Figure 6. AUC and DeLong test for proposed BNCNN and other models for (a) three-way and (b) two-way classification.
Figure 6. AUC and DeLong test for proposed BNCNN and other models for (a) three-way and (b) two-way classification.
Pathogens 12 00017 g006
Table 1. Dataset description.
Table 1. Dataset description.
DatasetCOVID-19NormalViral PneumoniaSubtotal
Training28928153107612,121
Validation36210191341514
Testing36210201351518
Subtotal361610,192134515,153
Table 2. A summary of the parameters used in the proposed BNCNN model.
Table 2. A summary of the parameters used in the proposed BNCNN model.
Layer TypeOutput ShapeParameter TypeNo. of Parameters
Convolution 2D(None, 150, 150, 16)Trainable448
Batch Normalization(None, 150, 150, 16)Trainable + Non-trainable32 + 32
Max. Pooling 2D(None, 75, 75, 16) 0
Convolution 2D(None, 75, 75, 32)Trainable4640
Batch Normalization(None, 75, 75, 32)Trainable + Non-trainable64 + 64
Max. Pooling 2D(None, 37, 37, 32) 0
Convolution 2D(None, 37, 37, 64)Trainable18,496
Batch Normalization(None, 37, 37, 64)Trainable + Non-trainable128 + 128
Max. Pooling 2D(None, 37, 37, 64) 0
Convolution 2D(None, 18, 18, 128)Trainable73,856
Batch Normalization(None, 18, 18, 128)Trainable + Non-trainable256 + 256
Max. Pooling 2D(None, 9, 9, 128) 0
Flatten(None, 10368) 0
Dense(None, 256)Trainable2,654,464
Batch Normalization(None, 256)Trainable + Non-trainable512 + 512
Dropout(None, 256) 0
Dense(None, 128)Trainable32,896
Batch Normalization(None, 128)Trainable + Non-trainable256 + 256
Dropout(None, 128) 0
Softmax(None, 3)/(None, 2)Trainable387/258
Total trainable2,786,435/2,786,306
Total non-trainable1248
Total2,787,683/2,787,554
Table 3. Hyper-parameter settings of Adam optimizer for different models.
Table 3. Hyper-parameter settings of Adam optimizer for different models.
ModelLearning RateBatch SizeEpochs
VGG-160.00011630
VGG-190.00011630
Inception-V30.001830
ResNet-500.001830
BNCNN0.00011630
Table 4. Evaluation based on three-way class classification.
Table 4. Evaluation based on three-way class classification.
Model
(Trainable/Total Parameters)
DatasetAccuracySenPPVF1
VGG-16
(165,379/14,880,835)
Training96.6896.8196.8896.84
Validation97.3397.3097.3197.30
Testing96.6493.0196.8194.87
VGG-19
(264,195/20,288,579)
Training84.1982.9484.8683.89
Validation89.2088.9289.6889.30
Testing88.3485.3789.1287.21
Inception-V3
(427,523/54,765,027)
Training87.3387.5087.0387.26
Validation91.7692.2491.1991.71
Testing89.6089.9489.4689.70
ResNet-50
(13,141,507/36,729,987)
Training87.1185.0784.0284.54
Validation87.7886.3688.3787.35
Testing85.7083.6086.3884.97
Proposed BNCNN
(2,786,435/2,787,683)
Training96.3296.3096.5296.41
Validation97.4997.4397.4897.45
Testing96.8493.0697.4095.18
Table 5. Evaluation based on two-way class classification.
Table 5. Evaluation based on two-way class classification.
Model
(Trainable/Total Parameters)
PhaseAccSenPPVF1
VGG-16
(165,250/14,880,706)
Training96.6196.6196.6396.62
Validation98.0198.0198.0198.01
Testing94.2894.2994.2994.29
VGG-19
(263,682/20,288,066)
Training89.5489.4389.5489.48
Validation92.6192.6292.6192.61
Testing92.5692.5492.5592.54
Inception-V3
(427,394/54,764,898)
Training87.3387.5087.0387.26
Validation91.7692.2491.1991.72
Testing89.6089.9489.4689.70
ResNet-50
(13,141,378/36,729,858)
Training84.7584.7584.8884.81
Validation89.2089.3489.1289.23
Testing89.0889.1888.7888.98
Proposed BNCNN
(2,786,306/2,787,554)
Training97.2897.2897.2397.25
Validation98.5598.5298.5398.52
Testing99.2799.4598.8399.14
Table 6. Accuracy comparison between the existing models and the proposed BNCNN model.
Table 6. Accuracy comparison between the existing models and the proposed BNCNN model.
Model3-Way Classification2-Way Classification
Ozturk et al. (2020) [21]8798
Khan et al. (2020) [22]95N/A
Apostolopoulos and Mpesiana (2020) [23]93.48N/A
Wang et al. (2020) [24]92.4N/A
Sethy and Behera (2020) [25]95.38N/A
Horry et al. (2020) [26]86N/A
Rahimzadeh and Attar (2020) [27]94.299.1
Song et al. (2021) [28]81N/A
Hussain et al. (2021) [29]91.4N/A
Chen (2021) [30]85N/A
Vinod et al. (2021) [31]96.8N/A
Anter et al. (2021) [32]N/A96
Basha et al. (2021) [33]N/A98.7
Proposed BNCNN model96.8499.27
Table 7. Holm’s test for comparing BNCNN with other pre-trained models based on the testing accuracy.
Table 7. Holm’s test for comparing BNCNN with other pre-trained models based on the testing accuracy.
3-Way Classification 2-Way Classification
Competing Modelp-ValueCorrected p-ValueNull Hypothesis Competing Modelp-ValueCorrected p-ValueNull Hypothesis
VGG-160.09560.5735Reject VGG-160.05970.2731Reject
VGG-190.11650.5825Reject VGG-190.00490.0489Reject
Inception-V30.00110.0105Reject Inception-V30.02320.1852Reject
ResNet-500.00440.0397Reject ResNet-500.01860.1677Reject
Table 8. Sensitivity analysis for different models using the DeLong test.
Table 8. Sensitivity analysis for different models using the DeLong test.
Competing ModelAUC95% CIp-Value
3-way classificationVGG-160.89[0.85–0.92]0.0412
VGG-190.87[0.82–0.93]0.0431
Inception-V30.84[0.78–0.89]0.0216
ResNet-500.85[0.79–0.90]0.0312
2-way classificationVGG-160.90[0.84–0.95]0.0274
VGG-190.87[0.81–0.93]0.0296
Inception-V30.85[0.80–0.89]0.0012
ResNet-500.88[0.82–0.92]0.0132
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Al-Shourbaji, I.; Kachare, P.H.; Abualigah, L.; Abdelhag, M.E.; Elnaim, B.; Anter, A.M.; Gandomi, A.H. A Deep Batch Normalized Convolution Approach for Improving COVID-19 Detection from Chest X-ray Images. Pathogens 2023, 12, 17. https://doi.org/10.3390/pathogens12010017

AMA Style

Al-Shourbaji I, Kachare PH, Abualigah L, Abdelhag ME, Elnaim B, Anter AM, Gandomi AH. A Deep Batch Normalized Convolution Approach for Improving COVID-19 Detection from Chest X-ray Images. Pathogens. 2023; 12(1):17. https://doi.org/10.3390/pathogens12010017

Chicago/Turabian Style

Al-Shourbaji, Ibrahim, Pramod H. Kachare, Laith Abualigah, Mohammed E. Abdelhag, Bushra Elnaim, Ahmed M. Anter, and Amir H. Gandomi. 2023. "A Deep Batch Normalized Convolution Approach for Improving COVID-19 Detection from Chest X-ray Images" Pathogens 12, no. 1: 17. https://doi.org/10.3390/pathogens12010017

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop