CNN Models Using Chest X-Ray Images for COVID-19 Detection: A Survey

CNN Models Using Chest X-Ray Images for COVID-19 Detection: A Survey

Naçima Mellal* Sofiane Zaidi

Department of Networks & Telecommunications, Research Laboratory of Artificial Intelligence and Autonomous Objects, University of Oum El Bouaghi, Oum El Bouaghi 04000, Algeria

Department of Mathematics and Computer Science, Research Laboratory on Computer Science’s Complex Systems (RELA(CS)2), University of Oum El Bouaghi, Oum El Bouaghi 04000, Algeria

Corresponding Author Email: 
nassima.mellal@univ-oeb.dz
Page: 
1565-1576
|
DOI: 
https://doi.org/10.18280/ria.370621
Received: 
26 July 2023
|
Revised: 
5 August 2023
|
Accepted: 
10 September 2023
|
Available online: 
27 December 2023
| Citation

© 2023 IIETA. This article is published by IIETA and is licensed under the CC BY 4.0 license (http://creativecommons.org/licenses/by/4.0/).

OPEN ACCESS

Abstract: 

The COVID-19 pandemic, which began in 2019, has spread globally, causing substantial human suffering and economic disruption. A collaborative global effort is essential to combat this disease. Artificial Intelligence has played a pivotal role in this battle, providing numerous deep learning strategies to automate the detection of COVID-19. Among these strategies, Convolutional Neural Network (CNN) models have emerged as a particularly potent tool for COVID-19 detection through the analysis of medical images. The present paper provides a comprehensive survey of various CNN models that have been developed for the classification of X-ray images in the context of COVID-19. These models have been categorized into three groups for the purpose of this review. The first category is centered on models that utilize transfer learning from pre-trained CNN models. The second one consists of Custom CNN Models that have been developed from scratch. The final category, known as Hybrid CNN Models, integrates elements from both of the previous categories. Outlined with details regarding the dataset size, the number of classes considered, the architecture of the model, and the criteria used for performance evaluation, encompassing accuracy, sensitivity, and specificity. This review thus provides a comprehensive overview of the current landscape of CNN models for COVID-19 detection using X-Ray images, offering valuable insights for future research in this critical area.

Keywords: 

convolutional neural network (CNN), classification, COVID-19, chest X-Ray images

1. Introduction

COVID-19 (Coronavirus disease 2019) is a contagious virus that has been spreading all over the world since December 2019 in Wuhan (China). It is caused by a virus called by the International Committee on Taxonomy of Viruses(ICT) SARS-CoV-2 [1]. There are various methods to identify patients infected with COVID-19, the most effective test is Reverse Transcription Polymerase Chain Reaction (RT-PCR). In this method, a miniature target sequence of nucleic acids which consists of a fragment of DNA is copied many times to make it easier to detect. As a consequence, the test is positive when this amplification is detected (using a fluorophore-labeled probe). While RT-PCR tests have been widely used to detect COVID-19 and have played a crucial role in the pandemic response, they have certain limitations, particularly in terms of accuracy and scalability [2].

The accuracy of RT-PCR tests can vary depending on when the test is administered during the infection. They are generally more sensitive in the early stages of infection when viral loads are higher but may produce false negatives in later stages or if the sample quality is suboptimal. RT-PCR tests can sometimes yield false positives, especially if there is contamination during sample collection or processing. So RT-PCR test is not completely accurate and needs a second test for the diagnosis confirmation. Thus, RT-PCR test is time-consuming. It takes more than a day to get results. It is also costly. RT-PCR test is resource-intensive and can be relatively costly to administer, especially when large-scale testing is required. RT-PCR tests are not efficiently scalable because it requires specialized laboratory equipment, managed by professional technicians. Scaling up testing capacity can be challenging, especially during surges in cases, leading to delays in testing and reporting which makes it an expensive test and difficult to scale.

Other methods that are useful for detecting COVID-19 include clinical examination, pathological tests, and radiography. Due to their easy accessibility, many clinicians prefer to use chest imaging for the diagnosis of COVID-19. The role of chest imaging in assessing complications, disease progression, and prognosis of COVID-19 has been discussed by Inui et al. [3]. Also, the cochrane COVID-19 Diagnostic Test Accuracy Group [4] considered chest imaging as a good diagnostic tool for detecting and managing COVID-19 pneumonia. They have conducted different studies to evaluate the diagnostic accuracy of chest imaging in people of any age with suspected COVID-19. Their studies assessed chest CT, chest X-Ray, or ultrasound of the lungs for the diagnosis of COVID-19, using a reference standard that included RT-PCR.

Among medical imaging modalities, chest X-Ray was generally preferred for different reasons. Chest X-Ray images are widely available and can be easily taken in most healthcare settings. They are less expensive compared to other imaging techniques like computed tomography (CT) scans. As COVID-19 is an emergency, X-Ray images are very beneficial because they provide quick results compared to RT-PCR tests, and they do not expose the patient to as much radiation as other imaging techniques like CT scans. X-Ray images can be useful in monitoring the progress of the disease and assessing the response to treatment. Cozzi et al. [5] investigated the principal radiological features of COVID-19 by describing the most important chest X-Ray findings in a selected cohort of patients. They also correlated the radiological appearance with the RT-PCR test and the outcome of the patients.

Accelerate diagnostic means using automatic tools that provide rapid results. The AI techniques have proven their effectiveness in different stages: the acquisition of medical images, then segmentation, and finally in the diagnosis of COVID-19. To illustrate the latest advances in medical imaging and radiology in the fight against COVID-19, Shi et al. [6] have illustrated the integration of AI with X-Rays, which are widely used in frontline hospitals.

Huang et al. [7] have provided a discussion of the challenges and perspectives of machine learning (ML) and deep learning (DL) in the detection of COVID-19. As a result, the researchers found that an AI model can be as accurate as experienced physicians in the diagnosis of COVID-19.

In general, deep learning has shown massive potential for healthcare applications. Specifically, for medical data analysis and diagnosis through medical image processing [8]. Due to their ability to learn complex features and patterns from the images, deep learning models, especially convolutional neural networks (CNNs) have shown promising results in detecting COVID-19 from X-Ray images and have the potential to assist professionals in diagnosing the disease accurately [9].

However, some challenges need to be addressed, such as the lack of publicly available data for training and evaluating COVID-19 detection models which limit the ability to build accurate models. Another issue is the variation in X-Ray image quality, and not all images will be clear or well-defined, leading to issues with classification. Also, the generalization challenge, using X-ray images to detect COVID-19 may limit the generalization capability of the model when it comes to identifying the virus in other parts of the body.

Consequently, the main intention behind this survey is to provide an overview of CNNs that have contributed to tackling these challenges. We have organized the reviewed CNN Models into three main categories. The first one focuses on transfer learning from pre-trained CNN models. In the second, the model is developed from scratch and called the Custom CNN Model. The last category couples the two previous ones to constitute the Hybrid CNN Model. A summary for each study is presented by giving the dataset size, the number of classes, the model architecture, and the performance evaluation criteria (Accuracy, Sensitivity, and Specificity).

The remainder of this paper is organized as follows. Section 2 presents the description of related work, where we classify CNN models through three categories: 1- Transfer Learning from pre-trained models; 2- Custom model, and 3- Hybrid model. Section 3 is devoted to some preliminary notions dealing with image classification, Datasets, CNN model architecture. Finally, the paper is concluded in Section 4 with a discussion of the related works.

2. Related Work

Deep learning has developed good solutions with high accuracy for medical image segmentation and classification in the health sector [10-13]. Many deep learning techniques based on medical images chest X-Ray have been developed to detect COVID-19 [14]. Jain et al. [15] achieved good results with their model in detecting COVID-19. They used a dataset containing 1215 images, including 250 COVID-19 ones. They got 98.93% of accuracy, 98.66% of specificity, 96.39% of precision, and 98.15% F-1 score.

Ucar and Korkmaz [16] obtained 98.26% accuracy employing a dataset of 5949 chest X-Ray images containing 76 COVID-19 images. Wang et al. [17] suggested a deep learning model drawing out visual characteristics from CT Scan images categorizing Covid19. They exploited a Dataset of 1065 computed tomography scans with 740 images of patients suffering from viral pneumonia, and 325 images are of COVID-19 patients. The model attains a good accuracy of 79.3%. The model is used to distinguish between COVID-19 and other typical viral pneumonia.

CNN models perform feature extraction of medical images (chest X-Rays) of patients suspected to be infected with the virus and classification. CNNs extract relevant features using convolutional layers that apply filters on the images, reduce the dimensionality of feature maps with pooling layers, and use fully connected layers to perform classification by generating class probabilities. The model's weights are optimized during training to minimize the loss and regularized to prevent overfitting. Transfer learning is another popular technique used in CNNs to improve feature extraction. It involves using pre-trained models that have already learned relevant features from large datasets and fine-tuning them on a specific task. This helps to speed up the training process and improve the accuracy of the model.

Different approaches for X-Ray image classification based on CNN are proposed. Some are focusing on transfer learning from pre-trained CNN models. Others are developed from scratch (custom CNN models). Other works couple these two approaches and constitute an approach called the « hybrid CNN Model». In this section, we present an overview of related work, where we organize them through three categories: 1- Transfer Learning from pre-trained models; 2- Custom model, and 3- Hybrid model.

2.1 Transfer learning from pre-trained CNN models

A pre-trained model is a model that is trained on large datasets to extract high-level features from images such as VGG, Inception, Resnet, etc. Using a pre-trained model is justified for several motives. Firstly, expensive computation power is required if the training models are large and the datasets are sizeable. The second reason is that the process of training in large models can be slow. It can take several weeks. Finally, a pre-trained model can gain time in order to generalize the network and accelerate the convergence. Adapting a pre-trained model and applying it to a new dataset or task by fine-tuning is the principle of the Transfer Learning technique.

In this section, we give an overview of related studies that focus on transfer learning (TL) from pre-trained CNN models with an X-Ray dataset.

Traditional, parallel convolutional layers, and residual connections are three ideas integrated in the developed model DCNN (deep convolutional neural network) [18]. The proposed model accomplished an 86.6% of F1-score when trained from scratch. It got 89.4% using the transfer learning from different domains of the targeted dataset, and it obtained 97.6%with transfer learning from the same domain of the targeted dataset. However, there is a crucial issue according to the source data type utilized by the TL compared to the target dataset. This in terms of data features, sizes, and characteristics. Also, it has been demonstrated that the transfer learning from different domains has a little effect on the performance, as lightweight models trained from scratch perform close to transferred standard models [19].

The Table 1 ranks recent research and exploits several metrics, such as Accuracy, Sensitivity, and Specificity, to compare the performance of models.

In each study, the first column of the table illustrates the best pre-trained model in terms of accuracy when compared to the others.

2.2 Custom CNN models

Custom CNN models are deep learning architectures that are designed from scratch to solve a specific problem or task. They are constructed using a series of layers and tailored to the specific requirements of the problem at hand. Compared to pre-trained models, custom CNN models can have a much higher level of control over the architecture’s structure, allowing for more flexibility and customization.There were several contributions in which authors presented their own CNN models without dependending on pre-trained models.The Table 2 presents some of these studies aiming to classify chest X-Ray images as COVID-19 or other classes.

Table 1. Overview of models based on TL from pre-trained CNN models

Best Model

& Year

Number of

Pre-Trained Models

Dataset

Classification

Accuracy (%)

Sensitivity (%) (Recall)

Specificity (%)

MobiNet V2 [20] 2020

Five (5):GG19, MobiNetV2, Inception, xCeption, Inception ResNetV

1427 CHEST X-RAYimages including 224 Covid images

2 classes: (Covid, Normal)

96.78

98.66

96.46

Inception-ResNetV2 [21] 2020

Seven (7): VGG16, VGG19, DenseNet20, Inception_ResNetV2, InceptionV3, tesnet50, and MobileNet_V2

6087 imagesincluding 1724 Covid images

3 classes: (bacterial, neumonia, Covid19, normal)

92.18

92.11

96.06

Googlenet [22] 2020

Three (3): Alexnet, Googlenet, and Restnet18

307 CHEST X-RAYimages including69 images of COVID-19

4 classes: (Covid19, normal, pneumonia bacterial, pneumonia virus)

80.56

80.56

/

VGG16 [23] 2020

Nine (9)

/

3 classes: (Covid19, Normal,peumonia bacteria

95.88

/

/

VGG16 [24] 2020

/

8474CHEST X-RAYimages

3 classes: (Covid 19, infected pneumonia, no-COVID-19infected pneumonia, normal (non-pneumonia))

94.50

98.40

98.00

VGG16 [25] 2020

/

396 CHEST X-RAYimages including 132 of COVID-19

3 classes: (Covid-19;healthy,normal

86.00

86.00

93.00

Inception-ResNetV2 [26] 2020

AlexNet GoogleNetVgg16

Vgg19 ResNet18

ResNet50 ResNet101 InceptionV3 InceptionResNetv2 SqueezeNet Densenet201Xception

6290 CHEST X-RAYimages

4 classes: (normal, bacterial, C0VID-19, viral infections (non Covid)

/

95.89

99.52

ResNet50V2 [27] 2020

5 (VGG16 ResNet50V2

XceptionMobileNetDenseNet121)

678 CHEST X-RAYimages

4 classes: (normal, bacterial, COVID-19, viral)

98.15

98.26

98.89

VGG16 [28] 2021

VGG16, VGG19, ResNet, DenseNet, and InceptionV3

120 CHEST X-RAYImages

2 classes:(Negative; Positive)

80.00

80.00

/

Densnet161 [29] 2021

3 (VGG16Densnet161ResNet50)

1431CHEST X-RAYImages

2 classes:(Covid-19; pneumonia)

97.10

100

/

vgg19 [30] 2022

One (vgg19)

3797 CHEST X-RAYImages

3 classes: (Covid; pneumonia,healthy)

97.11

97.00

/

Table 2. Custom CNN models

Model & Year

Dataset

Classification

Accuracy (%)

Sensitivity (Recall) (%)

Specificity (%)

CovXNet [31] 2020

5856 CHEST X-RAYimages including 305 images of Covid-19

4 classes: (Covid-19, Normal, viral, bacterial pneumonia)

90.00

89.00

89.00

CNN-LSTM [32] 2020

4575 CHEST X-RAYimages

3 classes: (Covid, normal, pneumonia)

99.40

99.90

99.20

CoroNet [33] 2020

1300 CHEST X-RAYimages including 290 images of Covid-19

4 classes: (Covid, Pneumonia bacterial, pneumonia viral, normal)

89.60

89.92

96.40

multi-CNN [34] 2020

950 CHEST X-RAYimages including 453 images of Covid-19

2 classes:(Covid, Normal)

91.16

98.50

/

COVID-NET [35] 2020

13,975 CHEST X-RAYimages

3 classes: (Normal, Pneumonia, Covid)

93.30

91.00

/

Covid-xnet [36] 2020

/

2 classes: (normal, Covid)

94.43

92.53

96.30

DarkCovidNet [37] 2020

1125 CHEST X-RAYimages

including 125 COVID-19 images.

2 classes: (Covid, no-findings)

98.08

95.13

95.30

[38] 2020

/

2 classes: (Covid, normal)

93.00

/

/

CNN [39] 2021

1184 CHEST X-RAYimages

2 classes: (Covid-19; normal)

98.65

98.49

98.82

 

OptCoNet [40] 2021

900 CHEST X-RAYimages

3 classes: (Covid, normal, pneumonia)

97.78

97.75

96.25

2019NCOV [41] 2021

9544 Chest X-rays Images

2 classes: (Covid, normal)

93.80

99.50

/

SARS-Net [42] 2022

13975 CHEST X-RAYimages

3 classes:(Normal, Pneumonia, Covid)

97.60

92.90

/

2dCNN-BiCuDNNLSTM [43] 2022

6863 CHEST X-RAY

images (1000 COVID-19 patients, 3863 normal cases, and 2000 pneumonia patients)

3 classes: (Covid, normal, pneumonia)

93.00

100

/

RT2022 [44] 2022

5255 CHEST X-RAYimages

2 classes: (Covid, normal)

90.60

/

/

2.3 Hybrid CNN models

Hybrid CNNs are models which combine predesigned CNNs with other machine learning techniques such as Support Vector Machine (SVM). The idea is simply to add new blocks to the pre-trained CNN model. For example, In the study [45], the process of Covid-19 detection using chest X-Ray is operated on three popular CNN models: VGG19, MobileNet, and InceptionV3 by adding two mechanisms: ConvLSTM and SE Block. ConvLSTM means Convolutional Long Short-Term Memory which is a layer used to encode the spatial dependency among the feature mapsreceived from the CNN's last layer of convolution and also to improve the model's image representational capability.The other block is squeeze-and-excitation block. It is then added and used to assign weights to significant local features. These two mechanisms are employed to improve the classification strength of CNN models through VGG19 + ConvLSTM + SE block, Inception V3 + ConvLSTM + SE block, and MobileNet + ConvLSTM + SE block. Accuracies using only VGG16, Inception V3, and MobileNet are respectively: 96.01%, 95.22%, and 96.98%. The accuracies after adding blocks are respectively: 97.88%, 97.23%, and 97.80%.

The Table 3 presents a brief overview of significant hybrid CNN models.

Table 3. Hybrid CNN models

Model

& Year

Dataset

Classification

Used

Pre-Trained Models

Accuracy (%)

Sensitivity (Recall) (%)

Specificity (%)

Covid ConvLSTM

[45] 2020

/

/

VGG19,

97.88

98.67

98.00

InceptionV3,

97.23

MobileNet

97.80

/ [46]2020

3440

3 classes:

(Covid-19, Pneumonia,N ormal)

Vgg16

96.91

98.33

98.68

/ [47] 2020

4273 Pneumonia CHEST X-RAY images, 1182 COVID-19 and 1583Normal

3 classes: (Covid, Pneumonia, Normal)

VGG-16

90.19

94.16

/

VGG-16 & SVM

93.15

93.15

VGG-16 & Bagging

90.19

92.16

VGG-16 & Ada Boot

90.19

90.10

HOG+CNN [48] 2021

6432 CHEST X-RAY images

3 classes:(COVID-19 positive, pneumonia positive,

Normal)

/

96.74

/

/

/ [49] 2021

/

/

VGG16+ SVM

99.82

99.82

/

/ [50] 2022

6432 X-ray images

3 classes: - COVID-19(positive), non-COVID-19 induced pneumonia (negative), normal (negative))

VGG-16

94.00

93.00

87.00

ResNet50

97.00

96.00

99.00

MobileNetV2

96.00

95.00

91.00

/ [51] 2022

912 CHEST X-RAY

images for regular people, and 912 CHEST X-RAY images for

COVID-19 infected people

2 classes:(Covid; normal)

InceptionResnet-v2

100

100

100

 

Inception-v3

99.00

98.00

100

 

ResNet50

98.00

97.00

100

ResNet101

100

100

100

/ [52] 2022

2613 CHEST X-RAY images

2 classes:(Covid; normal)

ResNet50

98.00

100

99.00

 

EfficientNetB0,

99.00

99.00

98.00

HDCNN [53] 2022

6000 CHEST X-RAY Images

3 classes: (Covid, pneumonia, normal)

/

98.00

97.00

97.00

3. CNNs Materials and Techniques

Deep learning techniques have the potential for modeling large sets of complex data. Building and training accurate and efficient deep learning models are based on key concepts. First, the neural network architecture is the foundation of a deep learning model. It is composed of layers of interconnected artificial neurons that process the input data and generate output predictions. The design of the architecture is crucial to the model's performance. Other concepts must be considered carefully such as the selection of appropriate activation functions used to introduce non-linearity into the output of a neural network. Also, the regularization techniques are used to reduce variance in the model by introducing restrictions that prevent it from overfitting to the training data. The accuracy and efficiency of the deep learning model can significantly differ depending on the hyperparameters that are selected.Hyperparameter tuning is a critical step to fine-tune the performance of the model but is also a challenging and time-consuming task.

CNN is one of the most widely used models in the domain of deep learning, because the appropriate features of input data are automatically extracted without any intervention of humans. CNN covers a wide range of topics: we mention computer vision [54], speech processing [55], Face Recognition [56], etc. CNNs are commonly used for image classification and have been successfully applied to medical image classification tasks. They use filter kernels to extract relevant features from images, and multiple layers for image classification.

The present section is devoted to discussing the basic bricks to build a CNN model that deals with Covid-19 classification from X-Ray images. It introduces the important characteristics of an X-Ray image. Then explains why we need a large dataset of X-Ray images and different techniques used for preprocessing data. Additionally, it displays the basic CNN architecture for image classification and the different categories of CNN models. Popular CNN models are summarized in each category. The process of image classification using CNN models, in general, is presented. The section is finalized by talking about metrics that will be used to evaluate how well the CNN model performs once built.

3.1 X-Ray characteristics

An electromagnetic pulse called an X-Ray is frequently used in medicine to image several organs, including the bones and the lungs [57]. Techniques based on chest X-Rays offer noninvasive disease diagnosis. The difference in tissue chest X-Ray attenuation produces a two-dimensional contrast image, commonly known as an X-Ray. A patient's chest X-Rays are often obtained from a variety of positions and angles having regard to the panel of the source and the detector. Radiographic analysis chest X-Ray is one of the simplest and most basic detection methods that are widely accessible and inexpensive. Chest X-Ray can play an important role in the diagnosis of patients suspected of having SARS-CoV-2. Figure 1 gives three examples of chest X-Ray images of the infected persons and three others for normal persons. The identification process of Covid-19 is based on the presence of a white patchy shadow in the lungs [58].

Figure 1. Examples of chest X-Ray images of normal vs. infected persons

3.2 Dataset

To combat the pandemic of COVID-19, open source data and methods are needed to enable the global scientific community to collaborate on research that is verifiable and transparent [59, 60]. As a result, a set of electronic medical repositories are developed to help researchers locate the appropriate resource.In our context, we are interested in public chest X-Ray datasets. Several datasets are developed such as PADCHEST1 [61]. It is reported at San Juan hospital in Spain by radiologists, the project (COVID chest X-Ray) [62] has the aim of collecting X-ray images showing COVID-19 from online and others sources [63]. In general, most of the available X-Ray datasets for COVID-19 classification are collected from Kaggle and Github platforms, such as: Covid-19 radiography database [64], COVID 19-image-dataset [65], COVID-chest X-Ray-dataset [66]. However, these datasets are limited by the number of lung images with COVID-19 infection. And this is not efficient for training a deep learning model. The model may overfit the data. Therefore, a hybrid dataset combining different repositories is used to train the existing works [67].

The dataset is split into different sets, once it is determined. The purpose of splitting is to avoid overfitting. In deep learning, overfitting occurs when the capacity to learn is so large that the network is learning wrong features instead of meaningful patterns. The original dataset is usually divided into three sets: training, validation, and test sets.The model is trained to employ the training set. The validation set contains examples utilized to vary parameters in the learning process. The final model evaluates the testing set and compares it to the previous datasets. Generally, there are no rules that specify how to partition the data. This may depend on the predictor's number in a predictive model or the original data pool size. Each dataset split should contain subfolders. The number of subfolders depends on the number of output classifications (binary, ternary, ...), for example, in binary classification, we find for each dataset two subfolders: Covid and normal.

3.3 Preprocessing techniques

Preprocessing techniques are methods aiming to transform data before it is delivered to the machine learning or deep learning algorithm. Training a CNN on unprocessed images can lead to poor classification results, which can be interpreted in terms of overfitting issues. Overfitting is considered to be a major problem in deep learning.To reduce this problem, an important step must be added as a preprocessing step. It is called data augmentation, which is the process of artificially increasing the amount of data by generating new data points from existing data [68]. Data augmentation techniques include:

Resizing: To standardize the dataset, all the images are resized into a fixed size such as 224×224 or 299×299.

Flipping or Rotating: To increase the sample size of the datasets, horizontal and vertical flipping are mainly used.

Scaling or Cropping: To reduce the redundancy, because not all parts of the images need to be used.

Brightness or Intensity: To increase or decrease the images brightness.

Traditional image preprocessing techniques such as Contrast Stretching, Histogram Equalization, Smoothing, Sharpening, Adaptive Winner Filter, Histogram Enhancement, and Color Space Conversion, are used in some works [69-72]. As a consequence, the preprocessing step is essential to improve data quality and accurate AI Models.

3.4 Basic CNN architecture for classification

The structure of CNNs is inspired by neurons in the brains of humans or animals. It works in a similar way to a conventional neural network [73]. The CNN composition consists of a series of connected layers of three types: convolutional layers, pooling layers, and fully connected layers. The CNN architecture is based on a combination of these layers [74]. The role of the convolutional layer is to subjugate the input data (image) to the convolution operation to produce a series of output features.

Pooling layers can reduce the size of the input image which comes from the convolution results without losing important features. This reduces the number of parameters and computations in the network. This leads to identifying patterns in the data more quickly and accurately in the network.

Finally, the last type of layers is Fully Connected (FC) layers. They are usually positioned before the output layer and form the final layers of a CNN structure. After feature extraction, comes the role of fully connected layers to split the data into different classes, so as to classify the input (see Figure 2 [75]).

Figure 2. CNN architecture

Usually, a complementing layer is added in the CNN architecture: The dropout layer. It is a technique that drops some neurons toward the next layer and doesn't modify all others. The dropped neuron does not participate in the back-propagation or the forward-propagation during the training process, but it is a part of the full-scale network to perform predictions during the testing process. In the prediction process, a technique called the Activation Function is adopted. It selects a neuron that should be activated or not. This means that it decides whether the neuron’s input to the network is significant or not. ReLU (Rectified Linear Unit), Softmax, Sigmoid, and tanH (Hyperbolic tangent) are the most used activation functions.

RelU: It converts all values of the input to positive numbers. The central advantage of ReLU compared to others is the lower computational load, because it does not activate, simultaneously, all the neurons. 

Softmax: It is usually applied as a final activation function in a neural network. It aims at normalizing a network’s output to a probability distribution over a set of classes (predicted output classes).

Sigmoid: It is used for binary classification in the CNN model. The input of the sigmoid function is real numbers, and the output value is limited between 0 and 1.

Tanh: It is equivalent to the sigmoid function. Its input is real numbers, but the output is limited to between -1 and 1.

3.5 Popular CNN Models Categories

CNN architectures have undergone numerous recent developments. Bhatt et al. [76] summarized 8 categories of CNN architectures: (1) Spatial exploitation, (2) multi-path, (3) depth, (4) breadth, (5) dimension, (6) channel boosting, (7) feature-map exploitation, and (8) attention-based CNN.

Spatial Exploitation (SE): In order to improve performance, spatial filters have been used. Spatial filters typically consist of a small matrix of weights (usually 3×3 or 5×5) applied to each pixel in the image. The strength of the feature detected in the image is determined by the weights in the filter. Convolutional operations are affected by the size of the filter. Small filters are usually used to extract fine-grained information, while large filters are used to extract coarse-grained information.

Multi-path (MP): Shortcut connections or numerous pathways allow the flow of information between layers by evading some in-between levels [77]. Through cross-layer connectivity, the network is divided into sections. These pathways extend the gradient to lower layers, which resolve the vanishing gradient issue.

Depth (Dp): Network depth is a key factor in determining the performance of a network and its ability to learn [78-81]. Deep networks are more effective at representing specific classes of functionality than shallow systems.

Breadth (B): besides depth, width is a key factor in network learning process. It is confirmed that using ReLU activation functions in neural networks needs to be wide enough to maintain a universal approximation property, while at the same time increasing in depth [82]. However, the main issue is that a set of layers may fail to learn features even if the depth is raised. Furthermore, if the maximum width of a network is not greater than the input dimension, then it is not possible for any deep network to approximate a class of continuous functions on a compact set [83]. Consequently, to address this issue, researchers focused on wide and thin designs than deep and narrow ones.

Dimension (Dm): separable convolutions (called also depth-wise separable) are introduced to improve the efficiency of normal convolutions. The separable convolutions use point-wise to encode the spatial one, and depth-wise to encode the channel-wise.

Channel boosting (CB): The channel boosting is called also as the input channel dimension. It adds supplementary learners in the CNN for the enhancement of the network’s representation [84]. The input representation is an influential parameter in network learning. Network performance can be hampered by poor diversity and information about the class in the input.

Feature-map exploitation (FE): The selection of feature maps can have a significant impact on the improvement of network generalization. Different feature extraction steps are achieved, allowing for diverse types of features. However, excessive feature sets may provide a noise effect, leading to the over-fitting of the network [85].

Attention (A): In CNN, improving representation and overcoming computational limitations are the aims of using the concept of attention. It also gives intelligence to a CNN in order to differentiate elements even in complex scenarios and busy backdrops.

The Table 4 lists some CNN models basing on each cited category [33, 41].

Table 4. Summary of popular CNN models in each category

CNN Model

& Year

Category

Dataset

Main Findings

Depth

Parameter

Error Rate

Input Size

LeNet1998

SE

MNIST

1st CNN architecture

 

0.060M

0.95

/

AlexNet2012

SE

Image Net

Dropout + ReLU

8

60M

16.4

227×227×3

ZfNet2014

SE

Image Net

Intermediate layers visualization

8

60 M

11.7

224×224×3

VGG2014

SE

Image Net

Small filter size + increase depth

16 – 19

138 M

7.3

224×224×3

GoogleNet2015

SE

Image Net

different filter sizes + increase depth +block concept

22

4 M

6.7

224×224×3

Inception-V3

Dp+B

Image Net

Small filter size+ better feature representation

48

23.6 M

3.5

229×229×3

High way2015

Dp+MP

CIFAR-10

1st idea multipath

19-32

2.3 M

7.76

32×32×3

Inception ResNet

Dp+B+MP

Image Net

Residual links

164

55.8 M

3.52

229×229×3

Residual Attention

2017

A

CIFAR-10CIFAR-100

1st model to

introduce attention

mechanism

452

8.6 M

3.90

20.4

40×40×3

Squeeze-and-excitationNetworks2017

FE

ImageNet

Modeled interdependencies

between channels

152

27.5

2.3

229 ×229×3

224×224×3

320×320×3

Channel Boosted

CNN 2018

CB

/

Boosted original channelswith additional informationby artificial channels

/

/

/

/

EdgeNet 2019

Dm

/

Introduced concept ofvisual intelligence at theedge

/

/

/

/

DiceNET 2021

Dm

/

Introduceddimension-based CNN, including height, width,and depth

/

/

/

 

3.6 CNN model performance

The evaluation of a CNN algorithm's classification performance is important. Generally,Accuracy, Recall/Sensitivity, Precision, Specificity, F1 Score, confusion matrix, Area Under Curve (AUC) are the most used metrics for the evaluation. To test a CNN model, a variety of these evaluation metrics are mixed.

Metrics formulas base on the terms TP, TN, FP, FN, where:

TP means True positive, the number of cases that were correctly predicted as positive;

TN denotes True Negative, the number of cases that were correctly predicted as negative;

FP signifies False Positive, the number of cases that were incorrectly predicted as positive;

FN means False Negative, the number of cases that were incorrectly predicted as negative.

The accuracy is a metric used to measure the percentage of the true and negative cases that are correctly predicted. It is calculated by the formula:

$Accuracy =\frac{T P+T N}{T P+T N+F P+F N}$      (1)

The sensitivity (called also Recall) is a measure of a model's ability to correctly identify true positives. It is given by the formula:

$Recall =\frac{\mathrm{TP}}{T P+F N}$      (2)

The specificity is the percentage of negative cases correctly predicted. It is calculated by the formula:

$Specifity =\frac{T N}{T N+F P}$      (3)

The precision measures the proportion of positive prediction results that are correct.

$Precision =\frac{T P}{T P+F P}$      (4)

There are additional metrics, such as the confusion matrix and the Area under the Curve (AUC), that the present paper does not mention.

3.7 Process of image classification using CNN models

The image classification technique is used for classifying or predicting the category of a particular object in an image. The main goal of this technique is identifying accurately the features in an image. In this study, we summarize the process of image classification based on the CNN model in 7 main steps (see Figure 3). The learning stage includes step 1 to step 5, while the inference one is determined by step 6 and step 7.

Figure 3. Process of image classification using CNN

Step 1: Choosing the Dataset, a large training dataset is essential;

Step 2: Preprocessing the dataset using different techniques;

Step 3: Construction of the CNN model;

Step 4: Training and testing the model on the chosen dataset (in step 1);

Step 5: The classification performance of a CNN algorithm is evaluated using metrics such as accuracy, specificity, F1 score, and sensitivity.

Step 6: Preprocessing the X-Ray image to be classified to enhance it and reduce noise or unwanted details;

Step 7: CNN classification of the introduced image.

4. Discussion and Conclusion

In literature, a lot of works demonstrate that the CNN model is a powerful method to classify Covid-19 using X-Ray images. We have reviewed significant research papers since the emergence of the virus. Based on the literature and after analyzing studies in this field, we have classified CNN Models into three categories. The first one focuses on transfer learning from pre-trained CNN models. In the second, the CNN model is developed from scratch and called the Custom CNN Model. The last category couples the two previous ones to constitute the Hybrid CNN Model. We have presented a summary of each model in each category by giving the dataset size, the number of classes, the model architecture, and the performance evaluation criteria (Accuracy, Sensitivity, and Specificity). As deep learning models need a large amount of data, dataset size is considered a critical factor in determining the classification performance of the CNN model. The majority of CNN models have achieved high performance for Covid-19 classification. Without ignoring that most research applies different techniques for preprocessing data to enhance their quality.

We further observed that models based on transfer learning from pre-trained models are the most used category in terms of performance and this is for multiple reasons. A pre-trained model that has learned general features can be used to improve the accuracy of the model on a new dataset. By transferring the knowledge acquired from a pre-trained model, transfer learning decreases the training time as well.Since the model has already learned the features, it requires less time to learn new features and patterns in the new domain. Fewer computational resources are needed to train a new model and achieve comparable results. Finally, Transfer learning allows models to be more easily scaled to new applications, platforms, and tasks, allowing for more efficient use of resources and increased flexibility. Without ignoring the great efforts of the developed models in the two other categories.

Custom CNN models can provide high accuracy if they are designed correctly and trained on large datasets. Still, they tend to require more computational power and time to train and optimize compared to pre-trained models and transfer learning techniques.

Hybrid CNN models can be challenging to design and optimize, and finding the optimal combination and tune hyperparameters can be time-consuming. However, these models can lead to superior performance and generalization ability compared to both transfer learning and custom CNN models.

The proposed models in the three categories have the same goal which is the contribution to reduce the ongoing pandemic's impact by improving diagnosis accuracy and assisting research efforts.

We remarked, in our survey, that CNN Models for COVID-19 detection acquire data from a wide range of sources (labs, hospitals, universities, research centers, etc.). So developing a health data collection platform is a necessity that enhances the possibilities for storing, sharing, and analyzing health data among researchers in several fields. Through real-time data acquisition and AI machine learning, researchers have access to huge, diverse, and high-quality datasets. The platform can help them uncover new patterns, create novel therapies, interventions, and prevention methods, and make informed judgments, all of which can reduce illness burden and improve patient outcomes through improved medical research.

We noticed, also, that the proposed models are not applicable or beneficial in hospitals or clinics, while most of these efforts and research aim to help the radiologist to make better decisions in taackling Covid-19. CNNs have the potential for early detection and diagnosis. CNNs have been used to monitor areas to ensure that people maintain social distancing. CNNs can accelerate the vaccine development process by using computational models to predict which parts of the virus are most prone to inducing an immune response. This helps researchers design and optimize vaccine candidates that can potentially protect against COVID-19.

Finally, we aim with this paper to provide important insights for CNN models based on X-Ray images to help researchers develop a standardized, reliable, and comprehensive system for COVID-19 diagnosis and control that can scale up globally and help mitigate the impact of the pandemic on public health and the economy.

  References

[1] Rabi, F.A., Al Zoubi, M.S., Kasasbeh, G.A., Salameh, D.M., Al-Nasser, A. D. (2020). SARS-CoV-2 and coronavirus disease 2019: What we know so far. Pathogens, 9(3): 231.  https://doi.org/10.3390/pathogens9030231

[2] Arevalo-Rodriguez, I., Buitrago-Garcia, D., Simancas-Racines, D., et al. (2020). False-negative results of initial RT-PCR assays for COVID-19: A systematic review. PloS one, 15(12), e0242958. https://doi.org/10.1371/journal.pone.0242958

[3] Inui, S., Gonoi, W., Kurokawa, R., Nakai, Y., Watanabe, Y., Sakurai, K., Ishida, M., Fujikawa, A., Abe, O. (2021). The role of chest imaging in the diagnosis, management, and monitoring of coronavirus disease 2019 (COVID-19). Insights into imaging, 12:155. https://doi.org/10.1186/s13244-021-01096-1 

[4] Cochrane COVID-19 Diagnostic Test Accuracy Group. (2022). Thoracic imaging tests for the diagnosis of COVID-19. Cochrane Database of Systematic Reviews, 2022(5): CD013639. https://doi.org/10.1002/14651858.CD013639.pub5

[5] Cozzi, D., Albanesi, M., Cavigli, E., Moroni, C., Bindi, A., Luvarà, S., Lucarini, S., Busoni, S., Mazzoni, L.N., Miele, V. (2020). Chest X-ray in new Coronavirus Disease 2019 (COVID-19) infection: Findings and correlation with clinical outcome. La Radiologia Medica, 125: 730-737. https://doi.org/10.1007/s11547-020-01232-9

[6] Shi, F., Wang, J., Shi, J., Wu, Z., Wang, Q., Tang, Z.Y., He, K.L., Shi, Y.H., Shen, D.G. (2020). Review of artificial intelligence techniques in imaging data acquisition, segmentation, and diagnosis for COVID-19. IEEE Reviews in Biomedical Engineering, 14: 4-15. https://doi.org/10.1109/RBME.2020.2987975 

[7] Huang, S., Yang, J., Fong, S., Zhao, Q. (2021). Artificial intelligence in the diagnosis of COVID-19: Challenges and perspectives. International Journal of Biological Sciences, 17(6): 1581. https://doi.org/10.7150/ijbs.58855

[8] Waring, J., Lindvall, C., Umeton, R. (2020). Automated machine learning: Review of the state-of-the-art and opportunities for healthcare. Artificial Intelligence in Medicine, 104: 101822. https://doi.org/10.1016/j.artmed.2020.101822

[9] Mahalingam, S.G., Pandraju, S. (2021). Unsupervised convolutional filter learning for COVID-19 classification. Revue d'Intelligence Artificielle, 35(5): 425-429. https://doi.org/10.18280/ria.350509

[10] Razzak, M.I., Naz, S., Zaib, A. (2018). Deep learning for medical image processing: Overview, challenges and the future. Classification in BioApps: Automation of Decision Making, pp. 323-350. https://doi.org/10.1007/978-3-319-65981-7_12

[11] Kermany, D.S., Goldbaum, M., Cai, W., et al. (2018). Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell, 172(5): 1122-1131. https://doi.org/10.1016/j.cell.2018.02.010 

[12] Lakhani, P., Sundaram, B. (2017). Deep learning at chest radiography: Automated classification of pulmonary tuberculosis by using convolutional neural networks. Radiology, 284(2): 574-582. https://doi.org/10.1148/radiol.2017162326

[13] Quadri, R., Deshpande, A. (2022). Deep learning-based segmentation and classification of COVID-19 infection severity levels from CT scans. Revue d'Intelligence Artificielle, 36(1): 41-48. https://doi.org/10.18280/ria.360105

[14] Chandra, T.B., Verma, K., Singh, B.K., Jain, D., Netam, S.S. (2021). Coronavirus disease (COVID-19) detection in chest X-ray images using majority voting based classifier ensemble. Expert systems with applications, 165: 113909. https://doi.org/10.1016/j.eswa.2020.113909

[15] Jain, G., Mittal, D., Thakur, D., Mittal, M.K. (2020). A deep learning approach to detect Covid-19 coronavirus with X-Ray images. Biocybernetics and Biomedical Engineering, 40(4): 1391-1405. https://doi.org/10.1016/j.bbe.2020.08.008

[16] Ucar, F., Korkmaz, D. (2020). COVIDiagnosis-Net: Deep Bayes-SqueezeNet based diagnosis of the coronavirus disease 2019 (COVID-19) from X-ray images. Medical Hypotheses, 140: 109761. https://doi.org/10.1016/j.mehy.2020.109761

[17] Wang, S., Kang, B., Ma, J., Zeng, X.J., Xiao, M.M., Guo, J., Cai, M.J., Yang, J.Y., Li, Y.D., Meng, X.F., Xu, B. (2021). A deep learning algorithm using CT images to screen for Corona Virus Disease (COVID-19). European Radiology, 31: 6096-6104. https://doi.org/10.1007/s00330-021-07715-1

[18] Alzubaidi, L., Fadhel, M.A., Al-Shamma, O., Zhang, J., Santamaría, J., Duan, Y., R. Oleiwi, S. (2020). Towards a better understanding of transfer learning for medical imaging: A case study. Applied Sciences, 10(13): 4523. https://doi.org/10.3390/app10134523

[19] Raghu, M., Zhang, C., Kleinberg, J., Bengio, S. (2019). Transfusion: Understanding transfer learning for medical imaging. In33rd Conference on Neural Information Processing Systems (NeurIPS 2019), Vancouver, Canada.

[20] Apostolopoulos, I.D., Mpesiana, T.A. (2020). Covid-19: Automatic detection from x-ray images utilizing transfer learning with convolutional neural networks. Physical and Engineering Sciences in Medicine, 43: 635-640. https://doi.org/10.1007/s13246-020-00865-4

[21] El Asnaoui, K., Chawki, Y. (2021). Using X-ray images and deep learning for automated detection of coronavirus disease. Journal of Biomolecular Structure and Dynamics, 39(10): 3615-3626. https://doi.org/10.1080/07391102.2020.1767212

[22] Loey, M., Smarandache, F., M. Khalifa, N.E. (2020). Within the lack of chest COVID-19 X-ray dataset: A novel detection model based on GAN and deep transfer learning. Symmetry, 12(4): 651. https://doi.org/10.3390/sym12040651

[23] Makris, A., Kontopoulos, I., Tserpes, K. (2020). COVID-19 detection from chest X-Ray images using deep learning and convolutional neural networks. In 11th Hellenic Conference on Artificial Intelligence, Athens Greece, pp. 60-66. https://doi.org/10.1145/3411408.3411416 

[24] Heidari, M., Mirniaharikandehei, S., Khuzani, A.Z., Danala, G., Qiu, Y., Zheng, B. (2020). Improving the performance of CNN to predict the likelihood of COVID-19 using chest X-ray images with preprocessing algorithms. International Journal of Medical Informatics, 144: 104284. https://doi.org/10.1016/j.ijmedinf.2020.104284

[25] Civit-Masot, J., Luna-Perejón, F., Domínguez Morales, M., Civit, A. (2020). Deep learning system for COVID-19 diagnosis aid using X-ray pulmonary images. Applied Sciences, 10(13): 4640. https://doi.org/10.3390/app10134640

[26] Majeed, T., Rashid, R., Ali, D., Asaad, A. (2020). Covid-19 detection using CNN transfer learning from X-ray images. Physical and Engineering Sciences in Medicine, 43: 1289-1303. https://doi.org/10.1007/s13246-020-00934-8

[27] Shorfuzzaman, M., Masud, M. (2020). On the detection of Covid-19 from chest X-ray images using CNN-based transfer learning. Computers, Materials & Continua, 64(3): 1359-1381. https://doi.org/10.32604/cmc.2020.011326

[28] Sahinbas, K., Catak, F. O. (2021). Transfer learning-based convolutional neural network for COVID-19 detection with X-ray images. In Data science for COVID-19, pp. 451-466. https://doi.org/10.1016/B978-0-12-824536-1.00003-4

[29] Polat, Ç., Karaman, O., Karaman, C., Korkmaz, G., Balcı, M.C., Kelek, S.E. (2021). COVID-19 diagnosis from chest X-ray images using transfer learning: Enhanced performance by debiasing dataloader. Journal of X-ray Science and Technology, 29(1): 19-36. https://doi.org/10.3233/XST-200757 

[30] Chakraborty, S., Paul, S., Hasan, K.A. (2022). A transfer learning-based approach with deep CNN for Covid-19-and pneumonia-affected chest X-ray image classification. SN Computer Science, 3: 17. https://doi.org/10.1007/s42979-021-00881-5

[31] Mahmud, T., Rahman, M.A., Fattah, S.A. (2020). CovXNet: A multi-dilation convolutional neural network for automatic COVID-19 and other pneumonia detection from chest X-ray images with transferable multi-receptive feature optimization. Computers in Biology and Medicine, 122: 103869. https://doi.org/10.1016/j.compbiomed.2020.103869

[32] Islam, M.Z., Islam, M.M., Asraf, A. (2020). A combined deep CNN-LSTM network for the detection of novel coronavirus (COVID-19) using X-ray images. Informatics in Medicine Unlocked, 20: 100412. https://doi.org/10.1016/j.imu.2020.100412

[33] Khan, A.I., Shah, J.L., Bhat, M.M. (2020). CoroNet: A deep neural network for detection and diagnosis of COVID-19 from chest X-ray images. Computer methods and Programs in Biomedicine, 196: 105581. https://doi.org/10.1016/j.cmpb.2020.105581

[34] Abraham, B., Nair, M.S. (2020). Computer-aided detection of COVID-19 from X-ray images using multi-CNN and Bayesnet classifier. Biocybernetics and Biomedical Engineering, 40(4): 1436-1445. https://doi.org/10.1016/j.bbe.2020.08.005

[35] Wang, L., Lin, Z.Q., Wong, A. (2020). Covid-net: A tailored deep convolutional neural network design for detection of COVID-19 cases from chest X-ray images. Scientific Reports, 10(1): 19549. https://doi.org/10.1038/s41598-020-76550-z

[36] Duran-Lopez, L., Dominguez-Morales, J.P., Corral-Jaime, J., Vicente-Diaz, S., Linares-Barranco, A. (2020). COVID-XNet: A custom deep learning system to diagnose and locate COVID-19 in chest X-ray images. Applied Sciences, 10(16): 5683. https://doi.org/10.3390/app10165683

[37] Ozturk, T., Talo, M., Yildirim, E.A., Baloglu, U.B., Yildirim, O., Acharya, U.R. (2020). Automated detection of COVID-19 cases using deep neural networks with X-ray images. Computers in Biology and Medicine, 121: 103792. https://doi.org/10.1016/j.compbiomed.2020.103792

[38] Medhi, K., Jamil, M., Hussain, M.I. (2020). Automatic detection of COVID-19 infection from chest X-ray using deep learning. medrxiv, 2020-05. https://doi.org/10.1101/2020.05.10.20097063 

[39] Bushra, K.F., Ahamed, M.A., Ahmad, M. (2021). Automated detection of COVID-19 from X-ray images using CNN and Android mobile. Research on Biomedical Engineering, 37(3): 545-552. https://doi.org/10.1007/s42600-021-00163-2

[40] Goel, T., Murugan, R., Mirjalili, S., Chakrabartty, D. K. (2021). OptCoNet: An optimized convolutional neural network for an automatic diagnosis of COVID-19. Applied Intelligence, 51: 1351-1366. https://doi.org/10.1007/s10489-020-01904-z

[41] Basantwani, N., Kumar, A., Gangwar, S., Olkha, A., Mathur, G. (2021). COVID-19 detection android app based on chest X-rays & CT scans. INFOCOMP Journal of Computer Science, 20(1): 91-100.

[42] Kumar, A., Tripathi, A.R., Satapathy, S.C., Zhang, Y.D. (2022). SARS-Net: COVID-19 detection from chest X-rays by combining graph convolutional network and convolutional neural network. Pattern Recognition, 122: 108255. https://doi.org/10.1016/j.patcog.2021.108255

[43] Kanwal, A., Chandrasekaran, S. (2022). 2DCNN-bicudnnlstm: Hybrid deep-learning-based approach for classification of COVID-19 X-ray images. Sustainability, 14(11): 6785. https://doi.org/10.3390/su14116785

[44] Mellal, N., Bouzekri, K., Sabri, M., Abdennebi, S. (2022). Android app based on CNN for COVID-19 detection using chest X-ray images. In 2022 4th International Conference on Pattern Analysis and Intelligent Systems (PAIS), El Bouaghi, Algeria, pp. 1-6. https://doi.org/10.1109/PAIS56586.2022.9946916 

[45] Dey, S., Bhattacharya, R., Malakar, S., Schwenker, F., Sarkar, R. (2022). CovidConvLSTM: A fuzzy ensemble model for COVID-19 detection from chest X-rays. Expert Systems with Applications, 206: 117812. https://doi.org/10.1016/j.eswa.2022.117812

[46] Bayoudh, K., Hamdaoui, F., Mtibaa, A. (2020). Hybrid-COVID: A novel hybrid 2D/3D CNN based on cross-domain adaptation approach for COVID-19 screening from chest X-ray images. Physical and Engineering Sciences in Medicine, 43: 1415-1431. https://doi.org/10.1007/s13246-020-00957-1

[47] Dubey, R.K. (2020). Deep learning based hybrid models for prediction of COVID-19 using chest X-ray. TechRxiv. https://doi.org/10.36227/techrxiv.12839204.v1

[48] Rahman, M.M., Nooruddin, S., Hasan, K.A., Dey, N.K. (2021). HOG+ CNN Net: Diagnosing COVID-19 and pneumonia by deep neural network from chest X-Ray images. SN Computer Science, 2: 371. https://doi.org/10.1007/s42979-021-00762-x

[49] Alawad, W., Alburaidi, B., Alzahrani, A., Alflaj, F. (2021). A comparative study of stand-alone and hybrid CNN models for COVID-19 detection. International Journal of Advanced Computer Science and Applications, 12(6): 877-883. https://doi.org/10.14569/IJACSA.2021.01206102

[50] Ting, P., Kasam, A., Lan, K. (2022). Applications of convolutional neural networks in chest X-ray analyses for the detection of COVID-19. Annals of Biomedical Science and Engineering, 6: 1-7. https://doi.org/10.29328/journal.abse.1001015 

[51] Shoaib, M.R., Emara, H.M., Elwekeil, M., El-Shafai, W., Taha, T.E., El-Fishawy, A.S., El-Rabaie, M., El-Samie, F.E.A. (2022). Hybrid classification structures for automatic COVID-19 detection. Journal of Ambient Intelligence and Humanized Computing, 13(9): 4477-4492. https://doi.org/10.1007/s12652-021-03686-9

[52] Chola, C., Mallikarjuna, P., Muaad, A.Y., Bibal Benifa, J.V., Hanumanthappa, J., Al-antari, M.A. (2021,). A hybrid deep learning approach for COVID-19 diagnosis via CT and X-ray medical images. Computer Sciences & Mathematics Forum, 2(1): 13. https://doi.org/10.3390/IOCA2021-10909

[53] Kumar, M., Shakya, D., Kurup, V., Suksatan, W. (2022). COVID-19 prediction through X-ray images using transfer learning-based hybrid deep learning approach. Materials Today: Proceedings, 51: 2520-2524. https://doi.org/10.1016/j.matpr.2021.12.123 

[54] Khan, S., Rahmani, H., Shah, S.A.A., Bennamoun, M., Medioni, G., Dickinson, S. (2018). A guide to convolutional neural networks for computer vision. San Rafael: Morgan & Claypool Publishers.

[55] Pandey, A., Wang, D. (2019). A new framework for CNN-based speech enhancement in the time domain. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 27(7), 1179-1188. https://doi.org/10.1109/TASLP.2019.2913512

[56] Wang, M., Deng, W. (2021). Deep face recognition: A survey. Neurocomputing, 429: 215-244. https://doi.org/10.1016/j.neucom.2020.10.081

[57] Çallı, E., Sogancioglu, E., van Ginneken, B., van Leeuwen, K.G., Murphy, K. (2021). Deep learning for chest X-ray analysis: A survey. Medical Image Analysis, 72: 102125. https://doi.org/10.1016/j.media.2021.102125

[58] Zhao, D., Yao, F., Wang, L., Zheng, L., Gao, Y.J., Ye, J., Guo, F.,Zhao, H., Gao, R.B (2020). A comparative study on the clinical features of coronavirus 2019 (COVID-19) pneumonia with other pneumonias. Clinical Infectious Diseases, 71(15): 756-761. https://doi.org/10.1093/cid/ciaa247

[59] Alimadadi, A., Aryal, S., Manandhar, I., Munroe, P.B., Joe, B., Cheng, X. (2020). Artificial intelligence and machine learning to fight COVID-19. Physiological Genomics, 52(4): 200-202. https://doi.org/10.1152/physiolgenomics.00029.2020

[60] Pham, Q.V., Nguyen, D. C., Huynh-The, T., Hwang, W.J., Pathirana, P.N. (2020). Artificial intelligence (AI) and big data for coronavirus (COVID-19) pandemic: A survey on the state-of-the-arts. IEEE Access, 8: 130820-130839. https://doi.org/10.1109/ACCESS.2020.3009328

[61] Bustos, A., Pertusa, A., Salinas, J.M., De La Iglesia-Vaya, M. (2020). Padchest: A large chest X-ray image dataset with multi-label annotated reports. Medical Image Analysis, 66: 101797. https://doi.org/10.1016/j.media.2020.101797

[62] Cohen, J.P., Morrison, P., Dao, L. (2020). COVID-19 image data collection. arXiv preprint arXiv: 2003.11597. https://doi.org/10.48550/arXiv.2003.11597

[63] Cohen, J.P., Morrison, P., Dao, L., Roth, K., Duong, T. Q., Ghassemi, M. (2020). COVID-19 image data collection: Prospective predictions are the future. arXiv preprint arXiv: 2006.11988. https://doi.org/10.48550/arXiv.2006.11988 

[64] COVID-19 Radiography Database. https://www.kaggle.com/datasets/tawsifurrahman/Covid19-radiography-database.

[65] COVID-19 Image Dataset. https://www.kaggle.com/datasets/pranavraikokte/Covid19-image-dataset.

[66] Covid-Chestxray-Dataset. https://github.com/ieee8023/Covid-chestxray-dataset.

[67] Al-Antari, M.A., Hua, C.H., Bang, J., Lee, S. (2021). Fast deep learning computer-aided diagnosis of COVID-19 based on digital chest X-ray images. Applied Intelligence, 51(5): 2890-2907. https://doi.org/10.1007/s10489-020-02076-6

[68] Shah, F.M., Joy, S.K.S., Ahmed, F., Hossain, T., Humaira, M., Ami, A.S., Paul, S., Jim, M.A.R.K., Ahmed, S. (2021). A comprehensive survey of COVID-19 detection using medical images. SN Computer Science, 2(6): 434. https://doi.org/10.1007/s42979-021-00823-1

[69] Tahir, A., Qiblawey, Y., Khandakar, A., Rahman, T., Khurshid, U., Musharavati, F., Kiranyaz, S., Chowdhury, M.E.H. (2021). Coronavirus: Comparing COVID-19, SARS and MERS in the eyes of AI. Preprint from arXiv.

[70] Al-Karawi, D., Al-Zaidi, S., Polus, N., Jassim, S. (2020). Machine learning analysis of chest CT scan images as a complementary digital test of coronavirus (COVID-19) patients. MedRxiv, 2020-04. https://doi.org/10.1101/2020.04.13.20063479  

[71] Lv, D., Qi, W., Li, Y., Sun, L., Wang, Y. (2020). A cascade network for detecting COVID-19 using chest X-rays. arXiv preprint arXiv: 2005.01468. https://doi.org/10.48550/arXiv.2005.01468 

[72] Zhang, J., Xie, Y., Li, Y., Shen, C., Xia, Y. (2020). Covid-19 screening on chest x-ray images using deep learning based anomaly detection. arXiv preprint arXiv:2003.12338. https://doi.org/10.48550/arXiv.2003.12338 

[73] Alzubaidi, L., Zhang, J., Humaidi, A.J., Al-Dujaili, A., Duan, Y., Al-Shamma, O., Santamaría, J., Fadhel, M.A., Al-Amidie, M., Farhan, L. (2021). Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. Journal of big Data, 8: 53. https://doi.org/10.1186/s40537-021-00444-8

[74] O'Shea, K., Nash, R. (2015). An introduction to convolutional neural networks. arXiv preprint arXiv:1511.08458. https://doi.org/10.48550/arXiv.1511.08458 

[75] Balaji, S. (2020). Binary Image classifier CNN using TensorFlow. Techiepedia. https://medium.com/techiepedia/binary-image-classifier-cnn-using-tensorflow-a3f5d6746697.

[76] Bhatt, D., Patel, C., Talsania, H., Modi, K., Pandya, S., Ghayvat, H. (2021). CNN variants for computer vision: History, architecture, application, challenges and future scope. Electronics, 10(20): 2470. https://doi.org/10.3390/electronics10202470

[77] Hu, J., Shen, L., Albanie, S., Sun, G., Wu, E. (2019). Squeeze-and-excitation networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 42(8): 2011-2023. https://doi.org/10.1109/TPAMI.2019.2913372

[78] Simonyan, K., Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv: 1409.1556. https://doi.org/10.48550/arXiv.1409.1556

[79] Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z. (2016). Rethinking the inception architecture for computer vision. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, pp. 2818-2826. https://doi.org/10.1109/CVPR.2016.308

[80] Szegedy, C., Liu, W., Jia, Y., et al. (2015). Going deeper with convolutions. In 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, pp. 1-9. https://doi.org/10.1109/CVPR.2015.7298594 

[81] Ioffe, S., Szegedy, C. (2015). Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv:1502.03167. https://doi.org/10.48550/arXiv.1502.03167 

[82] Nguyen, Q., Mukkamala, M.C., Hein, M. (2018). Neural networks should be wide enough to learn disconnected decision regions. arXiv preprint arXiv: 1803.00094. https://doi.org/10.48550/arXiv.1803.00094

[83] He, K., Gkioxari, G., Dollár, P., Girshick, R. (2018). Mask R-CNN. IEEE Transactions on Pattern Analysis and Machine Intelligence, 42(2): 386-397. https://doi.org/10.1109/TPAMI.2018.2844175

[84] Khan, A., Sohail, A., Zahoora, U., Qureshi, A.S. (2020). A survey of the recent architectures of deep convolutional neural networks. Artificial Intelligence Review, 53: 5455-5516. https://doi.org/10.1007/s10462-020-09825-6

[85] Kawaguchi, K., Huang, J., Kaelbling, L.P. (2019). Effect of depth and width on local minima in deep learning. Neural Computation, 31(7): 1462-1498. https://doi.org/10.1162/neco_a_01195