Next Article in Journal
Understanding the Role of Persistent Organic Pollutants and Stress in the Association between Proximity to the World Trade Center Disaster and Birth Outcomes
Next Article in Special Issue
Identifying Factors That Predict Behavioral Intention to Stay under Lockdown during the SARS-CoV-2 Pandemic Using a Structural Equation Model
Previous Article in Journal
Feelings of Loneliness: Understanding the Risk of Suicidal Ideation in Adolescents with Internet Addiction. A Theoretical Model to Answer to a Systematic Literature Review, without Results
Previous Article in Special Issue
Diverse Manifestations of COVID-19: Some Suggested Mechanisms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Efficient Deep Learning Model to Detect COVID-19 Using Chest X-ray Images

1
School of Computing Sciences and Computer Engineering, The University of Southern Mississippi, Hattiesburg, MS 39406, USA
2
Department of Epidemiology and Biostatistics, School of Public Health, College of Health Sciences, Jackson State University, Jackson, MS 39213, USA
*
Author to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2022, 19(4), 2013; https://0-doi-org.brum.beds.ac.uk/10.3390/ijerph19042013
Submission received: 10 December 2021 / Revised: 25 January 2022 / Accepted: 9 February 2022 / Published: 11 February 2022

Abstract

:
The tragic pandemic of COVID-19, due to the Severe Acute Respiratory Syndrome coronavirus-2 or SARS-CoV-2, has shaken the entire world, and has significantly disrupted healthcare systems in many countries. Because of the existing challenges and controversies to testing for COVID-19, improved and cost-effective methods are needed to detect the disease. For this purpose, machine learning (ML) has emerged as a strong forecasting method for detecting COVID-19 from chest X-ray images. In this paper, we used a Deep Learning Method (DLM) to detect COVID-19 using chest X-ray (CXR) images. Radiographic images are readily available and can be used effectively for COVID-19 detection compared to other expensive and time-consuming pathological tests. We used a dataset of 10,040 samples, of which 2143 had COVID-19, 3674 had pneumonia (but not COVID-19), and 4223 were normal (not COVID-19 or pneumonia). Our model had a detection accuracy of 96.43% and a sensitivity of 93.68%. The area under the ROC curve was 99% for COVID-19, 97% for pneumonia (but not COVID-19 positive), and 98% for normal cases. In conclusion, ML approaches may be used for rapid analysis of CXR images and thus enable radiologists to filter potential candidates in a time-effective manner to detect COVID-19.

1. Introduction

The Severe Acute Respiratory Syndrome Coronavirus-2 (SARS-CoV-2) has created the greatest pandemic of COVID-19 in nearly a century and has already traumatized the whole world [1,2,3]. The pandemic reportedly started in the City of Wuhan, China in late December, 2019 [1,2,3]. In two years, as of 9 December 2021, over 268 million cases and 5.3 million deaths have been caused by the pandemic worldwide [4]. In the United States alone, there were over 50.4 million cases and 814,000 deaths caused by COVID-19. Lung involvement is the most common noticeable manifestation of the disease, ranging from asymptomatic disease or mild pneumonia, to severe disease associated with hypoxia, critical disease associated with shock, respiratory failure and multi-organ failure or death [5]. An early and accurate detection of the COVID-19 virus is one of the cornerstones for containing the spread and minimizing the death toll of the disease.
In an early stage of the outbreak, due to the lack of knowledge of the virus genome, the detection of the disease was rudimentary. COVID-19 RT-PCR is a real-time reverse transcription polymerase chain reaction test. It offers a qualitative detection of nucleic acid from SARS-CoV-2 which is present in the respiratory specimens during the infection [6]. RT-PCR is used as the gold standard for the detection of COVID-19. Many challenges, however, still remain in the COVID-19 detection research. The RT-PCR sensitivity rate is 60–70%, and the test is expensive [7]. Antigen detection tests, or commonly known as rapid COVID tests, are fairly accurate in detecting positive cases but the false negative rate of the test is high and the sensitivity of the test is low (60.5%) [7].
Two other simpler methods in the detection of COVID-19 include the use of chest X-ray (CXR) and chest CT-scan [8,9,10,11,12,13,14,15,16]. Artificial Intelligence (AI) can be used in medical image processing where we can leverage AI for automatic detection and further guidance. We propose to detect COVID-19 from CXR images using an AI approach called a Deep Learning Model (DLM). Deep Learning methods have revolutionized the field of AI, and due to their potential application in medical diagnosis and prognosis systems, millions of human lives could be saved [17]. Radiographic images such as CXR are inexpensive and less time consuming to produce as results compared to other clinical or laboratory modalities. COVID-19 positive is characterized primarily by patches of ground-glass opacity and consolidations [14]. Therefore, CXR dataset is useful to test algorithms for detecting COVID-19 and other pulmonary disorders. Successful machine learning approaches could allow for rapid evaluations of chest X-ray images and thus enable radiologists to filter potential candidates in a time-effective manner [15,16]. Studies using deep neural networks have shown the effectiveness of the method in the diagnosis of pneumonia [15]. Another study of deep artificial neural networks which used more than 470,300 chest X-rays from hospitalized patients in three hospitals was able to detect abnormal cases with a positive predictive value of 94% [16]. In the later study, the average reporting delay was reduced by 76% from 11.2 to 2.7 days for critical X-ray findings (p < 0.001) in the model compared with historical data [16].
According to reports from the Centers for Disease Control and Prevention (CDC) [18], the Delta variant of the coronavirus (B.1.617.2) is spreading more rapidly and causing more infections compared with the original SARS-CoV-2 strain. Currently, the use vaccines remains the most effective method to reduce the risk of severe illnesses, transmission, hospitalization, and deaths due to COVID-19. However, CDC cautioned that effectiveness of the vaccines against new variants that may arise, including the new coronavirus variant Omicron (B.1.1.529) is not clear at this point in time [19].
Detection of COVID-19 using CXR has certain advantages over other methods. Currently, two types of diagnostic tests—molecular (RT-PCR) tests for the detection of the virus’s genetic material, and antigen tests for the detection of a specific viral agent—are available at the point of care [20]. For many countries, these laboratory tests continue to be costly, time consuming, and are not readily available for a mass population use. CXR image analysis shows promise in making the process of diagnosis faster, cheaper and more readily accessible compared to the other procedures [8,9,14,21]. Because of the multiple mutations of the virus and the emergence of newer variants such as the Delta variant and the Omicron coronavirus variant (B.1.1.529), the search for an early and accurate method of detection of COVID-19 is crucial. If an automated AI model using chest X-rays is readily available to the health professionals, it will be a substantial improvement in the clinical management of the disease.
For these purposes, we aimed to evaluate the effectiveness of the Deep Learning Model (DLM) using CXR in screening COVID-19 cases. We compared our results with previously reported models in predicting COVID-19.

2. Materials and Methods

2.1. Data Acquisition

The research was conducted from September 2020 to December 2021. To design a machine learning model as depicted in Figure 1, it takes a few basic steps to follow, of which the most important and formidable task was to gather a sufficient number of reliable sample data. The data collection was done from open repositories on the Internet including Kaggle and GitHub where CXR data are contributed from different hospitals and other institutions around the world [22,23,24,25]. The data consisted of: (1) CXR images of COVID-19 infected patients, (2) CXR images of samples with pneumonia but not having COVID-19 infections, and (3) normal CXR images. Our model used medical image segmentation to get more accurate feature vector component which was then processed by deep neural network.

2.2. Preprocessing and Image Segmentation

After data acquisition, we used the following preprocessing steps with the CXR images. Firstly, we discarded CXR images that did not have labels from radiologists. Then after splitting into training, validation and test datasets, we applied multiple transformations including annotation, labeling, normalization and augmentation to each image in each dataset [26]. Preprocessing makes the data more robust and better suitable for Deep Learning Models.
In order to obtain segmented lung regions of the CXR images, we used pulmonary contour masks [27,28], and FC-DenseNet103 semantic segmentation algorithm [29] as depicted in Figure 2. Initially, the figures consisted of many unwanted regions outside the lung contour. As the convolution neural network model does the feature engineering based on the input images fed to the system, it is always important to use the most designed region. It could be done only by the use of the segmentation network. In our model, segmentation helped to get the lung contours which is our region of interest. It not only allows the Deep Learning Model to accurately learn the actual infected area but also makes the model more robust. Thus, our model was fed with the segmented lung contours only.

2.3. Processes Used in the Deep Learning Model (DLM)

Figure 1 is a representation of the steps involved in the current DLM classification model. For this process, we used different kinds of deep neural network architecture to improve the model which can filter out the error and generate the optimum results. As the first step, the original images are fed to the system, then after preprocessing, we fed those images for segmentation. After segmentation, we fed the segmented lung contour to the deep neural network model and finally got the classification result.

2.4. Description of Deep Learning Models

Because of segmentation, the locality of the interest area is more exposed to the deep neural network model. It helps to get the understanding of the severity of the pulmonary lesions by visualizing the affected areas and damages of lung tissues as appeared from CXRs. Thus, it helps to detect the disease severity and its prognosis.
Our DLM model is an extract of different well known deep learning pretrained models, such as ResNet18 [30], AlexNet [31], DenseNet [32], VGG16 [33] etc. These are well known Deep Learning Models or the benchmark deep learning neural network which provide standard trained values. These pretrained models get trained values using the most reliable ImageNet dataset, which contains a total of 14,197,122 images. The ImageNet Large Scale Visual Recognition Challenge (ILSVRC) is a benchmark which helps in developing new models.
In the model, it is a deep learning code which was taking the optimum vector components to generate the model. ResNet18 is a pre-trained model which has been used in our proposed DLM to enhance the efficiency. We used the optimum value, especially the learned weightage parameter values of ResNet18. The pneumonia study of Rajpurkar et al. [34], who developed a deep learning, convolutional neural network model using CXR images prior to the COVID-19 pandemic inspired our research. The model [34] had a Receiver Operating Characteristic Curve (ROC) value of 0.76 in a multiclass classification model using 14 types of pneumonia. In our study, we also used the concept of multiclass classification model with 3 classes.
ResNet18, used in our model, introduces the concept of skip connection in the deep layers which helps to solve the problem of vanishing gradient. Each ResNet block is either 2-layers deep used in small networks (such as ResNet 18, 34) or 3-layers deep (such as ResNet 50, 101, 152), as shown in Figure 3.

2.5. Training, Validation, Testing, and Augmentation of Images

The inputs used to train and test the system were those lung X-rays and the outputs were classified three groups: COVID-19, pneumonia (but not COVID positive), or normal cases (not COVID-19 or pneumonia) (Figure 4). The original CXR images were resized to 224 × 224 to be compatible with the pretrained model used in this research. All the images in the dataset used for training, validation, and testing were leveled as 0 (COVID-19 positive), 1 (Pneumonia, but not COVID-19 positive) and 2 (Normal). We had used data augmentation to increase each data sample. It is a machine learning teaching which uses vertical and horizontal flip and increase the data sample with the deep learning framework.
The images were horizontally flipped, magnified, and rotated for augmentation. After all these processes, we had an increased amount of data image samples in training, validation, and testing, while maintaining the balancing property of the sample.
The DLM model used in our work uses ResNet18 Deep Learning Model’s pretrained weightages. It helps a lot for training processes, basically minimizing the forward passes in the deep learning network. Our Deep Learning Model as shown in Figure 4 consists of 46 layers. 14 convolution layers, 23 hidden layers, 4 Max-pooling layers, 2 average polling layers, 2 dropout layers, and one SoftMax layer. The main deep layers are divided into two groups: one group of layers perform operation related to ‘feature learning’ and another group of layers are used for classification. In feature learning, there are a lot of concepts which are used depending upon the research. Our proposed DLM uses deep convolution layers, which consist of a set of convolution operation and pooling operation. In the classification part of the model, there are flatten layers, fully connected layers, and the activation function to get the desired output.

2.6. Definitions and Statistical Formulae

The term “sensitivity” (Table 1) tells us out of the total number of people who have the disease (e.g., COVID-19), the number that are correctly classified as having the disease by the model. Specificity of the model is its ability to determine the healthy cases correctly. Positive predictive value (or precision) of the model means out of the total who are detected as disease positive (such as COVID-19 positive) by the model, how many of them are in fact disease positive. Accuracy of the model is the total persons correctly identified (true diseases positive plus true disease negative) out of the total people tested. F1 score is a harmonic mean of two factors i.e., precision and recall (or sensitivity). The performance of the model was measured by several factors including sensitivity, specificity, accuracy and F1 score. False positive fraction and true positive fraction were used to demonstrate the ROC curve.
Table 1. Calculation of the Screening Test Statistics.
Table 1. Calculation of the Screening Test Statistics.
Based on the Gold Standard
Disease PresentDisease AbsentTotal
Predicted Model PositiveTrue positive (TP)False positive (FP)TP + FP
Predicted Model NegativeFalse negative (FN)True negative (TN)FN + TN
TotalTP + FNFP + TNTP + FP + FN + TN
Sensitivity   or   Recall = TP TP + FN
Specificity = TN FP + TN
Positive   Predictive   Value   or   Precision = TP TP + FP
Accuracy = TP + TN TP + FP + FN + TN
The   weighted   average   of   precision   and   recall ,   known   as   F 1   Score = 2 Recall Precision Recall + Precision
False   Positive   Fraction = FP FP + TN
True   Positive   Fraction = TP TP + FN

2.7. ROC Curve

A Receiver Operating characteristic Curve or ROC curve is a graphical representation of the performance of a classification model. This curve used True Positive Rate (TPR) (sensitivity) in the y-axis, and False Positive Rate (FPR) (1—specificity) in the x-axis. The value of the area under the curve (AUC) ranges from 0 to 1, of which a value of 0 means the predictions are 100% wrong, and a value of 1 means predictions are 100% correct. Any value closer to 1.0 is a very accurate model.

2.8. Confusion Matrix

A Confusion matrix is a table showing the number of cases in each classification that were predicted correctly by the model out of the total cases in the target class. In our study, the represents a 3 × 3 matrix because we used three classifications in the model: COVID-19, pneumonia, and healthy cases.

3. Results

Our proposed model used a large dataset of 10,040 samples, of which 2143 had COVID-19, 3674 had pneumonia, and 4223 were normal (not COVID-19 or pneumonia). Due to augmentation, there are horizontal, vertical, and rotational flip; in these processes the image sample was increased by 65% as shown in Table 2.
Of the dataset, 13,251 (80%) were used for training the model and the remaining 3313 (20%) were used for testing the model. From the training dataset, 1325 images (10%) were used for validation.

3.1. Validity Data of the DLM Model

The proposed Deep Learning (DLM) model had a detection accuracy of 96.43% and a sensitivity of 93.68%. We have trained all models using batches of size 50 for 200 epochs using the Stochastic Gradient Descent (SGD) method to reduce the loss function.
Figure 5 shows the confusion matrix (left) and AUC (right). The AUC was 99% for COVID-19, 97% for pneumonia, and 98% for normal cases.
From the Figure 5a, the confusion matrix shows that out of 3535 COVID-19 cases 3520 were correctly detected and 12 images detected as a normal cases and 3 as pneumonia cases. Out of 6967 normal cases, 6893 were detected correctly, 21 detected as a COVID-19 and 53 detected as pneumonia. Also, out of 6062 pneumonia cases, 6017 were detected correctly but 24 was detected as COVID-19 and 31 as normal cases. From the Figure 5b, the AUC clearly shows the high prediction rates for the DLM model.

3.2. Comparison of Performance

A comparison of performance, in terms of accuracy, sensitivity, specificity, and F1-score of our proposed DLM model and other models is presented in Table 3. We analyzed some of the previous research data and showed a comparative analysis with the proposed model in the table. The research data reported here were published since mid-2020. The number of CXR images used was not available in one study [35]. Two studies [36,37], which consisted of large CXR samples, used a deep convolutional neural network model, whereas another study with over 2900 samples [38] used a hybrid model. Compared to the other models, the proposed DLM model showed some improvement in all parameters of performance matrices, including accuracy, sensitivity, specificity and F1-score (weighted average of precision and recall).

4. Discussion

The proposed DLM model in this study was efficient in the screening of cases with COVID-19, pneumonia, and normal (not COVID-19 or pneumonia). The area under the ROC curves in cases who had COVID-19 (0.99), pneumonia (0.97), or were normal (0.98) showed high accuracy of the results as predicted by the DLM model.
As indicated by the performance indicators, such as accuracy, sensitivity, specificity and F1-score, the proposed DLM outperformed all other models. The ROC curve also showed the high prediction rates for COVID-19, pneumonia, and health states in our DLM model. A similar high accuracy of 99% was reported by a recent study [43], which was achieved by enhanced VGG16 in the detection of CXR images from COVID-19 and pneumonia patients. Several reports have emphasized the importance of using CXR images in the COVID-19 diagnosis using AI [44,45,46,47].
The Convolutional Neural Network (CNN) is the one of the most prominent domain of AI and machine learning (ML). It is widely used in different types of classification problem as well as in image classification problems [48,49,50]. Rahaman and colleagues (2020) used 15 different CNN models in a transfer learning process [51]. Another study using the DenseNet201 model [52] yielded an overall accuracy of 92% with a sensitivity of 94% for COVID-19 screening using chest X-rays.
Due to the advancement of different deep neural network, such as ResNet18, ResNet50 [30], AlexNet [31], DenseNet [32], and VGG16 [33], we extracted techniques from these models, and were able to make our model more error-free, sophisticated, and accurate. Similar to our study, another recent study [53] applied the method of preprocessing for lung segmentation, which removed the unwanted surroundings of the lung X-rays and kept the desired lung tissues only. By this process, the authors achieved a high detection accuracy rate of 97% for COVID-19. We used the similar process of preprocessing and then trained the classification model for the learning scheme. We thoroughly evaluated the problems associated with data imbalance in some of the previously reported studies [54,55,56,57,58]. Wang et al. [36] developed a machine learning model, called COVID-Net using 13,975 CXR images. They had a total of 8066 normal cases and 5538 patients with non-COVID-19 pneumonia, whereas, they had only 266 COVID-19 patient cases. This makes their dataset having the issue of data imbalance. Minaee et al. [40] developed their model using 5000 CXR images. They used a transfer learning technique to use the pretrained weightages in their model. The main disadvantages of this kind of transfer learning weightages is that the original deep learning network did not actually use medical dataset, more specially COVID-19 dataset; so the weightages may lead to bias prediction to some extent.
Prior to COVID-19, Rajpurkar et al. [34] described a model named as CheXNet, where they used deep neural network to classify 14 kinds of pneumonia disease and a large number of samples (more than 100,000 X-rays) for the model. It was a very promising research work which could greatly influence the study of COVID-19 research using ML. However, several other studies suffered from issues of low samples. For example, Khan et al. [41] used CoroNet, which is a deep convolutional neural network model. The model consisted of 310 normal images, 284 images with COVID-19, 327 cases with viral pneumonia, and 330 cases with bacterial pneumonia. The problem of small sample size was identified in some other studies [42]. If we begin with very few COVID-19 samples and comparatively large number of CXR images for normal and pneumonia samples, the data are not balanced, and the weights of the other samples may lead to biased predictions. Our study included a large sample of 10,040 which increased to 16,574 after augmentation.

4.1. Limitations

Most of the model built are limited with the amount of COVID-19 images, especially at the beginning of the pandemic because of unavailability or limited access to publicly available records. For this reasons, we initially could retrieve only 2143 number of COVID-19 images which was further increased to 3535. A limitation of the study is the lack of validation, i.e., use of the program in a different setting or context.

4.2. Strengths

In this study, chest X-rays were preferred over chests CT scan images. The reason behind this was that X-ray machines are readily available in most of the hospitals. Also, the CT scan machines are costlier that X-ray machines. The number of CXR images in our study was quite large to accurately measure the model performances. The performance indicators such as accuracy, sensitivity, specificity, and F1 scores of the study model were high.

5. Conclusions

In this research work, an automated analysis of CXR was achieved using deep learning-based approaches. The model had a high sensitivity, specificity, and accuracy regarding the identification of cases with COVID-19, pneumonia, and normal subjects. The results of high performance of the model make it a useful tool to screen COVID-19 cases using CXR. Promising results of the DLM model indicate that this model can be helpful for the medical professionals globally. The research potential of this study is of significant interest as we are still living in the pandemic scenario and new variants of COVID-19 are emerging as substantial threats of infectiousness and/or virulence. Should we have automated detection techniques available and health professionals get the information in a timely manner, the diagnosis and the management of COVID-19 will be more efficient.

Author Contributions

Conceptualization: S.C. and B.M.; Data collection, Data preprocessing, Data augmentation, and Model development: S.C.; Original draft preparation: S.C. and A.K.M.; Review and editing, B.M. and A.K.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable as the study did not involve humans or animals.

Informed Consent Statement

Not applicable as the study did not involve humans.

Data Availability Statement

The model, data, and methods used in the research are presented in sufficient detail so that other researchers can replicate the work. The secondary data for this study are extracted from publicly available sources.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Huang, C.; Wang, Y.; Li, X.; Ren, L.; Zhao, J.; Hu, Y.; Zhang, L.; Fan, G.; Xu, J.; Gu, X.; et al. Clinical features of patients infected with 2019 novel coronavirus in Wuhan, China. Lancet 2020, 395, 497–506. [Google Scholar] [CrossRef] [Green Version]
  2. Fauci, A.S.; Lane, H.C.; Redfield, R.R. COVID-19—Navigating the uncharted. N. Engl. J. Med. 2020, 382, 1268–1269. [Google Scholar] [CrossRef] [PubMed]
  3. Wu, F.; Zhao, S.; Yu, B.; Chen, Y.M.; Wang, W.; Song, Z.G.; Hu, Y.; Tao, Z.-W.; Tian, J.-H.; Pei, Y.-Y.; et al. A new coronavirus associated with human respiratory disease in China. Nature 2020, 579, 265–269. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. WorldOMeter. Coronavirus. Available online: https://www.worldometers.info/coronavirus/ (accessed on 10 December 2021).
  5. Gavriatopoulou, M.; Korompoki, E.; Fotiou, D.; Ntanasis-Stathopoulos, I.; Psaltopoulou, T.; Kastritis, E.; Terpos, E.; Dimopoulos, M.A. Organ-specific manifestations of COVID-19 infection. Clin. Exp. Med. 2020, 20, 493–506. [Google Scholar] [CrossRef]
  6. Wang, W.; Xu, Y.; Gao, R.; Lu, R.; Han, K.; Wu, G.; Tan, W. Detection of SARS-Cov-2 in different types of clinical specimens. JAMA 2020, 323, 1843–1844. [Google Scholar] [CrossRef] [Green Version]
  7. Brihn, A.; Chang, J.; OYong, K.; Balter, S.; Terashita, D.; Rubin, Z.; Yeganeh, N. Diagnostic performance of an antigen test with RT-PCR for the detection of SARS-CoV-2 in a hospital setting—Los Angeles County, California, June-August 2020. MMWR Morb. Mortal. Wkly. Rep. 2021, 70, 702–706. [Google Scholar] [CrossRef]
  8. Ai, T.; Yang, Z.; Hou, H.; Zhan, C.; Chen, C.; Lv, W.; Tao, Q.; Sun, Z.; Xia, L. Correlation of chest CT and RT-PCR testing for coronavirus disease 2019 (COVID-19) in China: A report of 1014 cases. Radiology 2020, 296, E32–E40. [Google Scholar] [CrossRef] [Green Version]
  9. Fan, L.; Li, D.; Xue, H.; Zhang, L.; Liu, Z.; Zhang, B.; Zhang, L.; Yang, W.; Xie, B.; Duan, X.; et al. Progress and prospect on imaging diagnosis of COVID-19. Chin. J. Acad. Radiol. 2020, 3, 4–13. [Google Scholar] [CrossRef] [Green Version]
  10. Bai, H.X.; Hsieh, B.; Xiong, Z.; Halsey, K.; Choi, J.W.; Tran, T.M.L.; Pan, I.; Shi, L.-B.; Wang, D.-C.; Mei, J.; et al. Performance of radiologists in differentiating COVID-19 from viral pneumonia on chest CT. Radiology 2020, 296, 200823. [Google Scholar] [CrossRef]
  11. He, X.; Yang, X.; Zhang, S.; Zhao, J.; Zhang, Y.; Xing, E.; Xie, P. Sample-efficient deep learning for COVID-19 diagnosis based on CT scans. IEEE Trans. Med. Imaging. 2020. Available online: https://www.medrxiv.org/content/10.1101/2020.04.13.20063941v1.full.pdf (accessed on 10 December 2021).
  12. Gunraj, H.; Wang, L.; Wong, A. COVIDNet-CT: A tailored deep convolutional neural network design for detection of COVID-19 cases from chest CT images. Front. Med. 2020. Available online: https://www.frontiersin.org/articles/10.3389/fmed.2020.608525/full (accessed on 10 December 2021).
  13. Harmon, S.A.; Sanford, T.H.; Xu, S.; Turkbey, E.B.; Roth, H.; Xu, Z.; Yang, D.; Myronenko, A.; Anderson, V.; Amalou, A.; et al. Artificial intelligence for the detection of COVID-19 pneumonia on chest CT using multinational datasets. Nat. Commun. 2020, 11, 4080. [Google Scholar] [CrossRef]
  14. Das, D.; Santosh, K.C.; Pal, U. Truncated inception net: COVID-19 outbreak screening using chest X-rays. Phys. Eng. Sci. Med. 2020, 43, 915–925. [Google Scholar] [CrossRef] [PubMed]
  15. Jaiswal, A.K.; Tiwari, P.; Kumar, S.; Gupta, D.; Khanna, A.; Rodrigues, J.P.C. Identifying pneumonia in chest X-rays: A deep learning approach. Measurement 2019, 145, 511–518. [Google Scholar] [CrossRef]
  16. Annarumma, M.; Withey, S.J.; Bakewell, R.J.; Pesce, E.; Goh, V. Automated triaging of adult chest radiographs with deep artificial neural networks. Radiology 2019, 291, 272. [Google Scholar] [CrossRef] [Green Version]
  17. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  18. Centers for Disease Control and Prevention. Delta Variant: What We Know about the Science. 2021. Available online: https://www.cdc.gov/coronavirus/2019-ncov/variants/delta-variant.html?s_cid=11512:cdc%20delta%20variant:sem.ga:p:RG:GM:gen:PTN:FY21 (accessed on 9 December 2021).
  19. Centers for Disease Control and Prevention. Omicron Variant: What You Need to Know. 2021. Available online: https://www.cdc.gov/coronavirus/2019-ncov/variants/omicron-variant.html (accessed on 9 December 2021).
  20. US Food & Drug Administration. Coronavirus Testing Basics. Available online: https://www.fda.gov/media/138094/download (accessed on 9 December 2021).
  21. Shiraishi, J.; Katsuragawa, S.; Ikezoe, J.; Matsumoto, T.; Kobayashi, T.; Komatsu, K.-I.; Matsui, M.; Fujita, H.; Kodera, Y.; Doi, K. Development of a digital image database for chest radiographs with and without a lung nodule. Am. J. Roentgenol. 2020, 174, 71–74. [Google Scholar] [CrossRef] [PubMed]
  22. Praveen. Corona Hack-Chest X-ray-Dataset. Available online: https://www.kaggle.com/praveengovi/coronahack-chest-xraydataset (accessed on 8 December 2021).
  23. Cohen, J.P.; Morrison, P.; Dao, L. COVID-19 Image Data Collection. arXiv 2020, arXiv:2003.11597. Available online: https://arxiv.org/abs/2003.11597 (accessed on 8 December 2021).
  24. Paul, C.J. COVID-19 Image Data Collection. 2020. Available online: https://github.com/ieee8023/covid-chestxray-dataset (accessed on 7 December 2021).
  25. Paul, M. Kaggle Chest X-ray Images (Pneumonia) Dataset. 2020. Available online: https://www.kaggle.com/paultimothymooney/chest-xray-pneumonia (accessed on 8 December 2021).
  26. Mittal, A.; Kumar, D.; Mittal, M.; Saba, T.; Abunadi, I.; Rehman, A.; Roy, S. Detecting pneumonia using convolutions and dynamic capsule routing for chest X-ray images. Sensors 2020, 20, 1068. [Google Scholar] [CrossRef] [Green Version]
  27. Ginneken, B.V.; Stegmann, M.B.; Loog, M. Segmentation of anatomical structures in chest radiographs using supervised methods: A comparative study on a public database. Med. Image Anal. 2006, 10, 19–40. [Google Scholar] [CrossRef] [Green Version]
  28. Jaeger, S.; Candemir, S.; Antani, S.; Wang, Y.X.J.; Lu, P.X.; Thoma, G. Two public chest X-ray datasets for computer aided screening of pulmonary diseases. Quant. Imag. Med. Surg. 2014, 4, 475. [Google Scholar]
  29. Jegou, S.; Drozdzal, M.; Vazquez, D.; Romero, A.; Bengio, Y. The one hundred layers tiramisu: Fully convolutional DenseNets for semantic segmentation. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 1175–1183. [Google Scholar] [CrossRef] [Green Version]
  30. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  31. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  32. Huang, G.; Liu, Z.; Maaten, L.V.D.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar] [CrossRef] [Green Version]
  33. Zhang, X.; Zou, J.; He, K.; Sun, J. Accelerating very deep convolutional networks for classification and detection. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 1943–1955. [Google Scholar] [CrossRef] [Green Version]
  34. Rajpurkar, P.; Irvin, J.; Zhu, K.; Yang, B.; Mehta, H.; Duan, T.; Ding, D.; Bagul, A.; Langlotz, C.; Shpanskaya, K.; et al. CheXNet: Radiologist-level pneumonia detection on chest X-rays with deep learning. arXiv 2017, arXiv:1711.05225. Available online: http://arxiv.org/abs/1711.05225 (accessed on 8 December 2021).
  35. Das, N.N.; Kumar, N.; Kaur, M.; Kumar, V.; Singh, D. Automated deep transfer learning-based approach for detection of COVID-19 infection in chest X-rays. Ing. Rech. Biomed. 2020; in press. [Google Scholar] [CrossRef]
  36. Wang, L.; Lin, Z.Q.; Wong, A. COVID-Net: A tailored deep convolutional neural network design for detection of COVID-19 cases from chest X-ray images. Sci. Rep. 2020, 10, 19549. [Google Scholar] [CrossRef]
  37. Narin, A.; Kaya, C.; Pamuk, Z. Automatic detection of coronavirus disease (COVID-19) using X-ray images and deep convolutional neural networks. Pattern Anal. Appl. 2021, 24, 1207–1220. [Google Scholar] [CrossRef] [PubMed]
  38. Altan, A.; Karasu, S.S. Recognition of COVID-19 disease from X-ray images by hybrid model consisting of 2D curvelet transform, chaotic salp swarm algorithm and deep learning technique. Chaos Solitons Fractals 2020, 140, 110071. [Google Scholar] [CrossRef] [PubMed]
  39. Ozturk, T.; Talo, M.; Yildirim, E.A.; Baloglu, U.B.; Yildirim, O.; Acharya, U.R. Automated detection of COVID-19 cases using deep neural networks with X-ray images. Comput. Biol. Med. 2020, 121, 103792. [Google Scholar] [CrossRef]
  40. Minaee, S.; Kafish, R.; Sonka, M.; Yazdani, S.; Soufi, G.J. Deep-COVID: Predicting COVID-19 from chest X-ray images using deep transfer learning. Med. Image Anal. 2020, 65, 101794. [Google Scholar] [CrossRef]
  41. Khan, A.I.; Shah, J.L.; Bhat, M.M. CoroNet: A deep neural network for detection and diagnosis of COVID-19 from chest X-ray images. Comput. Meth. Prog. Biomed. 2020, 196, 105581. [Google Scholar] [CrossRef]
  42. Civit-Masot, J.; Luna-Perejón, F.; Morales, M.D.; Civit, A. Deep learning system for COVID-19 diagnosis aid using X-ray pulmonary images. Appl. Sci. 2020, 10, 4640. [Google Scholar] [CrossRef]
  43. Yang, D.; Martinez, C.; Visuña, L.; Khandhar, H.; Bhatt, C.; Carretero, J. Detection and analysis of COVID-19 in medical images using deep learning techniques. Sci. Rep. 2021, 11, 19638. [Google Scholar] [CrossRef] [PubMed]
  44. Islam, M.Z.; Islam, M.M.; Asraf, A. A combined deep CNN-LSTM network for the detection of novel coronavirus (COVID-19) using X-ray images. Informat. Med. Unlocked. 2020, 20, 100412. [Google Scholar] [CrossRef] [PubMed]
  45. Pereira, R.M.; Bertolini, D.; Teixeira, L.O.; Silla, C.N.; Costa, Y.M.G. COVID-19 identification in chest X-ray images on at and hierarchical classification scenarios. Comput. Meth. Prog. Bio. 2020, 194, 105532. [Google Scholar] [CrossRef] [PubMed]
  46. Hemdan, E.E.D.; Shouman, M.A.; Karar, M.E. COVIDXNet: A framework of deep learning classifiers to diagnose COVID-19 in X-ray images. arXiv 2020, arXiv:2003.11055. Available online: http://arxiv.org/abs/2003.11055 (accessed on 8 December 2021).
  47. Panwar, H.; Gupta, P.K.; Siddiqui, M.K.; Morales-Menendez, R.; Bhardwaj, P.; Singh, V. A deep learning and grad-CAM based color visualization approach for fast detection of COVID-19 cases using chest X-ray and CT-Scan images. Chaos Solitons Fractals 2020, 140, 110190. [Google Scholar] [CrossRef]
  48. LeCun, Y.; Kavukcuoglu, K.; Farabet, C. Convolutional networks and applications in vision. In Proceedings of the ISCAS 2010—2010 IEEE International Symposium on Circuits and Systems: Nano-Bio Circuit Fabrics and Systems, Paris, France, 30 May–2 June 2010; IEEE: Piscataway, NY, USA, 2010; pp. 253–256. [Google Scholar] [CrossRef] [Green Version]
  49. Chakraborty, S. Category identification technique by a semantic feature generation algorithm. In Deep Learning for Internet of Things Infrastructure; CRC Press: Boca Raton, FL, USA, 2021; pp. 129–144. [Google Scholar]
  50. Zhang, K.; Liu, X.; Shen, J.; Li, Z.; Sang, Y.; Wu, X.; Zha, Y.; Liang, W.; Wang, C.; Wang, K.; et al. Clinically applicable AI system for accurate diagnosis, quantitative measurements, and prognosis of COVID-19 pneumonia using computed tomography. Cell 2020, 181, 1423–1433. [Google Scholar] [CrossRef]
  51. Rahaman, M.M.; Li, C.; Yao, Y.; Kulwa, F.; Rahman, M.A.; Wang, Q.; Qi, S.; Kong, F.; Zhu, X.; Zhao, X. Identification of COVID-19 samples from chest X-ray images using deep learning: A comparison of transfer learning approaches. J. X-ray Sci. Technol. 2020, 28, 821–839. [Google Scholar] [CrossRef]
  52. Manokaran, J.; Zabihollahy, F.; Hamilton-Wright, A.; Ukwatta, E. Detection of COVID-19 from chest X-ray images using transfer learning. J. Med. Imaging 2021, 8, 017503. [Google Scholar] [CrossRef]
  53. Arias-Garzón, D.; Alzate-Grisales, J.A.; Orozco-Arias, S.; Arteaga-Arteaga, H.B.; Bravo-Ortiz, M.A.; Mora-Rubio, A.; Saborit-Torres, J.M.; Serrano, J.M.; Tabares-Soto, R.; Vayá, M.D.L.I. COVID-19 detection in X-ray images using convolutional neural networks. Mach. Learn. Appl. 2021, 6, 100138. [Google Scholar] [CrossRef]
  54. Ebenuwa, S.H.; Sharif, M.S.; Alazab, M.; Nemrat, A.A. Variance ranking attributes selection techniques for binary classification problem in imbalance data. IEEE Access 2019, 7, 24649–24666. [Google Scholar] [CrossRef]
  55. Khushi, M.; Shaukat, K.; Alam, T.A.; Hameed, I.A.; Uddin, S.; Luo, S.; Yang, X.; Reyes, M.C. A comparative performance analysis of data resampling methods on imbalance medical data. IEEE Access 2021, 9, 109960–109975. [Google Scholar] [CrossRef]
  56. Jegierski, H.; Saganowski, S. An “Outside the Box” solution for imbalanced data classification. IEEE Access 2020, 8, 125191–125209. [Google Scholar] [CrossRef]
  57. Alam, T.; Ahmed, C.F.; Zahin, S.A.; Khan, M.A.H.; Islam, M.T. An effective recursive technique for multi-class classification and regression for imbalanced data. IEEE Access 2019, 7, 127615–127630. [Google Scholar] [CrossRef]
  58. Maillo, J.; Triguero, I.; Herrera, F. Redundancy and complexity metrics for big data classification: Towards smart data. IEEE Access 2020, 8, 87918–87928. [Google Scholar] [CrossRef]
Figure 1. A Block Diagram Representation of the Deep Learning Model (DLM).
Figure 1. A Block Diagram Representation of the Deep Learning Model (DLM).
Ijerph 19 02013 g001
Figure 2. Detail Layer Architecture Representation of ResNet18 Deep Learning Neural Network Model [30].
Figure 2. Detail Layer Architecture Representation of ResNet18 Deep Learning Neural Network Model [30].
Ijerph 19 02013 g002
Figure 3. ResNet 2 and ResNet 3-layer block.
Figure 3. ResNet 2 and ResNet 3-layer block.
Ijerph 19 02013 g003
Figure 4. Proposed Deep Learning Model (DLM) Architecture.
Figure 4. Proposed Deep Learning Model (DLM) Architecture.
Ijerph 19 02013 g004
Figure 5. (a) Confusion matrix of the sample data; (b) Area under the ROC curve for the model.
Figure 5. (a) Confusion matrix of the sample data; (b) Area under the ROC curve for the model.
Ijerph 19 02013 g005
Table 2. Dataset Sample details before and after data augmentation.
Table 2. Dataset Sample details before and after data augmentation.
Stage of DataCOVID-19PneumoniaNormalTotal
Before augmentation21433674422310,040
After augmentation35356072696716,574
Table 3. Comparison of accuracy, sensitivity, specificity, and F1 score of the proposed model and other models.
Table 3. Comparison of accuracy, sensitivity, specificity, and F1 score of the proposed model and other models.
Study [Ref]No. Images UsedAI Method UsedAccuracySensitivitySpecificityF1-Score
Proposed Model10,040Deep Learning Model96.4393.6899.093.0
Das et al. [35]UnknownDeep transfer learning92.4191.2992.089.0
Wang et al. [36]13,975Deep convolutional neural network92.0490.4194.087.0
Narin et al. [37]7486Deep convolutional neural network91.2689.2493.086.0
Altan et al. [38]2905A hybrid model, having 2D Curvelet transformation, a Salp swarm algorithm (SSA) and deep learning 91.8592.4291.090.0
Ozturk et al. [39]1000Deep neural network93.4092.1289.090.0
Minaee et al. [40]5,000Deep transfer learning90.4992.0891.088.0
Khan et al. [41]1300Deep neural network89.60 84.9887.086.0
Civit-Masot et al. [42]396VGG16-based (convolutional) Deep Learning Model94.5292.1196.092.0
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chakraborty, S.; Murali, B.; Mitra, A.K. An Efficient Deep Learning Model to Detect COVID-19 Using Chest X-ray Images. Int. J. Environ. Res. Public Health 2022, 19, 2013. https://0-doi-org.brum.beds.ac.uk/10.3390/ijerph19042013

AMA Style

Chakraborty S, Murali B, Mitra AK. An Efficient Deep Learning Model to Detect COVID-19 Using Chest X-ray Images. International Journal of Environmental Research and Public Health. 2022; 19(4):2013. https://0-doi-org.brum.beds.ac.uk/10.3390/ijerph19042013

Chicago/Turabian Style

Chakraborty, Somenath, Beddhu Murali, and Amal K. Mitra. 2022. "An Efficient Deep Learning Model to Detect COVID-19 Using Chest X-ray Images" International Journal of Environmental Research and Public Health 19, no. 4: 2013. https://0-doi-org.brum.beds.ac.uk/10.3390/ijerph19042013

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop