Next Article in Journal
Diagnosis of Peritoneal Carcinomatosis of Colorectal Origin Based on an Innovative Fuzzy Logic Approach
Previous Article in Journal
Isolated Sphenoid Sinusitis: Anatomical Features for Choosing a Method of Treatment, a Case-Control Study
 
 
diagnostics-logo
Article Menu

Article Menu

Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

COVLIAS 1.0Lesion vs. MedSeg: An Artificial Intelligence Framework for Automated Lesion Segmentation in COVID-19 Lung Computed Tomography Scans

by
Jasjit S. Suri
1,2,*,
Sushant Agarwal
2,3,
Gian Luca Chabert
4,
Alessandro Carriero
5,
Alessio Paschè
4,
Pietro S. C. Danna
4,
Luca Saba
4,
Armin Mehmedović
6,
Gavino Faa
7,
Inder M. Singh
1,
Monika Turk
8,
Paramjit S. Chadha
1,
Amer M. Johri
9,
Narendra N. Khanna
10,
Sophie Mavrogeni
11,
John R. Laird
12,
Gyan Pareek
13,
Martin Miner
14,
David W. Sobel
13,
Antonella Balestrieri
4,
Petros P. Sfikakis
15,
George Tsoulfas
16,
Athanasios D. Protogerou
17,
Durga Prasanna Misra
18,
Vikas Agarwal
18,
George D. Kitas
19,20,
Jagjit S. Teji
21,
Mustafa Al-Maini
22,
Surinder K. Dhanjil
23,
Andrew Nicolaides
24,
Aditya Sharma
25,
Vijay Rathore
23,
Mostafa Fatemi
26,
Azra Alizad
27,
Pudukode R. Krishnan
28,
Ferenc Nagy
29,
Zoltan Ruzsa
30,
Mostafa M. Fouda
31,
Subbaram Naidu
32,
Klaudija Viskovic
6 and
Manudeep K. Kalra
33
add Show full author list remove Hide full author list
1
Stroke Diagnostic and Monitoring Division, AtheroPoint™, Roseville, CA 95661, USA
2
Advanced Knowledge Engineering Centre, GBTI, Roseville, CA 95661, USA
3
Department of Computer Science Engineering, PSIT, Kanpur 209305, India
4
Department of Radiology, Azienda Ospedaliero Universitaria (A.O.U.), 09124 Cagliari, Italy
5
Department of Radiology, “Maggiore della Carità” Hospital, University of Piemonte Orientale (UPO), Via Solaroli 17, 28100 Novara, Italy
6
University Hospital for Infectious Diseases, 10000 Zagreb, Croatia
7
Department of Pathology, Azienda Ospedaliero Universitaria (A.O.U.), 09124 Cagliari, Italy
8
The Hanse-Wissenschaftskolleg Institute for Advanced Study, 27753 Delmenhorst, Germany
9
Department of Medicine, Division of Cardiology, Queen’s University, Kingston, ON K7L 3N6, Canada
10
Department of Cardiology, Indraprastha APOLLO Hospitals, New Delhi 110076, India
11
Cardiology Clinic, Onassis Cardiac Surgery Center, 17674 Athens, Greece
12
Heart and Vascular Institute, Adventist Health St. Helena, St Helena, CA 94574, USA
13
Minimally Invasive Urology Institute, Brown University, Providence, RI 02912, USA
14
Men’s Health Center, Miriam Hospital, Providence, RI 02906, USA
15
Rheumatology Unit, National Kapodistrian University of Athens, 15772 Athens, Greece
16
Department of Surgery, Aristoteleion University of Thessaloniki, 54124 Thessaloniki, Greece
17
Cardiovascular Prevention and Research Unit, Department of Pathophysiology, National & Kapodistrian University of Athens, 15772 Athens, Greece
18
Department of Immunology, Sanjay Gandhi Postgraduate Institute of Medical Sciences, Lucknow 226014, India
19
Academic Affairs, Dudley Group NHS Foundation Trust, Dudley DY1 2HQ, UK
20
Arthritis Research UK Epidemiology Unit, Manchester University, Manchester M13 9PL, UK
21
Ann and Robert H. Lurie Children’s Hospital of Chicago, Chicago, IL 60611, USA
22
Allergy, Clinical Immunology and Rheumatology Institute, Toronto, ON L4Z 4C4, Canada
23
AtheroPoint LLC, Roseville, CA 95661, USA
24
Vascular Screening and Diagnostic Centre, University of Nicosia Medical School, Nicosia 2408, Cyprus
25
Division of Cardiovascular Medicine, University of Virginia, Charlottesville, VA 22908, USA
26
Department of Physiology and Biomedical Engineering, Mayo Clinic College of Medicine and Science, Rochester, MN 55905, USA
27
Department of Radiology, Mayo Clinic College of Medicine and Science, Rochester, MN 55905, USA
28
Neurology Department, Fortis Hospital, Bangalore 560076, India
29
Internal Medicine Department, University of Szeged, 6725 Szeged, Hungary
30
Invasive Cardiology Division, University of Szeged, 6725 Szeged, Hungary
31
Department of Electrical and Computer Engineering, Idaho State University, Pocatello, ID 83209, USA
32
Electrical Engineering Department, University of Minnesota, Duluth, MN 55812, USA
33
Department of Radiology, Massachusetts General Hospital, 55 Fruit Street, Boston, MA 02114, USA
*
Author to whom correspondence should be addressed.
Submission received: 16 April 2022 / Revised: 18 May 2022 / Accepted: 19 May 2022 / Published: 21 May 2022
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
Background: COVID-19 is a disease with multiple variants, and is quickly spreading throughout the world. It is crucial to identify patients who are suspected of having COVID-19 early, because the vaccine is not readily available in certain parts of the world. Methodology: Lung computed tomography (CT) imaging can be used to diagnose COVID-19 as an alternative to the RT-PCR test in some cases. The occurrence of ground-glass opacities in the lung region is a characteristic of COVID-19 in chest CT scans, and these are daunting to locate and segment manually. The proposed study consists of a combination of solo deep learning (DL) and hybrid DL (HDL) models to tackle the lesion location and segmentation more quickly. One DL and four HDL models—namely, PSPNet, VGG-SegNet, ResNet-SegNet, VGG-UNet, and ResNet-UNet—were trained by an expert radiologist. The training scheme adopted a fivefold cross-validation strategy on a cohort of 3000 images selected from a set of 40 COVID-19-positive individuals. Results: The proposed variability study uses tracings from two trained radiologists as part of the validation. Five artificial intelligence (AI) models were benchmarked against MedSeg. The best AI model, ResNet-UNet, was superior to MedSeg by 9% and 15% for Dice and Jaccard, respectively, when compared against MD 1, and by 4% and 8%, respectively, when compared against MD 2. Statistical tests—namely, the Mann–Whitney test, paired t-test, and Wilcoxon test—demonstrated its stability and reliability, with p < 0.0001. The online system for each slice was <1 s. Conclusions: The AI models reliably located and segmented COVID-19 lesions in CT scans. The COVLIAS 1.0Lesion lesion locator passed the intervariability test.

1. Introduction

Severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) is an infectious disease that poses a concern to humans worldwide. The World Health Organization (WHO) proclaimed COVID-19 (the novel coronavirus disease) as a global pandemic on 11 March 2020. COVID-19 is a rapidly spreading illness worldwide, yet hospital resources are limited. As of 1 December 2021, COVID-19 had led to the infection of 260 million people and 5.2 million deaths worldwide [1]. COVID-19 has clearly shown to have several molecular pathways [2], leading to myocardial injury [3], diabetes [4], pulmonary embolism [5], and thrombosis [6]. Due to the lack of an effective vaccine or medication, early detection of COVID-19 is critical to saving many lives and safeguarding frontline workers. Most medical staff have become infected due to their frequent contact with patients, significantly aggravating the already dire healthcare situation.
The early detection of COVID-19 is critical to saving many lives and protecting frontline workers, due to the lack of an appropriate vaccination or therapy. RT-PCR, or “reverse transcription-polymerase chain reaction”, is one of the gold standards for the detection of COVID-19 [7,8]. Furthermore, since the RT-PCR test is slow—causing delays in report generation—and has low sensitivity [9], there is a need for better detection methods. However, imaging-based diagnosis, including ultrasound [10], chest X-ray [11], and chest computed tomography (CT) [12], is becoming more popular in detecting and managing infection with COVID-19 [13,14]. CT has demonstrated great sensitivity and repeatability in the diagnosis of COVID-19, and for body imaging in general [15]. It is a significant and trustworthy complement to RT-PCR testing in identifying the disease [16,17,18]. The main imaging advantage of CT [15,19,20] imaging is capturing anomalies such as ground-glass opacity (GGO) [21,22], consolidation, and other opacities seen in the CT of a COVID-19 patient [23]. The anomaly of GGO is a prevalent feature in most chest CT lung images [14,24,25,26]. Due to time constraints and the sheer volume of studies, most radiologists use a judgmental and semantic approach to evaluate the COVID-19 lesions with different opacities. Furthermore, the manual and semi-automated assessment is subjective, slow, and time-consuming [27,28,29,30]. As a result, rapid and error-free detection and real-time prognostic solutions are required for early COVID-19 illness to improve the speed of diagnosis.
Artificial intelligence (AI) has accelerated research and development in almost every field, including healthcare imaging [31,32,33]. The ability of AI techniques to replicate what is done manually has made detection and diagnosis of this disease faster [34,35,36,37,38,39,40,41,42,43,44,45,46]. The AI techniques try to accurately mimic the human brain using deep neural networks. This makes them suitable for solving medical imaging problems. Deep learning (DL) is an extension of AI that uses dense layers to deliver completely automatic feature extraction, classification, and segmentation [47,48,49,50,51,52,53].
DL has advantages, but it also has drawbacks and unknowns, such as optimization of the learning rate, determining the number of epochs, preventing overfitting, handling large datasets, and functioning in a multiresolution framework [54]. This is also known as hyperparameter tuning, which is the most crucial task when accurately training a DL model. Recently published studies by Suri et al. prove that using hybrid DL (HDL) models over solo DL [55,56] models in the medical domain can help to learn complex imaging features quickly and accurately [57,58,59]. Transfer learning can also be adapted for knowledge transfer from one model to another. This process helps train the DL models faster, and with fewer images [60,61]. The proposed study utilizes SDL and HDL models to segment COVID-19-based lesions in CT lung images. To prove the robustness of the AI systems, we postulate two conditions as the hypotheses: (a) the performance of the AI model benchmarked against two manual delineations must be within 10% of one another, and (b) the HDL model outperforms the SDL model in terms of performance. Figure 1 depicts the global COVLIAS 1.0Lesion system for COVID-19-based lesion segmentation using AI models, consisting of volume acquisition, online segmentation, and benchmarking against MedSeg, along with performance evaluation.
The main contributions of this study are as follows: (1) The proposed study consists of a combination of solo DL and HDL to tackle the lesion location for faster segmentation. One DL and four HDL models—namely, PSPNet, VGG-SegNet, ResNet-SegNet, VGG-UNet, and ResNet-UNet—were trained by an expert radiologist. (2) The training scheme adopted a fivefold cross-validation strategy on a cohort of 3000 images selected from a set of 40 COVID-19-positive individuals. Performance evaluation was carried out using systems such as (a) Dice similarity, (b) Jaccard index, (c) Bland–Altman plots, and (d) regression plots. (3) COVLIAS 1.0Lesion was benchmarked against the online MedSeg system, demonstrating COVLIAS 1.0Lesion to be superior to MedSeg when compared against Manual Delineation 1 and Manual Delineation 2. (4) The proposed interobserver variability study used tracings from two trained radiologists as part of the validation. (5) Statistical tests—namely, the Mann–Whitney test, paired t-test, and Wilcoxon test—demonstrated its stability and reliability, along with the p-values. (6) The online system for each slice was <1 s.
The layout of this lesion segmentation study is as follows: In Section 2, we present the patient demographics and types of AI architectures. The results of the experimental protocol using the AI architectures, along with the performance evaluation, are shown in Section 3. The in-depth discussion is elaborated in Section 4, where we present our findings, benchmarking tables, strengths, weaknesses, and extensions of our study. The study concludes in Section 5.

2. Methods

2.1. Demographics and Baseline Characteristics

Approximately 3000 CT images (collected from 40 patients from Croatia) were used to create the training cohort (Figure 2). The patients had a mean age of 66 (SD 7.988), with 35 males (71.4 %) and the remainder females. In the cohort, the average GGO and consolidation scores were 2 and 1.2, respectively. Out of the 40 patients who participated in this study, all had a cough, 85% had dyspnoea, 28% had hypertension, 14% were smokers, and none had a sore throat, diabetes, COPD, or cancer. None of them were admitted to the intensive care unit (ICU) or died due to COVID-19 infection.

2.2. Image Acquisition and Data Preparation

This proposed study used a Croatian cohort of 40 COVID-19-positive patients. The retrospective cohort study was conducted from 1 March to 31 December 2020, at the University Hospital for Infectious Diseases in Zagreb, Croatia. All patients over the age of 18 who agreed to participate in the study had a positive RT-PCR test for the SARS-CoV-2 virus, underwent thoracic MDCT during their hospital stay, and met at least one of the following criteria: hypoxia (oxygen saturation below 92%), tachypnea (respiratory rate above 22 per minute), tachycardia (pulse rate > 100), or hypotension (systolic blood pressure 100 mmHg) prior to starting the study. The UHID Ethics Committee gave their consent. The acquisition was carried out using a 64-detector scanner from FCT Speedia HD (from Fujifilm Corporation, Tokyo, Japan, 2017), while the acquisition protocol consisted of a single full inspiratory breath-hold for collection of CT scans of the thorax in the craniocaudal direction.
Researchers used Hitachi Ltd.’s (Tokyo, Japan) Whole-Body X-ray CT System with Supria Software, and a typical imaging method to view the images (System Software Version: V2.25, Copyright Hitachi, Ltd., 2017). When scanning, the following values were used: wide focus, 120 kV tube voltage, 350 mA tube current, and 0.75 s rotation speed in the IntelligentEC (automatic tube-current modulation) mode. We followed the standardized protocol for reconstruction as adopted in our previous studies where, for multi-recon options, the field of view was 350 mm, the slice thickness was 5 mm (0.625 × 64), and the table pitch was 1.3281. We selected a slice thickness of 1.25 mm and a recon index of 1 mm for picture filter 22 (lung standard) with the Intelli IP Lv.2 iterative algorithm (WW1600/WL600). Furthermore, for picture filter 31 (mediastinal), with the Lv.3 Intelli IP iterative algorithm (WW450/WL45), the slice thickness was 1.25 mm and the recon index was 1 mm.
Scanned areas were chosen based on the presence of no metallic objects and reasonable image quality without artefacts or blurriness caused by the movement of the patients during the conduction of the scan. Each patient’s CT volume in this cohort consisted of ~300 slices. The senior radiologist (K.V.) carefully selected ~70 CT slices (512 × 512 px2) that preserved most of the lung region (only accounting for about 20% of the total CT slices). Figure 3 and Figure 4 show the annotated lesions from tracers 1 and 2, respectively, in red, with the raw CT image as the background.

2.3. The Deep Learning Models

The proposed study consists of a combination of solo deep learning (DL) and hybrid DL (HDL) models to tackle the lesion location and lesion segmentation more quickly. It was recently shown that the combination of two DL models has more feature-extraction power compared to the solo DL models; this motivation brought the innovation of combining two solo DL models. This study therefore implemented four HDL models—namely, VGG-SegNet, ResNet-SegNet, VGG-UNet, and ResNet-UNet—that were trained by an expert radiologist. This was then also benchmarked against a solo DL model, namely, PSPNet.
By replacing the kernel filter in the initial layer with 11- and 5-sized filters, the VGGNet architecture was meant to reduce training time [62]. VGGNet was extremely efficient and speedy, but it had a problem in optimization due to vanishing gradients. During backpropagation, it resulted in training with substantially less or no weights, because it was multiplied by the gradient at each epoch, and the update to the initial layers was very modest. Residual Network, or ResNet [63], was created to address this issue. A new link called the “skip connection” was created in this architecture, allowing gradients to bypass a limited number of layers and thereby resolve the issue of the vanishing gradient problem. Furthermore, during the backpropagation step, another modification to the network—namely, an identity function—kept the local gradient value at a non-zero quantity.
The HDL models were designed by combining one DL (i.e., VGG or ResNet, in our study) with another DL (i.e., UNet or SegNet, in our study), thereby producing a superior network with the advantages of both parent networks. The VGG-SegNet, VGG-UNet, ResNet-SegNet, and ResNet-UNet architectures employed in this research are made up of three parts: an encoder, a decoder, and a pixel-wise softmax classifier. The details of the SDL and HDL models are discussed in the following sections.

2.3.1. PSPNet—Solo DL Model

The pyramid scene parsing network (PSPNet) [64] is a semantic segmentation network that takes into account the image’s overall context. PSPNet includes four sections to its design (Figure 5): (1) input, (2) feature map, (3) pyramid pooling module, and (4) output [65,66]. The segmented image is sent into the network, which then uses a set of dilated convolution and pooling blocks to extract the feature map. The network’s heart is the pyramid pooling module, which helps capture the global context of the image/feature map constructed in the previous stage. This section is divided into four sections, each with its own scaling capabilities. This module’s scaling options are 1, 2, 3, and 6, with 1 × 1 scaling assisting in the acquisition of spatial data, and thereby increasing the resolution of the acquired features. The higher-resolution features are captured by the 6 × 6 scaling. All of the outputs from these four components are pooled at the end of this module using global average pooling. The global average pooling output is sent to a collection of convolutional layers in the final section. Finally, the output binary mask generates the collection of prediction classes. The main advantage of PSPNet is the global feature extraction using the pyramid pooling strategy.

2.3.2. Two SegNet-Based HDL Model Designs—VGG-SegNet and ResNet-SegNet

The VGG-SegNet architecture used in this study (Figure 6) consists of three components: an encoder, a decoder, and a pixel-wise softmax classifier at the end. It consists of 16 convolution (conv) layers (green in color) compared to the 13 in the SegNet [67] design (VGG backbone). The difference between ResNet-SegNet (Figure 7) and VGG-SegNet (Figure 6) is in the encoder and decoder parts. The VGG is replaced by ResNet [63] architecture in the encoder part of the architecture. Skip connection in VGG-SegNet is shown by the horizontal lines running from encoder to decoder in Figure 7, which help in retaining the features. To overcome the vanishing gradient problem, a new link known as the “skip connection” (Figure 7) was invented in this architecture, allowing the gradients to bypass a set number of levels [68,69]. This consists of conv blocks and identity blocks (Figure 7). The conv block consists of three serial 1 × 1, 3 × 3, and 1 × 1 convolution blocks in parallel to a 1 × 1 convolution block, which is then added in the end. The identity block is similar to the conv block, except that it uses skip connection. Since VGG is faster and SegNet is a basic segmentation network, this segmentation process is relatively faster; thus, VGG-SegNet is more advantageous compared to SegNet alone. On the other hand, ResNet-SegNet is more accurate, since it has a greater number of layers, and prevents the vanishing gradient problem.

2.3.3. Two UNet-Based HDL Model Designs: VGG-UNet and ResNet-UNet

VGG-UNet (Figure 8) and ResNet-UNet (Figure 9) are based on the classic UNet structure, which consists of encoder (downsampling) and decoder (upsampling) components. The VGG-19 [62,70,71,72] and ResNet-51 [58,63,73,74] models replace the downsampling encoder in VGG-UNet and ResNet-UNet, respectively. These architectures are better than the traditional UNet [75], since each level’s traditional convolution blocks are changed by the VGG and ResNet blocks in VGG-UNet and ResNet-UNet, respectively. Note that skip connection in VGG-UNet is shown by the horizontal lines running from encoder to decoder in Figure 8, which help in retaining the features, similar to Figure 7 in VGG-SegNet. To overcome the vanishing gradient problem, a new link known as the “skip connection” (Figure 9) was invented in this architecture, allowing gradients to bypass a set number of levels [68,69]. This consists of conv blocks and identity blocks (Figure 9). This is very similar to ResNet-SegNet, as shown in Figure 7. The conv block consists of three serial 1 × 1, 3 × 3, and 1 × 1 convolution blocks in parallel to a 1 × 1 convolution block, which is then added in the end. The identity block is similar to the conv block, except that it uses skip connection. The key advantage of VGG-UNet over UNet is its higher speed of operation, while ResNet-UNet offers better accuracy and avoids the vanishing gradient problem due to new skip connections.

2.4. Loss Function for SDL and HDL Models

The new models adopted the cross-entropy (CE) loss functions during the model generation [76,77,78]. If α C E represents the CE loss function, pr i represents the classifier’s probability used in the AI model, x i represents the input gold standard label 1, and (1 − x i) represents the gold standard label 0, then the loss function can be expressed mathematically as shown in Equation (1):
α C E = [ ( x i × log   pr i ) + ( 1 x i ) × log ( 1 pr i ) ]
where × represents the product of the two terms.

2.5. Experimental Protocol

The AI models’ accuracy was determined using a standardized cross-validation (CV) technique. Using the AI framework, our group produced a number of CV-based protocols of various types. We adopted a fivefold cross-validation protocol consisting of 80% training (2400 scans), while the remaining 20% were training data (600 CT scans). The choice of the fivefold cross-validation was due to the mild COVID-19 conditions. Five folds were created in such a way that each fold had the opportunity to have a distinct test set. The K5 protocol included an internal validation mechanism in which 10% of the data were considered for validation.
The AI systems’ accuracy was determined by comparing anticipated output to ground-truth pixel values. Because the output lung mask was either black or white, these readings were interpreted as binary (0 or 1) integers. Finally, the sum of these binary integers was divided by the total number of pixels in the image. Using the standardized symbols for truth tables for the determination of accuracy, we used TP, TN, FN, and FP to denote true positive, true negative, false negative, and false positive, respectively. The AI systems’ accuracy can be mathematically expressed as shown in Equation (2):
Accuracy   ( % ) = ( TP + TN TP + FN + TN + FP ) × 100

3. Results and Performance Evaluation

3.1. Results

This proposed study is an improvement on the previously published COVLIAS 1.0Lung system with lesion segmentation. This study uses a cohort of 3000 images for a set of 40 COVID-19-positive patients, with five AI models utilizing a fivefold CV technique. The training was carried out on one set of manual delineation from a senior radiologist. Figure 10 shows the accuracy and the loss plot using the best AI model (ResNet-UNet) out of the five models used in this proposed study. Figure 11 shows the overlay of the AI-predicted lesions (green) in rows 3–7 against manual delineation (red, row 2), with raw CT images (row 1) as the background. Figure A1, Figure A2, Figure A3, Figure A4 and Figure A5 show the outputs from PSPNet, VGG-SegNet, ResNet-SegNet, VGG-UNet, and ResNet-UNet, respectively. Figure A6 shows the visual lesion overlays of MedSeg (green) vs. MD (red).

3.2. Performance Evaluation

This proposed study uses (1) the Dice similarity coefficient (DSC) [79,80], (2) Jaccard index (JI) [81], (3) Bland–Altman (BA) plots [82,83], and (4) receiver operating characteristics (ROC) [84,85,86] for the five AI models against MD 1 and MD 2 for performance evaluation. The same five metrics are used for MedSeg to validate the five AI models against it. Figure 12, Figure 13, Figure 14, Figure 15 and Figure 16 show the cumulative frequency distribution (CFD) plots for DSC and JI from PSPNet, VGG-SegNet, ResNet-SegNet, VGG-UNet, and ResNet-UNet, respectively, and depict the score at an 80% threshold. The CFD plots for DSC and JI are shown in Figure 17, which shows the output from the MedSeg model used for validating the COVLIAS 1.0Lesion system. This study also uses manual delineation from two trained radiologists (K.V. and G.L.) to validate the results of the five AI models and MedSeg. Figure 18 shows lesions detected by the best AI model (ResNet-UNet) and MedSeg, along with MD by two trained radiologists (K.V. and G.L.).
Table 1 presents the DSC and JI scores for five AI models using MD 1 and MD 2. The left-hand side of the table shows statistical computation using MD 1, while the right-hand side of the table shows the statistical computation using MD 2. The first five rows are the five AI models. The percentage difference is the difference between the AI model and the MedSeg model. As can be seen, the five AI models (ResNet-SegNet, PSPNet, VGG-SegNet, VGG-UNet, and ResNet-UNet) are all better than MedSeg, by 1%, 4%, 4%, 5%, and 9%, respectively. The mean Dice similarity for all five models is 0.8, which is better than that of MedSeg by 5%. The same is true for the Jaccard index where, as can be seen, the five AI models (ResNet-SegNet, PSPNet, VGG-SegNet, VGG-UNet, and ResNet-UNet) are all better than MedSeg, by 2%, 5%, 6%, 8%, and 15%, respectively. The mean JI is 0.66 which is better than that of MedSeg by 7%. Thus, in summary, both the Dice similarity and Jaccard index in all five AI models are better than those of the MedSeg model.
We also used another manual delineation system (G.L.), labelled as MD 2. The behavior was consistent with that of MD 2. The Dice similarity in the five AI models was superior to that of MedSeg by 4%, 0%, 4%, 1%, and 4%, respectively. Similarly, the JI was superior to that of MedSeg by 5%, 2%, 8%, 3%, and 8%, respectively. The mean Dice similarity using MD 2 was superior by 3%, while the mean Jaccard index was superior by 5%, thus proving our hypothesis. Figure 19, Figure 20, Figure 21, Figure 22 and Figure 23 show the correlation coefficient (CC) plots for the five AI models against MD 1 and MD 2. The plots also show the CC values of all of the plots with p < 0.0001. Finally, we also present the benchmarking against MedSeg in Figure 24, against MD 1 and MD 2. Table 2 presents the CC scores for the five AI models, along with the means of these AI models and MedSeg against MD 1 and MD 2, and the percentage difference between the results of the AI models and MedSeg.

3.3. Statistical Validation

To assess the system’s dependability and stability, standard tests—namely, paired t-tests [87,88], Mann–Whitney tests [89,90,91], and Wilcoxon tests [92]—were utilized. MedCalc software was used for the statistical analysis (Osteen, Belgium) [93,94]. To validate the system described in the study, we supplied 13 potential combinations for the five AI models and MedSeg against MD 1 and MD 2. Table 3 displays the Mann–Whitney test, paired t-test, and Wilcoxon test findings. Using the varying threshold strategy, one can compute COVLIAS’s diagnostic performance using receiver operating characteristics (ROC). The ROC curve and area under the curve (AUC) values for the five (two new and three old) AI models are depicted in Figure 25, with AUC values more than ~0.85 and ~0.75 for MD 1 and MD 2, respectively. The BA computation strategy [95,96] was used to demonstrate the consistency of two methods. We show the mean and standard deviation of the lesion area for the AI models (Figure 26, Figure 27, Figure 28, Figure 29 and Figure 30) and MedSeg (Figure 31), plotted against MD 1 and MD 2.

4. Discussion

This proposed study presents automated lesion detection in an AI framework using SDL and HDL models—namely, (1) PSPNet, (2) VGG-SegNet, (3) ResNet-SegNet, (4) VGG-UNet, and (5) ResNet-UNet—trained using a fivefold cross-validation strategy using a set of 3000 manually delineated images. As part of the benchmarking strategy, we compared the five AI models against MedSeg. As part of the variability study, we utilized the lesion annotations from another tracer to validate the results of the five AI models and MedSeg. We used four kinds of metric for evaluation of the five AI models, namely, (1) DSC, (2) JI, (3) BA plots, and (4) ROC. The best AI model, ResNet-UNet, was superior to MedSeg by 9% and 15% for Dice similarity and Jaccard index, respectively, when compared against MD 1, and by 4% and 8%, respectively, when compared against MD 2. Statistical tests—namely, the Mann–Whitney test, paired t-test, and Wilcoxon test—demonstrated its stability and reliability. The training, testing, and evaluation of the AI model were carried out using NVIDIA’s DGX V100. Multi-GPU training was used to speed up the process. The online system for each slice was <1 s. Table 2 shows the CC values of all of the AI models against MD 1 and MD 2; furthermore, it also presents a benchmark against MedSeg. The results show consistency, where ResNet-UNet is the best model amongst all of the AI models. It is ~14% and ~2% better than MedSeg for MD 1 and MD 2, respectively.
The primary attributes used for comparison of the five models are shown in Table 4, including (1) the backbone of the segmentation model, (2) the total number of parameters in the AI models (in millions), (3) the number of neural network layers, (4) the size of the final saved model used in COVLIAS 1.0, (5) the training time of the models, (6) the batch size used while training the network, and (7) the online prediction time per image for COVLIAS 1.0. ResNet-UNet was the AI model with the highest number of NN layers and the largest model size; due to this, it took the maximum amount of time to train the network.

4.1. Short Note on Lesion Annotation

Ground-truth annotation is always a challenge in AI [97,98]. In our scenario, in certain CT slices, the lesions overlapped, making it difficult to ensure precise lesion annotations. Some opacities are borderline, and the radiologist’s decision may be highly subjective, resulting in false positives or false negatives. When it is difficult to notice and differentiate opacities in patients with COVID-19, or with cardiac disorders, emphysema, fibrosis, or autoimmune diseases with pulmonary manifestation, the differences in experience are particularly significant for the annotation of complex investigations [99,100,101,102,103,104,105].

4.2. Explanation and Effectiveness of the AI-Based COVLIAS System

The proposed study uses five AI-based models—PSPNet, VGG-SegNet, ResNet-SegNet, VGG-UNet, and ResNet-UNet—for COVID-19-based lesion detection, and presents a comparison against an existing system in the same domain, known as MedSeg. This proposed study uses (1) DSC (Equation (3)), (2) JI (Equation (4)), (3) BA plots, and (4) ROC curves for the five AI models against MD 1 (or GS 1) and MD 2 (or GS 2) for performance evaluation, to prove the effectiveness of the AI-based COVLIAS system. The same five metrics were used for MedSeg against MD1 and MD2 to validate the five AI-based COVLIAS models against it.
D S C = 2 × | X Y | | X | + | Y |
J I = 2 × | X Y | | X Y |
where X is the set of pixels of the image 1, ground-truth, or manually delineated image, and Y is the set of pixels of the image 2 or AI-predicted image from COVLIAS 1.0Lesion.

4.3. Benchmarking

Several studies have been published that use deep learning algorithms based on chest CT imaging to identify and segment COVID-19 lesions [73,106,107,108]. However, most investigations lack lesion area measurement, transparency overlay generation, HDL utilization, and interobserver analysis. Our benchmarking analysis consists of 12 studies that use solo deep learning (DL) models and hybrid DL models for lesion detection [109,110,111,112,113,114,115,116,117,118,119,120]. Table 5 shows the benchmarking table, consisting of 21 attributes and 13 studies.
Ding et al. [109] presented MT-nCov-Net which is a multitasking DL network that includes segmentation of both lungs and lesions in CT scans, based on Res2Net50 [121] as its backbone. This study used five different CT image databases, totaling more than 36,000 images. Augmentation techniques such as random flipping, rotation, cropping, and Gaussian blurring were also applied. The Dice similarity was 0.86. Hou et al. [110] demonstrated the use of an improvised Canny edge detector [122,123] for CT images to detect COVID-19 lesions using a dataset of about 800 CT images. Lizzi et al. [112] adopted UNet by cascading it for COVID-19-based lesion segmentation on CT images. Various augmentation techniques—such as zooming, rotation, Gaussian noise, elastic deformation, and motion blur—were used in this study. The Dice similarity coefficient (DSC) was 0.62, which is lower compared to the 0.86 of Ding et al. [109]. ResNet-50 and XceptionNet [124] were used as the backbone of the DR-ML network demonstrated by Qi et al. [113]. This study used ~2400 CT images, with rotation, reflection, and translation as image augmentation techniques. DSC was not reported in this study, but it had an AUC of 0.94. Paluru et al. [114] presented a combination of UNet and ENet, named Anam-Net. It was designed for COVID-19-based lesion segmentation from lung CT images. The model was trained using a cohort of ~4300 images, and the input image to this model had to be a segmented lung. Anam-Net was benchmarked against ENet, UNet++, SegNet, LEDNet, etc. There was no augmentation reported, and the DSC was 0.77. The authors demonstrated an Android application and a deployment on an edge device for Anan-Net to perform COVID-19-based lesion segmentation. Zhang et al. [115] demonstrated CoSinGAN—the only generative adversarial network (GAN) of its kind for COVID-19-based lesion segmentation. Only ~700 CT lung images were used by this GAN in the training process, with no augmentation techniques. The DSC was 0.75 for CoSinGAN, and was benchmarked against other models. Singh et al. [111] modified the basic UNet architecture for lesion detection and heatmap generation. LungINFseg, a modified UNet architecture, was developed using a cohort of 1800 CT lung images with some augmentation techniques, and it reported a DSC of 0.8. The results of the modified UNet were benchmarked against some previously published segmentation networks, such as FCN [125], UNet, SegNet, Inf-Net [126], MIScnn [127,128], etc. The use of UNet with a multiresolution approach was demonstrated by Amyar et al. [117] for lesion detection and classification using 449 COVID-19-positive images. The authors reported an accuracy of 94% and DSC of 0.88, with no augmentation techniques. In only the classification framework, the model performance was benchmarked against some previously published studies. Budak et al. [116] used SegNet with attention gates to solve the problem of lesion segmentation for COVID-19 patients. Hounsfield unit windowing was also used as part of image pre-processing, with different loss functions to deal with small lesions. A cohort consisting of 69 patients was used in this study, where the author only reported a DSC of 0.89. A 10-fold CV protocol on 250 images with the UNet model was demonstrated by Cai et al. [118], with a DSC of 0.77. The authors presented lung and lesion segmentation using the same model. They also proposed a method to predict the duration of intensive care unit (ICU) stay based on the findings of the lesion segmentation. Ma et al. [119] also used the standard UNet architecture on a set of 70 patients for 3D CT volume segmentation. Model optimization was also carried out during the training process, and a DSC of 0.67 was reported in the study. The authors benchmarked the performance of the model with other studies in the same domain. Lastly, Kuchana et al. [120] used a cohort of 50 patients for lung and lesion segmentation with UNet and Attention UNet. During the training process, the authors optimized the hyperparameters, and a 0.84 DSC was reported by the model. Arunachalam et al. [129] recently presented a lesion segmentation system based on a two-stage process. Stage I consisted of region-of-interest estimation using region-based convolutional neural networks (RCNNs), while Stage II was used for bounding-box generation. The performance parameters for the training, validation, and test sets were 0.99, 0.931, and 0.8, respectively. The RCNN was primarily for COVID-19 lesion detection, coupled with automated bounding-box estimation for mask generation.

4.4. Strengths, Weaknesses, and Extension

This is the first pilot study for the localization and segmentation of COVID-19 lesions in CT scans of COVID-19 patients, under the class of COVLIAS 1.0. The main strengths were the design of five AI models that were benchmarked against MedSeg—the current industry standard. Furthermore, we demonstrated that COVLIAS 1.0Lesion is superior to MedSeg using manual lesion tracings MD 1 and MD 2, where MD 1 was used for training and MD 2 was used for evaluation of the AI models. The system was evaluated using several performance metrics.
Despite the encouraging results, the study could not include more than one observer (MD 1) for manual delineation, due to factors such as cost, time, and availability of a radiologist during the pandemic. During lesion segmentation, the image analysis component that changes the HU values could affect the training process; therefore, in-depth analysis was needed [130,131,132]. This is currently beyond the scope of our current objectives.
Several extensions can be attempted in the future. (1) Multiresolution techniques [133,134] embedded with advanced stochastic image-processing methods could be adapted to improve the speed of the system [135,136]. (2) A big data framework could be adopted, whereby multiple sources of information can be used in a deep learning framework [137]. (3) Our study tested interobserver variability by considering two different observers (MD1 and MD2). Our assumption for intraobserver analysis consisted of very subtle changes, as per our previous conducted studies [46,58,138,139,140]; we therefore did not consider it crucial to conduct intraobserver studies, due to lack of funding and the radiologists’ time constraints. Thus, intraobserver analysis could be conducted as part of future research. [58,96,138]. (4) Furthermore, there could be an additional step involved where, first, the lung is segmented, and then this segmented lung is used for analyzing the lesions [141,142]. This should help to increase the DSC and JI of the AI system. (5) The addition of lung segmentation does, however, increase the system’s time and computational cost. One could use the joint lesion segmentation and classification in a multiclass framework such as classification of GGO, consolidations, and crazy paving, using tissue-characterization approaches [56,143]. (6) One could also conduct multiethnic and multi-institutional studies for lung lesion segmentation, as attempted in other modalities [144]. (7) One could understand the lesion distribution in different COVID-19 symptom categories—i.e., high-COVID-19-symptom lesions vs. low-COVID-19-symptom lesions—as tried in other diseases [36]. (8) Since SDL and HDL strategies have been adapted for lesion segmentation, it is very likely that it can have a bias in AI [54] and, therefore, can be studied for lesion segmentation. (9) Several new ideas have emerged that need shape, position, and scale, and such techniques require spatial attention, channel attention, and scale-based solutions. Recently, advanced solutions have been tried for different applications, such as human activity recognition (HAR) [145]. Methods such as RNN or LSTM can also be incorporated in the skip connection of the UNet or hybrid UNet, which can be used for superior feature map selection [146]. Systems could also be designed where the high-risk lesions (high-valued GGO) and low-risk lesions (low-valued GGO) can be combined using ideas such as deep transfer networks [147]. Furthermore, increased loss function could be explored as part of training the AI models [148,149,150,151,152]. (10) As part of the extension to the system design, one could compare other kinds of cross-validation protocols, such as 2-fold, 3-fold, 4-fold, 10-fold, and jack-knife (JK) protocols such as training equals testing. Examples of such protocols can be seen in our previous studies [45,59,60,153,154,155]. Even though our design had a fivefold protocol, our experiences have shown slight variations in performance with the changes in cross-validation results.

5. Conclusions

The proposed study presents a comparison between COVLIAS 1.0Lesion and MedSeg for lesion segmentation in 3000 CT scans taken from 40 COVID-19 patients. COVLIAS 1.0Lesion (Global Biomedical Technologies, Inc., Roseville, CA, USA) consists of a combination of solo deep learning (DL) and hybrid DL (HDL) models to tackle the lesion location and segmentation more quickly. One DL and four HDL models—namely, PSPNet, VGG-SegNet, ResNet-SegNet, VGG-UNet, and ResNet-UNet—were trained by an expert radiologist. The training scheme adopted a fivefold cross-validation strategy for performance evaluation. As part of the validation, it used tracings from two trained radiologists. The best AI model, ResNet-UNet, was superior to MedSeg by 9% and 15% for Dice similarity and Jaccard index, respectively, when compared against MD 1, and by 4% and 8%, respectively, when compared against MD 2. Other error metrics, such as correlation coefficient plots for lesion area errors and Bland–Altman plots, showed a close correlation with the manual delineations. Statistical tests such as the paired t-test, Mann–Whitney test, and Wilcoxon test were used to demonstrate the stability and reliability of the AI system. The online system for each slice was <1 s. To conclude, our pilot study demonstrated the AI model’s reliability in locating and segmenting COVID-19 lesions in CT scans; however, multicenter data need to be collected and experimented with.

Author Contributions

Conceptualization, J.S.S., S.A., N.N.K. and M.K.K.; Data curation, G.L.C., A.C., A.P., P.S.C.D., L.S., A.M., G.F., M.T., P.R.K., F.N., Z.R. and K.V.; Formal analysis, J.S.S.; Funding acquisition, M.M.F.; Investigation, J.S.S., I.M.S., P.S.C., A.M.J., N.N.K., S.M., J.R.L., G.P., D.W.S., P.P.S., G.T., A.D.P., D.P.M., V.A., J.S.T., M.A.-M., S.K.D., A.N., A.S., A.A., F.N., Z.R., M.M.F. and K.V.; Methodology, J.S.S., S.A. and A.B.; Project administration, J.S.S. and M.K.K.; Software, S.A. and L.S.; Supervision, S.N.; Validation, J.S.S., G.L.C., L.S., A.M., M.M.F., K.V. and M.K.K.; Visualization, S.A. and V.R.; Writing—original draft, S.A.; Writing—review & editing, J.S.S., G.L.C., A.C., A.P., P.S.C.D., L.S., A.M., I.M.S., M.T., P.S.C., A.M.J., N.N.K., S.M., J.R.L., G.P., M.M., D.W.S., A.B., P.P.S., G.T., A.D.P., D.P.M., V.A., G.D.K., J.S.T., M.A.-M., S.K.D., A.N., A.S., V.R., M.F., A.A., M.M.F., S.N., K.V. and M.K.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Compliance with Ethical Standards: For Italian Dataset: IRB for the retrospective analysis of CT lung in patients affected by COVID19 granted by the Hospital of Novara to Professor Alessandro Carriero, Co-author of the research you are designing in the artificial intelligence application in the detection and risk stratification of COVID patients. Ethic Committee Name: Assessment of diagnostic performance of Computed Tomography in patients affected by SARS COVID 19 Infection. Approval Code: 131/20. Approval: authorized by the Azienda Ospedaliero Universitaria Maggiore della Carità di Novara on 25 June 2020. For Croatian Dataset: Ethic Committee Name: The use of artificial intelligence for multislice computer tomography (MSCT) images in patients with adult respiratory diseases syndrome and COVID-19 pneumonia. Approval Code: 01-2239-1-2020. Approval: authorized by the University Hospital for Infectious Diseases “Fran Mihaljevic”, Zegreb, Mirogojska 8. On 9 November 2020. Approved to Klaudija Viskovic.

Informed Consent Statement

Written informed consent, including consent for publication, was obtained from the patient.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Figure A1. Results of visual lesion overlays showing PSPNet (green) vs. MD 1 (red).
Figure A1. Results of visual lesion overlays showing PSPNet (green) vs. MD 1 (red).
Diagnostics 12 01283 g0a1
Figure A2. Results of visual lesion overlays showing VGG-SegNet (green) vs. MD 1 (red).
Figure A2. Results of visual lesion overlays showing VGG-SegNet (green) vs. MD 1 (red).
Diagnostics 12 01283 g0a2
Figure A3. Results of visual lesion overlays showing ResNet-SegNet (green) vs. MD 1 (red).
Figure A3. Results of visual lesion overlays showing ResNet-SegNet (green) vs. MD 1 (red).
Diagnostics 12 01283 g0a3
Figure A4. Results of visual lesion overlays showing VGG-UNet (green) vs. MD 1 (red).
Figure A4. Results of visual lesion overlays showing VGG-UNet (green) vs. MD 1 (red).
Diagnostics 12 01283 g0a4
Figure A5. Results of visual lesion overlays showing ResNet-UNet (green) vs. MD 1 (red).
Figure A5. Results of visual lesion overlays showing ResNet-UNet (green) vs. MD 1 (red).
Diagnostics 12 01283 g0a5
Figure A6. Results of visual lesion overlays showing MedSeg (green) vs. MD 1 (red).
Figure A6. Results of visual lesion overlays showing MedSeg (green) vs. MD 1 (red).
Diagnostics 12 01283 g0a6

References

  1. WHO Coronavirus (COVID-19) Dashboard. Available online: https://covid19.who.int/ (accessed on 24 January 2022).
  2. Saba, L.; Gerosa, C.; Fanni, D.; Marongiu, F.; La Nasa, G.; Caocci, G.; Barcellona, D.; Balestrieri, A.; Coghe, F.; Orru, G.; et al. Molecular pathways triggered by COVID-19 in different organs: ACE2 receptor-expressing cells under attack? A review. Eur. Rev. Med. Pharmacol. Sci. 2020, 24, 12609–12622. [Google Scholar] [CrossRef] [PubMed]
  3. Cau, R.; Bassareo, P.P.; Mannelli, L.; Suri, J.S.; Saba, L. Imaging in COVID-19-related myocardial injury. Int. J. Cardiovasc. Imaging 2021, 37, 1349–1360. [Google Scholar] [CrossRef] [PubMed]
  4. Viswanathan, V.; Puvvula, A.; Jamthikar, A.D.; Saba, L.; Johri, A.M.; Kotsis, V.; Khanna, N.N.; Dhanjil, S.K.; Majhail, M.; Misra, D.P. Bidirectional link between diabetes mellitus and coronavirus disease 2019 leading to cardiovascular disease: A narrative review. World J. Diabetes 2021, 12, 215. [Google Scholar] [CrossRef] [PubMed]
  5. Cau, R.; Pacielli, A.; Fatemeh, H.; Vaudano, P.; Arru, C.; Crivelli, P.; Stranieri, G.; Suri, J.S.; Mannelli, L.; Conti, M.; et al. Complications in COVID-19 patients: Characteristics of pulmonary embolism. Clin. Imaging 2021, 77, 244–249. [Google Scholar] [CrossRef]
  6. Fanni, D.; Saba, L.; Demontis, R.; Gerosa, C.; Chighine, A.; Nioi, M.; Suri, J.S.; Ravarino, A.; Cau, F.; Barcellona, D.; et al. Vaccine-induced severe thrombotic thrombocytopenia following COVID-19 vaccination: A report of an autoptic case and review of the literature. Eur. Rev. Med. Pharmacol. Sci. 2021, 25, 5063–5069. [Google Scholar] [CrossRef]
  7. Gibson, U.E.; Heid, C.A.; Williams, P.M. A novel method for real time quantitative RT-PCR. Genome Res. 1996, 6, 995–1001. [Google Scholar] [CrossRef] [Green Version]
  8. Bustin, S.; Benes, V.; Nolan, T.; Pfaffl, M. Quantitative real-time RT-PCR–A perspective. J. Mol. Endocrinol. 2005, 34, 597–601. [Google Scholar] [CrossRef] [Green Version]
  9. Fang, Y.; Zhang, H.; Xie, J.; Lin, M.; Ying, L.; Pang, P.; Ji, W. Sensitivity of Chest CT for COVID-19: Comparison to RT-PCR. Radiology 2020, 296, E115–E117. [Google Scholar] [CrossRef]
  10. Leighton, T.G. What is ultrasound? Prog. Biophys. Mol. Biol. 2007, 93, 3–83. [Google Scholar] [CrossRef]
  11. Jacobi, A.; Chung, M.; Bernheim, A.; Eber, C. Portable chest X-ray in coronavirus disease-19 (COVID-19): A pictorial review. Clin. Imaging 2020, 64, 35–42. [Google Scholar] [CrossRef]
  12. Sluimer, I.; Schilham, A.; Prokop, M.; Van Ginneken, B. Computer analysis of computed tomography scans of the lung: A survey. IEEE Trans. Med. Imaging 2006, 25, 385–405. [Google Scholar] [CrossRef] [PubMed]
  13. Giannitto, C.; Sposta, F.M.; Repici, A.; Vatteroni, G.; Casiraghi, E.; Casari, E.; Ferraroli, G.M.; Fugazza, A.; Sandri, M.T.; Chiti, A.; et al. Chest CT in patients with a moderate or high pretest probability of COVID-19 and negative swab. Radiol. Med. 2020, 125, 1260–1270. [Google Scholar] [CrossRef] [PubMed]
  14. Cau, R.; Falaschi, Z.; Pasche, A.; Danna, P.; Arioli, R.; Arru, C.D.; Zagaria, D.; Tricca, S.; Suri, J.S.; Karla, M.K.; et al. Computed tomography findings of COVID-19 pneumonia in Intensive Care Unit-patients. J. Public Health Res. 2021, 10. [Google Scholar] [CrossRef]
  15. Saba, L.; Suri, J.S. Multi-Detector CT Imaging: Principles, Head, Neck, and Vascular Systems; CRC Press: Boca Raton, FL, USA, 2013; Volume 1. [Google Scholar]
  16. Dangis, A.; Gieraerts, C.; De Bruecker, Y.; Janssen, L.; Valgaeren, H.; Obbels, D.; Gillis, M.; Van Ranst, M.; Frans, J.; Demeyere, A.; et al. Accuracy and Reproducibility of Low-Dose Submillisievert Chest CT for the Diagnosis of COVID-19. Radiol. Cardiothorac. Imaging 2020, 2, e200196. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Majidi, H.; Niksolat, F. Chest CT in patients suspected of COVID-19 infection: A reliable alternative for RT-PCR. Am. J. Emerg. Med. 2020, 38, 2730–2732. [Google Scholar] [CrossRef] [PubMed]
  18. Chung, M.; Bernheim, A.; Mei, X.; Zhang, N.; Huang, M.; Zeng, X.; Cui, J.; Xu, W.; Yang, Y.; Fayad, Z.A.; et al. CT Imaging Features of 2019 Novel Coronavirus (2019-nCoV). Radiology 2020, 295, 202–207. [Google Scholar] [CrossRef] [Green Version]
  19. Wu, X.; Hui, H.; Niu, M.; Li, L.; Wang, L.; He, B.; Yang, X.; Li, L.; Li, H.; Tian, J. Deep learning-based multi-view fusion model for screening 2019 novel coronavirus pneumonia: A multicentre study. Eur. J. Radiol. 2020, 128, 109041. [Google Scholar] [CrossRef]
  20. Pathak, Y.; Shukla, P.K.; Tiwari, A.; Stalin, S.; Singh, S. Deep transfer learning based classification model for COVID-19 disease. Irbm 2020, 43, 87–92. [Google Scholar] [CrossRef]
  21. Salehi, S.; Abedi, A.; Balakrishnan, S.; Gholamrezanezhad, A. Coronavirus Disease 2019 (COVID-19): A Systematic Review of Imaging Findings in 919 Patients. AJR Am. J. Roentgenol. 2020, 215, 87–93. [Google Scholar] [CrossRef]
  22. Cozzi, D.; Cavigli, E.; Moroni, C.; Smorchkova, O.; Zantonelli, G.; Pradella, S.; Miele, V. Ground-glass opacity (GGO): A review of the differential diagnosis in the era of COVID-19. Jpn. J. Radiol. 2021, 39, 721–732. [Google Scholar] [CrossRef]
  23. Xie, X.; Zhong, Z.; Zhao, W.; Zheng, C.; Wang, F.; Liu, J. Chest CT for Typical Coronavirus Disease 2019 (COVID-19) Pneumonia: Relationship to Negative RT-PCR Testing. Radiology 2020, 296, E41–E45. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Gozes, O.; Frid-Adar, M.; Greenspan, H.; Browning, P.D.; Zhang, H.; Ji, W.; Bernheim, A.; Siegel, E. Rapid ai development cycle for the coronavirus (COVID-19) pandemic: Initial results for automated detection & patient monitoring using deep learning ct image analysis. arXiv 2020, arXiv:2003.05037. [Google Scholar]
  25. Shalbaf, A.; Vafaeezadeh, M. Automated detection of COVID-19 using ensemble of transfer learning with deep convolutional neural network based on CT scans. Int. J. Comput. Assist. Radiol. Surg. 2021, 16, 115–123. [Google Scholar]
  26. Yang, X.; He, X.; Zhao, J.; Zhang, Y.; Zhang, S.; Xie, P. COVID-CT-dataset: A CT scan dataset about COVID-19. arXiv 2020, arXiv:2003.13865, 490. [Google Scholar]
  27. Alqudah, A.M.; Qazan, S.; Alquran, H.; Qasmieh, I.A.; Alqudah, A. COVID-2019 detection using x-ray images and artificial intelligence hybrid systems. Poitiers 2020, 2, 1. [Google Scholar] [CrossRef]
  28. Aslan, M.F.; Unlersen, M.F.; Sabanci, K.; Durdu, A. CNN-based transfer learning–BiLSTM network: A novel approach for COVID-19 infection detection. Appl. Soft Comput. 2021, 98, 106912. [Google Scholar] [CrossRef]
  29. Wu, Y.-H.; Gao, S.-H.; Mei, J.; Xu, J.; Fan, D.-P.; Zhang, R.-G.; Cheng, M.-M. Jcs: An explainable COVID-19 diagnosis system by joint classification and segmentation. IEEE Trans. Image Process. 2021, 30, 3113–3126. [Google Scholar] [CrossRef]
  30. Xu, X.; Jiang, X.; Ma, C.; Du, P.; Li, X.; Lv, S.; Yu, L.; Ni, Q.; Chen, Y.; Su, J. A deep learning system to screen novel coronavirus disease 2019 pneumonia. Engineering 2020, 6, 1122–1129. [Google Scholar] [CrossRef]
  31. Winston, P.H. Artificial Intelligence; Addison-Wesley Longman Publishing Co., Inc.: Boston, MA, USA, 1992. [Google Scholar]
  32. Hamet, P.; Tremblay, J. Artificial intelligence in medicine. Metabolism 2017, 69, S36–S40. [Google Scholar] [CrossRef]
  33. Ramesh, A.; Kambhampati, C.; Monson, J.R.; Drew, P. Artificial intelligence in medicine. Ann. R. Coll. Surg. Engl. 2004, 86, 334. [Google Scholar] [CrossRef] [Green Version]
  34. Acharya, R.U.; Faust, O.; Alvin, A.P.; Sree, S.V.; Molinari, F.; Saba, L.; Nicolaides, A.; Suri, J.S. Symptomatic vs. asymptomatic plaque classification in carotid ultrasound. J. Med. Syst. 2012, 36, 1861–1871. [Google Scholar] [CrossRef] [PubMed]
  35. Acharya, U.; Vinitha Sree, S.; Mookiah, M.; Yantri, R.; Molinari, F.; Zieleźnik, W.; Małyszek-Tumidajewicz, J.; Stępień, B.; Bardales, R.; Witkowska, A. Diagnosis of Hashimoto’s thyroiditis in ultrasound using tissue characterization and pixel classification. Proc. Inst. Mech. Eng. Part H J. Eng. Med. 2013, 227, 788–798. [Google Scholar] [CrossRef] [PubMed]
  36. Acharya, U.R.; Faust, O.; S., V.S.; Alvin, A.P.; Krishnamurthi, G.; Seabra, J.C.; Sanches, J.; Suri, J.S. Understanding symptomatology of atherosclerotic plaque by image-based tissue characterization. Comput. Methods Programs Biomed. 2013, 110, 66–75. [Google Scholar] [CrossRef] [PubMed]
  37. Acharya, U.R.; Faust, O.; Sree, S.V.; Alvin, A.P.C.; Krishnamurthi, G.; Sanches, J.; Suri, J.S. Atheromatic™: Symptomatic vs. asymptomatic classification of carotid ultrasound plaque using a combination of HOS, DWT & texture. In Proceedings of the 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Boston, MA, USA, 30 August–3 September 2011; pp. 4489–4492. [Google Scholar]
  38. Acharya, U.R.; Mookiah, M.R.; Vinitha Sree, S.; Afonso, D.; Sanches, J.; Shafique, S.; Nicolaides, A.; Pedro, L.M.; Fernandes, E.F.J.; Suri, J.S. Atherosclerotic plaque tissue characterization in 2D ultrasound longitudinal carotid scans for automated classification: A paradigm for stroke risk assessment. Med. Biol. Eng. Comput. 2013, 51, 513–523. [Google Scholar] [CrossRef]
  39. Acharya, U.R.; Saba, L.; Molinari, F.; Guerriero, S.; Suri, J.S. Ovarian tumor characterization and classification: A class of GyneScan™ systems. In Proceedings of the 2012 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, San Diego, CA, USA, 28 August–1 September 2012; pp. 4446–4449. [Google Scholar]
  40. Biswas, M.; Kuppili, V.; Edla, D.R.; Suri, H.S.; Saba, L.; Marinhoe, R.T.; Sanches, J.M.; Suri, J.S. Symtosis: A liver ultrasound tissue characterization and risk stratification in optimized deep learning paradigm. Comput. Methods Programs Biomed. 2018, 155, 165–177. [Google Scholar] [CrossRef]
  41. Molinari, F.; Liboni, W.; Pavanelli, E.; Giustetto, P.; Badalamenti, S.; Suri, J.S. Accurate and automatic carotid plaque characterization in contrast enhanced 2-D ultrasound images. In Proceedings of the 2007 29th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Lyon, France, 22–26 August 2007; pp. 335–338. [Google Scholar]
  42. Pareek, G.; Acharya, U.R.; Sree, S.V.; Swapna, G.; Yantri, R.; Martis, R.J.; Saba, L.; Krishnamurthi, G.; Mallarini, G.; El-Baz, A. Prostate tissue characterization/classification in 144 patient population using wavelet and higher order spectra features from transrectal ultrasound images. Technol. Cancer Res. Treat. 2013, 12, 545–557. [Google Scholar] [CrossRef]
  43. Saba, L.; Sanagala, S.S.; Gupta, S.K.; Koppula, V.K.; Johri, A.M.; Khanna, N.N.; Mavrogeni, S.; Laird, J.R.; Pareek, G.; Miner, M.; et al. Multimodality carotid plaque tissue characterization and classification in the artificial intelligence paradigm: A narrative review for stroke application. Ann. Transl. Med. 2021, 9, 1206. [Google Scholar] [CrossRef]
  44. Agarwal, M.; Saba, L.; Gupta, S.K.; Johri, A.M.; Khanna, N.N.; Mavrogeni, S.; Laird, J.R.; Pareek, G.; Miner, M.; Sfikakis, P.P. Wilson disease tissue classification and characterization using seven artificial intelligence models embedded with 3D optimization paradigm on a weak training brain magnetic resonance imaging datasets: A supercomputer application. Med. Biol. Eng. Comput. 2021, 59, 511–533. [Google Scholar] [CrossRef]
  45. Saba, L.; Biswas, M.; Suri, H.S.; Viskovic, K.; Laird, J.R.; Cuadrado-Godia, E.; Nicolaides, A.; Khanna, N.N.; Viswanathan, V.; Suri, J.S. Ultrasound-based carotid stenosis measurement and risk stratification in diabetic cohort: A deep learning paradigm. Cardiovasc. Diagn. Ther. 2019, 9, 439–461. [Google Scholar] [CrossRef]
  46. Biswas, M.; Kuppili, V.; Saba, L.; Edla, D.R.; Suri, H.S.; Sharma, A.; Cuadrado-Godia, E.; Laird, J.R.; Nicolaides, A.; Suri, J.S. Deep learning fully convolution network for lumen characterization in diabetic patients using carotid ultrasound: A tool for stroke risk. Med. Biol. Eng. Comput. 2019, 57, 543–564. [Google Scholar] [CrossRef]
  47. Hesamian, M.H.; Jia, W.; He, X.; Kennedy, P. Deep learning techniques for medical image segmentation: Achievements and challenges. J. Digit. Imaging 2019, 32, 582–596. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. Ker, J.; Wang, L.; Rao, J.; Lim, T. Deep learning applications in medical image analysis. IEEE Access 2017, 6, 9375–9389. [Google Scholar] [CrossRef]
  49. Razzak, M.I.; Naz, S.; Zaib, A. Deep learning for medical image processing: Overview, challenges and the future. In Classification in BioApps; Springer: Berlin/Heidelberg, Germany, 2018; pp. 323–350. [Google Scholar]
  50. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; Van Der Laak, J.A.; Van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  51. Shen, D.; Wu, G.; Suk, H.-I. Deep learning in medical image analysis. Annu. Rev. Biomed. Eng. 2017, 19, 221–248. [Google Scholar] [CrossRef] [Green Version]
  52. Zhou, T.; Ruan, S.; Canu, S. A review: Deep learning for medical image segmentation using multi-modality fusion. Array 2019, 3, 100004. [Google Scholar] [CrossRef]
  53. Fourcade, A.; Khonsari, R. Deep learning in medical image analysis: A third eye for doctors. J. Stomatol. Oral Maxillofac. Surg. 2019, 120, 279–288. [Google Scholar] [CrossRef]
  54. Suri, J.S.; Agarwal, S.; Gupta, S.; Puvvula, A.; Viskovic, K.; Suri, N.; Alizad, A.; El-Baz, A.; Saba, L.; Fatemi, M.; et al. Systematic Review of Artificial Intelligence in Acute Respiratory Distress Syndrome for COVID-19 Lung Patients: A Biomedical Imaging Perspective. IEEE J. Biomed. Health Inform. 2021, 25, 4128–4139. [Google Scholar] [CrossRef]
  55. Jain, P.K.; Sharma, N.; Giannopoulos, A.A.; Saba, L.; Nicolaides, A.; Suri, J.S. Hybrid deep learning segmentation models for atherosclerotic plaque in internal carotid artery B-mode ultrasound. Comput. Biol. Med. 2021, 136, 104721. [Google Scholar] [CrossRef]
  56. Jena, B.; Saxena, S.; Nayak, G.K.; Saba, L.; Sharma, N.; Suri, J.S. Artificial intelligence-based hybrid deep learning models for image classification: The first narrative review. Comput. Biol. Med. 2021, 137, 104803. [Google Scholar] [CrossRef]
  57. Suri, J.S. Imaging Based Symptomatic Classification and Cardiovascular Stroke Risk Score Estimation. U.S. Patent 2011/0257545 A1, 21 October 2011. [Google Scholar]
  58. Suri, J.S.; Agarwal, S.; Elavarthi, P.; Pathak, R.; Ketireddy, V.; Columbu, M.; Saba, L.; Gupta, S.K.; Faa, G.; Singh, I.M.; et al. Inter-Variability Study of COVLIAS 1.0: Hybrid Deep Learning Models for COVID-19 Lung Segmentation in Computed Tomography. Diagnostics 2021, 11, 2025. [Google Scholar] [CrossRef]
  59. Suri, J.S.; Agarwal, S.; Pathak, R.; Ketireddy, V.; Columbu, M.; Saba, L.; Gupta, S.K.; Faa, G.; Singh, I.M.; Turk, M. COVLIAS 1.0: Lung Segmentation in COVID-19 Computed Tomography Scans Using Hybrid Deep Learning Artificial Intelligence Models. Diagnostics 2021, 11, 1405. [Google Scholar] [CrossRef] [PubMed]
  60. Saba, L.; Agarwal, M.; Patrick, A.; Puvvula, A.; Gupta, S.K.; Carriero, A.; Laird, J.R.; Kitas, G.D.; Johri, A.M.; Balestrieri, A.; et al. Six artificial intelligence paradigms for tissue characterisation and classification of non-COVID-19 pneumonia against COVID-19 pneumonia in computed tomography lungs. Int. J. Comput. Assist. Radiol. Surg. 2021, 16, 423–434. [Google Scholar] [CrossRef] [PubMed]
  61. Saba, L.; Agarwal, M.; Sanagala, S.S.; Gupta, S.K.; Sinha, G.; Johri, A.; Khanna, N.; Mavrogeni, S.; Laird, J.; Pareek, G. Brain MRI-based Wilson disease tissue classification: An optimised deep transfer learning approach. Electron. Lett. 2020, 56, 1395–1398. [Google Scholar] [CrossRef]
  62. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  63. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  64. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar]
  65. Jeong, C.; Yang, H.S.; Moon, K. Horizon detection in maritime images using scene parsing network. Electron. Lett. 2018, 54, 760–762. [Google Scholar] [CrossRef]
  66. Zhang, Z.; Gao, S.; Huang, Z. An automatic glioma segmentation system using a multilevel attention pyramid scene parsing network. Curr. Med. Imaging 2021, 17, 751–761. [Google Scholar] [CrossRef]
  67. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  68. Mazaheri, G.; Mithun, N.C.; Bappy, J.H.; Roy-Chowdhury, A.K. A Skip Connection Architecture for Localization of Image Manipulations. In Proceedings of the CVPR Workshops, Long Beach, CA, USA, 16–20 June 2019; pp. 119–129. [Google Scholar]
  69. Liu, F.; Ren, X.; Zhang, Z.; Sun, X.; Zou, Y. Rethinking skip connection with layer normalization. In Proceedings of the 28th International Conference on Computational Linguistics, Barcelona, Spain, 8–13 December 2020; pp. 3586–3598. [Google Scholar]
  70. Mateen, M.; Wen, J.; Song, S.; Huang, Z. Fundus image classification using VGG-19 architecture with PCA and SVD. Symmetry 2018, 11, 1. [Google Scholar] [CrossRef] [Green Version]
  71. Wen, L.; Li, X.; Li, X.; Gao, L. A new transfer learning based on VGG-19 network for fault diagnosis. In Proceedings of the 2019 IEEE 23rd International Conference on Computer Supported Cooperative Work in Design (CSCWD), Porto, Portugal, 6–8 May 2019; pp. 205–209. [Google Scholar]
  72. Xiao, J.; Wang, J.; Cao, S.; Li, B. Application of a novel and improved VGG-19 network in the detection of workers wearing masks. In Proceedings of the Journal of Physics: Conference Series, Sanya, China, 20–22 February 2020; p. 012041. [Google Scholar]
  73. Chaddad, A.; Hassan, L.; Desrosiers, C. Deep CNN models for predicting COVID-19 in CT and X-ray images. J. Med. Imaging 2021, 8, 014502. [Google Scholar] [CrossRef]
  74. Karasawa, H.; Liu, C.-L.; Ohwada, H. Deep 3d convolutional neural network architectures for alzheimer’s disease diagnosis. In Proceedings of the Asian Conference on Intelligent Information and Database Systems, Dong Hoi City, Vietnam, 19–21 March 2018; pp. 287–296. [Google Scholar]
  75. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  76. Shore, J.; Johnson, R. Properties of cross-entropy minimization. IEEE Trans. Inf. Theory 1981, 27, 472–482. [Google Scholar] [CrossRef] [Green Version]
  77. De Boer, P.-T.; Kroese, D.P.; Mannor, S.; Rubinstein, R.Y. A tutorial on the cross-entropy method. Ann. Oper. Res. 2005, 134, 19–67. [Google Scholar] [CrossRef]
  78. Jamin, A.; Humeau-Heurtier, A. (Multiscale) cross-entropy methods: A review. Entropy 2019, 22, 45. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  79. Chu, Z.; Wang, L.; Zhou, X.; Shi, Y.; Cheng, Y.; Laiginhas, R.; Zhou, H.; Shen, M.; Zhang, Q.; de Sisternes, L. Automatic geographic atrophy segmentation using optical attenuation in OCT scans with deep learning. Biomed. Opt. Express 2022, 13, 1328–1343. [Google Scholar] [CrossRef] [PubMed]
  80. Basar, S.; Waheed, A.; Ali, M.; Zahid, S.; Zareei, M.; Biswal, R.R. An Efficient Defocus Blur Segmentation Scheme Based on Hybrid LTP and PCNN. Sensors 2022, 22, 2724. [Google Scholar] [CrossRef]
  81. Eelbode, T.; Bertels, J.; Berman, M.; Vandermeulen, D.; Maes, F.; Bisschops, R.; Blaschko, M.B. Optimization for medical image segmentation: Theory and practice when evaluating with dice score or jaccard index. IEEE Trans. Med. Imaging 2020, 39, 3679–3690. [Google Scholar] [CrossRef]
  82. Dewitte, K.; Fierens, C.; Stockl, D.; Thienpont, L.M. Application of the Bland–Altman plot for interpretation of method-comparison studies: A critical investigation of its practice. Clin. Chem. 2002, 48, 799–801. [Google Scholar] [CrossRef]
  83. Giavarina, D. Understanding bland altman analysis. Biochem. Med. 2015, 25, 141–151. [Google Scholar] [CrossRef] [Green Version]
  84. Hanley, J.A. Receiver operating characteristic (ROC) methodology: The state of the art. Crit. Rev. Diagn. Imaging 1989, 29, 307–335. [Google Scholar]
  85. Zweig, M.H.; Campbell, G. Receiver-operating characteristic (ROC) plots: A fundamental evaluation tool in clinical medicine. Clin. Chem. 1993, 39, 561–577. [Google Scholar] [CrossRef]
  86. Streiner, D.L.; Cairney, J. What’s under the ROC? An introduction to receiver operating characteristics curves. Can. J. Psychiatry 2007, 52, 121–128. [Google Scholar] [CrossRef] [Green Version]
  87. Xu, M.; Fralick, D.; Zheng, J.Z.; Wang, B.; Tu, X.M.; Feng, C. The differences and similarities between two-sample t-test and paired t-test. Shanghai Arch. Psychiatry 2017, 29, 184. [Google Scholar] [PubMed]
  88. Hsu, H.; Lachenbruch, P.A. Paired t test. In Wiley Statsref: Statistics Reference Online; Wiley Online Library: Hoboken, NJ, USA, 2014. [Google Scholar]
  89. Nachar, N. The Mann-Whitney U: A test for assessing whether two independent samples come from the same distribution. Tutor. Quant. Methods Psychol. 2008, 4, 13–20. [Google Scholar] [CrossRef]
  90. McKnight, P.E.; Najab, J. Mann-Whitney U Test. In The Corsini Encyclopedia of Psychology; Wiley: Hoboken, NJ, USA, 2010. [Google Scholar]
  91. Birnbaum, Z. On a use of the Mann-Whitney statistic. In Proceedings of the Third Berkeley Symposium on Mathematical Statistics and Probability, 1956; University of California: La Jolla, CA, USA, 1956; pp. 13–17. [Google Scholar]
  92. Cuzick, J. A Wilcoxon-type test for trend. Stat. Med. 1985, 4, 87–90. [Google Scholar] [CrossRef]
  93. Schoonjans, F.; Zalata, A.; Depuydt, C.; Comhaire, F. MedCalc: A new computer program for medical statistics. Comput. Methods Programs Biomed. 1995, 48, 257–262. [Google Scholar] [CrossRef]
  94. MedCalc. MedCalc Statistical Software. 2016. Available online: https://www.scirp.org/(S(lz5mqp453ed%20snp55rrgjct55))/reference/referencespapers.aspx?referenceid=2690486 (accessed on 24 January 2022).
  95. Riffenburgh, R.H.; Gillen, D.L. Contents. In Statistics in Medicine, 4th ed.; Academic Press: Cambridge, MA, USA, 2020; pp. ix–xvi. [Google Scholar]
  96. Saba, L.; Than, J.C.; Noor, N.M.; Rijal, O.M.; Kassim, R.M.; Yunus, A.; Ng, C.R.; Suri, J.S. Inter-observer variability analysis of automatic lung delineation in normal and disease patients. J. Med. Syst. 2016, 40, 142. [Google Scholar] [CrossRef] [PubMed]
  97. Zhang, L.; Tanno, R.; Xu, M.-C.; Jin, C.; Jacob, J.; Cicarrelli, O.; Barkhof, F.; Alexander, D. Disentangling human error from ground truth in segmentation of medical images. Adv. Neural Inf. Process. Syst. 2020, 33, 15750–15762. [Google Scholar]
  98. Foncubierta Rodríguez, A.; Müller, H. Ground truth generation in medical imaging: A crowdsourcing-based iterative approach. In Proceedings of the ACM Multimedia 2012 Workshop on Crowdsourcing for Multimedia, Online, 29 October 2012; pp. 9–14. [Google Scholar]
  99. Ng, M.-Y.; Lee, E.Y.; Yang, J.; Yang, F.; Li, X.; Wang, H.; Lui, M.M.-S.; Lo, C.S.-Y.; Leung, B.; Khong, P.-L. Imaging profile of the COVID-19 infection: Radiologic findings and literature review. Radiol. Cardiothorac. Imaging 2020, 2, e200034. [Google Scholar] [CrossRef] [Green Version]
  100. Johnson, K.D.; Harris, C.; Cain, J.K.; Hummer, C.; Goyal, H.; Perisetti, A. Pulmonary and extra-pulmonary clinical manifestations of COVID-19. Front. Med. 2020, 7, 526. [Google Scholar] [CrossRef]
  101. Mottaghi, A.; Roham, M.; Makiani, M.J.; Ranjbar, M.; Laali, A.; Rahimian, N.R. Verifying Extra-Pulmonary Manifestation of COVID-19 in Firoozgar Hospital 2020: An Observational Study. 2021. Available online: https://www.researchgate.net/publication/348593779_Verifying_Extra-Pulmonary_Manifestation_of_COVID-19_in_Firoozgar_Hospital_2020_An_Observational_Study (accessed on 24 January 2022).
  102. Bansal, M. Cardiovascular disease and COVID-19. Diabetes Metab. Syndr. Clin. Res. Rev. 2020, 14, 247–250. [Google Scholar] [CrossRef]
  103. Linschoten, M.; Peters, S.; van Smeden, M.; Jewbali, L.S.; Schaap, J.; Siebelink, H.-M.; Smits, P.C.; Tieleman, R.G.; van der Harst, P.; van Gilst, W.H. Cardiac complications in patients hospitalised with COVID-19. Eur. Heart J. Acute Cardiovasc. Care 2020, 9, 817–823. [Google Scholar] [CrossRef]
  104. Yohannes, A.M. COPD patients in a COVID-19 society: Depression and anxiety. Expert Rev. Respir. Med. 2021, 15, 5–7. [Google Scholar] [CrossRef] [PubMed]
  105. Team, C.C.-R.; Team, C.C.-R.; Team, C.C.-R.; Chow, N.; Fleming-Dutra, K.; Gierke, R.; Hall, A.; Hughes, M.; Pilishvili, T.; Ritchey, M. Preliminary estimates of the prevalence of selected underlying health conditions among patients with coronavirus disease 2019—United States, 12 February–28 March 2020. Morb. Mortal. Wkly. Rep. 2020, 69, 382–386. [Google Scholar]
  106. Gunraj, H.; Wang, L.; Wong, A. COVIDNet-CT: A Tailored Deep Convolutional Neural Network Design for Detection of COVID-19 Cases From Chest CT Images. Front. Med. 2020, 7, 608525. [Google Scholar] [CrossRef]
  107. Iyer, T.J.; Joseph Raj, A.N.; Ghildiyal, S.; Nersisson, R. Performance analysis of lightweight CNN models to segment infectious lung tissues of COVID-19 cases from tomographic images. PeerJ Comput. Sci. 2021, 7, e368. [Google Scholar] [CrossRef] [PubMed]
  108. Ranjbarzadeh, R.; Jafarzadeh Ghoushchi, S.; Bendechache, M.; Amirabadi, A.; Ab Rahman, M.N.; Baseri Saadi, S.; Aghamohammadi, A.; Kooshki Forooshani, M. Lung Infection Segmentation for COVID-19 Pneumonia Based on a Cascade Convolutional Network from CT Images. BioMed Res. Int. 2021, 2021, 5544742. [Google Scholar] [CrossRef]
  109. Ding, W.; Abdel-Basset, M.; Hawash, H.; Elkomy, O.M. MT-nCov-Net: A Multitask Deep-Learning Framework for Efficient Diagnosis of COVID-19 Using Tomography Scans. IEEE Trans. Cybern. 2021. [Google Scholar] [CrossRef]
  110. Hou, S.M.; Jia, C.L.; Hou, M.J.; Fernandes, S.L.; Guo, J.C. A Study on Weak Edge Detection of COVID-19′s CT Images Based on Histogram Equalization and Improved Canny Algorithm. Comput. Math Methods Med. 2021, 2021, 5208940. [Google Scholar] [CrossRef]
  111. Kumar Singh, V.; Abdel-Nasser, M.; Pandey, N.; Puig, D. LungINFseg: Segmenting COVID-19 Infected Regions in Lung CT Images Based on a Receptive-Field-Aware Deep Learning Framework. Diagnostics 2021, 11, 158. [Google Scholar] [CrossRef]
  112. Lizzi, F.; Agosti, A.; Brero, F.; Cabini, R.F.; Fantacci, M.E.; Figini, S.; Lascialfari, A.; Laruina, F.; Oliva, P.; Piffer, S. Quantification of pulmonary involvement in COVID-19 pneumonia by means of a cascade oftwo U-nets: Training and assessment on multipledatasets using different annotation criteria. arXiv 2021, arXiv:02566. [Google Scholar]
  113. Qi, S.; Xu, C.; Li, C.; Tian, B.; Xia, S.; Ren, J.; Yang, L.; Wang, H.; Yu, H. DR-MIL: Deep represented multiple instance learning distinguishes COVID-19 from community-acquired pneumonia in CT images. Comput. Methods Programs Biomed. 2021, 211, 106406. [Google Scholar] [CrossRef]
  114. Paluru, N.; Dayal, A.; Jenssen, H.B.; Sakinis, T.; Cenkeramaddi, L.R.; Prakash, J.; Yalavarthy, P.K.; Systems, L. Anam-Net: Anamorphic depth embedding-based lightweight CNN for segmentation of anomalies in COVID-19 chest CT images. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 932–946. [Google Scholar] [CrossRef] [PubMed]
  115. Zhang, P.; Zhong, Y.; Deng, Y.; Tang, X.; Li, X. CoSinGAN: Learning COVID-19 Infection Segmentation from a Single Radiological Image. Diagnostics 2020, 10, 901. [Google Scholar] [CrossRef] [PubMed]
  116. Budak, U.; Cibuk, M.; Comert, Z.; Sengur, A. Efficient COVID-19 Segmentation from CT Slices Exploiting Semantic Segmentation with Integrated Attention Mechanism. J. Digit. Imaging 2021, 34, 263–272. [Google Scholar] [CrossRef] [PubMed]
  117. Amyar, A.; Modzelewski, R.; Li, H.; Ruan, S. Multi-task deep learning based CT imaging analysis for COVID-19 pneumonia: Classification and segmentation. Comput. Biol. Med. 2020, 126, 104037. [Google Scholar] [CrossRef]
  118. Cai, W.; Liu, T.; Xue, X.; Luo, G.; Wang, X.; Shen, Y.; Fang, Q.; Sheng, J.; Chen, F.; Liang, T. CT Quantification and Machine-learning Models for Assessment of Disease Severity and Prognosis of COVID-19 Patients. Acad. Radiol. 2020, 27, 1665–1678. [Google Scholar] [CrossRef]
  119. Ma, J.; Wang, Y.; An, X.; Ge, C.; Yu, Z.; Chen, J.; Zhu, Q.; Dong, G.; He, J.; He, Z.; et al. Toward data-efficient learning: A benchmark for COVID-19 CT lung and infection segmentation. Med. Phys. 2021, 48, 1197–1210. [Google Scholar] [CrossRef]
  120. Kuchana, M.; Srivastava, A.; Das, R.; Mathew, J.; Mishra, A.; Kiran Khatter Multimedia Tools and Applications. AI aiding in diagnosing, tracking recovery of COVID-19 using deep learning on Chest CT scans. Multimed. Tools Appl. 2021, 80, 9161–9175. [Google Scholar] [CrossRef]
  121. Gao, S.-H.; Cheng, M.-M.; Zhao, K.; Zhang, X.-Y.; Yang, M.-H.; Torr, P. Res2net: A new multi-scale backbone architecture. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 652–662. [Google Scholar] [CrossRef] [Green Version]
  122. Ding, L.; Goshtasby, A. On the Canny edge detector. Pattern Recognit. 2001, 34, 721–725. [Google Scholar] [CrossRef]
  123. McIlhagga, W. The Canny edge detector revisited. Int. J. Comput. Vis. 2011, 91, 251–261. [Google Scholar] [CrossRef] [Green Version]
  124. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  125. Noh, H.; Hong, S.; Han, B. Learning deconvolution network for semantic segmentation. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1520–1528. [Google Scholar]
  126. Fan, D.-P.; Zhou, T.; Ji, G.-P.; Zhou, Y.; Chen, G.; Fu, H.; Shen, J.; Shao, L. Inf-net: Automatic COVID-19 lung infection segmentation from ct images. IEEE Trans. Med. Imaging 2020, 39, 2626–2637. [Google Scholar] [CrossRef] [PubMed]
  127. Müller, D.; Kramer, F. MIScnn: A framework for medical image segmentation with convolutional neural networks and deep learning. BMC Med. Imaging 2021, 21, 12. [Google Scholar] [CrossRef] [PubMed]
  128. Müller, D.; Rey, I.S.; Kramer, F. Automated chest ct image segmentation of COVID-19 lung infection based on 3d u-net. arXiv 2020, arXiv:2007.04774. [Google Scholar]
  129. Arunachalam, A.; Ravi, V.; Acharya, V.; Pham, T.D. Toward data-model-agnostic autonomous machine-generated data labeling and annotation platform: COVID-19 autoannotation use case. IEEE Trans. Eng. Manag. 2021, 1–12. [Google Scholar] [CrossRef]
  130. Schneider, U.; Pedroni, E.; Lomax, A. The calibration of CT Hounsfield units for radiotherapy treatment planning. Phys. Med. Biol. 1996, 41, 111. [Google Scholar] [CrossRef] [PubMed]
  131. Zurl, B.; Tiefling, R.; Winkler, P.; Kindl, P.; Kapp, K. Hounsfield units variations. Strahlenther. Onkol. 2014, 190, 88–93. [Google Scholar] [CrossRef]
  132. DenOtter, T.D.; Schubert, J. Hounsfield Unit; StatPearls Publishing: Treasure Island, FL, USA, 2019. [Google Scholar]
  133. Harten, A. Multiresolution representation of data: A general framework. SIAM J. Numer. Anal. 1996, 33, 1205–1256. [Google Scholar] [CrossRef]
  134. Rosenfeld, A. Multiresolution Image Processing and Analysis; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2013; Volume 12. [Google Scholar]
  135. Molinari, F.; Pattichis, C.S.; Zeng, G.; Saba, L.; Acharya, U.R.; Sanfilippo, R.; Nicolaides, A.; Suri, J.S. Completely automated multiresolution edge snapper—A new technique for an accurate carotid ultrasound IMT measurement: Clinical validation and benchmarking on a multi-institutional database. IEEE Trans. Image Process. 2011, 21, 1211–1222. [Google Scholar] [CrossRef]
  136. El-Baz, A.; Gimel’farb, G.; Suri, J.S. Stochastic Modeling for Medical Image Analysis, 1st ed.; CRC Press: Boca Raton, FL, USA, 2015. [Google Scholar]
  137. El-Baz, A.; Suri, J.S. Big Data in Multimodal Medical Imaging; CRC Press: Boca Raton, FL, USA, 2019. [Google Scholar]
  138. Saba, L.; Banchhor, S.K.; Araki, T.; Viskovic, K.; Londhe, N.D.; Laird, J.R.; Suri, H.S.; Suri, J.S. Intra- and inter-operator reproducibility of automated cloud-based carotid lumen diameter ultrasound measurement. Indian Heart J. 2018, 70, 649–664. [Google Scholar] [CrossRef]
  139. Than, J.C.; Saba, L.; Noor, N.M.; Rijal, O.M.; Kassim, R.M.; Yunus, A.; Suri, H.S.; Porcu, M.; Suri, J.S. Lung disease stratification using amalgamation of Riesz and Gabor transforms in machine learning framework. Comput. Biol. Med. 2017, 89, 197–211. [Google Scholar] [CrossRef]
  140. Zhang, S.; Suri, J.S.; Salvado, O.; Chen, Y.; Wacker, F.K.; Wilson, D.L.; Duerk, J.L.; Lewin, J.S. Inter-and Intra-Observer Variability Assessment of in Vivo Carotid Plaque Burden Quantification Using Multi-Contrast Dark Blood MR Images. Stud. Health Technol. Inform. 2005, 113, 384–393. [Google Scholar] [PubMed]
  141. El-Baz, A.; Gimelfarb, G.; Falk, R.; El-Ghar, M.A.; Suri, J. Appearance analysis for the early assessment of detected lung nodules. Lung Imaging Comput. Aided Diagn. 2011, 17, 395–404. [Google Scholar]
  142. Noor, N.M.; Than, J.C.; Rijal, O.M.; Kassim, R.M.; Yunus, A.; Zeki, A.A.; Anzidei, M.; Saba, L.; Suri, J.S. Automatic lung segmentation using control feedback system: Morphology and texture paradigm. J. Med. Syst. 2015, 39, 1–18. [Google Scholar] [CrossRef] [PubMed]
  143. Sanagala, S.S.; Nicolaides, A.; Gupta, S.K.; Koppula, V.K.; Saba, L.; Agarwal, S.; Johri, A.M.; Kalra, M.S.; Suri, J.S. Ten Fast Transfer Learning Models for Carotid Ultrasound Plaque Tissue Characterization in Augmentation Framework Embedded with Heatmaps for Stroke Risk Stratification. Diagnostics 2021, 11, 2109. [Google Scholar] [CrossRef]
  144. Molinari, F.; Meiburger, K.M.; Saba, L.; Acharya, U.R.; Ledda, G.; Zeng, G.; Ho, S.Y.S.; Ahuja, A.T.; Ho, S.C.; Nicolaides, A. Ultrasound IMT measurement on a multi-ethnic and multi-institutional database: Our review and experience using four fully automated and one semi-automated methods. Comput. Methods Programs Biomed. 2012, 108, 946–960. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  145. Shu, X.; Yang, J.; Yan, R.; Song, Y. Expansion-Squeeze-Excitation Fusion Network for Elderly Activity Recognition. IEEE Trans. Circuits Syst. Video Technol. 2022, 1. [Google Scholar] [CrossRef]
  146. Shu, X.; Zhang, L.; Qi, G.-J.; Liu, W.; Tang, J. Spatiotemporal co-attention recurrent neural networks for human-skeleton motion prediction. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3300–3315. [Google Scholar] [CrossRef]
  147. Shu, X.; Qi, G.-J.; Tang, J.; Wang, J. Weakly-shared deep transfer networks for heterogeneous-domain knowledge propagation. In Proceedings of the 23rd ACM International Conference on Multimedia, Online, 13 October 2015; pp. 35–44. [Google Scholar]
  148. Sudre, C.H.; Li, W.; Vercauteren, T.; Ourselin, S.; Jorge Cardoso, M. Generalised dice overlap as a deep learning loss function for highly unbalanced segmentations. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Berlin/Heidelberg, Germany, 2017; pp. 240–248. [Google Scholar]
  149. Xu, J.; Zhang, Z.; Friedman, T.; Liang, Y.; Broeck, G. A semantic loss function for deep learning with symbolic knowledge. In Proceedings of the International Conference on Machine Learning, Macau, China, 26–28 February 2018; pp. 5502–5511. [Google Scholar]
  150. Chen, L.; Qu, H.; Zhao, J.; Chen, B.; Principe, J.C. Efficient and robust deep learning with correntropy-induced loss function. Neural Comput. Appl. 2016, 27, 1019–1031. [Google Scholar] [CrossRef]
  151. Salehi, S.S.M.; Erdogmus, D.; Gholipour, A. Tversky loss function for image segmentation using 3D fully convolutional deep networks. In Machine Learning in Medical Imaging; Springer: Cham, Switzerland, 2017; pp. 379–387. [Google Scholar]
  152. de La Torre, J.; Puig, D.; Valls, A. Weighted kappa loss function for multi-class classification of ordinal data in deep learning. Pattern Recognit. Lett. 2018, 105, 144–154. [Google Scholar] [CrossRef]
  153. Jain, P.K.; Sharma, N.; Saba, L.; Paraskevas, K.I.; Kalra, M.K.; Johri, A.; Nicolaides, A.N.; Suri, J.S. Automated deep learning-based paradigm for high-risk plaque detection in B-mode common carotid ultrasound scans: An asymptomatic Japanese cohort study. Int. Angiol. A J. Int. Union Angiol. 2022, 41, 9–23. [Google Scholar] [CrossRef]
  154. Jamthikar, A.; Gupta, D.; Saba, L.; Khanna, N.N.; Araki, T.; Viskovic, K.; Mavrogeni, S.; Laird, J.R.; Pareek, G.; Miner, M. Cardiovascular/stroke risk predictive calculators: A comparison between statistical and machine learning models. Cardiovasc. Diagn. Ther. 2020, 10, 919. [Google Scholar] [CrossRef] [PubMed]
  155. Tandel, G.S.; Balestrieri, A.; Jujaray, T.; Khanna, N.N.; Saba, L.; Suri, J.S. Multiclass magnetic resonance imaging brain tumor classification using artificial intelligence paradigm. Comput. Biol. Med. 2020, 122, 103804. [Google Scholar] [CrossRef] [PubMed]
Figure 1. AI system workflow for comparing COVLIAS 1.0Lesion against MedSeg.
Figure 1. AI system workflow for comparing COVLIAS 1.0Lesion against MedSeg.
Diagnostics 12 01283 g001
Figure 2. Raw CT images from the Croatia dataset.
Figure 2. Raw CT images from the Croatia dataset.
Diagnostics 12 01283 g002
Figure 3. Manual delineation overlays (red) from tracer 1 on raw CT images.
Figure 3. Manual delineation overlays (red) from tracer 1 on raw CT images.
Diagnostics 12 01283 g003
Figure 4. Manual delineation overlays (red) from tracer 2 on raw CT images.
Figure 4. Manual delineation overlays (red) from tracer 2 on raw CT images.
Diagnostics 12 01283 g004
Figure 5. PSPNet’s architecture.
Figure 5. PSPNet’s architecture.
Diagnostics 12 01283 g005
Figure 6. VGG-SegNet HDL model’s architecture.
Figure 6. VGG-SegNet HDL model’s architecture.
Diagnostics 12 01283 g006
Figure 7. ResNet-SegNet HDL model’s architecture.
Figure 7. ResNet-SegNet HDL model’s architecture.
Diagnostics 12 01283 g007
Figure 8. VGG-UNet’s architecture.
Figure 8. VGG-UNet’s architecture.
Diagnostics 12 01283 g008
Figure 9. ResNet-UNet’s architecture.
Figure 9. ResNet-UNet’s architecture.
Diagnostics 12 01283 g009
Figure 10. Training accuracy and loss plot for the best AI model (ResNet-UNet).
Figure 10. Training accuracy and loss plot for the best AI model (ResNet-UNet).
Diagnostics 12 01283 g010
Figure 11. Row 1: raw CT; row 2: MD 1 (gold standard); rows 3–7: overlay images—AI (green) over MD (red). The 5 AI models are PSPNet (row 3), VGG-SegNet (row 4), ResNet-SegNet (row 5), VGG-UNet (row 6), and ResNet-UNet (row 7).
Figure 11. Row 1: raw CT; row 2: MD 1 (gold standard); rows 3–7: overlay images—AI (green) over MD (red). The 5 AI models are PSPNet (row 3), VGG-SegNet (row 4), ResNet-SegNet (row 5), VGG-UNet (row 6), and ResNet-UNet (row 7).
Diagnostics 12 01283 g011
Figure 12. Cumulative frequency plots for Dice (left) and Jaccard (right) for PSPNet when computed against MD 1.
Figure 12. Cumulative frequency plots for Dice (left) and Jaccard (right) for PSPNet when computed against MD 1.
Diagnostics 12 01283 g012
Figure 13. Cumulative frequency plot for Dice (left) and Jaccard (right) for VGG-SegNet when computed against MD 1.
Figure 13. Cumulative frequency plot for Dice (left) and Jaccard (right) for VGG-SegNet when computed against MD 1.
Diagnostics 12 01283 g013
Figure 14. Cumulative frequency plot for Dice (left) and Jaccard (right) for ResNet-SegNet when computed against MD 1.
Figure 14. Cumulative frequency plot for Dice (left) and Jaccard (right) for ResNet-SegNet when computed against MD 1.
Diagnostics 12 01283 g014
Figure 15. Cumulative frequency plot for Dice (left) and Jaccard (right) for VGG-UNet when computed against MD 1.
Figure 15. Cumulative frequency plot for Dice (left) and Jaccard (right) for VGG-UNet when computed against MD 1.
Diagnostics 12 01283 g015
Figure 16. Cumulative frequency plot for Dice (left) and Jaccard (right) for ResNet-UNet when computed against MD 1.
Figure 16. Cumulative frequency plot for Dice (left) and Jaccard (right) for ResNet-UNet when computed against MD 1.
Diagnostics 12 01283 g016
Figure 17. Cumulative frequency plot for Dice (left) and Jaccard (right) for MedSeg when computed against MD 1.
Figure 17. Cumulative frequency plot for Dice (left) and Jaccard (right) for MedSeg when computed against MD 1.
Diagnostics 12 01283 g017
Figure 18. Lesions detected by the best AI model (ResNet-UNet) vs. MedSeg vs. MD 1 vs. MD 2.
Figure 18. Lesions detected by the best AI model (ResNet-UNet) vs. MedSeg vs. MD 1 vs. MD 2.
Diagnostics 12 01283 g018
Figure 19. Correlation coefficient plots for (left) PSPNet vs. MD 1 and (right) PSPNet vs. MD 2.
Figure 19. Correlation coefficient plots for (left) PSPNet vs. MD 1 and (right) PSPNet vs. MD 2.
Diagnostics 12 01283 g019
Figure 20. Correlation coefficient plots for (left) VGG-SegNet vs. MD 1 and (right) VGG-SegNet vs. MD 2.
Figure 20. Correlation coefficient plots for (left) VGG-SegNet vs. MD 1 and (right) VGG-SegNet vs. MD 2.
Diagnostics 12 01283 g020
Figure 21. Correlation coefficient plots for (left) ResNet-SegNet vs. MD 1 and (right) ResNet-SegNet vs. MD 2.
Figure 21. Correlation coefficient plots for (left) ResNet-SegNet vs. MD 1 and (right) ResNet-SegNet vs. MD 2.
Diagnostics 12 01283 g021
Figure 22. Correlation coefficient plots for (left) VGG-UNet vs. MD 1 and (right) VGG-UNet vs. MD 2.
Figure 22. Correlation coefficient plots for (left) VGG-UNet vs. MD 1 and (right) VGG-UNet vs. MD 2.
Diagnostics 12 01283 g022
Figure 23. Correlation coefficient plots for (left) ResNet-UNet vs. MD 1 and (right) ResNet-UNet vs. MD 2.
Figure 23. Correlation coefficient plots for (left) ResNet-UNet vs. MD 1 and (right) ResNet-UNet vs. MD 2.
Diagnostics 12 01283 g023
Figure 24. Correlation coefficient plots for (left) MedSeg vs. MD 1 and (right) MedSeg vs. MD 2.
Figure 24. Correlation coefficient plots for (left) MedSeg vs. MD 1 and (right) MedSeg vs. MD 2.
Diagnostics 12 01283 g024
Figure 25. ROC for COVLIAS (5 AI models) vs. MedSeg using MD 1 (left) and MD 2 (right).
Figure 25. ROC for COVLIAS (5 AI models) vs. MedSeg using MD 1 (left) and MD 2 (right).
Diagnostics 12 01283 g025
Figure 26. BA plot for PSPNet using MD 1 (left) vs. MD 2 (right).
Figure 26. BA plot for PSPNet using MD 1 (left) vs. MD 2 (right).
Diagnostics 12 01283 g026
Figure 27. BA plot for VGG-SegNet using MD 1 (left) vs. MD 2 (right).
Figure 27. BA plot for VGG-SegNet using MD 1 (left) vs. MD 2 (right).
Diagnostics 12 01283 g027
Figure 28. BA plot for ResNet-SegNet using MD 1 (left) vs. MD 2 (right).
Figure 28. BA plot for ResNet-SegNet using MD 1 (left) vs. MD 2 (right).
Diagnostics 12 01283 g028
Figure 29. BA plot for VGG-UNet using MD 1 (left) vs. MD 2 (right).
Figure 29. BA plot for VGG-UNet using MD 1 (left) vs. MD 2 (right).
Diagnostics 12 01283 g029
Figure 30. BA plot for ResNet-UNet using MD 1 (left) vs. MD 2 (right).
Figure 30. BA plot for ResNet-UNet using MD 1 (left) vs. MD 2 (right).
Diagnostics 12 01283 g030
Figure 31. BA plot for MedSeg using MD 1 (left) vs. MD 2 (right).
Figure 31. BA plot for MedSeg using MD 1 (left) vs. MD 2 (right).
Diagnostics 12 01283 g031
Table 1. Dice similarity coefficient and Jaccard index when computed against MedSeg.
Table 1. Dice similarity coefficient and Jaccard index when computed against MedSeg.
MD 1MD 2
Dice% Diff *Jaccard% Diff *Dice% Diff *Jaccard% Diff *
ResNet-SegNet0.771%0.632%0.744%0.605%
PSPNet0.794%0.655%0.770%0.642%
VGG-SegNet0.794%0.666%0.804%0.688%
VGG-UNet0.805%0.678%0.781%0.653%
ResNet-UNet0.839%0.7115%0.804%0.688%
Mean of AI0.805%0.667%0.783%0.655%
MedSeg0.76-0.62-0.77-0.63-
DSC: Dice similarity coefficient; JI: Jaccard index; * % Diff = absolute (COVLIAS − MedSeg)/MedSeg.
Table 2. Correlation coefficient plot: 5 AI models vs. MD 1 and 5 AI models vs. MD 2.
Table 2. Correlation coefficient plot: 5 AI models vs. MD 1 and 5 AI models vs. MD 2.
MD 1MD 2
CC% Diff *CC% Diff *
ResNet-SegNet0.9011%0.802%
PSPNet0.9011%0.811%
VGG-SegNet0.792%0.794%
VGG-UNet0.810%0.811%
ResNet-UNet0.9214%0.802%
Mean AI0.868%0.802%
MedSeg0.81-0.82-
CC: Correlation coefficient; * % Diff = absolute (COVLIAS − MedSeg)/MedSeg.
Table 3. Statistical tests for the 5 AI models and MedSeg against MD 1 and MD 2.
Table 3. Statistical tests for the 5 AI models and MedSeg against MD 1 and MD 2.
Paired t-TestMann-WhitneyWilcoxon
PSPNet vs. MD 1p < 0.0001p < 0.0001p < 0.0001
PSPNet vs. MD 2p < 0.0001p < 0.0001p < 0.0001
VGG-SegNet vs. MD 1p < 0.0001p < 0.0001p < 0.0001
VGG-SegNet vs. MD 2p < 0.0001p < 0.0001p < 0.0001
ResNet-SegNet vs. MD 1p < 0.0001p < 0.0001p < 0.0001
ResNet-SegNet vs. MD 2p < 0.0001p < 0.0001p < 0.0001
VGG-UNet vs. MD 1p < 0.0001p < 0.0001p < 0.0001
VGG-UNet vs. MD 2p < 0.0001p < 0.0001p < 0.0001
ResNet-UNet vs. MD 1p < 0.0001p < 0.0001p < 0.0001
ResNet-UNet vs. MD 2p < 0.0001p < 0.0001p < 0.0001
MedSeg vs. MD 1p < 0.0001p < 0.0001p < 0.0001
MedSeg vs. MD 2p < 0.0001p < 0.0001p < 0.0001
MD 1 vs. MD 2p < 0.0001p < 0.0001p < 0.0001
Table 4. Parameters for the five AI models.
Table 4. Parameters for the five AI models.
SNAttributesPSP-NetVGG-SegNetVGG-UNetResNet-SegNetResNet-UNet
1Backbone-encoderNAVGG-16VGG-16Res-50Res-50
2# Parameters~4.4 M~11.6 M~12.4 M~15 M~16.5 M
3# NN layers543336160165
4Model size (MB)50133142171188
5Batch size88444
6Training time *~15~50~54~60~63
7Prediction time<1 s<1 s<1 s<1 s<1 s
* In minutes; MB: megabytes; M: million; NN: neural network; Res: ResNet.
Table 5. Benchmarking table.
Table 5. Benchmarking table.
A1A2A3A4A5A6A7A8A9A10A11A12A13A14A15A16A17A18
AuthorYearModelClassifier# Patients# Img# GT TracingsFocusObjectiveModalityOpt&Augm#DSCACCAUCRad *CEBench
Ding et al. [109]2021MT-nCov-NetRes2Net50189364858Segm.LesionCT0.8699.610.923
Hou et al. [110]2021Improved Canny edge detectorNA271812NANALesionCT🗶🗶🗶🗶🗶🗶🗶
Lizzi et al. [112]2021Cascaded UNetNANANANAClass. + Segm.LesionCT0.6293🗶1🗶
Qi et al. [113]2021DR-MIL(ResNet-50 and Xception24124101NANACT🗶🗶950.943🗶
Paluru et al. [114]2021Anam-Netcustom (UNet + ENet)6943391Segm.LesionCT🗶0.7798🗶🗶
Zhang et al. [115]2020CoSinGANNA707041Class. + Segm.LesionCT0.75🗶🗶🗶🗶
Singh et al. [111]2021LungINFsegModified UNet2018001Heatmap + Segm.LesionCT0.880🗶🗶🗶
Amyar et al. [117]2020UNetNA136913691Class. + Segm.LesionCT🗶0.88940.97🗶
Budak et al. [116]2021A-SegNetNA694731Segm.LesionCT🗶0.89🗶🗶🗶🗶🗶
Cai et al. [118]2020UNetNA992501Class. + Segm.Lung + lesion + predict ICU stayCT🗶0.77🗶🗶🗶🗶
Ma et al. [119]2021UNetNA70NA1Segm.LesionCT🗶0.67🗶🗶2
Kuchana et al. [120]2020UNet and attention UNet,NA509291Segm.Lung + lesionCT🗶0.84🗶🗶1🗶🗶
Suri et al. [proposed]2021PSPNet,
VGG-SegNet
ResNet-SegNet
VGG-UNet
ResNet-UNet
VGG,
ResNet
4030002Segm.LesionCT🗶0.79
0.79
0.77
0.80
0.83
0.95
0.96
0.95
0.97
0.98
0.95
0.94
0.87
0.91
0.87
2
* Rad: radiologist; Augm#: augmentation; Opt&: optimization; CE: clinical evaluation; Bench: benchmarking; # Img: number of images.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Suri, J.S.; Agarwal, S.; Chabert, G.L.; Carriero, A.; Paschè, A.; Danna, P.S.C.; Saba, L.; Mehmedović, A.; Faa, G.; Singh, I.M.; et al. COVLIAS 1.0Lesion vs. MedSeg: An Artificial Intelligence Framework for Automated Lesion Segmentation in COVID-19 Lung Computed Tomography Scans. Diagnostics 2022, 12, 1283. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics12051283

AMA Style

Suri JS, Agarwal S, Chabert GL, Carriero A, Paschè A, Danna PSC, Saba L, Mehmedović A, Faa G, Singh IM, et al. COVLIAS 1.0Lesion vs. MedSeg: An Artificial Intelligence Framework for Automated Lesion Segmentation in COVID-19 Lung Computed Tomography Scans. Diagnostics. 2022; 12(5):1283. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics12051283

Chicago/Turabian Style

Suri, Jasjit S., Sushant Agarwal, Gian Luca Chabert, Alessandro Carriero, Alessio Paschè, Pietro S. C. Danna, Luca Saba, Armin Mehmedović, Gavino Faa, Inder M. Singh, and et al. 2022. "COVLIAS 1.0Lesion vs. MedSeg: An Artificial Intelligence Framework for Automated Lesion Segmentation in COVID-19 Lung Computed Tomography Scans" Diagnostics 12, no. 5: 1283. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics12051283

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop