Next Article in Journal
Differential Diagnosis of Hepatic Mass with Central Scar: Focal Nodular Hyperplasia Mimicking Fibrolamellar Hepatocellular Carcinoma
Next Article in Special Issue
Comparison between Deep Learning and Conventional Machine Learning in Classifying Iliofemoral Deep Venous Thrombosis upon CT Venography
Previous Article in Journal
Clinical Evaluation of a New Electronic Periodontal Probe: A Randomized Controlled Clinical Trial
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid Deep Learning Model for Endoscopic Lesion Detection and Classification Using Endoscopy Videos

1
Department of Computer Science, University of Gujrat, Gujrat 50700, Pakistan
2
Department of Information Sciences, University of Education Lahore, Lahore 41000, Pakistan
3
Centre for Smart Systems, AI and Cybersecurity, Staffordshire University, Stoke-on-Trent ST4 2DE, UK
4
Department of Computer Science, College of Computers and Information Technology, Taif University, P. O. Box 11099, Taif 21944, Saudi Arabia
*
Author to whom correspondence should be addressed.
Submission received: 28 November 2021 / Revised: 22 December 2021 / Accepted: 23 December 2021 / Published: 26 December 2021
(This article belongs to the Special Issue Artificial Intelligence in Clinical Medical Imaging Analysis)

Abstract

:
In medical imaging, the detection and classification of stomach diseases are challenging due to the resemblance of different symptoms, image contrast, and complex background. Computer-aided diagnosis (CAD) plays a vital role in the medical imaging field, allowing accurate results to be obtained in minimal time. This article proposes a new hybrid method to detect and classify stomach diseases using endoscopy videos. The proposed methodology comprises seven significant steps: data acquisition, preprocessing of data, transfer learning of deep models, feature extraction, feature selection, hybridization, and classification. We selected two different CNN models (VGG19 and Alexnet) to extract features. We applied transfer learning techniques before using them as feature extractors. We used a genetic algorithm (GA) in feature selection, due to its adaptive nature. We fused selected features of both models using a serial-based approach. Finally, the best features were provided to multiple machine learning classifiers for detection and classification. The proposed approach was evaluated on a personally collected dataset of five classes, including gastritis, ulcer, esophagitis, bleeding, and healthy. We observed that the proposed technique performed superbly on Cubic SVM with 99.8% accuracy. For the authenticity of the proposed technique, we considered these statistical measures: classification accuracy, recall, precision, False Negative Rate (FNR), Area Under the Curve (AUC), and time. In addition, we provided a fair state-of-the-art comparison of our proposed technique with existing techniques that proves its worthiness.

1. Introduction

The stomach is a muscular organ that helps to digest food. It is located on the left side of the upper abdomen. It takes food from the esophagus through a muscular valve called the lower esophageal sphincter [1]. The stomach performs three tasks. Firstly, it stores the food that we eat. Secondly, it mixes the food and discharges gastric juices that help break down and digest food. Then it moves that mixture into the small intestine. Sometimes, it can be affected by different diseases, such as gastritis, gastroparesis, diarrhea, peptic ulcers, Crohn’s disease, etc. Gastritis is the swelling of the stomach lining. Intense gastritis may come on abruptly. Chronic gastritis happens gradually. As indicated by the Cleveland Clinic, 8 out of 1000 individuals have intense gastritis, and 2 out of 10,000 have chronic gastritis. The symptoms of gastritis are vomiting, hiccups, indigestion, etc. The significant causes behind gastritis are stress, viral infection, etc. A peptic ulcer is another critical disease of the stomach. If the internal lining of the stomach is destroyed, then the patients will likely be positive for a stomach disease. Most are situated in the top layer of the inward covering. An ulcer that goes entirely through the stomach lining is a hole and requires prompt clinical consideration. Fatigue, vomiting, and feeling hungry soon after eating are the symptoms of a peptic ulcer. If the disease is not diagnosed in the early stages, it can convert into cancer. One of the fatal stomach cancers is gastric cancer. It is the fourth most common cancer, as well as it is the fourth primary cause of death for cancer patients worldwide [2].
According to a survey on stomach cancer, 26% of men and 11% of women have this type of cancer globally. In men, stomach cancer is ranked second below lung cancer, and in women, it is ranked third below lung and breast cancer [3]. Generally, stomach cancer starts with an ulcer and gastritis-type complaints. It can affect the lymph nodes and peripheral organs. Only healthy food and early detection of infection can help to overcome this disease.
In 2012, more than 951,000 cases of gastric cancer were diagnosed, which was 7% of the total cancers [4]. Thereafter, more than 950,000 patients with gastric cancer are being diagnosed every year [2]. According to the American Cancer Society’s 2019 report on stomach cancer in the United States (US), about 27,510 cases of stomach cancer were diagnosed, of which 17230 were men and the remaining cases were women. About 11,140 died due to this cancer (6800 women and 4340 men). According to the World Health Organization’s 2018 report on stomach cancer in Pakistan, about 4154 cases of stomach cancer were diagnosed, and about 3923 died due to stomach cancer. This means that approximately 95% of patients with stomach cancer died. These stats are alarming. From the past until 1930, stomach cancer was the prominent cause of death in the US. Over the last ten years, stomach cancer’s decreasing percentage has been 1.5. The reasons for the decline were the use of fresh vegetables and fruits and the use of a proper antibiotic that kills helicobacter pylori (H pylori) bacteria [5]. Another reason was the early detection of stomach diseases.
Deep-learning-based techniques are playing a vital role in the production of authentic and more detailed results of stomach diseases. The results obtained from the computer vision or deep-learning-based approaches are more accurate than manually obtained results. As stomach diseases are increasing day by day, it is very important to develop a reliable system that automatically detects stomach diseases at an early stage. Image-enhanced endoscopy (IEE) is an excellent method to detect gastrointestinal neoplasm, such as narrowband imaging (NBI) with magnifying endoscopy, which is very beneficial for the detection of esophageal and pharyngeal cancers. However, IEE efficiency is not effective for detecting gastric cancer due to its low brightness. Moreover, gastric cancer is hard to detect because of the background mucosal change in gastritis [6]. In other techniques, computerized tomography (CT) scans and a special type of X-ray exam called a barium swallow are also present. Furthermore, there are also various computer-aided design (CAD) systems present which help us to diagnose different stomach diseases in the early stages using endoscopic videos or images. Endoscopy is a procedure that allows a surgeon to look inside a human’s body without making any incisions. There are two types of endoscopy. The first is simple endoscopy and the second is capsule endoscopy.
The structure of a human stomach is discussed in detail by [7] and is illustrated in Figure 1.
The early detection of diseases helps greatly to reduce the death rate. For early detection, we need a CAD system that helps us to detect healthy and unhealthy tissues accurately because the detection of diseases from the endoscopy video is time-consuming and hectic for doctors due to a large number of frames in the video. Selection of the frames containing abnormalities is also a difficult task because typically only 5% of the frames contain abnormalities [8]. Several researchers [3,6,9,10,11,12,13,14,15,16] have suggested different solutions for the automated detection of stomach diseases. However, due to the resemblance of different symptoms including color, shape, texture, etc., it is challenging to accurately classify the type of infection. Most of the previous work was conducted on the detection of a single disease/infection [1,11,16,17,18]. Accurate classification of four significant diseases (gastritis, esophagitis, peptic ulcers, and bleeding) and healthy images using a single framework is still challenging. That is why we need an automatic system that can classify healthy and unhealthy tissues with more accuracy. In this paper, we proposed a methodology for detecting and classifying four major stomach diseases. The proposed approach for stomach disease detection:
  • The proposed methodology comprises seven significant steps: data acquisition, preprocessing of data, transfer learning of deep models, feature extraction, feature selection, hybridization, and classification.
  • We selected two different CNN models (VGG19 and Alexnet) to extract features and then used transfer learning on the feature vectors before using them as feature extractors. In feature selection, heuristic GA was used to select rich information from the extracted feature vectors.
  • The fusion was performed on the selected features of both models using a serial-based approach. Finally, the best features were provided to multiple machine learning classifiers for detection and classification.

2. Related Work

Computer-aided diagnosis (CAD) plays a vital role in medical imaging, allowing accurate results to be obtained in minimal time [13,19,20,21]. Various computer techniques are being used in medical imaging for the automatic detection of diseases. Artificial intelligence (AI) and its applications such as machine learning and deep learning are the best examples [13,19,22,23].
Machine learning is a technique in which a computer takes help from data to improve its performance without explicit instructions [19]. Supervised and unsupervised learning are two types of machine learning. Machine learning demands labeled data at the start, and it takes time to tune the features [24]. On the other hand, deep learning uses neural networks that can automatically select the features by assigning weights to neurons. In 2006, deep learning came as a new and upgraded face of machine learning [25]. Deep learning is used in many fields, such as computer vision, natural language processing, bioinformatics, and medical imaging [26]. In the last few years, it changed every bit of technology. It allows the machine to recognize images and voices by identifying their patterns. The main goal is that now the machine can learn quickly without providing special attention to data [25].
In the recent past, many researchers used machine learning and deep learning to perform clinical tasks, such as analyzing a massive volume of unstructured data, processing images, identifying diseases, etc [3,12,19,23,27]. Medical imaging gained much attention from researchers, due to its importance in the world. Disease detection and its classification are two very hot topics in medical imaging. There are many segmentation and classification methods based on machine learning and computer-based techniques that are used in medical imaging. Whereas stomach disease detection is a favorite among all researchers, there are various stomach disease detection, segmentation, and classification methods that are already implemented, and every method consists of various steps, such as noise removal [28], segmentation [29], extraction of features [1,30], feature selection, feature fusion [31], and classification [1,15,27].
To improve the quality of frames/images, preprocessing is a mandatory step. Some researchers use narrow-band imaging (NBI) and chromo-endoscopy using indigo carmine spraying [2,6,10]. Interestingly, some researchers follow the pre-step before preprocessing, such as [10], who used the inclusion and exclusion criteria for images before preprocessing. Included images needed to have standard white-light and narrow-band imaging and images that needed to be magnified or had poor quality were excluded. Based on these criteria, they selected 13,584 images of gastric cancer from 69 consecutive patients with 77 gastric cancer lesions. The authors in [3] used a gradient-based technique to detect the edges of the images. This authors first changed the images from color to grayscale, then applied the derivative mask on the grayscale images to obtain the horizontal (fa) and vertical (fb) gradient separately. The formulas are shown in Equations (1) and (2).
f a   ( a ,   b ) = I   ( a + 1 ,   b )     I   ( a 1 ,   b )
f b   ( a ,   b ) = I   ( a ,   b + 1 )     I   ( a ,   b 1 )
For the segmentation of images, some researchers used region growing (RG), statistical region merging (SRM), and statistical region merging with region growing (SRMWRG) techniques [13]. In RG, the regions should have smooth boundaries and a single pixel can not belong to more than one region. The RG method produced a good result as compared to the other two. SRMWRG also showed good results, but they were not good as the RG results. On the other hand, SRM produced the worst results when compared to the other techniques.
Scholars widely use machine learning and deep learning techniques to develop models for detecting stomach diseases. The authors in [17] developed a system to detect an ulcer using two descriptors, the complete local binary pattern (CLBP) and the global local oriented edge magnitude pattern (Global LOEMP), to obtain texture features and color features. The accuracy of the proposed method (94.07% using SVM) was high compared with the other existing methods at that time. Feature extraction is an important step in computer-aided diagnosis systems. It plays a significant role in obtaining better results from the stomach disease detection process. Several researchers used different techniques for the extraction of features, such as discrete wavelet transform (DWT) [32], the complete local binary pattern (CLBP), and color Global LOEMP [17]. Feature dimension reduction is conducted by cropping the image to obtain a closer look at the infected area [33,34]. In [33], the authors reduced the dimensions of features into lower dimensions using DFT-NCA methods.
Classification of frames/images into healthy and unhealthy parts is the most critical step of a system. Computer-aided diagnosis systems use different machine learning and deep learning classifiers to classify images. There are different classifiers, such as support vector machine (SVM), decision tree (DT), convolutional neural network (CNN), k-nearest neighbors (KNN), discrete Fourier transforms (DFT), naïve Bayes(NB), artificial neural network (ANN), random forest (RF), and discriminant analysis (LDA). In [33], the authors used the naïve Bayes classifier in an extension of his previous work that produced 90.27% accurate results in their proposed systems. In [35], the authors used different classifiers such as KNN, SVM, and DT for the detection of gastric deviations. SVM outperformed the other classifiers with an accuracy of 87.2%.
The convolutional neural network (CNN) used by [6,10,12] gave 92.2%, 86.7%, and 97.25% accuracy, respectively. CNN is a deep learning model that consists of multiple layers. The deep convolutional neural network (DCNN) was used by [25] for the classification of gastric cancer images. The dataset consisted of over 3000 images and the DCNN produced excellent results, with an accuracy of 96.88%. The ANN and RF were used by [22] in their proposed model that gave 96.26% accuracy for the classification of stomach cancer images. A machine learning classifier SVM was used by multiple researchers for the classification of stomach diseases [4,17,21]. The obtained results were 96.07%, 96.3%, and 98.95%, respectively. In their proposed methodology of ulcer detection, [36] used VGG16 for features extraction and SVM for classification purposes. This combination provided 98.4% accuracy on their privately collected dataset.
BPNN is another technique that is widely used by researchers for the detection and classification of stomach diseases. The authors in [1] used BPNN for a stomach disorder and [18] used the same method for Crohn’s disease detection. They acheived 87.5% and 97.67% accuracy, respectively. In [29], the authors applied ResNet and LSTM methods to a huge dataset that consists of more than 50,000 images for multiple disease detection. ResNet and LSTM produced 97.05% accuracy, which was excellent for such a vast dataset. The previous work is discussed in detail in Table 1, below.

3. Materials and Methods

The proposed methodology consists of seven steps, such as data acquisition, preprocessing, feature extraction using VGG19 and Alexnet, transfer learning, feature selection using a genetic algorithm, feature fusion, and classification. In the first step, we collected a dataset from a medical specialist. The dataset was in rough form. We normalized our collected data by taking the input of a gastrologist. In the second step, we applied different filters to our collected dataset, now available in images. After that, we selected two convolutional neural networks, named Alexnet and VGG19, to extract features. The best thing about these models was that there was no need to develop a system from scratch, as both models were already trained on thousands of Imagenet datasets and could be classified into 1000 different classes. We performed transfer learning on these models that helped in the modification process. After transfer learning, our models were modified and ready to extract features from our dataset.
We used the genetic algorithm for feature selection purposes. As we extracted features with the two models, we needed to select features separately. One feature at a time, the genetic algorithm selected features separately from the extracted features of the two models. After that, we hybridized the selected features by the serial-based approach. We classified our images into five different classes, using machine learning classifiers in the final step.
To observe the performance of our models, we calculated classification results using machine learning classifiers after feature extraction and the feature selection process. To obtain the absolute accuracy of the proposed methodology, we provided the hybridized selected features to different machine learning classifiers. We observed that the proposed cubic SVM outclassed the other classifiers with an accuracy of 99.8% in the proposed methodology. The flow diagram of the proposed technique is shown in Figure 2.

3.1. Data Acquisition

We collected our dataset of endoscopic videos from a medical specialist in Sargodha. Initially, our dataset consisted of 118 videos of 80 patients. We selected 50 videos based on quality and converted them into frames using MATLAB [45]. After that, we labeled our dataset with the help of a medical expert and made sure the finalized dataset was in a normalized form. Our final dataset consisted of five classes. The names of the classes were gastritis, ulcer, bleeding, esophagitis, and healthy. There were 527, 519, 514, 519, and 511 images, respectively, from each class. Furthermore, we split our data into training and testing sets using a cross-validation technique with ten folds.

3.2. Data Preprocessing

Image processing plays a vital role in providing enhanced images with more information in medicine. The quality of images/frames is a significant factor in obtaining maximum accuracy. If an image/frame contains a blur or noise or is low in quality, it is necessary to improve the quality of image to obtain a better result. Firstly, we reduced the size of the images from 381 × 321 to 256 × 256. After that, we used some filters to enhance image quality. We used morphological top-hat and bottom-hat filters to binarize images for spot target detection. They calculated the morphological opening or closing of an image and then subtracted the obtained value from the original image [46]. The mathematical representations of the filters are shown in Equations (3) and (4).
F = { ( x ,   f ( x ) )   |   x P ,   P     E 2 }
B = { ( m ,   b ( m ) )   |   m S ,   S     E 2 }
Input image (F) and structuring element (B) are used above. The formulas are displayed in Equations (5) and (6).
( F B )   ( x ) = s u p   { f   ( x m ) + b ( m ) }
( F B )   ( x ) = i n f   { f   ( x + m ) b ( m ) }
These filters helped us extract the dark features, and we also obtained image details, such as edge, size, and surface.
As in medical imaging, the denoising of images was an essential part of the process. With denoising, we could obtain more accurate information than could be obtained from the original images. It was also essential to retain the important features as much as possible while denoising the image. Keeping this in mind, we used a median 3D filter to remove noise from the images.
Median filtering is a nonlinear method to denoise the images. It is generally utilized as it is highly successful at eliminating noise while protecting the edges of the original image. The median filter works by moving from one single pixel to another and taking the median value of neighboring pixels and replacing it with every pixel at the time of movement [28]. The mathematical expression is shown in Equations (7) and (8).
M e d i a n   ( X ) = X ( k + 1 ) = X ( m ) ;   N = 2 k + 1
½   { X ( k ) + X ( k + 1 ) } ;   N = 2 k
Since the median value should be the estimation of one of the pixels in the area, the median filter does not make new unreasonable pixel estimates when the filter rides an edge. Therefore, the median filter is vastly improved at safeguarding sharp edges than the mean filter.
The condition of an image after every filter was applied is represented in Figure 3.

3.3. Feature Extraction Using Alexnet and VGG19

Feature extraction is an essential step to building a model. Several features, such as color, shape, and geometry and speeded-up robust features (SURF) can be used to develop a system. In the proposed methodology, we used transferred Alexnet for feature extraction purposes. We extracted deep features, as they have the most details, and this detail is significant in the field of medical imaging [47]. These features have details, such as color, texture, edges, etc. Color features help to calculate the evidence of the infected region. Alexnet is a CNN that consists of eight layers [23]. It used RELU instead of the tanh function, which helped it to complete training much faster than a CNN using tanh. Alexnet also provided a process for data augmentation and dropout that helped the model to reduce the overfitting problem. The standard formulas of the classification layer and the RELU layer for CNN are displayed in Equations (9) and (10).
G   [ m ,   n ] = ( f × h ) [ m ,   n ] = j K h   [ j ,   k ]   f   [ m     j ,   n k ]
f ( x ) = m a x ( 0 , x ) f ( x ) = m a x ( 0 , x )
We also used VGG19 for feature extraction. VGG19 is a series of convolutional layers followed by some fully connected layers. It is a deep CNN that consists of 19 layers and was trained using the Imagenet dataset [27]. We passed the data through transferred VGG19, and all layers used three × three filters, making it small and easy to interrupt. The input layer to the last max-pooling layer model worked as a feature extractor, and the rest of the network was regarded as a classification model. After performing transfer learning, we collected our final results on an FC7 fully connected layer. The standard formulas of the batch normalization layer and the softmax layer are displayed in Equations (11) and (12).
B a t c h   M e a n = 1 m i = 1 m x
S o f t m a x   ( x ) i = e x p   ( x i )   / j = 1 n exp ( x j )

3.4. Transfer Learning

AlexNet and VGG19 demonstrated their great characterization capacity. However, their training was time-consuming. Moreover, our dataset was not large enough to be suitable to train such a huge volume of deep networks [27]. We used transfer learning to solve this problem. We changed the last three layers of AlexNet with the following layers: a fully connected layer with five nodes, a softmax layer, and a classification layer. Furthermore, we divided our structure into two portions: the first is the pre-trained network, and the second is the transferred network. The parameters in the pre-trained network were prepared on ImageNet with a large dataset and proved effective, so these parameters needed a minor modification. The parameters used by the transferred network were a tiny part of the whole network, which is good for our small dataset compared to the vast dataset. The activation was applied to the seventh FC layer to extract features. The size of the layer that we obtained as a resultant vector of FC7 was 1 × 4096.
The transfer learning of networks is expressed below in Figure 4.

3.5. Feature Selection using a Genetic Algorithm

Feature selection is an essential step in deep learning models. The accuracy of a classifier is needed to select the relevant features, and it requires much computational work. We used a genetic algorithm for feature selection, to keep these points in mind. It operated on the set of individuals, which was called population. Each generation created a new population by checking their level of fitness. We separately selected features for both models (VGG19 and Alexnet) from their extracted vectors in the proposed methodology. We used 50 chromosomes, the number of generations was 100, the crossover rate was 0.8, and the mutation rate was 0.01.
A state diagram for the selection of features with the genetic algorithm is shown in Figure 5.

3.6. Hybridization and fusion of features

Hybridization and fusion features are the processes where we integrated multiple features to obtain prominent features. This process also improved the discriminative power. We used the serial-based approach for this purpose. It was different from other fusion techniques as it used a complex vector rather than a super vector to combine two different features.

3.7. Classification

Classification is the final stage for a model where the goal is to predict the label [1]. In the proposed methodology we used the multi-class classification technique to classify the input image, based on the selected features. There are many classifiers that we used in the proposed methodology, such as decision tree, naïve Bayes, k-nearest neighbor, cubic support vector machine, bagged tree, cosine k-nearest neighbor, fine tree, and coarse tree. The proposed technique produced the best result for Cubic SVM on our dataset.

4. Results

4.1. Performance Matrix

Stomach disease detection and classification can be assessed with the help of different performance metrics. The disease classification performance measure is highly dependent on the detection rate, the ratio between the total number of pixels and infected pixels. The detection rate is also known as the probability of detection. There are some important terms, such as recall, precision, AUC, positive predictive value (PPV), negative predictive value (NPV), false positive rate (FPR), false discovery rate (FDR), and accuracy, which wereused by several researchers to analyze the stomach disease classification results [13,20,48]. There are some performance metrics given below in Table 2.

4.2. Feature Extraction Accuracy

In this section, we discuss the results of feature extraction using both the VGG19 and Alexnet models. We applied different machine learning classifiers, named cubic SVM, quadratic SVM, linear SVM, fine KNN, cosine KNN, fine tree, bagged tree, coarse tree, and naïve Bayes, to obtain the accuracy. We observed that cubic SVM outclassed all other classifiers, showing 99.9% accuracy, followed by fine KNN and cosine KNN with accuracies of 99.8% and 98.2%, respectively. The classification results that we obtained using VGG19 model features are displayed in Table 3.
In addition, we used Alexnet for feature extraction purposes. We applied different machine learning classifiers, named cubic SVM, quadratic SVM, linear SVM, fine KNN, cosine KNN, fine tree, bagged tree, coarse tree, and naïve Bayes, to obtain the accuracy. We observed that fine KNN showed the best result, with 99.9% accuracy, as compared to all other classifiers, followed by cubic SVM and cosine KNN, with accuracies of 99.8% and 98.2%, respectively. The precision and f1 score of fine KNN was 99.8%. Moreover, we compared the results using recall, precision, f1 score, etc. as we mentioned earlier. The results of the classification using Alexnet model features are displayed in Table 4.

4.3. Feature Selection Accuracy

We used a genetic algorithm for feature selection. Firstly, we applied the genetic algorithm on the extracted features of Alexnet, and after that, we applied the genetic algorithm again on the extracted features of the VGG19 model separately. We discussed the results of both models here. We applied different machine learning classifiers, named cubic SVM, quadratic SVM, linear SVM, fine KNN, cosine KNN, fine tree, bagged tree, coarse tree, and naïve Bayes, to obtain the accuracy.
We observed that fine KNN outclassed all other classifiers, showing 99.9% accuracy, followed by cubic SVM and cosine KNN, with an accuracy of 99.8% and 98.2%, respectively. Moreover, we compared the results using recall, precision, and f1 score, as we mentioned earlier. The recall and f1 score of fine KNN was 99.8%. The classification results of the genetic algorithm on Alexnet extracted features are given in Table 5.
We utilized Alexnet and VGG19 to extract features. We separately applied GA on both models to select the features from the extracted features of those two models. Here we will discuss the classification results using the GA for VGG19 model features. We observed that fine KNN outclassed all other classifiers, showing 99.8% accuracy, followed by cubic SVM and quadratic SVM with 99.7% and 99.7%, respectively. The recall and f1 scores of fine KNN were 99.8% and 99.69%, respectively. The performance of the fine tree was comparatively low compared to the other classifiers with an accuracy of 87.7%, and the recall and f1 scores were 87.4%. The results of classification using the GA for VGG19 model features are displayed in Table 6.

4.4. Classification Accuracy

We applied different machine learning classifiers and compared them to obtain better accuracy. The machine learning classifiers were cubic SVM, quadratic SVM, linear SVM, fine KNN, cosine KNN, fine tree, bagged tree, coarse tree, and naïve Bayes. After hybridizing our features, we gave those features to the classifiers mentioned above to obtain the best final accuracy of our model. We observed that cubic SVM outclassed all other classifiers, with 99.8% accuracy, followed by bagged tree and linear SVM, with 98.8% and 98.6%, respectively. The precision obtained using cubic SVM was 99.8%, which was 0.54% higher than the second-best classifier. The accuracy of the coarse tree was 75.9%, and coarse KNN, at 90.7%, was comparatively low compared to the other classifiers. The classification results of fused features are displayed in Table 7.
We analyzed that cubic SVM provided 100% results for every class (gastritis, ulcer, bleeding, esophagitis, and healthy) except for bleeding. In bleeding image classification, our classifiers produced 99.5% true positive results, and only 0.5% of the results were false positives. This means that 99.5% of our classifiers detected the disease correctly while only 0.5% of our classifiers produced false-positive results when there was no disease.
The confusion matrix of cubic SVM on fused features is displayed in Figure 6.
The applied classifier Cubic SVM produced 100% results for all classes (gastritis, ulcer, bleeding, esophagitis, and healthy) when no diseases were found except for gastritis. In gastritis images, the confusion matrix shows that 99.5% was the accuracy of detection of no disease; only 0.5% of the results were false negatives. This means that our classifier wrongly predicted only 0.5% of the results when there was a disease, but our classifier judged that as no disease. So, as a result, our accuracy fell from 100% to 99.5%.
The confusion matrix of cubic SVM using fused features is displayed in Figure 7.

5. Discussion

A detailed discussion of the proposed methodology is described in this section. There are five major steps involved in our work, as shown in Figure 2.
The proposed methodology consists of seven steps, such as data acquisition, preprocessing, feature extraction using VGG19 and Alexnet, transfer learning, feature selection using a genetic algorithm, feature fusion, and classification. In the first step, we collected a dataset from a medical specialist. The dataset was in rough form. We normalized our collected data by taking the input of a gastrologist. In the second step, we applied different filters to our collected dataset, which was now available in images. After that, we selected two convolutional neural networks, named Alexnet and VGG19, to extract features. The best thing about these models was that there was no need to develop a system from scratch, as both models were already trained on thousands of imagenet datasets and could classify into 1000 different classes. We performed transfer learning on these models that helped in the modification process. After transfer learning, our models were modified and were ready to extract the features of our dataset.
We used the genetic algorithm for feature selection purposes. As we extracted features with two models, we needed to select features separately. One at a time, the genetic algorithm selected features separately from the extracted features of the two models. After that, we hybridized the selected features by the serial-based approach. We classified our images into five different classes using machine learning classifiers in the final step.
To observe the performance of our models, we calculated classification results using machine learning classifiers after feature extraction and the feature selection process. To obtain the final accuracy of the proposed methodology, we provided the hybridized selected features to different machine learning classifiers. The analysis of these classifiers was conducted through well-known performance metrics, including sensitivity, precision, accuracy, f1 score, true-negative rate (TNR), true-positive rate (TPR), false-negative rates (FNR), and false-positive rates (FPR). We observed that the proposed cubic SVM outclassed the other classifiers, with an accuracy of 99.8%.
We also provided the same dataset that we used in the proposed methodology to the pre-trained Alexnet and VGG19 models for the sake of comparison. We observed that the Alexnet model produced better results than the VGG19 model for our dataset. The obtained accuracies on our dataset were 87.7% for Alexnet and 86.4 % for VGG19. The results were relatively low compared to the proposed hybrid approach that produced 99.8% results on the same dataset.
In the end, we compared our results with the existing state-of-the-art techniques, in terms of accuracies. We concluded that the proposed hybrid technique helped improve the results. The comparison is shown in Table 8.

6. Conclusions and Future Work

Manual classification of endoscopic lesions in the stomach is a challenging task, and there is a need to have CAD for efficient and accurate results. This paper proposed a deep-learning-based model following the feature fusion framework to classify four different stomach diseases. We selected two different CNN models (VGG19 and Alexnet) to extract features and then used transfer learning on the feature vectors before using them as feature extractors. In feature selection, a heuristic GA selected rich information from the extracted feature vectors. The hybridization helped to improve the accuracy of the proposed model, due to strong predictor values. We obtained 99.8% accuracy using the cubic SVM classifier on the given dataset. The proposed framework selected only 50% of the best features, which is a limiting aspect and can be tackled in future work. Furthermore, a study can be performed compare the histopathological aspect of the lesion after numerous comparisons of the processed images with the microscopic images. This method could be credible for medical practice.

Author Contributions

Conceptualization, M.S.A., M.I.U.L., M.H. and H.T.R.; methodology, M.S.A., M.I.U.L., M.H., H.T.R., B.A., H.A. and S.W.; software, M.S.A., M.I.U.L., M.H. and H.T.R.; writing—original draft preparation, M.S.A., M.I.U.L., M.H., H.T.R., B.A., H.A. and S.W.; supervision, M.I.U.L., M.H. and H.T.R.; funding acquisition, B.A. and H.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Taif University Researchers Supporting Project number (TURSP-2020/306), Taif University, Taif, Saudi Arabia.

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Not Applicable.

Data Availability Statement

Not Applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dewi, A.K.; Novianty, A.; Purboyo, T.W. Stomach disorder detection through the Iris Image using Backpropagation Neural Network. In Proceedings of the 2016 International Conference on Informatics and Computing (ICIC), Mataram, Indonesia, 28–29 October 2016; pp. 192–197. [Google Scholar] [CrossRef]
  2. Zhao, A.-J.; Qian, Y.-Y.; Sun, H.; Hou, X.; Pan, J.; Liu, X.; Zhou, W.; Chen, Y.-Z.; Jiang, X.; Li, Z.-S.; et al. Screening for gastric cancer with magnetically controlled capsule gastroscopy in asymptomatic individuals. Gastrointest. Endosc. 2018, 88, 466–474.e1. [Google Scholar] [CrossRef]
  3. Korkmaz, M.F. Artificial Neural Network by using HOG Features HOG_LDA_ANN. 2017; 327–332. [Google Scholar]
  4. Li, C.; Shi, C.; Zhang, H.; Chen, Y.; Zhang, S. Multiple instance learning for computer aided detection and diagnosis of gastric cancer with dual-energy CT imaging. J. Biomed. Inform. 2015, 57, 358–368. [Google Scholar] [CrossRef] [Green Version]
  5. Balakrishnan, M.; George, R.; Sharma, A.; Graham, D.Y. Changing Trends in Stomach Cancer Throughout the World. Curr. Gastroenterol. Rep. 2017, 19, 1–10. [Google Scholar] [CrossRef]
  6. Sakai, Y.; Takemoto, S.; Hori, K.; Nishimura, M.; Ikematsu, H.; Yano, T.; Yokota, H. Automatic detection of early gastric cancer in endoscopic images using a transferring convolutional neural network. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; pp. 4138–4141. [Google Scholar]
  7. Brandstaeter, S.; Fuchs, S.; Aydin, R.C.; Cyron, C.J. Mechanics of the stomach: A review of an emerging field of biomechanics. GAMM-Mitteilungen 2019, 42, 1–17. [Google Scholar] [CrossRef] [Green Version]
  8. Nawarathna, R.; Oh, J.; Muthukudage, J.; Tavanapong, W.; Wong, J.; de Groen, P.C.; Tang, S.J. Abnormal image detection in endoscopy videos using a filter bank and local binary patterns. Neurocomputing 2014, 144, 70–91. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Lee, T.-C.; Lin, Y.-H.; Uedo, N.; Wang, H.-P.; Chang, H.-T.; Hung, C.-W. Computer-aided diagnosis in endoscopy: A novel application toward automatic detection of abnormal lesions on magnifying narrow-band imaging endoscopy in the stomach. In Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Osaka, Japan, 3–7 July 2013; pp. 4430–4433. [Google Scholar] [CrossRef]
  10. Hirasawa, T.; Aoyama, K.; Tanimoto, T.; Ishihara, S.; Shichijo, S.; Ozawa, T.; Ohnishi, T.; Fujishiro, M.; Matsuo, K.; Fujisaki, J.; et al. Application of artificial intelligence using a convolutional neural network for detecting gastric cancer in endoscopic images. Gastric Cancer 2018, 21, 653–660. [Google Scholar] [CrossRef] [Green Version]
  11. Korkmaz, S.A.; Akcicek, A.; Binol, H.; Korkmaz, M.F. Recognition of the stomach cancer images with probabilistic HOG feature vector histograms by using HOG features. In Proceedings of the 2017 IEEE 15th International Symposium on Intelligent Systems and Informatics (SISY), Subotica, Serbia, 14–16 September 2017; pp. 000339–000342. [Google Scholar]
  12. Yu, J.-S.; Chen, J.; Xiang, Z.; Zou, Y.-X. A hybrid convolutional neural networks with extreme learning machine for WCE image classification. In Proceedings of the 2015 IEEE International Conference on Robotics and Biomimetics (ROBIO), Zhuhai, China, 6–9 December 2015; pp. 1822–1827. [Google Scholar]
  13. Yasar, A.; Saritas, I.; Korkmaz, H. Computer-Aided Diagnosis System for Detection of Stomach Cancer with Image Processing Techniques. J. Med. Syst. 2019, 43, 99. [Google Scholar] [CrossRef] [PubMed]
  14. Li, L.; Chen, Y.; Shen, Z.; Zhang, X.; Sang, J.; Ding, Y.; Yang, X.; Li, J.; Chen, M.; Jin, C.; et al. Convolutional neural network for the diagnosis of early gastric cancer based on magnifying narrow band imaging. Gastric Cancer 2020, 23, 126–132. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Maghsoudi, O.H.; Alizadeh, M.; Mirmomen, M. A computer aided method to detect bleeding, tumor, and disease regions in Wireless Capsule Endoscopy. In Proceedings of the 2016 IEEE Signal Processing in Medicine and Biology Symposium (SPMB), Philadelphia, PA, USA, 3 December 2016; pp. 1–6. [Google Scholar] [CrossRef]
  16. de Groen, P.C. Using artificial intelligence to improve adequacy of inspection in gastrointestinal endoscopy. Tech. Innov. Gastrointest. Endosc. 2020, 22, 71–79. [Google Scholar] [CrossRef]
  17. Charfi, S.; El Ansari, M. Computer-aided diagnosis system for ulcer detection in wireless capsule endoscopy videos. In Proceedings of the 2017 International Conference on Advanced Technologies for Signal and Image Processing (ATSIP), Fez, Morocco, 22–24 May 2017; pp. 1–5. [Google Scholar]
  18. Ahmed, S.S.; Dey, N.; Ashour, A.S.; Sifaki-Pistolla, D.; Bălas-Timar, D.; Balas, V.E.; Tavares, J.M.R.S. Effect of fuzzy partitioning in Crohn’s disease classification: A neuro-fuzzy-based approach. Med. Biol. Eng. Comput. 2017, 55, 101–115. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Alagappan, M.; Brown, J.R.G.; Mori, Y.; Berzin, T.M. Artificial intelligence in gastrointestinal endoscopy: The future is almost here. World J. Gastrointest. Endosc. 2018, 10, 239–249. [Google Scholar] [CrossRef] [PubMed]
  20. Kanesaka, T.; Lee, T.-C.; Uedo, N.; Lin, K.-P.; Chen, H.-Z.; Lee, J.-Y.; Wang, H.-P.; Chang, H.-T. Computer-aided diagnosis for identifying and delineating early gastric cancers in magnifying narrow-band imaging. Gastrointest. Endosc. 2018, 87, 1339–1344. [Google Scholar] [CrossRef]
  21. Hassan, A.R.; Haque, M.A. Computer-aided gastrointestinal hemorrhage detection in wireless capsule endoscopy videos. Comput. Methods Programs Biomed. 2015, 122, 341–353. [Google Scholar] [CrossRef]
  22. Korkmaz, S.A.; Binol, H. Classification of molecular structure images by using ANN, RF, LBP, HOG, and size reduction methods for early stomach cancer detection. J. Mol. Struct. 2018, 1156, 255–263. [Google Scholar] [CrossRef]
  23. Lu, S.; Lu, Z.; Zhang, Y.-D. Pathological brain detection based on AlexNet and transfer learning. J. Comput. Sci. 2019, 30, 41–47. [Google Scholar] [CrossRef]
  24. Liu, C.; Qi, L.; Feng, Q.-X.; Sun, S.-W.; Zhang, Y.-D.; Liu, X.-S. Performance of a machine learning-based decision model to help clinicians decide the extent of lymphadenectomy (D1 vs. D2) in gastric cancer before surgical resection. Abdom. Radiol. 2019, 44, 3019–3029. [Google Scholar] [CrossRef]
  25. Garcia, E.; Hermoza, R.; Castanon, C.B.; Cano, L.; Castillo, M.; Castanneda, C. Automatic Lymphocyte Detection on Gastric Cancer IHC Images Using Deep Learning. In Proceedings of the 2017 IEEE 30th International Symposium on Computer-Based Medical Systems (CBMS), Thessaloniki, Greece, 22–24 June 2017; pp. 200–204. [Google Scholar]
  26. Sharma, H.; Zerbe, N.; Klempert, I.; Hellwich, O.; Hufnagl, P. Deep convolutional neural networks for automatic classification of gastric carcinoma using whole slide images in digital histopathology. Comput. Med. Imaging Graph. 2017, 61, 2–13. [Google Scholar] [CrossRef]
  27. Xiao, J.; Wang, J.; Cao, S.; Li, B. Application of a Novel and Improved VGG-19 Network in the Detection of Workers Wearing Masks. J. Phys. Conf. Ser. 2020, 1518, 012041. [Google Scholar] [CrossRef]
  28. Das, S.; Mohan, A. Medical Image Enhancement Techniques by Bottom Hat and Median Filtering. Int. J. Electron. Commun. Comput. Eng. 2014, 5, 347–351. [Google Scholar]
  29. Owais, M.; Arsalan, M.; Choi, J.; Mahmood, T.; Park, K.R. Artificial Intelligence-Based Classification of Multiple Gastrointestinal Diseases Using Endoscopy Videos for Clinical Diagnosis. J. Clin. Med. 2019, 8, 986. [Google Scholar] [CrossRef] [Green Version]
  30. Muto, M.; Yao, K.; Kaise, M.; Kato, M.; Uedo, N.; Yagi, K.; Tajiri, H. Magnifying endoscopy simple diagnostic algorithm for early gastric cancer (MESDA-G). Dig. Endosc. 2016, 28, 379–393. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Jiang, Y.-G.; Wu, Z.; Tang, J.; Li, Z.; Xue, X.; Chang, S.-F. Modeling Multimodal Clues in a Hybrid Deep Learning Framework for Video Classification. IEEE Trans. Multimed. 2018, 20, 3137–3147. [Google Scholar] [CrossRef] [Green Version]
  32. Charfi, S.; El Ansari, M. Computer-aided diagnosis system for colon abnormalities detection in wireless capsule endoscopy images. Multimed. Tools Appl. 2018, 77, 4047–4064. [Google Scholar] [CrossRef]
  33. Korkmaz, S.A.; Esmeray, F. A New Application Based on GPLVM, LMNN, and NCA for Early Detection of the Stomach Cancer. Appl. Artif. Intell. 2018, 32, 1–17. [Google Scholar] [CrossRef]
  34. Sumiyama, K. Past and current trends in endoscopic diagnosis for early stage gastric cancer in Japan. Gastric Cancer 2016, 20, 20–27. [Google Scholar] [CrossRef]
  35. Ali, H.; Yasmin, M.; Sharif, M.; Rehmani, M.H. Computer assisted gastric abnormalities detection using hybrid texture descriptors for chromoendoscopy images. Comput. Methods Programs Biomed. 2018, 157, 39–47. [Google Scholar] [CrossRef] [PubMed]
  36. Khan, M.A.; Kadry, S.; Alhaisoni, M.; Nam, Y.; Zhang, Y.-D.; Rajinikanth, V.; Sarfaraz, M.S. Computer-Aided Gastrointestinal Diseases Analysis From Wireless Capsule Endoscopy: A Framework of Best Features Selection. IEEE Access 2020, 8, 132850–132859. [Google Scholar] [CrossRef]
  37. Caroppo, A.; Leone, A.; Siciliano, P. Deep transfer learning approaches for bleeding detection in endoscopy images. Comput. Med. Imaging Graph. 2021, 88, 101852. [Google Scholar] [CrossRef]
  38. Park, J.; Jang, B.G.; Kim, Y.W.; Park, H.; Kim, B.-H.; Kim, M.J.; Ko, H.; Gwak, J.M.; Lee, E.J.; Chung, Y.R.; et al. A Prospective Validation and Observer Performance Study of a Deep Learning Algorithm for Pathologic Diagnosis of Gastric Tumors in Endoscopic Biopsies. Clin. Cancer Res. 2021, 27, 719–728. [Google Scholar] [CrossRef]
  39. Klang, E.; Barash, Y.; Margalit, R.Y.; Soffer, S.; Shimon, O.; Albshesh, A.; Ben-Horin, S.; Amitai, M.M.; Eliakim, R.; Kopylov, U. Deep learning algorithms for automated detection of Crohn’s disease ulcers by video capsule endoscopy. Gastrointest. Endosc. 2020, 91, 606–613.e2. [Google Scholar] [CrossRef]
  40. Gong, J.; Ma, D.; Zhang, Y.; Jiang, J.; Pei, M. Image Classification to Diagnose Chronic Atrophic Gastritis Based on Squeeze-and-Excitation Block. In Proceedings of the 2020 IEEE 5th International Conference on Cloud Computing and Big Data Analytics (ICCCBDA), Chengdu, China, 10–13 April 2020; pp. 336–340. [Google Scholar] [CrossRef]
  41. Ma, B.; Guo, Y.; Hu, W.; Yuan, F.; Zhu, Z.; Yu, Y.; Zou, H. Artificial Intelligence-Based Multiclass Classification of Benign or Malignant Mucosal Lesions of the Stomach. Front. Pharmacol. 2020, 11, 572372. [Google Scholar] [CrossRef] [PubMed]
  42. Luo, S.; Fan, Y.; Chang, W.; Liao, H.; Kang, H.; Huo, L. Classification of human stomach cancer using morphological feature analysis from optical coherence tomography images. Laser Phys. Lett. 2019, 16, 095602. [Google Scholar] [CrossRef]
  43. Kim, J.H.; Nam, H.S.; Choi, C.W.; Kang, D.H.; Kim, H.W.; Park, S.; Kim, S.J.; Hwang, S.H.; Lee, S.H. Risk factors associated with difficult gastric endoscopic submucosal dissection: Predicting difficult ESD. Surg. Endosc. 2017, 31, 1617–1626. [Google Scholar] [CrossRef]
  44. Horiuchi, Y.; Fujisaki, J.; Yamamoto, N.; Shimizu, T.; Miyamoto, Y.; Tomida, H.; Omae, M.; Ishiyama, A.; Yoshio, T.; Hirasawa, T.; et al. Accuracy of diagnostic demarcation of undifferentiated-type early gastric cancers for magnifying endoscopy with narrow-band imaging: Endoscopic submucosal dissection cases. Gastric Cancer 2015, 19, 515–523. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. MATLAB, version 9.4 (R2018a); The MathWorks Inc.: Natick, MA, USA, 2018.
  46. Zeng, M.; Li, J.; Peng, Z. The design of Top-Hat morphological filter and application to infrared target detection. Infrared Phys. Technol. 2006, 48, 67–76. [Google Scholar] [CrossRef]
  47. Liaqat, A.; Khan, M.A.; Shah, J.H.; Sharif, M.; Yasmin, M.; Fernandes, S.L. Automated ulcer and bleeding classification from WCE images using multiple features fusion and selection. J. Mech. Med. Biol. 2018, 18, 1–25. [Google Scholar] [CrossRef]
  48. Khan, M.A.; Khan, M.A.; Ahmed, F.; Mittal, M.; Goyal, L.M.; Hemanth, D.J.; Satapathy, S.C. Gastrointestinal diseases segmentation and classification based on duo-deep architectures. Pattern Recognit. Lett. 2020, 131, 193–204. [Google Scholar] [CrossRef]
Figure 1. Structure of the human stomach and gastric wall [7].
Figure 1. Structure of the human stomach and gastric wall [7].
Diagnostics 12 00043 g001
Figure 2. The proposed methodology for the detection and classification of stomach diseases.
Figure 2. The proposed methodology for the detection and classification of stomach diseases.
Diagnostics 12 00043 g002
Figure 3. The results of preprocessing after the use of filters.
Figure 3. The results of preprocessing after the use of filters.
Diagnostics 12 00043 g003
Figure 4. Transfer learning of networks.
Figure 4. Transfer learning of networks.
Diagnostics 12 00043 g004
Figure 5. A diagram to show the process of feature selection using a genetic algorithm.
Figure 5. A diagram to show the process of feature selection using a genetic algorithm.
Diagnostics 12 00043 g005
Figure 6. Confusion matrix (true positive rates) of cubic SVM using fused features.
Figure 6. Confusion matrix (true positive rates) of cubic SVM using fused features.
Diagnostics 12 00043 g006
Figure 7. Confusion matrix (false discovery rates) of cubic SVM using fused features.
Figure 7. Confusion matrix (false discovery rates) of cubic SVM using fused features.
Diagnostics 12 00043 g007
Table 1. Existing studies on stomach disease detection and classification.
Table 1. Existing studies on stomach disease detection and classification.
Techniques/MethodsDiseaseDatasetAccuracyReference
Inception V3 and SVMBleeding2352 Images97.71%[37]
CNNGastric Cancer2434 Images95%[38]
VGG16 and SVMUlcer and Bleeding6000 Images98.4%[36]
CNNGastric Cancer2088 Images90.91%[14]
CNNUlcer17640 Images96.7%[39]
CNNGastritis5470 Images88.15%[40]
DCNNGastric Cancer763 Images96%[41]
ResNet and LSTMUlcer and Crohn’s52,471 Frames97.05%[29]
SVM and RFStomach cancer3106 Images96.36%[42]
CNNGastric Cancer13,584 Images92.2%[10]
CNNGastric Cancer1000 Images87.6%[6]
DFT and NBStomach Cancer900 Images90.27%[33]
SVM and MLPColon Abnormalities1670 Images96.5%[32]
CNNGastric Cancer3257 Images96.88%[25]
ANN, RF, LBP, and HOGStomach Cancer180 Images96.29%[22]
SVMGastric Cancer207 Images96.3%[20]
ANNStomach Cancer270 Images88.9%[11]
SVM and MLPUlcer2333 Images94.07%[17]
BPNNStomach Disorder40 Images87.5%[1]
ESDGastric Neoplasms1052 Patients93.3%[43]
BPNNCrohn’s Disease387 Patients97.67%[18]
ME-NBIGastric Cancer76 Patients81.6%[44]
SVMGastrointestinal Hemorrhage2920 Images98.95%[21]
CNN and ELMDigestion Disease25 Examinations97.25%[12]
Table 2. Performance matrices.
Table 2. Performance matrices.
Sr. NoPerformance MatrixFormulaDescription
1. Recall Re = Pt Pt + Nf Where,
Pt = True Positives
Pf = False positives
Nf = False Negatives
Nt = True Negatives
2. Precision Pr = Nt Nt + Pf
3. PPV PPV = Pt Pt + Pf
4. NPV NPV = Nt Nt + Nf
5. FPR FPR = Pf Pf + Nt
6. FDR FDR = Pf Pf + Pt
Table 3. Classification results using VGG19 model features.
Table 3. Classification results using VGG19 model features.
ClassifierRecall (%)Precision (%)F1 Score (%)FPRAUCAccuracy (%)FNR (%)
Fine Tree88.688.888.70.0280.94688.711.4
Cubic SVM99.899.899.80199.90.2
Fine KNN99.699.699.60199.80.4
Cosine KNN98.298.298.20.006198.21.8
Bagged Tree97.897.297.40.006197.22.2
Linear SVM97.496.696.90.01196.62.6
Coarse Tree7363.667.90.0680.9173.127
Table 4. Classification results using Alexnet model features.
Table 4. Classification results using Alexnet model features.
ClassifierRecall (%)Precision (%)F1 Score (%)FPRAUCAccuracy (%)FNR (%)
Fine Tree92.492.892.590.00160.96492.77.6
Cubic SVM99.699.699.60199.80.4
Fine KNN99.899.899.80199.90.2
Cosine KNN98.29898.10.004198.21.8
Bagged Tree9898980.002198.12
Linear SVM98.297.297.690.006197.21.8
Coarse KNN86.486.686.50.0340.96486.413.6
Table 5. Classification results using GA for Alexnet model features.
Table 5. Classification results using GA for Alexnet model features.
ClassifierRecall (%)Precision (%)F1 Score (%)FPRAUCAccuracy (%)FNR (%)
Fine Tree90.89190.890.0220.95490.99.2
Cubic SVM99.699.699.800199.80.4
Fine KNN99.899.899.80199.90.2
Cosine KNN98.49897.600.004198.21.6
Bagged Tree97.497.497.560.006197.32.6
Linear SVM96.696.899.690.008196.73.4
Quadratic SVM99.699.699.80199.80.4
Table 6. Classification results using the GA for VGG19 model features.
Table 6. Classification results using the GA for VGG19 model features.
ClassifierRecall (%)Precision (%)F1 Score (%)FPRAUCAccuracy (%)FNR (%)
Fine Tree87.487.687.490.030.74287.712.6
Cubic SVM99.499.699.490199.70.6
Fine KNN99.899.699.6900.9999.80.2
Cosine KNN98.698.498.490.004198.51.4
Bagged Tree97.297.297.20.004197.32.8
Linear SVM97.697.499.490.006197.62.4
Quadratic SVM99.699.699.60199.70.4
Table 7. Classification results of fused features.
Table 7. Classification results of fused features.
ClassifierRecall (%)Precision (%)F1 Score (%)FPRAUC Accuracy (%)FNR (%)
Fine Tree90.3390.4390.340.0240.96290.39.67
Cosine KNN99.2699.2699.260.002199.30.74
Bagged Tree98.8498.8498.840.004198.81.16
Linear SVM98.6298.6498.630.004198.61.38
Coarse Tree75.8479.7477.740.060.91475.924.16
Cubic SVM99.899.899.80199.80.2
Naïve Bayes96.1696.2496.190.0080.97696.23.84
Coarse KNN90.7491.691.160.0240.9890.79.26
Table 8. Comparison of the proposed methodology with the state-of-the-art approaches.
Table 8. Comparison of the proposed methodology with the state-of-the-art approaches.
Author/YearTechniques/MethodsDiseaseDatasetResults
[39]CNNUlcer17,640 Images96.7%
[41]DCNNGastric Cancer763 Images96%
[36]VGG16 and SVMUlcer6000 Images98.4%
[37]Inception V3 and SVMBleeding2352 Images97.71%
Proposed MethodologyVGG19, Alexnet, and Cubic SVMUlcer, Bleeding, Esophagitis, and Gastritis2600 Images99.8%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ayyaz, M.S.; Lali, M.I.U.; Hussain, M.; Rauf, H.T.; Alouffi, B.; Alyami, H.; Wasti, S. Hybrid Deep Learning Model for Endoscopic Lesion Detection and Classification Using Endoscopy Videos. Diagnostics 2022, 12, 43. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics12010043

AMA Style

Ayyaz MS, Lali MIU, Hussain M, Rauf HT, Alouffi B, Alyami H, Wasti S. Hybrid Deep Learning Model for Endoscopic Lesion Detection and Classification Using Endoscopy Videos. Diagnostics. 2022; 12(1):43. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics12010043

Chicago/Turabian Style

Ayyaz, M Shahbaz, Muhammad Ikram Ullah Lali, Mubbashar Hussain, Hafiz Tayyab Rauf, Bader Alouffi, Hashem Alyami, and Shahbaz Wasti. 2022. "Hybrid Deep Learning Model for Endoscopic Lesion Detection and Classification Using Endoscopy Videos" Diagnostics 12, no. 1: 43. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics12010043

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop