Next Article in Journal
Traffic Vibration Signal Analysis of DAS Fiber Optic Cables with Different Coupling Based on an Improved Wavelet Thresholding Method
Previous Article in Journal
Monitoring System for Operating Variables in Incubators in the Neonatology Service of a Highly Complex Hospital through the Internet of Things (IoT)
Previous Article in Special Issue
Saliency Map and Deep Learning in Binary Classification of Brain Tumours
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Deep Learning Architectures for Detection and Classification of Diabetic Retinopathy

by
Cheena Mohanty
1,
Sakuntala Mahapatra
2,*,
Biswaranjan Acharya
3,*,
Fotis Kokkoras
4,
Vassilis C. Gerogiannis
4,
Ioannis Karamitsos
5 and
Andreas Kanavos
6
1
Department of Electronics and Telecommunication, Biju Patnaik University of Technology, Rourkela 769012, Odisha, India
2
Department of Electronics and Telecommunication Engineering, Trident Academy of Technology, Bhubaneswar 751016, Odisha, India
3
Department of Computer Engineering-AI, Marwadi University, Rajkot 360003, Gujarat, India
4
Department of Digital Systems, University of Thessaly, 41500 Larissa, Greece
5
Department of Graduate and Research, Rochester Institute of Technology, Dubai 341055, United Arab Emirates
6
Department of Informatics, Ionian University, 49100 Corfu, Greece
*
Authors to whom correspondence should be addressed.
Submission received: 21 April 2023 / Revised: 7 June 2023 / Accepted: 14 June 2023 / Published: 19 June 2023
(This article belongs to the Special Issue Machine and Deep Learning in Sensing and Imaging)

Abstract

:
Diabetic retinopathy (DR) is a common complication of long-term diabetes, affecting the human eye and potentially leading to permanent blindness. The early detection of DR is crucial for effective treatment, as symptoms often manifest in later stages. The manual grading of retinal images is time-consuming, prone to errors, and lacks patient-friendliness. In this study, we propose two deep learning (DL) architectures, a hybrid network combining VGG16 and XGBoost Classifier, and the DenseNet 121 network, for DR detection and classification. To evaluate the two DL models, we preprocessed a collection of retinal images obtained from the APTOS 2019 Blindness Detection Kaggle Dataset. This dataset exhibits an imbalanced image class distribution, which we addressed through appropriate balancing techniques. The performance of the considered models was assessed in terms of accuracy. The results showed that the hybrid network achieved an accuracy of 79.50%, while the DenseNet 121 model achieved an accuracy of 97.30%. Furthermore, a comparative analysis with existing methods utilizing the same dataset revealed the superior performance of the DenseNet 121 network. The findings of this study demonstrate the potential of DL architectures for the early detection and classification of DR. The superior performance of the DenseNet 121 model highlights its effectiveness in this domain. The implementation of such automated methods can significantly improve the efficiency and accuracy of DR diagnosis, benefiting both healthcare providers and patients.

1. Introduction

The early detection and diagnosis of diabetes-related diseases are crucial in any healthcare system, as they enable improved treatment and care. Diabetes often coexists with various comorbidities in many patients. It is a metabolic disease characterized by elevated blood sugar levels; and, over time, it affects multiple organs in the human body [1]. In 2019, an estimated 463 million people worldwide were affected by diabetes [2]. Developing countries, such as China and India, are currently experiencing a diabetes epidemic [3]. According to the International Diabetes Federation [4], China, India, and the USA were the top three nations with the highest rates of diabetes in 2019; this trend is projected to continue from 2030 to 2045, with China and India bearing the greatest burden of the disease. Additionally, diabetes appears to be more prevalent in men (9.0%) than in women (7.9%) [5].
As previously mentioned, diabetes has a significant impact on various organs of the body, including the eyes. The ocular manifestation of diabetes is known as diabetic retinopathy (DR). DR primarily affects the eye retina, which is responsible for capturing light and transmitting signals to the brain. Like other parts of the body, the retina is also supplied with blood vessels. The elevated blood glucose levels in diabetes can lead to damage and alteration of the retinal blood vessels, resulting in fluid leakage [6]. Consequently, these pathological changes can lead to vision loss and, in some cases, permanent visual impairment [7]. Given these implications, DR is a highly significant and serious disease.
According to statistics [4], out of the 463 million individuals diagnosed with diabetes worldwide, approximately 30% also experience DR. This disease demonstrates a progressive nature, highlighting the importance of timely diagnosis and treatment to prevent permanent vision loss [8]. DR can be categorized into two subtypes: nonproliferative DR (NPDR) and proliferative DR (PDR). NPDR encompasses lesions such as microaneurysms and exudates [9]. Patients with DR exhibit various signs, as illustrated in Figure 1.
Regular screening of the retina is crucial for detecting DR, especially considering its asymptomatic nature in the early stages. Ophthalmologists typically rely on manual grading for this purpose. During the screening process, the patient’s pupil is dilated using specific eye drops, enabling the ophthalmologist to examine the retina using specialized lenses and lighting techniques. However, it is important to note that this procedure demands special expertise and experience to ensure accurate diagnosis.
With advancements in healthcare technologies, computer-aided diagnosis (CAD) systems have emerged as valuable tools for the rapid diagnosis of various diseases, including DR [10]. These systems offer cost-effective methods for retina screening; specifically, they can be extensively employed in the analysis of color fundus images for DR diagnosis. CAD systems enable experts to differentiate patients requiring more intensive treatment from those with less severe forms of the disease [11]. The underlying concept of CAD systems involves the expedited diagnosis of DR by analyzing features such as vessel segmentation, optic disk characteristics, and lesion segmentation from color fundus images. Subsequently, these images can be classified using various classifiers [12].
The evolution of artificial neural networks (ANNs) has paved the way for the development of more advanced machine learning (ML) techniques, such as deep learning (DL). In today’s healthcare systems, the role of artificial intelligence (AI) has become increasingly vital [13]. DL approaches, specifically in the field of medical image analysis and processing, have demonstrated their effectiveness [14,15]. DL has been successfully applied in not only the detection of DR but also ivarious other diseases, including the detection of COVID-19 [16]. DL approaches excel in extracting intricate patterns from data by leveraging filters and utilizing the given dataset to its full potential. The robustness of these models stems from their ability to handle vast amounts of data by employing deep neural networks [17].
DL models and, in particular, convolutional neural networks (CNNs), have emerged as effective solutions for DR classification problems, as supported by several studies [18,19,20]. CNNs make use of various pretrained models, including VGG16 [21], AlexNet [21], ResNet50 [22], and Inception-v3 [22]. The primary advantage of these models lies in their utilization of transfer learning [23,24,25], which allows models that have been previously trained on specific tasks to be repurposed for classification or detection in the context of another problem. This approach is particularly applicable in DL models, which often handle extensive datasets for training purposes.
The classification problems related to DR can be broadly categorized into two types: binary and multiclass. Binary classification focuses on distinguishing between a diseased retina and a healthy retina in color fundus images, as supported by the research [26,27]. On the other hand, multiclass classification approaches aim to grade the images into five different categories: Class 0—non DR, Class 1—mild DR, Class 2—moderate DR, Class 3—severe DR, and Class 4—proliferative DR [28,29]. Various publicly available datasets of retina images have been utilized in DL approaches for DR detection and diagnosis. Some notable examples include the DRIVE dataset [30], STARE [31], EyePACS dataset [32], ROC dataset [33], and Messidor dataset [34].
In this paper, we present our study focusing on experimentation with two powerful deep neural network models for the purpose of DR detection and classification. We conducted our experiments using the APTOS 2019 Blindness Detection dataset [35]. In both models, we employed the technique of transfer learning by utilizing pretrained feature detectors and training only the classifier component. More specifically, the two deep neural network approaches we utilized in our study were as follows:
  • Hybrid model: This model is a combination of the VGG16 architecture, as the feature detector; and the XGBoost algorithm, as the classifier. By leveraging the strengths of both the DL architecture and the gradient boosting classifier, we aimed to enhance the overall performance of the system.
  • DenseNet 121 Model: This model is based on the DenseNet 121 architecture, which is known for its dense connectivity pattern and efficient feature extraction capabilities. We employed this model to further explore its effectiveness in the context of DR detection and classification.
Furthermore, in order to enhance the performance these two models, we took steps to address the imbalanced nature of the selected dataset and applied preprocessing techniques to the considered images. In particular, to ensure a balanced representation of different classes, we trained the classifiers using a dataset that had been appropriately balanced. Additionally, we employed various preprocessing methods to enhance the quality and suitability of the images for our models’ training and evaluation processes.
The rest of this paper is organized as follows: Section 2 discusses recent research related to the detection, gradation, and classification of DR, providing an overview of relevant studies and approaches. Section 3 describes the dataset used in this study, specifically the APTOS 2019 Blindness Detection dataset, including its characteristics and relevant details. Section 4 presents the proposed methods, detailing the preprocessing steps applied to the images, introducing the hybrid model (combining VGG16 as the feature detector and XGBoost as the classifier) and the DenseNet 121 model. It also provides the experimental details, including software and parameter settings. Section 5 discusses the obtained results, analyzing the performance and accuracy of the proposed models and comparing them with existing methods. Finally, Section 6 concludes the paper, summarizing the key findings, suggesting future directions, and highlighting the contributions of this research.

2. Related Studies

Automated systems have emerged as potential solutions to facilitate early diagnosis and prevent permanent blindness in DR cases, eliminating the challenges associated with manual grading, which requires specialized expertise and can be burdensome for patients. To address these problems, researchers have explored various approaches for the classification of DR. This section presents a brief survey of models employed in this research field.
Rocha et al. [36] focused on addressing challenges in medical image analysis, including low contrast, poor lighting, and noise levels. Their study utilized the VGG16 network to classify retina fundus images into relevant categories. They employed publicly available datasets such as DDR, IDRiD, and EyePACS/Kaggle for image classification. The preprocessing steps included resizing the images and removing those with low contrast, followed by data augmentation, class balancing, hyperparameter adjustment, and image classification using the VGG16 network. Among the three databases, DDR exhibited the best performance, as measured via accuracy, precision, specificity, sensitivity, and the F1 score.
In their study, Khan et al. [37] focused on the challenge of reducing the model training time and convergence time in DR classification. They introduced a spatial pyramid pooling layer and network-in-network concepts alongside the VGG16 model. Preprocessing steps, such as resizing, cropping, normalization, and augmentation, were applied to the fundus retina images. The VGG16 network, spatial pyramid pooling layer, and network-in-network were stacked together. The spatial pyramid pooling layer connected the last convolutional layer of VGG16 with its first fully connected part, addressing concerns regarding cropping and information loss. The network-in-network layer was added on top to capture the nonlinear patterns within the datasets. The network-in-network part was initialized using the Xavier method, and fine-tuning was performed on the fully connected layers of VGG16. The results showed an AUC of 0.95 with 52% fewer parameters on the Kaggle dataset, and comparisons were made with the methods of other researchers.
Al-Antary et al. [38] proposed a CNN called multiscale attention network (MSA-Net) for detecting retina damage while dealing with high-level features. The authors utilized a multilevel and multiscale representation approach. Initially, the retina images were preprocessed using the APTOS and EyePACS datasets. A RESNET was employed as an encoder for feature extraction and to address the vanishing gradient issue. The extracted features encompassed local features and semantic information. To integrate these features, two multilevel representations were employed, combining mid-level and high-level features. A multiscale representation was utilized to ensure a uniform size, as the resolution of the features varied. The MSA-Net was then introduced to prioritize the relevant parts and differentiate the abnormalities in the retina images. The method achieved impressive performance on the APTOS database, with 98.1% accuracy, 98.3% sensitivity, 98.2% specificity, and an F1 score of 0.982, outperforming the other models. On the EyePACS database, the approach achieved an accuracy of 87.5%, sensitivity of 90.6%, specificity of 78.7%, and an F1 score of 0.767.
In [39], the authors proposed a hybrid model called E-DenseNet for the early diagnosis of DR. The motivation behind this model was to address research challenges in using a CNN for DR detection from retina images. Conventional CNNs may not accurately distinguish different types of lesions with distinct features. Hence, the E-DenseNet model was developed by stacking the Eyenet model on top of the DenseNet model, creating a customized hybrid architecture. The model was evaluated on four different datasets (EyePACS, IRiRD, Messidor, and APTOS 2019) spanning from 2006 to 2019 for detecting and classifying different grades of DR. The E-DenseNet model achieved impressive performance with an average accuracy of 91.2%, specificity of 69%, sensitivity of 96%, dice similarity coefficient of 92.45%, quadratic kappa score of 0.883, and a calculation time of 3.5 min.
Furthermore, Das et al. [40] conducted a comprehensive review of over one hundred research papers focused on the diagnosis of DR. The review highlighted various ML methods and their associated challenges in DR detection and diagnosis. It emphasized the superiority of DL architectures over traditional ML approaches in terms of feature extraction and image classification. In another study by Shaila et al. [41], a DL CNN model incorporating ResNet and VGG16 was developed for early DR detection. Texture analysis was performed on both balanced and imbalanced Kaggle datasets, and a combination of DL models was used for classification. The results demonstrated the model’s ability to accurately classify different stages of DR compared with other methods. Another approach involved the development of an intelligent system using case-based reasoning, as presented in Barman et al. [42]. This system employs retina image processing, feature extraction, and similarity-based case retrieval using the Euclidean distance measure to detect DR. All these studies highlight the effectiveness of DL models and the application of advanced techniques such as texture analysis and case-based reasoning in improving the accuracy and efficiency of DR detection and classification.
Challa et al. [43] proposed a deep All-CNN network for the diagnosis and grading of DR. In their study, they preprocessed images from the Kaggle dataset by applying Gaussian filters to enhance blending and remove retinal boundaries. The preprocessed images were then fed into the All-CNN network, which consists of ten convolution layers and a softmax layer for classification. The model achieved an accuracy of 86.64%, a loss of 0.46, and an average F1 score of 0.6318 across all five different stages of DR. Furthermore, other researchers [44,45] reviewed the contributions of numerous studies in the field of DR detection and classification, highlighting the implementation of both ML and DL models in these endeavors.
In recent years, several research studies have explored innovative techniques and applications in various fields. Federated learning has emerged as a promising approach to address the challenge of insufficient training data while maintaining data privacy. Authors [46] proposed a federated learning framework that allows multiple users to collaboratively train models locally without sharing sensitive data, enhancing performance and avoiding data privacy concerns. Additionally, in the domain of face swap deep fakes, Zhao et al. [47] introduced the conditional weighting transfer Wasserstein autoencoder, which enables effective knowledge transfer between multiple source domains. Biometrics, particularly face recognition, has attracted significant attention due to its uniqueness, stability, versatility, and difficulty to counterfeit, leading to its wide application [48]. Hyperspectral imagery has been recognized as valuable in remote sensing applications, including object classification, hyperspectral unmixing, anomaly detection, and change detection [49]. In the field of medical imaging, Ban et al. [50] proposed a novel 2D/3D registration model based on spatial histograms and tested it on X-ray and CT images. Other works [51,52] focused on Twitter sentiment analysis for the classification of user sentiments in tweets about COVID-19 on Twitter and implemented sentiment analysis using seven different deep learning models based on LSTM neural networks. Deep learning techniques have the potential to improve accuracy, and the work in [53] emphasizes the importance of employing the most up-to-date methods in the aviation industry. These references highlight the advancements and diverse applications of various techniques in different domains, contributing to the development of robust and effective solutions.
Table 1 provides an overview of the techniques proposed in relevant studies and summarizes the techniques employed in the current study.

3. Dataset Description

This section focuses on the dataset used in our study and describes how the dataset was balanced. The APTOS 2019 Blindness Detection Database [35] was employed, which consists of 3662 retinal images captured under various lighting conditions. The dataset was collected from the Aravind Eye Hospital in India. The retinal images in the dataset are categorized into five classes representing different severity levels of DR: Class 0 corresponds to non-DR, Class 1 corresponds to mild DR, Class 2 corresponds to moderate DR, Class 3 corresponds to severe DR, and Class 4 corresponds to proliferative DR, as outlined in Table 2. The distribution of the samples across these severity levels is presented in Table 3, indicating the number of images in each class. The dataset was balanced to ensure a sufficient representation of each severity level, which is crucial for effectively training and evaluating the models.
The APTOS 2019 Blindness Detection database exhibits a significant class imbalance, as illustrated in Figure 2. To address this issue, dataset-balancing techniques [54,55] were applied as a crucial preprocessing step. Balancing the training and testing datasets involved adjusting the ratios to minimize the disparities between the classes. This ensured that each class had a more equal representation, enabling more effective training and evaluation of the models.
Figure 3 displays the balanced training dataset, where each class is represented by an approximately equal number of samples. On the other hand, Figure 4a,b illustrate the imbalanced and balanced testing datasets, respectively. In the imbalanced testing dataset (Figure 4a), the class distribution reflects the original dataset; while in the balanced testing dataset (Figure 4b), the classes were adjusted to achieve a more balanced representation.

4. Methodology

In this section, we provide a detailed description of the proposed method. Our approach involves the implementation and experimentation of two different deep learning (DL) models: a hybrid model based on the combination of VGG16 and XGBoost classifier and a DenseNet 121 model.
The hybrid model was designed to leverage the strengths of both VGG16, a popular convolutional neural network (CNN) architecture known for its effectiveness in feature extraction; and XGBoost, a powerful gradient boosting algorithm widely used for classification tasks. By combining these two models, we aimed to enhance the overall performance of the system.
The second model, DenseNet 121, is a deep CNN architecture known for its dense connectivity pattern, which allows for efficient information flow between layers. This model has shown promising results in various image classification tasks and was well suited for our objective of DR detection and classification.

4.1. Image Preprocessing

The images in the considered dataset were collected from rural parts of India under diverse conditions, leading to a lack of uniformity among them. Utilizing these images in their raw form would not have yielded the desired results. Therefore, preprocessing was necessary to enhance the images before feeding them into the neural network model. The application of various preprocessing techniques, as depicted in Figure 5, aimed to standardize and optimize the images, ensuring improved quality and facilitating accurate analysis and classification.
Figure 6 showcases a selection of sample images included in the dataset.
The size of the images in the dataset was not uniform, as they were collected from different places. To standardize the input, a series of preprocessing steps was applied to the images:
  • First, all images were uniformly resized to a fixed dimension of 224 × 224 pixels, as shown in Figure 7a. This resizing step ensured that all images had the same size, facilitating consistent analysis.
  • Additionally, a Gaussian blur filter was applied to reduce noise and enhance image quality.
  • Finally, the Ben Graham procedure [56] was utilized to further improve image quality and accuracy. This involved cropping the images to their region of interest, as depicted in Figure 7b.
These preprocessing steps enable the neural network models to receive standardized and optimized input images, and thus improving the overall performance of the system.

4.2. Modeling

To detect unhealthy retina images, we developed two distinct DL models that are capable of accomplishing this task. The first model is a hybrid model, while the second model is based on the DenseNet 121 architecture. In this section, we discuss each model in detail, starting with the hybrid model and then proceeding to the DenseNet 121 model.

HybridModel: VGG16 and XGBoost Classifier

The VGG16 and XGBoost classifiers were combined to create a hybrid model for detection and classification of diabetic retinopathy (DR). The VGG16 network, which is a powerful pretrained convolutional neural network (CNN), was used as the base model for image classification [57,58]. It has 16 layers of processing, including convolutional and max pooling layers [59], as depicted in Figure 8.
To enhance the performance of the VGG16 model, an XGBoost classifier was employed. XGBoost is a boosting decision tree classifier that optimizes a cost function through gradient descent [60]. Unlike traditional ensemble classifiers that adjust the weights of the training set [61], XGBoost boosts a weak model to improve its predictive power.
In the hybrid model, the output from the VGG16 network is fed as the input to the XGBoost classifier, allowing for a combination of the powerful feature extraction capabilities of VGG16 with the gradient boosting capabilities of XGBoost. This integration aims to enhance the classification accuracy and performance of the model in detecting and classifying DR.
More specifically, the VGG16 CNN consists of 13 convolutional layers and 3 fully connected layers. Each convolutional layer has a kernel size of 3 × 3 and uses ReLU activation. The number of channels increases gradually from 64 to 512 in the deeper layers. Max pooling layers with a 2 × 2 window and stride of 2 are applied after certain convolutional blocks to downsample the feature maps.
After the feature extraction process with VGG16, the extracted features are fed into the XGBoost classifier. XGBoost is a gradient boosting algorithm that uses an ensemble of decision trees. The decision trees are sequentially trained, with each subsequent tree trying to correct the mistakes made by the previous trees. The number of decision trees and other hyperparameters of the XGBoost classifier were optimized through cross-validation.
To combine the VGG16 CNN and XGBoost classifier, we used a two-step approach. First, we trained the VGG16 CNN on the training data and obtained the output features from the last fully connected layer. These features served as the input to the XGBoost classifier, which was separately trained using the labeled data. During inference, the input image was first passed through the VGG16 CNN to extract features, and then these features were used as input to the trained XGBoost classifier for the final prediction. By combining the strengths of both deep learning and gradient boosting techniques, the hybrid model aims to leverage the feature extraction capabilities of CNNs and the powerful ensemble learning of XGBoost to improve the overall classification performance.

4.3. DenseNet 121 Model

The DenseNet is a type of convolutional neural network (CNN) that enables deeper network architectures by connecting each layer to every other layer in a feed-forward fashion [62]. In the DenseNet model, each layer receives inputs from all preceding layers and passes its feature maps to all subsequent layers, resulting in a dense connectivity pattern. This dense connectivity allows for efficient information flow and promotes feature reuse throughout the network [63]. The architecture of the DenseNet 121 model is illustrated in Figure 9.
The DenseNet architecture consists of basic convolutional and pooling layers, dense blocks, and transition layers. The model begins with a convolutional block that applies a sliding window of size 7 × 7 to the input image, producing 64 output layers/filters. This block uses a stride of 2, resulting in a downsampled feature map. It is followed by a max pooling layer with a 3 × 3 sliding window and a stride of 2, further reducing the spatial dimensions of the feature map.
Within the DenseNet architecture, there are multiple dense blocks, each consisting of a sequence of operations. The convolutional blocks within each dense block follow a specific pattern. They begin with a batch normalization layer to standardize the input, followed by a ReLU activation function to introduce nonlinearity, and then a Conv2D layer to perform the convolution operation. In DenseNet 121, this sequence of batch normalization, ReLU activation, and Conv2D is repeated 6 times in the first dense block, 12 times in the second dense block, 24 times in the third dense block, and 16 times in the final dense block.
The transition layers in DenseNet reduce the number of channels in the feature maps. The transition layers are placed immediately after each dense block. They consist of a 1 × 1 convolutional layer followed by a 2 × 2 average pooling layer with a stride of 2. This combination progressively reduces the number of channels by half from one dense block to the next. In DenseNet 121, the number of channels is reduced from 256 to 128, then to 64, and finally to 32.
Finally, in the DenseNet architecture, there is a global average pooling layer that performs spatial pooling across the entire feature map, resulting in a fixed-length vector representation [64]. This pooling operation aggregates the feature maps into a compact representation that captures the most salient information. Finally, a fully connected layer is used for classification, followed by a softmax activation function to generate the class probabilities.

4.4. Experimental Details

The considered models were trained and tested for binary classification using the TensorFlow [65] and Scikit-Learn [66] libraries in Python programming language. The Adam optimizer [67] was used for optimizing the training process. Table 4 shows the different parameter settings for the hybrid model and the DenseNet 121 model.
The initialization of hyperparameters is a critical aspect in deep learning models as it can significantly impact their performance and convergence. In our study, we followed the established practices for initializing hyperparameters based on the characteristics of the specific models employed.
The learning rate is a configurable hyperparameter used in the training of neural networks. It determines the amount that the weights are updated during each iteration, also referred to as the step size. The learning rate acts as a scale factor for the gradients computed during backpropagation, influencing the speed and quality of convergence. A higher learning rate can result in faster convergence, but it may also risk overshooting the optimal solution. Conversely, a lower learning rate can ensure more precise weight updates, but it may require more training epochs to reach convergence.
The batch size is another important hyperparameter that determines the number of samples processed before the model is updated. In each training iteration, the batch size specifies the subset of data samples used to compute the gradients and update the model’s parameters. Choosing an appropriate batch size involves balancing computational efficiency and the quality of weight updates. Smaller batch sizes provide more frequent updates, leading to faster convergence but with higher computational overhead. On the other hand, larger batch sizes can leverage parallelism and optimize computational efficiency but may result in less frequent weight updates and potentially slower convergence.
Furthermore, the number of epochs defines the number of complete passes through the training dataset during training. Each epoch consists of multiple iterations, where the model updates its weights based on the gradients computed from the batched data. The number of epochs is typically set based on the convergence behavior of the model and the desired level of training. More complex tasks or models may require a greater number of epochs to reach convergence, while simpler tasks or models may converge faster.
The initialization basis for these hyperparameters was carefully considered in our study, taking into account the characteristics of the deep learning models and the requirements of the diabetic retinopathy classification task.
The models were evaluated using the accuracy metric, which is calculated as follows:
A c c u r a c y = T P + T N T P + T N + F P + F N
where T P represents the true positives (the number of correct predictions of unhealthy retina images), T N represents the true negatives (the number of correctly predicted healthy retina images), F P represents the false positives (the number of healthy retina images incorrectly predicted as unhealthy), and F N represents the false negatives (the number of unhealthy retina images incorrectly predicted as healthy). This metric provides an overall measure of the models’ performance in correctly classifying the retina images.

5. Results and Discussion

The models were trained and tested on the APTOS dataset, with 80% of the data used for training and 20% used for testing. The training process was conducted over 50 epochs. The hybrid model achieved an output accuracy of approximately 80%, while the DenseNet 121 model achieved an impressive overall accuracy of 97.30%. The accuracy and loss curves for the DenseNet 121 model are shown in Figure 10a,b, respectively.
The performance of the two proposed models, the hybrid model and the DenseNet 121 model, was evaluated and compared in terms of accuracy. The results of the comparison are presented in Figure 11. From the figure, it can be observed that the DenseNet 121 model outperformed the hybrid model in terms of accuracy. The DenseNet 121 model achieved an accuracy of 97.30%, while the hybrid model achieved an accuracy of approximately 80%.
The higher accuracy of the DenseNet 121 model can be attributed to its dense connectivity pattern, which allows for effective information propagation throughout the network. This enables the model to capture and utilize important features from all preceding layers, leading to more accurate predictions.
The performance of the proposed hybrid model and DenseNet 121 model was compared with that of several existing methods using different datasets, including APTOS, EyePACS, and Messidor. The comparison of the models with other researchers’ methods is presented in Table 5. The proposed hybrid model performed better and obtained an accuracy of 79.50% compared with an accuracy of 75.61% of CNN [68] on the APTOS 2019 Blindness Detection Kaggle Dataset. Moreover, the DenseNet 121 model achieved a higher accuracy score of 97.30% than other existing techniques, such as Inception V3 [69], CNN [70], Inception ResNet V2 [71], and GoogleNet [72], on the APTOS 2019 Blindness Detection dataset.

Discussion

While our approach in this study incorporates the VGG model as a foundation, it is important to note that we made specific structural updates and modifications to adapt it to the task of medical image classification, particularly for diabetic retinopathy detection.
Medical image classification poses unique challenges and requires specialized considerations due to the complexity and intricacy of medical imaging data. The specific structural updates we made to the VGG model for medical image classification include:
  • Preprocessing: Medical images often require specific preprocessing steps such as normalization, resizing, and data augmentation techniques tailored to the characteristics of the medical imaging data. These preprocessing steps help with improving the robustness and generalization of the model.
  • Transfer learning: Given the limited availability of labeled medical image datasets, transfer learning becomes crucial. We leveraged transfer learning by initializing the VGG model with pretrained weights on large-scale image datasets and fine tuning it on our specific medical image dataset. This transfer of knowledge from general image classification tasks to the medical domain helps with learning relevant features and patterns.
  • DenseNet architecture: In addition to the VGG model, we also employed the DenseNet architecture, which has shown promising performance in various medical image analysis tasks. DenseNet introduces dense connections between layers, facilitating feature reuse and gradient flow throughout the network. This architecture helps with capturing more intricate details and dependencies within the medical images.
  • Class imbalance handling: Class imbalance is a common challenge in medical image classification tasks, where certain classes have significantly fewer samples than others. To address this, we employed techniques such as data augmentation, class weighting, and sampling strategies to balance the class distribution during training, ensuring that the model effectively learned from all classes.
These specific structural updates and adaptations are essential for enhancing the performance and relevance of the neural network models in the context of medical image classification. By tailoring the architecture and incorporating domain-specific considerations, we can effectively leverage the power of advanced neural network models to accurately and reliably analyze medical images.

6. Conclusions and Future Work

The early detection of diabetic retinopathy is crucial in preventing vision loss caused by diabetes mellitus. Computer-aided diagnosis (CAD) systems have significantly simplified the process of regular eye screening for diabetic patients [73]. With the advancements in CAD systems, deep learning neural networks, such as the ones examined in this study, have emerged as powerful tools for retinal image classification. In this study, we investigated two deep learning models, namely, a hybrid model (a combination of VGG16 and XGBoost Classifier) and a DenseNet 121 model. An essential step in our approach was to balance the training and testing datasets of the APTOS 2019 Blindness Detection database. While the hybrid model did not yield satisfactory results, the DenseNet 121 model demonstrated superior classification accuracy. We also compared our proposed models with existing methods on the same dataset, and the results revealed that the DenseNet 121 model achieved an impressive accuracy of 97.30%, outperforming all other compared architectures. Although the hybrid model achieved an accuracy of 79.50%, it still performed better than the CNN model [68]. Furthermore, the DenseNet 121 model exhibited fast classification capabilities, making it suitable for real-time medical applications.
In terms of future work, several techniques can be explored to further enhance the performance of the proposed models in retinal image classification tasks [74]. One significant contribution would involve the development of an application that can assist medical experts and even patients in the early detection of diabetic retinopathy. Such an application would not only prevent vision loss but also save valuable therapy time and costs [75]. By leveraging the power of deep learning models such as DenseNet 121, this application could provide efficient and reliable diagnoses, benefiting individuals at risk of developing diabetic retinopathy.

Author Contributions

Conceptualization, C.M., S.M., B.A., F.K., V.C.G., I.K. and A.K.; Methodology, C.M., S.M., B.A., F.K., V.C.G., I.K. and A.K.; Data curation, C.M., S.M., B.A., F.K., V.C.G., I.K. and A.K.; Writing—original draft, C.M., S.M., B.A., F.K., V.C.G., I.K. and A.K.; Writing—review & editing, C.M., S.M., B.A., F.K., V.C.G., I.K. and A.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Diabetes-PAHO/WHO Pan American Health Organization. Available online: https://www.paho.org/en/topics/diabetes (accessed on 10 April 2023).
  2. Saeedi, P.; Petersohn, I.; Salpea, P.; Malanda, B.; Karuranga, S.; Unwin, N.; Colagiuri, S.; Guariguata, L.; Motala, A.A.; Ogurtsova, K.; et al. Global and Regional Diabetes Prevalence Estimates for 2019 and Projections for 2030 and 2045: Results from the International Diabetes Federation Diabetes Atlas. Diabetes Res. Clin. Pract. 2019, 157, 107843. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Pradeepa, R.; Mohan, V. Epidemiology of Type 2 Diabetes in India. Indian J. Ophthalmol. 2021, 69, 2932. [Google Scholar] [PubMed]
  4. IDF Diabetes Atlas. Available online: https://diabetesatlas.org/atlas/ninth-edition (accessed on 27 March 2023).
  5. Chandrasekharan Kartha, C.; Ramachandran, S.; Pillai, R.M. Mechanisms of Vascular Defects in Diabetes Mellitus; Advances in Biochemistry in Health and Disease; Springer: Berlin/Heidelberg, Germany, 2017. [Google Scholar]
  6. Sallam, A. Diabetic Retinopathy Update. Egypt. Retin. J. 2014, 2, 1. [Google Scholar] [CrossRef]
  7. Abramoff, M.D.; Fort, P.E.; Han, I.C.; Jayasundera, K.T.; Sohn, E.H.; Gardner, T.W. Approach for a Clinically Useful Comprehensive Classification of Vascular and Neural Aspects of Diabetic Retinal Disease. Investig. Ophthalmol. Vis. Sci. 2018, 59, 519–527. [Google Scholar] [CrossRef] [PubMed]
  8. Yadav, P.; Singh, S.V.; Nada, M.; Dahiya, M. Impact of Severity of Diabetic Retinopathy on Quality of Life in Type 2 Indian Diabetic Patients. Int. J. Community Med. Public Health 2021, 8, 207–211. [Google Scholar] [CrossRef]
  9. Salmon, J.F. Kanski’s Clinical Ophthalmology: A Systematic Approach; Elsevier: Amsterdam, The Netherlands, 2019. [Google Scholar]
  10. Memari, N.; Abdollahi, S.; Ganzagh, M.M.; Moghbel, M. Computer-assisted Diagnosis (CAD) System for Diabetic Retinopathy Screening using Color Fundus Images using Deep Learning. In Proceedings of the IEEE Student Conference on Research and Development (SCOReD), Online, 27–29 September 2020; pp. 69–73. [Google Scholar]
  11. Asiri, N.M.; Hussain, M.; Adel, F.A.; Alzaidi, N. Deep Learning based Computer-Aided Diagnosis Systems for Diabetic Retinopathy: A Survey. Artif. Intell. Med. 2019, 99, 101701. [Google Scholar] [CrossRef] [Green Version]
  12. Carrera, E.V.; González, A.; Carrera, R. Automated Detection of Diabetic Retinopathy using SVM. In Proceedings of the IEEE XXIV International Conference on Electronics, Electrical Engineering and Computing (INTERCON), Cusco, Peru, 15–18 August 2017; pp. 1–4. [Google Scholar]
  13. Davenport, T.; Kalakota, R. The Potential for Artificial Intelligence in Healthcare. Future Healthc. J. 2019, 6, 94. [Google Scholar] [CrossRef] [Green Version]
  14. Liu, X.; Gao, K.; Liu, B.; Pan, C.; Liang, K.; Yan, L.; Ma, J.; He, F.; Zhang, S.; Pan, S.; et al. Advances in Deep Learning-Based Medical Image Analysis. Health Data Sci. 2021, 2021, 8786793. [Google Scholar] [CrossRef]
  15. Puttagunta, M.; Subban, R. Medical Image Analysis based on Deep Learning Approach. Multimed. Tools Appl. 2021, 80, 24365–24398. [Google Scholar] [CrossRef]
  16. Basu, S.; Mitra, S.; Saha, N. Deep Learning for Screening COVID-19 using Chest X-ray Images. In Proceedings of the IEEE Symposium Series on Computational Intelligence (SSCI), Canberra, Australia, 1–4 December 2020; pp. 2521–2527. [Google Scholar]
  17. Chen, W.; Yang, B.; Li, J.; Wang, J. An Approach to Detecting Diabetic Retinopathy Based on Integrated Shallow Convolutional Neural Networks. IEEE Access 2020, 8, 178552–178562. [Google Scholar] [CrossRef]
  18. Das, S.; Saha, S.K. Diabetic Retinopathy Detection and Classification using CNN tuned by Genetic Algorithm. Multimed. Tools Appl. 2022, 81, 8007–8020. [Google Scholar] [CrossRef]
  19. Raj, M.A.H.; Al Mamun, M.; Faruk, M.F. CNN Based Diabetic Retinopathy Status Prediction using Fundus Images. In Proceedings of the IEEE Region 10 Symposium (TENSYMP), Dhaka, Bangladesh, 5–7 June 2020; pp. 190–193. [Google Scholar]
  20. Reguant, R.; Brunak, S.; Saha, S. Understanding Inherent Image Features in CNN-based Assessment of Diabetic Retinopathy. Sci. Rep. 2021, 11, 9704. [Google Scholar] [CrossRef]
  21. Patel, S. Diabetic Retinopathy Detection and Classification using Pre-trained Convolutional Neural Networks. Int. J. Emerg. Technol. 2020, 11, 1082–1087. [Google Scholar]
  22. Aatila, M.; Lachgar, M.; Hrimech, H.; Kartit, A. Diabetic Retinopathy Classification Using ResNet50 and VGG-16 Pretrained Networks. Int. J. Comput. Eng. Data Sci. (IJCEDS) 2021, 1, 1–7. [Google Scholar]
  23. Savvopoulos, A.; Kanavos, A.; Mylonas, P.; Sioutas, S. LSTM Accelerator for Convolutional Object Identification. Algorithms 2018, 11, 157. [Google Scholar] [CrossRef] [Green Version]
  24. Sharma, C.; Parikh, S. Comparison of CNN and Pre-Trained Models: A Study. 2022. Available online: https://www.researchgate.net/publication/359850786_Comparison_of_CNN_and_Pre-trained_models_A_Study (accessed on 1 May 2023).
  25. Tuyen, D.N.; Tuan, T.M.; Son, L.H.; Ngan, T.T.; Giang, N.L.; Thong, P.H.; Hieu, V.V.; Gerogiannis, V.C.; Tzimos, D.; Kanavos, A. A Novel Approach Combining Particle Swarm Optimization and Deep Learning for Flash Flood Detection from Satellite Images. Mathematics 2021, 9, 2846. [Google Scholar] [CrossRef]
  26. Adriman, R.; Muchtar, K.; Maulina, N. Performance Evaluation of Binary Classification of Diabetic Retinopathy through Deep Learning Techniques using Texture Feature. Procedia Comput. Sci. 2021, 179, 88–94. [Google Scholar] [CrossRef]
  27. Ramchandre, S.; Patil, B.; Pharande, S.; Javali, K.; Pande, H. A Deep Learning Approach for Diabetic Retinopathy detection using Transfer Learning. In Proceedings of the IEEE International Conference for Innovation in Technology (INOCON), Bangluru, India, 6–8 November 2020; pp. 1–5. [Google Scholar]
  28. Dai, L.; Wu, L.; Li, H.; Cai, C.; Wu, Q.; Kong, H.; Liu, R.; Wang, X.; Hou, X.; Liu, Y.; et al. A Deep Learning System for Detecting Diabetic Retinopathy across the Disease Spectrum. Nat. Commun. 2021, 12, 3242. [Google Scholar] [CrossRef]
  29. Tymchenko, B.; Marchenko, P.; Spodarets, D. Deep Learning Approach to Diabetic Retinopathy Detection. arXiv 2020, arXiv:2003.02261. [Google Scholar]
  30. Mateen, M.; Wen, J.; Hassan, M.; Nasrullah, N.; Song, S.; Hayat, S. Automatic Detection of Diabetic Retinopathy: A Review on Datasets, Methods and Evaluation Metrics. IEEE Access 2020, 8, 48784–48811. [Google Scholar] [CrossRef]
  31. Leopold, H.A.; Orchard, J.; Zelek, J.S.; Lakshminarayanan, V. PixelBNN: Augmenting the PixelCNN with Batch Normalization and the Presentation of a Fast Architecture for Retinal Vessel Segmentation. J. Imaging 2019, 5, 26. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Voets, M.; Møllersen, K.; Bongo, L.A. Reproduction Study using Public Data of: Development and Validation of a Deep Learning Algorithm for Detection of Diabetic Retinopathy in Retinal Fundus Photographs. PLoS ONE 2019, 14, e0217541. [Google Scholar] [CrossRef] [PubMed]
  33. Niemeijer, M.; van Ginneken, B.; Cree, M.J.; Mizutani, A.; Quellec, G.; Sánchez, C.I.; Zhang, B.; Hornero, R.; Lamard, M.; Muramatsu, C.; et al. Retinopathy Online Challenge: Automatic Detection of Microaneurysms in Digital Color Fundus Photographs. IEEE Trans. Med Imaging 2010, 29, 185–195. [Google Scholar] [CrossRef] [PubMed]
  34. Messidor Project. Available online: https://www.adcis.net/en/third-party/messidor (accessed on 10 April 2023).
  35. APTOS 2019 Blindness Detection. Available online: https://www.kaggle.com/c/aptos2019-blindness-detection (accessed on 10 April 2023).
  36. Da Rocha, D.A.; Ferreira, F.M.F.; Peixoto, Z.M.A. Diabetic Retinopathy Classification using VGG16 Neural Network. Res. Biomed. Eng. 2022, 38, 761–772. [Google Scholar] [CrossRef]
  37. Khan, Z.; Khan, F.G.; Khan, A.; Rehman, Z.U.; Shah, S.; Qummar, S.; Ali, F.; Pack, S. Diabetic Retinopathy Detection Using VGG-NIN a Deep Learning Architecture. IEEE Access 2021, 9, 61408–61416. [Google Scholar] [CrossRef]
  38. Al-Antary, M.T.; Arafa, Y. Multi-Scale Attention Network for Diabetic Retinopathy Classification. IEEE Access 2021, 9, 54190–54200. [Google Scholar] [CrossRef]
  39. AbdelMaksoud, E.; Barakat, S.I.; Elmogy, M. A Computer-aided Diagnosis System for Detecting various Diabetic Retinopathy Grades based on a Hybrid Deep Learning Technique. Med. Biol. Eng. Comput. 2022, 60, 2015–2038. [Google Scholar] [CrossRef]
  40. Das, D.; Biswas, S.K.; Bandyopadhyay, S. A Critical Review on Diagnosis of Diabetic Retinopathy using Machine Learning and Deep Learning. Multimed. Tools Appl. 2022, 81, 25613–25655. [Google Scholar] [CrossRef]
  41. Shaila, S.G.; Lavanya, S.; Rajesh, T.M.; Bhuvana, D.S.; Deshpande, K. Early Detection of Diabetic Retinopathy Using Multimodal Approach. In Computer Vision and Robotics (CVR); Springer: Singapore, 2022; pp. 107–118. [Google Scholar]
  42. Barman, R.; Biswas, S.K.; Das, D.; Purkayastha, B.; Borah, M.D. Case-Based Expert System for Early Detection of Diabetic Retinopathy. In Intelligent Computing and Communication Systems; Springer: Singapore, 2021; pp. 259–267. [Google Scholar]
  43. Challa, U.K.; Yellamraju, P.; Bhatt, J.S. A Multi-class Deep All-CNN for Detection of Diabetic Retinopathy Using Retinal Fundus Images. In Proceedings of the 8th International Conference on Pattern Recognition and Machine Intelligence (PReMI), Tezpur, India, 17–20 December 2019; Lecture Notes in Computer Science. Springer: Cham, Switzerland, 2019; Volume 11941, pp. 191–199. [Google Scholar]
  44. Atwany, M.Z.; Sahyoun, A.; Yaqub, M. Deep Learning Techniques for Diabetic Retinopathy Classification: A Survey. IEEE Access 2022, 10, 28642–28655. [Google Scholar] [CrossRef]
  45. Wahid, F.F.; Raju, G. Diabetic Retinopathy Detection Using Convolutional Neural Network—A Study. In Data Science and Security (IDSCS); Springer: Singapore, 2021; pp. 127–133. [Google Scholar]
  46. Zhao, K.; Hu, J.; Shao, H.; Hu, J. Federated multi-source domain adversarial adaptation framework for machinery fault diagnosis with data privacy. Reliab. Eng. Syst. Saf. 2023, 236, 109246. [Google Scholar] [CrossRef]
  47. Zhao, K.; Jia, F.; Shao, H. A novel conditional weighting transfer Wasserstein auto-encoder for rolling bearing fault diagnosis with multi-source domains. Knowl.-Based Syst. 2023, 262, 110203. [Google Scholar] [CrossRef]
  48. Jin, B.; Cruz, L.; Gonçalves, N. Deep Facial Diagnosis: Deep Transfer Learning From Face Recognition to Facial Diagnosis. IEEE Access 2020, 8, 123649–123661. [Google Scholar] [CrossRef]
  49. Wang, S.; Hu, X.; Sun, J.; Liu, J. Hyperspectral anomaly detection using ensemble and robust collaborative representation. Inf. Sci. 2023, 624, 748–760. [Google Scholar] [CrossRef]
  50. Ban, Y.; Wang, Y.; Liu, S.; Yang, B.; Liu, M.; Yin, L.; Zheng, W. 2D/3D Multimode Medical Image Alignment Based on Spatial Histograms. Appl. Sci. 2022, 12, 8261. [Google Scholar] [CrossRef]
  51. Lyras, A.; Vernikou, S.; Kanavos, A.; Sioutas, S.; Mylonas, P. Modeling Credibility in Social Big Data using LSTM Neural Networks. In Proceedings of the 17th International Conference on Web Information Systems and Technologies (WEBIST), Online, 26–28 October 2021; pp. 599–606. [Google Scholar]
  52. Vernikou, S.; Lyras, A.; Kanavos, A. Multiclass sentiment analysis on COVID-19-related tweets using deep learning models. Neural Comput. Appl. 2022, 34, 19615–19627. [Google Scholar] [CrossRef]
  53. Kanavos, A.; Kounelis, F.; Iliadis, L.; Makris, C. Deep learning models for forecasting aviation demand time series. Neural Comput. Appl. 2021, 33, 16329–16343. [Google Scholar] [CrossRef]
  54. Mondal, S.; Mian, K.F.; Das, A. Deep Learning-based Diabetic Retinopathy Detection for Multiclass Imbalanced Data. In Recent Trends in Computational Intelligence Enabled Research; Elsevier: Amsterdam, The Netherlands, 2021; pp. 307–316. [Google Scholar]
  55. Saini, M.; Susan, S. Diabetic Retinopathy Screening using Deep Learning for Multi-class Imbalanced Datasets. Comput. Biol. Med. 2022, 149, 105989. [Google Scholar] [CrossRef]
  56. Graham, B. Kaggle Diabetic Retinopathy Detection Competition Report; University of Warwick: Coventry, UK, 2015; pp. 24–26. [Google Scholar]
  57. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  58. Step by Step VGG16 Implementation in Keras for Beginners. Available online: https://towardsdatascience.com/step-by-step-vgg16-implementation-in-keras-for-beginners-a833c686ae6c (accessed on 10 April 2023).
  59. Tammina, S. Transfer Learning using VGG-16 with Deep Convolutional Neural Network for Classifying Images. Int. J. Sci. Res. Publ. (IJSRP) 2019, 9, 143–150. [Google Scholar] [CrossRef]
  60. What Is XGBoost? Available online: https://www.nvidia.com/en-us/glossary/data-science/xgboost (accessed on 10 April 2023).
  61. Ren, X.; Guo, H.; Li, S.; Wang, S.; Li, J. A Novel Image Classification Method with CNN-XGBoost Model. In Proceedings of the 16th International Workshop on Digital Forensics and Watermarking (IWDW), Magdeburg, Germany, 23–25 August 2017; Volume 10431, pp. 378–390. [Google Scholar]
  62. Creating DenseNet 121 with TensorFlow. Available online: https://towardsdatascience.com/creating-densenet-121-with-tensorflow-edbc08a956d8 (accessed on 10 April 2023).
  63. Review: DenseNet—Dense Convolutional Network (Image Classification). Available online: https://towardsdatascience.com/review-densenet-image-classification-b6631a8ef803 (accessed on 10 April 2023).
  64. Zhang, K.; Guo, Y.; Wang, X.; Yuan, J.; Ding, Q. Multiple Feature Reweight DenseNet for Image Classification. IEEE Access 2019, 7, 9872–9880. [Google Scholar] [CrossRef]
  65. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems. arXiv 2016, arXiv:1603.04467. [Google Scholar]
  66. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  67. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  68. R., Y.; Sarobin, M.V.R.; Panjanathan, R.; Jasmine, S.G.; Anbarasi, L.J. Diabetic Retinopathy Classification Using CNN and Hybrid Deep Convolutional Neural Networks. Symmetry 2022, 14, 1932. [Google Scholar] [CrossRef]
  69. Kurup, G.; Jothi, J.A.A.; Kanadath, A. Diabetic Retinopathy Detection and Classification using Pretrained Inception-v3. In Proceedings of the IEEE International Conference on Smart Generation Computing, Communication and Networking (SMART GENCON), Pune, India, 29–30 October 2021; pp. 1–6. [Google Scholar]
  70. Kumar, R.R.; Pandian, R.; Jacob, T.P.; Pravin, A.; Indumathi, P. Detection of Diabetic Retinopathy Using Deep Convolutional Neural Networks. In Proceedings of the Computational Vision and Bio-Inspired Computing (ICCVBIC), Coimbatore, India, 25–26 November 2021; pp. 415–430. [Google Scholar]
  71. Gangwar, A.K.; Ravi, V. Diabetic Retinopathy Detection Using Transfer Learning and Deep Learning. In Evolution in Computational Intelligence—Frontiers in Intelligent Computing: Theory and Applications (FICTA); Advances in Intelligent Systems and Computing; Springer: Singapore, 2020; Volume 1176, pp. 679–689. [Google Scholar]
  72. Shi, B.; Zhang, X.; Wang, Z.; Song, J.; Han, J.; Zhang, Z.; Toe, T.T. GoogLeNet-based Diabetic-Retinopathy-Detection. In Proceedings of the 14th IEEE International Conference on Advanced Computational Intelligence (ICACI), Wuhan, China, 15–17 July 2022; pp. 246–249. [Google Scholar]
  73. Elsharkawy, M.; Sharafeldeen, A.; Soliman, A.; Khalifa, F.; Ghazal, M.; El-Daydamony, E.; Atwan, A.; Sandhu, H.S.; El-Baz, A. A Novel Computer-Aided Diagnostic System for Early Detection of Diabetic Retinopathy Using 3D-OCT Higher-Order Spatial Appearance Model. Diagnostics 2022, 12, 461. [Google Scholar] [CrossRef]
  74. Khalifa, N.E.M.; Loey, M.; Taha, M.H.N.; Mohamed, H.N.E.T. Deep Transfer Learning Models for Medical Diabetic Retinopathy Detection. Acta Inform. Med. 2019, 27, 327. [Google Scholar] [CrossRef]
  75. Kipli, K.; Hui, L.Y.; Tajudin, N.M.A.; Sapawi, R.; Sahari, S.K.; Mat, D.A.A.; Jalil, M.A.; Ray, K.; Kaiser, M.S.; Mahmud, M. Development of Mobile Application for Detection and Grading of Diabetic Retinopathy. In Trends in Electronics and Health Informatics (TEHI); Lecture Notes in Computer Science; Springer: Singapore, 2022; pp. 339–349. [Google Scholar]
Figure 1. Diabetic retinopathy.
Figure 1. Diabetic retinopathy.
Sensors 23 05726 g001
Figure 2. Imbalanced class distribution of the severity level of diabetic retinopathy.
Figure 2. Imbalanced class distribution of the severity level of diabetic retinopathy.
Sensors 23 05726 g002
Figure 3. The balanced training dataset.
Figure 3. The balanced training dataset.
Sensors 23 05726 g003
Figure 4. Imbalanced and balanced testing datasets.
Figure 4. Imbalanced and balanced testing datasets.
Sensors 23 05726 g004
Figure 5. Various image preprocessing techniques.
Figure 5. Various image preprocessing techniques.
Sensors 23 05726 g005
Figure 6. Input sample images.
Figure 6. Input sample images.
Sensors 23 05726 g006
Figure 7. A resized input image and an image cropped to the region of interest.
Figure 7. A resized input image and an image cropped to the region of interest.
Sensors 23 05726 g007
Figure 8. VGG 16 architecture.
Figure 8. VGG 16 architecture.
Sensors 23 05726 g008
Figure 9. DenseNet model architecture.
Figure 9. DenseNet model architecture.
Sensors 23 05726 g009
Figure 10. Accuracy and loss function of DenseNet 121 CNN model.
Figure 10. Accuracy and loss function of DenseNet 121 CNN model.
Sensors 23 05726 g010
Figure 11. Comparison between the hybrid and the DenseNet 121 model.
Figure 11. Comparison between the hybrid and the DenseNet 121 model.
Sensors 23 05726 g011
Table 1. Overview of relevant approaches and the proposed techniques.
Table 1. Overview of relevant approaches and the proposed techniques.
PaperProposed Techniques
[36]Removal of low-contrast images, data augmentation, and classification using VGG16 network.
[37]Utilization of spatial pyramid pooling layer and network-in-network layer in conjunction with VGG16 network.
[38]Introduction of a multiscale attention network (CNN) capable of detecting damages while handling high-level features.
[39]Development of E-DenseNet, an ensemble model combining EyeNet and DenseNet architectures.
[41]Texture analysis performed on balanced and imbalanced datasets using various CNN models.
[42]Implementation of an expert system utilizing case-based reasoning with retina image processing and feature extraction.
[43]Adoption of an All-CNN network consisting of ten convolution layers and a softmax layer.
Current approachTwo DL models are examined: a hybrid model (a combination of VGG16 and XGBoost Classifier) and one based on the DenseNet 121 architecture.
Table 2. Severity levels of diabetic retinopathy.
Table 2. Severity levels of diabetic retinopathy.
Class01234
ClassificationNon-DRMild DRModerate DRSevere DRProliferative DR
Table 3. Distribution of samples as per severity level of diabetic retinopathy.
Table 3. Distribution of samples as per severity level of diabetic retinopathy.
Severity LevelNumber of Samples
Class 0 (normal)1805
Class 1 (mild)370
Class 2 (moderate)999
Class 3 (severe)193
Class 4 (proliferative)295
Table 4. Parameters for the hybrid and DenseNet 121 models.
Table 4. Parameters for the hybrid and DenseNet 121 models.
Hybrid ModelDenseNet 121 Model
Batch Size1632
Initial Learning Rate0.010.01
Minimum Learning Rate0.00010.00005
Epochs5050
Table 5. Performance comparison of different architectures with the proposed models.
Table 5. Performance comparison of different architectures with the proposed models.
PaperDatasetArchitectureAccuracy
[69]APTOSInception V382
[70]APTOSCNN94.44
[71]APTOSInception Res Net V282.18
[72]APTOSGoogle Net97
[68]APTOSCNN75.61
Present workAPTOSHybrid79.50
Present workAPTOSDenseNet 12197.30
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mohanty, C.; Mahapatra, S.; Acharya, B.; Kokkoras, F.; Gerogiannis, V.C.; Karamitsos, I.; Kanavos, A. Using Deep Learning Architectures for Detection and Classification of Diabetic Retinopathy. Sensors 2023, 23, 5726. https://0-doi-org.brum.beds.ac.uk/10.3390/s23125726

AMA Style

Mohanty C, Mahapatra S, Acharya B, Kokkoras F, Gerogiannis VC, Karamitsos I, Kanavos A. Using Deep Learning Architectures for Detection and Classification of Diabetic Retinopathy. Sensors. 2023; 23(12):5726. https://0-doi-org.brum.beds.ac.uk/10.3390/s23125726

Chicago/Turabian Style

Mohanty, Cheena, Sakuntala Mahapatra, Biswaranjan Acharya, Fotis Kokkoras, Vassilis C. Gerogiannis, Ioannis Karamitsos, and Andreas Kanavos. 2023. "Using Deep Learning Architectures for Detection and Classification of Diabetic Retinopathy" Sensors 23, no. 12: 5726. https://0-doi-org.brum.beds.ac.uk/10.3390/s23125726

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop