Next Article in Journal
Weak Calibration of a Visible Light Positioning System Based on a Position-Sensitive Detector: Positioning Error Assessment
Next Article in Special Issue
GMSRI: A Texture-Based Martian Surface Rock Image Dataset
Previous Article in Journal
Construction and Application of Graphene Oxide-Bovine Serum Albumin Modified Extended Gate Field Effect Transistor Chiral Sensor
Previous Article in Special Issue
LeanNet: An Efficient Convolutional Neural Network for Digital Number Recognition in Industrial Products
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adversarial Attack and Defence through Adversarial Training and Feature Fusion for Diabetic Retinopathy Recognition

by
Sheeba Lal
1,
Saeed Ur Rehman
1,
Jamal Hussain Shah
1,
Talha Meraj
1,
Hafiz Tayyab Rauf
2,*,
Robertas Damaševičius
3,*,
Mazin Abed Mohammed
4 and
Karrar Hameed Abdulkareem
5
1
Department of Computer Science, COMSATS University Islamabad, Wah Campus, Wah Cantt 47040, Pakistan
2
Department of Computer Science, Faculty of Engineering & Informatics, University of Bradford, Bradford BD7 1DP, UK
3
Faculty of Applied Mathematics, Silesian University of Technology, 44-100 Gliwice, Poland
4
College of Computer Science and Information Technology, University of Anbar, Anbar 31001, Iraq
5
College of Agriculture, Al-Muthanna University, Samawah 66001, Iraq
*
Authors to whom correspondence should be addressed.
Submission received: 2 May 2021 / Revised: 31 May 2021 / Accepted: 4 June 2021 / Published: 7 June 2021
(This article belongs to the Special Issue Deep Learning Image Recognition Systems)

Abstract

:
Due to the rapid growth in artificial intelligence (AI) and deep learning (DL) approaches, the security and robustness of the deployed algorithms need to be guaranteed. The security susceptibility of the DL algorithms to adversarial examples has been widely acknowledged. The artificially created examples will lead to different instances negatively identified by the DL models that are humanly considered benign. Practical application in actual physical scenarios with adversarial threats shows their features. Thus, adversarial attacks and defense, including machine learning and its reliability, have drawn growing interest and, in recent years, has been a hot topic of research. We introduce a framework that provides a defensive model against the adversarial speckle-noise attack, the adversarial training, and a feature fusion strategy, which preserves the classification with correct labelling. We evaluate and analyze the adversarial attacks and defenses on the retinal fundus images for the Diabetic Retinopathy recognition problem, which is considered a state-of-the-art endeavor. Results obtained on the retinal fundus images, which are prone to adversarial attacks, are 99% accurate and prove that the proposed defensive model is robust.

1. Introduction

A rapidly growing computer vision domain leverages advanced innovation with comprehensive knowledge, while the developed techniques are used for a wide area of applications such as cancer detection [1,2,3], facial expression recognition [4], Parkinson’s disease diagnostics [5,6] and precision agriculture [7,8]. The success of computer vision is due to its more powerful ability to interpret image patterns than the human cognitive visual system. For example, artificial intelligence (AI) based image processing has transformed the field of medical diagnostics in the healthcare domain [9]. Radiomics is an evolving medical imaging field that utilizes a progression of subjective and quantitative examinations of high-throughput image highlights to acquire symptomatic, prescient, or prognostic data from clinical images [10,11]. Image data can take multiple formats, including multi-dimensional data from a 3D scanner or medical scanning devices. Advanced modalities are computed tomography (CT), magnetic resonance imaging (MRI), and nuclear/molecular imaging (which uses biomarkers for in vivo imaging) [12]. Moreover, automated computer vision methods are relevant for health and in-home medical diagnosis [13,14,15].
One such problem successfully addressed by computer vision based diagnostics is the recognition of Diabetic Retinopathy (DR). DR is a chronic disorder that causes blindness in individuals if untreated. A high glucose ratio in the blood causes changes in the retinal microvasculature, resulting in DR, which can lead to a total vision loss. DR is a cause of visual impairment globally that affects nearly 30% of diabetic patients [16]. Early detection of DR through retinal fundus images can avoid possible blindness due to disease. Previous studies have concentrated on the automated early identification of DR by color fundus photography and have produced spectacular classification results [17,18,19].
However, the accuracy and robustness of the deep learning model are frequently plagued by confidentiality of data [20,21]. Minor changes to input images have recently been shown to significantly alter the output of deep learning models [22]. These minor disruptions are an example of adversarial attacks (or adversarial perturbations), which mislead the model, cause it to predict the wrong label, and have drastic consequences for the performance of deep neural network models [23,24]. These models are vulnerable to adversarial examples, which pose a threat in real-world application scenarios [25,26].
Adversarial attacks are categorized as white box, black box and grey box attacks [27]. White-Box (WB) attack has both full information and access to the internal system model. The WB attack can use two iterative methods of the Fast-Gradient Sign Method and the Deep Fool approach, using a set classifier model to reduce its space for searching, and to produce a positive response to unseen adversarial data [28]. The attackers do not know the target model or network, input, and weights in a Black-Box (BB) attack [29]. For BB attacks, reference [30] generated a GenAttack gradient-free optimization algorithm with fewer probes while using Mixed National Institute of Standards and Technology (MNIST) [31], CIFAR-10 [32], and other datasets. Similarly, reference [33] introduced the gradient-based data augmentation technique and substituted ensemble training, which targeted BB attacks on the MNIST and GTSRB datasets with accurate results [34]. In machine learning, the robustness of the adversarial attack detection ability was enhanced by increasing the model capacity with more adversarial training and improved label leaking accuracy [35]. Contractive auto-encoder (CAE) deep neural networks work as a robust model against adversarial examples with high accuracy [36].
Some applications of adversarial attacks using pre-trained deep learning (DL) models in computer vision tasks include, e.g., visual classification [37], textual data system [38], privacy-preserving filter [39], object detector [40], image segmentation [41], natural language processing [42], data fusion [43], hybrid digital watermarking and text document retrieval [44], fingerprint liveness detection [45], person re-identification [46], time series classification [47], human activity recognition [48], face recognition [49], handwritten signature verification [50], and multi-objective reinforcement learning [51].
On the subject of image restoration, noise in an image is crucial. Speckle noise is a type of granular patterning that can be seen in radar coherent images. The Synthetic Aperture Radar image and spatial data both include a lot of speckle noise. In general, SN is the gritty salt-and-pepper pattern seen in radar imaging. It can even be considered a granular ‘noise’ that appears fundamentally in [52] ultrasound, synthetic aperture radar (SAR), active radar, and optical coherence tomography imaging, reducing their quality. Finally, it degrades the performance of critical image processing approaches such as detection, segmentation and classification [4]. A dynamic ultrasound video can be considered three-dimensional (3-D) images with moving parts. It presents a speckle technique for dynamic ultrasound called the 3-D Gabor-based anisotropic diffusion, which has two dimensions in the spatial domain and one in the temporal domain (GAD-3D) [5]. Three test models could be applied to generate synthetic images: radial polar, uniform grid and radial uniform. These synthetic images, which imitate the basic noise features of actual ultrasound images, might be useful for speckle experimentation [53]. Adversarial training is a method of demonstrating and defining the model as a threat by using examples of adversarial situations. In the training phase, it’s also essential to generate and then provide adverse examples from a complete and accurate optimization perspective at least. Whereas this strategy approximates a robust loss, which is precisely the goal we want to achieve, it is frequent to have a lot of the standard loss in the original data points (i.e., gradient measures as well) in that it increases the ‘task standard’ error’s efficiency slightly. Adversarial examples were first prepared using methods such as FGSM, I-FGSM, DeepFool and CW, and then used to train the target model to make it more resilient against an unknown adversarial attack using a diversity adversarial training approach. This technique reduces average attack success rates by 27.2 and 24.3 percent for various adversarial scenarios, while retaining 98.7 and 91.5 percent accuracies for the original data of the MNIST and FashionMNIST datasets, respectively [9]. Features represent the object’s numerical value that expresses the local and global function. The selection of the function features is normally dependent on the problem. Sometimes there are different results according to each feature. In certain cases, the use of a particular feature would be no more successful, so that a successful model is created by a mixing multiple feature. Many people have used different feature fusion techniques because when we fuse the features they have diverse results regarding the research problem. The mixing of characteristics from distinct layers or branches, known as feature fusion, is a common element in current network topologies. This, however, corresponds to iterative attentional feature fusion. On both the CIFAR-100 and ImageNet datasets, our models outperform state-of-the-art networks with fewer layers or parameters [10]. FFU-Net (Feature Fusion U-Net) enhances U-Net from the following characteristic points for diabetic retinopathy lesion segmentation. To decrease spatial loss of the fundus image, the network’s pooling layer is first superseded with a convolutional layer. Then, by fusing contextual channel attention (CCA) models, we combine the multiscale feature fusion (MSFF) block into the encoders, which also enables the network to learn multiscale features efficiently and to enhance the data produced [12]. Diabetic retinopathy is a chronic disorder that cannot be examined properly with normal vision, either aided or unaided, and it is also difficult to predict its density. For the diagnosis and classification of diabetic retinopathy, the key problem occurs when different sensitive sections of the eye, such as retina colors, irregular blood vessels, hard rough exudates, cotton wool spots and different adversarial attacks, are not detected properly. Much work has been done on DR classification and detection with high accuracy, but recently the concept of adversarial attacks has arisen. A small disruption is named an adversarial example/adversarial attack that misleads, with devastating effects, an informed profound neural network model and decreases its accuracy with respect to the correct label. Adversarial attacks against DNN are a serious security obstacle and they decrease accuracy, thus inventing new distance metrics for human perceptual systems and obtaining optimized results via a greedy algorithm [13]. Recently, most work done on adversarial attacks in medical imaging [16], such as stabilized medical image attacks [17], medical image classification [18,19], adversarial learning detecting erroneous diagnoses [20], adversarial heart attacks [21], segmentation of biomedical images [22] and defenses, included binary thresholding [23] using an adversarial attack to evaluate the durability of deep diagnostic models [24] and generative model defense [25] and a critical analysis of antagonistic threats as well as defense mechanisms in physiological computing [26]. Therefore, in this paper, we propose a new Speckle Noise (SN) attack using adversarial image generation, and two defensive methods against these attacks, including defensive adversarial training and feature fusion. The contribution of this research is as follows:
  • We evaluate and analyze the adversarial attacks and defenses on retinal fundus images, which is considered a state-of-the-art endeavor.
  • We propose a framework that contains a new SN attack, a defensive model against adversarial attacks, the adversarial training (AT), and a feature fusion strategy, which preserves the DR classification result with correct labelling.
  • We achieve accurate detection of DR from retinal fundus images using the proposed feature fusion approach.
The remaining paper’s organisation is as follows: Section 2 overviews related work. The proposed method is described in Section 3. Results and analysis are given in Section 4. The research is concluded in Section 5.

2. Related Work

During the last few decades, the medical image processing methods help in the early and efficient diagnosis of various severe aliments frequently detected in human beings. Recently, the advanced AI based algorithms have attained great importance with high accuracy in the classification of medical images and the detection of diseases in the medical field with productive results. Two-fold detection of DR using morphological procedures was introduced [54], which detects microaneurysms, exudate, blood vessels and second severity of its type using Support Vector Machine (SVM), but through adversarial attacks, their credibility has decreased. Deep radiomics performs well in medical imaging, but accuracy has deteriorated, and the incorrect label is based on minor disturbances (SP). In this regard, reference [55] introduced two novel attacks—Bracketed Exposure Fusion (BEF) and Convolution Bracketed Exposure Fusion (CBEF)—based on component-wise multiplicative fusion and element-wise convolutional for the detection of diabetic retinopathy (DR) by using the Eyepacs Dataset with high-quality images and transferal rates.
Universal perturbations attacks (UPA), which used iterative algorithms for targeted and non-targeted attacks, were proposed by [56], and achieved 80% accuracy in classification. Reference [57] presented two lightweight techniques, which used local perturbation and universal attacks. The sequential decision method for fixing the image reconstruction model is implemented using reinforcement learning [58]. The adversarial data augmentation approach proposed by [59] for medical image segmentation was designed for deep neural network (DNN) model training induced by a shared type of artifact in magnetic resonance imaging (MRI).
The adversarial augmentation approach was proposed in [60], which was used to generalize the model. Project gradient descent (PGD) or adverse synthetic nodule and adverse perturbation noise work detected the lung by false positive reduction (FPR). For malignancy prediction of lung nodules, reference [61] introduced an adversarial attack deep neural network ensemble methodology for classification using FGMS and 1-pixel attack, achieving 82.27% and 81.43% accuracy. The authors proposed a DL-based encryption and a decryption network (DLEDNet) [62] using an X-ray image dataset through region of interest (ROI) segmentation in an encrypted medical image. In medical imaging for adversarial training, reference [63] developed transfer learning and a self-supervision based procedure for adversarial training for pneumonia classification of X-ray images and MRI segmentation using PGD and fast gradient methods. The detailed comparison of recent related works with their dataset description is presented in Table 1.

Defenses against Adversarial Attacks

Reference [65] proposed defense against two groups: feature-level interpretation and model-level interpretation, input denoising, and model robustification. Two image transformations, Discrete Wavelet Transform (DWT) and Sine Transform (ST), were presented by [66] for classifying features with a SVM classifier. Some techniques demonstrated by [67] included dimensionality reduction, a characterization of the adversarial region, and combining input discretization with adversarial training. Activation transformations for the best and most robust defense against these attacks were also considered. Meng and Chen [68] proposed the MagNet DNN classifier, which performs classification and reformer networks against adversarial examples (AEs) on the manifold and standard examples. Another defense method against AEs is Hilbert-based Generative defense introduced by Bai et al. [69], which worked as a pixel CNN on different dimensions and improved their results more accurately. For weak or small adversarial attacks, for example, in crafted attacks in DNN background class, a training process works as a defense [70]. Reference [71] introduced pro-trace vectorization algorithms defense against adversarial attacks on the MNIST digits dataset. The defense obfuscated gradient-based approach [72] gives false sensitive security and was tested on different nine attacks with accurate results. Another defense for adversarial attacks on MNIST digits dataset proposed [64] a Mary EAD elastic-net attack with minimum distortion. Reference [73] suggested that the defense of the adversarial attack on a fuzzy unique image transformation (FUIT) method used down-sampling while using a chest X-ray and a CT image dataset for the diagnosis of COVID19 through a DNN model.

3. Methodology

The proposed methodology performs DR classification using original and perturbed images, and the accuracy is preserved by including different adversarial attacks, such as FGSM and SN DF, in which a speckle noise (SN) attack is a novel attack. The presence of these attacks decreased the model’s credibility and the wrong classification was made. To overcome this problem, adversarial training and feature fusion were proposed as two defensive strategies against adversarial attacks. We performed four distinct training sessions with fine-tuned transfer learning utilizing the Darknet53 model and outcomes in adversarial training. For robust results, we integrated deep and handcrafted features in feature fusion. The handcrafted features included HOG, FHOG(Sv) SFTA FST(Tv), LBP, FLBP(Uv) and FDARK53(xv), which increased the accuracy. All primary steps are shown in Figure 1 and explained in the subsections below.
In this proposed defense model block diagram, we used both the original and perturbed images dataset prepared by using a data augment technique and then resizing the data images into 224 × 224 × 3 . Perturbed images generated through FGSM and SN/MN and DEEP FOOL attacks were applied to the dataset and three types of perturbed images were generated. Adversarial attacks decreased the accuracy with abrupt changes and misclassified the model. To overcome this problem, we proposed the feature fusion defense, in which we combined the deep features using Drknet53 features extracted by applying fine-tuned transfer learning In the proposed model, we use conventional extraction methods of LBP, HOG, and SFTA, which extract robust, immutable features to image, translate and display. These attributes are based on appearance and composition, an essential component in defining the characteristics of the images of an individual’s eyes. This feature fusion works as a robust model against these adversaries, which fooled our network and maintained accuracy with high precision.

3.1. Data Augmentation and Pre Processing

A dataset with 1000 instances from Kaggle has been taken with three diabetic retinopathy classes: DR1, DR2, and DR3 fundus images with minor, moderate and severe conditions. We used a data augmentation approach that involved flipping and flopping at various rotations to construct a new dataset with 6497 images. Regional resolution is the lowest in the retinal images. Since the original dataset images are 2592 × 1728 too large and complicated to be processed further for this dimension, the time spent in the RGB (red/green/blue) channel to produce adversary attacks is reduced, and the images are resized to 224 × 224 × 3 .

3.2. Transfer Learning

Fine-tuned transfer learning is applied; it involves using the features learnt from one issue and a new related concern. The fine-tuning of the DarkNet-53 [76] model involves unfreezing the whole or part of these model structures and re-training it with a meagre learning rate on the new results. This could lead to significant changes by adapting the pre-trained functionality to the new data gradually. Fine-tuning is an interesting activity that entails unfreezing the entire model (or a section of it) and retraining it on new data with a very modest learning rate. By incrementally modifying the pretrained features to the new data, this has the capability for considerable improvements. Using a fully convolutional neural network, a multi-source adversarial transfer learning approach enables the development of a feature representation glucose prediction for diabetic persons. The evaluation is carried out by examining several transfer scenarios using three datasets with considerable inter and intra variation [29]. COVID-19 is diagnosed using Distant Domain Transfer Learning (DDTL) [30]. COVID-19 detection used fine-tuned convolutional neural networks and confined in chest X-ray images [31]. We used fine-tuned transfer learning in our proposed work, which creates a basic model and loads pre-trained weights into it. The FC layer was removed from DarkNet53 and was replaced with the ’new classoutput’ layer. Some convolutions layers were frozen. Various parameters and loss functions were optimized. It was run with a new dataset and the output of one (or more) layers was recorded from the basic model. Feature extraction is the term for this process, using the output as the basis for a new, more compact model.

3.3. Perturbed /Adversarial Image Generation

Many Deep Neural Network (DNN) adversaries have recently been revealed as the source of defects. In addition to the research entry, these disruptions are small and unnoticeable to humans, but the output of the network becomes unpredictable.

3.3.1. Fast Gradient Sign Method (FGSM) Attack Image Generation

The attack changes the entry data to optimize the loss based on the same back-propagated gradients rather than mitigating loss. In other terms, the attack uses the malfunction gradient to change the data to increase the loss [77]. FGSM is based on the standard networks’ principle implementing the gradient descent to set a minimum loss point. We can maximize the loss by only adding a small perturbation in the case of following the sign of gradient descent, described as:
I p r t = I + Δ y l ( I , Z t l ,
where I is an original image Iprt is adversarial image, ∈ is a multiplier to guarantee the perturbations are minor, ∂ are model parameters, l is the classification loss function and Ztl is a true label for original input I. Examples of the FGSM attacked images, which mislead the model, are presented in Figure 2.

3.3.2. Speckle Noise (SN) Attack Image Generation

Frequently known as multiplicative noise or speckle noise (MN/SN), multiplicative noise is less frequent than additive white Gaussian noise (AWGN). However, it is widely used in incoherent image acquisition, including radar and synthetic sonar depth of field, and primarily for medical imaging, using ultrasound and laser imaging techniques. The systematic interference of waves reflected from several primitive scatterers causes speckle to appear in synthetic aperture radar images. This generates pixel-to-pixel intensity variance, which appears as granular noise in SAR images [4]. Because of the system’s function, noise is more complex and challenging to cope with:
  • Each pixel of the original image is composed of noise components.
  • The noise of speckle is not usually distributed and similar to the Rayleigh and Gamma distributions described below:
S = ( I + n × I ) ,
where n is random noise with a mean of 0 and variance of s is uniformly distributed, s is set to 0.50 by default. The value of s might be anywhere between 0 and 1. The mean and variance parameters for the gaussian’, ssian’, 0.50 by localvar’ noise types are always supplied as if the image were of class double in the range, with 0 indicating no noise and 1 indicating a completely noisy image ( 0 , 1 ) .
The medical images are significantly degraded because of these images:
  • Noise is unavoidable in the process of data acquisition.
  • Low contrast due to the variations of lighting and a variety of other causes.
  • Random pixel values for individual pixels of an image can be created by multiplying speckle-noise.
The addition of SN attacks is given in Figure 3.

3.3.3. Deep Fool Attack Generation

Deep Fool (DF) is an opposing attack aimed at taking an example of the closest boundary. In contrast to rough extrapolations of an optimal distributive vector generated by FGSM, according to the authors, this method produced a subtle disturbance. The DF attack uses one loss gradient in l ( f ( k ) and y) as follows [78].
Δ ( I ; b ) = : m i n c c s u b j e c t t o b ( I + c ) b ( I ) .
Here, I is an original image, b is estimated label, c is minimal perturbation.
Deep-Fool describes optimization for a two-class problem as follows. Deep-Fool can have a simple solution for multi-class problems if the classifier is one-vs-all. Here, we mean the classification system of one-vs-all, taking into account two-class concerns, where n are the number of classes which are also the number of discrimination-related functions. However, the one-vs-all method does not apply to a linear machine because one-vs-all essentially manages a series of separating hyperplanes while one-vs-all does not apply. In contrast to rough extrapolations of optimal distributive vector generated by FGSM, according to the authors this method produced stubble disturbance.
The addition of DF attacks is given in Figure 4.

3.4. Proposed Defense against Adversarial Attacks

3.4.1. Adversarial Training (AT)

In the proposed defense method, we have done four adversarial training (AT) on a dataset. In the first training session, we take half of the original data and half of our FGSM adversarial images of three classes: DR1, DR2, DR3, and trained a new prepared dataset from scratch using deep network DarkNet-53 model through fine-tuning and transfer learning. The proposed model extracted the features and made predictions, which were further checked through testing. The testing is performed on the newly trained dataset using original and perturbed images. The accuracy measure increases when the testing is performed on originally trained data in the second adversarial training. In the second adversarial training, two parts of the dataset, images, were included in which half of images of the original dataset and half of speckle-noise (SN) attacks images data set of every class included DR1, DR2, and DR3. In the third training, images were included in which half of images of the original dataset and half of Deepfool (DF) attacks images data set of every class included DR1, DR2, and DR3.
We equally divided the whole data into four parts in this training in which original, the FGSM, SN, DF attacked images were included according to each class data images DR1, DR2, and DR3. This defense is more robust than the first one, because in this training, more data is given, and classifier learns to work best, and model fooling chances are less when compared to the first one. Through the testing process, we check can the defensive model accuracy and robustness.
  • Training 1: original + FGSM attacks images (AT1)
  • Training 2: original + SN attacks images (AT2)
  • Training 3: original + DF attacks images (AT3)
  • Training 4: original + FGSM + DF images (MAT)
Adversarial training architecture of all types of data sets is given in Figure 5 [34,35,52].

3.4.2. Feature Extraction and Feature Fusion Defense

In the proposed method, we use conventional extraction methods of Local Binary Pattern (LBP) [79], Histogram Oriented Gradient (HOG) [80], and Segmentation Based Fractal Texture Analysis (SFTA) [81], which extract robust, immutable features to image, translate, and display. These attributes are based on appearance and composition, which are essential components in defining the characteristics of the images of an individual’s eyes.

3.4.3. Local Binary Pattern (LBP)

LBP is a primary but very effective method that labels the image pixels by threshing every pixel region and takes the output as a binary number. LBPs is a part of the computer vision classification visual descriptor. The LBP descriptor by its specifications, represents the input image. For capturing images such as boundaries, spots, and flat regions used it. Feature vector F _ L B P is calculated as:
L B P x , y = P = 0 P 1 t g n p g c p 2 P
where g n p is the intensity of neighboring pixel, and g c p is the intensity of the central pixel t x , which can be defined as:
t x = 1 i f x > 0 0 i f x < 0 .

3.4.4. Histogram Oriented Gradient (HOG)

The strategy calculates gradient orientation instances in the located sections of an image. This approach is close to that of histograms of edge orientations, scale-invariant descriptors, and shape contexts, but differs in that this method is measured on a dense grid of continuously adjacent cells using local contrast normalization, which overlap for enhanced precision. The number of pixels is specified for each cell, and the histogram of the gradients is then computed for each cell. The Laplacian and Sobel operator give u the direction of HOG. The gradient of f is given as a column vector for a function f ( x , y ) at the coordinates ( x , y ) :
f = G x G y = f x f y
The magnitude of this vector is given by:
f = m a g ( f )
f x , y = t a n 1 f y x , y / f x x , y .
Feature vector F H O G is calculated through it.

3.4.5. Segmentation Based Fractal Texture Analysis (SFTA)

SFTA is an active texture-based extraction method. It gives a set of reliable features of an individual not overemphasised by scaling, rotation, and translation complications. The SFTA characteristics are immune to the “noise” effects of the image. Feature vector F S T is calculated through it. The SFTA functionality is sensitive to the image impact of “noise” SFTA transforms an image of an individual into a binary as an input (referred to below equations):
i b n y k , l = 1 i f r l w < i k , l r u p 0 , o t h e r w i s e
where i b n y k , l is a resulting binary image, an input image is denoted by i ( k , l ) and r l w and r u p upper and lower threshold values. SFTA utilizes a threshold value by multimodal Otsu algorithm [82]. SFTA then calculates the binary fractal calculation border area.
k , l = 1 i f k , l ϵ N 8 k , l ]
l b k , l = 0
l b k , l = 1 , 0 , o t h e r w i s e
where Δ ( k , l ) is consequential boundary image, l b k , l is a binary image and N [ ( k , l ) ] connected pixel value Δ ( k , l ) has value 1, if the corresponding l b k , l has value 1, and otherwise 0. SFTA generates an invariant image vector for scaling, rotation, or translation.

3.5. Deep Feature Extraction

The DarkNet-53 network model contains 53 layers, including input and output layers. Transfer learning is used through DarkNet-53. There are 184 layers in DarkNet-53, including one input layer, 53 convolutions layers, 53 Batch Normalization (BN) layers, 52 Leaky ReLU, 23 Addition, 1 Global Average pooling layer, and 1 classification output softmax layer. The image size of the input of a network is 256 × 256 . The detector module consists of several Conv layers clustered in blocks, up-sampling layers, and three Conv layers, which are linearly activated and allow detections at three different scales. There is no max-pooling layer present in DarkNet-53. Instead, it uses BN and leaky RELU layers for every convolution step.For deep features extraction using DNN, we used trained Darknet53 from starch to classify the different stages of diabetic retinopathy which included DR1, DR2, and DR3.Darknet53 is used in image processing many tasks included object detection, real-time object detection YOLO, image classification, segmentation, model compression, fruit classification [52] etc. In medical imaging darknet53 used for detection of covid-19 [34] computed aided covid-19 detection [36] for MRI scan brain tumor data augmentation [35] YOLO V3 has been used to identify red lesions in retinal fundus images [37]. smart medical autonomous distributed system for diagnosis [38], melanoma detection [39].

Feature Fusion (FF)

We have fused the hand-crafted features with deep features to obtain a single vector. A serial feature fusion approach is used in the proposed method. The feature vector obtained is more efficient, since it includes additional information than these, which we obtain by using a single extractor procedure.
Three function vectors like the HOG, SFTA, and LBP are allowed by the suggested method F H O G S v , F S T T v , F L B P U v deep characteristic allows for the DarkNet-53 F D A R K 53 x v . I × J is the dimensions of it. The classification os performed using the following equation:
F i * j = S v + T v + U v + F D a r k n e t 53 .

4. Results and Discussion

In this section, the experimental results are presented. Table 2 shows the experimental results obtained on the attacks’ images, three different attacked images, fast gradient sign method (FGSM) attack images, speckle noise (SN) attack images, and deep fool (DF) attacked images. The unexpected changes in accuracy results occur in this part of the experimentation.
Table 2 shows abrupt changes in the prediction, when testing the attacked images with the originally trained network; they wrongly predicted their classes with maximum accuracy all wrongly predicted values highlighted in this tables, which shows wrong labels of each class. When FGSM attacked images of class DR1 were tested, it predicted as belonging to class DR2 with 93.01% accuracy, while they are from class DR1 same as for other classes of DR, class DR3 attacked images were misclassified. When speckle-noise attacked images of class DR2 were tested, they are categorised as class DR3 with 79.09%. The other two classes images wrongly predicted with a high precision rate due to a speckle-noise attack. For DF, when class DR3 image was tested, it was predicted into class DR2 with 99.75% accuracy, while DR1, DR2 classes images were also wrongly labelled.
Table 3 shows the testing accuracy of FGSM attack Dataset images, SN attack images, and DF attack images (Adversarial Training AT1). In the first training session, we take a half of the original data and a half of our FGSM adversarial images of three classes: DR1, DR2, DR3, and trained a new dataset from scratch using DarkNet-53 through fine-tuning transfer learning. The model extracted the features and made predictions which checked through testing. When testing performed on this newly trained network using original and perturbed images, the accuracy measure increase compared to already testing performed on the trained network initially.
When the adversarial training AT1 network is tested through the FGSM attack images, the results are presented in Table 2, in which the FGSM attacked image was misclassified through maximum 93.01% is classified into DR2 class after this training correctly labelled with 88.86% in DR1 class, and 0% chance that it belongs to DR2 class same as in the other class images. In the SN and DF attacked images, some images were correctly labelled with 40% and 64.51% accuracy.
Table 4 shows the testing accuracy of the FGSM attacks dataset images, SN attacks images, and DF attacks images with Adversarial Training AT2. In the second adversarial training, two parts of the dataset, images were included in which a half of images of the original dataset and half of the speckle noise (SN) attacks images data set of every class included DR1, DR2, and DR3 classes, and network trained using fine-tuned transfer learning and testing performed results are shown in Table 4.
When adversarial training AT2 network was tested through attack images, the detected anomaly resolved all the SN attacked images classified with high 98.02%, 88.66%, 94.07% accuracy and the other two attacked images are also correctly classified.
Table 5 showed testing accuracy of FGSM attacks images, SN attack images, and DF attack images with Adversarial Training AT3. In the third training, images were included in which a half of images of the original dataset and a half of Deep-fool (DF) attacks image data set of every class included DR1, DR2, and DR3 classes. The network trained using fine-tuned transfer learning and testing performed on all types of attacked images and the result is shown in Table 5.
When adversarial training AT3 network tested through attack images, the anomaly present in the previous tables was resolved. Most of the FGSM and SN attacked images were correctly labeled with a maximum accuracy, and all the DF attacked images were correctly classified with 82.97%, 99.96%, and 100% accuracy.
Table 6 shows the testing accuracy of the FGSM attack dataset images, SN attack images, and DF attack images with Mixed Adversarial Training (MAT). We equally divided the whole data into four parts in which the original FGSM, SN, and DF attacked images were included according to each class of images DR1, DR2, and DR3. This defense is more robust than the previous scenarios, because more data is given in this training, and classifier learns to work best, and model fooling chances are less. Through the testing process, we check the accuracy and robustness of the defensive model.
Through mixed adversarial training (MAT), the results of MAT were obtained in which the majority or attacked images were correctly classified with high accuracy and precision, and the trained model became most robust. To incorporate and deal with the adversarial attacks, we performed the adversarial training on mixed data, and the testing results revealed that almost half of the labels were predicted correctly. Moreover, we also used the original, FGSM, SN, and DF datasets for adversarial training, which performs efficiently and labels the majority of the labels correctly. The summary of the defensive proposed model is presented in Table 7.

Feature Extraction and Feature Fusion Defense

A serial feature fusion approach is used in the proposed method. Three function vectors from HOG, SFTA, and LBP methods are allowed by the suggested method F H O G S v , F S T T v , F L B P U v deep characteristic allows for the DarkNet-53 F D A R K 53 x v . I × J is the dimensions of it. Our proposed feature fusion method achieved robust results on adversary images and classified correctly. The results obtained using feature fusion approaches are given in Table 8 and Table 9. Furthermore, the ROC curve and fusion scatter plots for the proposed method are visualized in Figure 6.

5. Conclusions

The attacks against artificial intelligence models can decrease model performance. We have proposed an adversarial training based defense against adverse disruptions in order to address this problem. The method improvement includes not simply an advanced understanding of image processing techniques, but also needs essential medical input, including expert knowledge related to diabetic retinopathy and its screening procedure, in addition to the eye fundus photography process.
To mitigate the impact of adversarial attacks, we have executed different kinds of adversarial training, through which the adverse effect is reduced, and with which the model cannot become fooled when compared to existing models. Results obtained are 92% correct and prove that the proposed defensive model is robust. Another defense model based on feature fusion was also proposed for adversarial attacks, in which deep and handcrafted features were fused, including the DarkNet-53 deep features and LBP, HOG, SFTA features, and the accuracy was increased by 99.9%.

Author Contributions

Conceptualization, S.U.R. and J.H.S.; Data curation, S.L.; Formal analysis, T.M., H.T.R. and R.D.; Funding acquisition, R.D.; Investigation, S.L., J.H.S., T.M. and H.T.R.; Methodology, S.U.R., J.H.S. and M.A.M.; Software, S.L.; Supervision, K.H.A.; Validation, S.L., S.U.R., M.A.M. and K.H.A.; Writing—original draft, S.L., S.U.R., J.H.S., T.M., H.T.R. and M.A.M.; Writing—review & editing, R.D. All authors have read and agreed to the published version of the manuscript.

Funding

No funding received for this research.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Albahli, S.; Rauf, H.T.; Arif, M.; Nafis, M.T.; Algosaibi, A. Identification of Thoracic Diseases by Exploiting Deep Neural Networks. Neural Netw. 2021, 5, 6. [Google Scholar]
  2. Albahli, S.; Rauf, H.T.; Algosaibi, A.; Balas, V.E. AI-driven deep CNN approach for multi-label pathology classification using chest X-Rays. PeerJ Comput. Sci. 2021, 7, e495. [Google Scholar] [CrossRef] [PubMed]
  3. Abdulsahib, A.A.; Mahmoud, M.A.; Mohammed, M.A.; Rasheed, H.H.; Mostafa, S.A.; Maashi, M.S. Comprehensive review of retinal blood vessel segmentation and classification techniques: Intelligent solutions for green computing in medical images, current challenges, open issues, and knowledge gaps in fundus medical images. Netw. Model. Anal. Health Inform. Bioinform. 2021, 10, 1–32. [Google Scholar]
  4. Canedo, D.; Neves, A.J.R. Facial Expression Recognition Using Computer Vision: A Systematic Review. Appl. Sci. 2019, 9, 4678. [Google Scholar] [CrossRef] [Green Version]
  5. Kour, N.; Sunanda; Arora, S. Computer-vision based diagnosis of Parkinson’s disease via gait: A survey. IEEE Access 2019, 7, 156620–156645. [Google Scholar] [CrossRef]
  6. Mohammed, M.A.; Elhoseny, M.; Abdulkareem, K.H.; Mostafa, S.A.; Maashi, M.S. A Multi-agent Feature Selection and Hybrid Classification Model for Parkinson’s Disease Diagnosis. ACM Trans. Multimid. Comput. Commun. Appl. 2021, 17, 1–22. [Google Scholar] [CrossRef]
  7. Rauf, H.T.; Lali, M.I.U.; Zahoor, S.; Shah, S.Z.H.; Rehman, A.U.; Bukhari, S.A.C. Visual features based automated identification of fish species using deep convolutional neural networks. Comput. Electron. Agric. 2019, 167, 105075. [Google Scholar] [CrossRef]
  8. Rauf, H.T.; Saleem, B.A.; Lali, M.I.U.; Khan, M.A.; Sharif, M.; Bukhari, S.A.C. A citrus fruits and leaves dataset for detection and classification of citrus diseases through machine learning. Data Brief 2019, 26, 104340. [Google Scholar] [CrossRef] [PubMed]
  9. Ahuja, A.S. The impact of artificial intelligence in medicine on the future role of the physician. PeerJ 2019, 7, e7702. [Google Scholar] [CrossRef] [PubMed]
  10. Van Timmeren, J.E.; Cester, D.; Tanadini-Lang, S.; Alkadhi, H.; Baessler, B. Radiomics in medical imaging—“How-to” guide and critical reflection. Insights Imaging 2020, 11, 1–16. [Google Scholar] [CrossRef]
  11. Mutlag, A.A.; Khanapi Abd Ghani, M.; Mohammed, M.A.; Maashi, M.S.; Mohd, O.; Mostafa, S.A.; Abdulkareem, K.H.; Marques, G.; de la Torre Díez, I. MAFC: Multi-agent fog computing model for healthcare critical tasks management. Sensors 2020, 20, 1853. [Google Scholar] [CrossRef] [Green Version]
  12. Lambin, P.; Leijenaar, R.T.; Deist, T.M.; Peerlings, J.; De Jong, E.E.; Van Timmeren, J.; Sanduleanu, S.; Larue, R.T.; Even, A.J.; Jochems, A. Radiomics: The bridge between medical imaging and personalized medicine. Nat. Rev. Clin. Oncol. 2017, 14, 749–762. [Google Scholar] [CrossRef] [PubMed]
  13. Kuziemsky, C.; Maeder, A.J.; John, O.; Gogia, S.B.; Basu, A.; Meher, S.; Ito, M. Role of Artificial Intelligence within the Telehealth Domain. Yearb. Med. Inform. 2019, 28, 035–040. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Zhou, X.; Ma, Y.; Zhang, Q.; Mohammed, M.A.; Damaševičius, R. A Reversible Watermarking System for Medical Color Images: Balancing Capacity, Imperceptibility, and Robustness. Electronics 2021, 10, 1024. [Google Scholar] [CrossRef]
  15. Mohammed, M.A.; Abdulkareem, K.H.; Mostafa, S.A.; Ghani, M.K.A.; Maashi, M.S.; Garcia-Zapirain, B.; Oleagordia, I.; Alhakami, H.; Al-Dhief, F.T. Voice pathology detection and classification using convolutional neural network model. Appl. Sci. 2020, 10, 3723. [Google Scholar] [CrossRef]
  16. Ruta, L.; Magliano, D.; Lemesurier, R.; Taylor, H.; Zimmet, P.; Shaw, J. Prevalence of diabetic retinopathy in Type 2 diabetes in developing and developed countries. Diabet. Med. 2013, 30, 387–398. [Google Scholar] [CrossRef]
  17. Gulshan, V.; Peng, L.; Coram, M.; Stumpe, M.C.; Wu, D.; Narayanaswamy, A.; Venugopalan, S.; Widner, K.; Madams, T.; Cuadros, J. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. Jama 2016, 316, 2402–2410. [Google Scholar] [CrossRef] [PubMed]
  18. Orujov, F.; Maskeliūnas, R.; Damaševičius, R.; Wei, W. Fuzzy based image edge detection algorithm for blood vessel detection in retinal images. Appl. Soft Comput. J. 2020, 94. [Google Scholar] [CrossRef]
  19. Ramasamy, L.; Padinjappurathu, S.; Kadry, S.; Damaševičius, R. Detection of diabetic retinopathy using a fusion of textural and ridgelet features of retinal images and sequential minimal optimization classifier. PeerJ Comput. Sci. 2021, 7, 456. [Google Scholar] [CrossRef]
  20. Tajbakhsh, N.; Jeyaseelan, L.; Li, Q.; Chiang, J.N.; Wu, Z.; Ding, X. Embracing imperfect datasets: A review of deep learning solutions for medical image segmentation. Med. Image Anal. 2020, 63, 101693. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Karimi, D.; Dou, H.; Warfield, S.K.; Gholipour, A. Deep learning with noisy labels: Exploring techniques and remedies in medical image analysis. Med. Image Anal. 2020, 65, 101759. [Google Scholar] [CrossRef] [PubMed]
  22. Qiu, S.; Liu, Q.; Zhou, S.; Wu, C. Review of Artificial Intelligence Adversarial Attack and Defense Technologies. Appl. Sci. 2019, 9, 909. [Google Scholar] [CrossRef] [Green Version]
  23. Gluck, T.; Kravchik, M.; Chocron, S.; Elovici, Y.; Shabtai, A. Spoofing Attack on Ultrasonic Distance Sensors Using a Continuous Signal. Sensors 2020, 20, 6157. [Google Scholar] [CrossRef]
  24. Zhou, X.; Xu, M.; Wu, Y.; Zheng, N. Deep Model Poisoning Attack on Federated Learning. Future Internet 2021, 13, 73. [Google Scholar] [CrossRef]
  25. Chakraborty, A.; Alam, M.; Dey, V.; Chattopadhyay, A.; Mukhopadhyay, D. Adversarial attacks and defences: A survey. arXiv 2018, arXiv:1810.00069. [Google Scholar]
  26. Edwards, D.; Rawat, D.B. Study of Adversarial Machine Learning with Infrared Examples for Surveillance Applications. Electronics 2020, 9, 1284. [Google Scholar] [CrossRef]
  27. Ren, K.; Zheng, T.; Qin, Z.; Liu, X. Adversarial Attacks and Defenses in Deep Learning. Engineering 2020, 6, 346–360. [Google Scholar] [CrossRef]
  28. Nazemi, A.; Fieguth, P. Potential adversarial samples for white-box attacks. arXiv 2019, arXiv:1912.06409. [Google Scholar]
  29. Lin, J.; Xu, L.; Liu, Y.; Zhang, X. Black-box adversarial sample generation based on differential evolution. J. Syst. Softw. 2020, 170, 110767. [Google Scholar] [CrossRef]
  30. Alzantot, M.; Sharma, Y.; Chakraborty, S.; Zhang, H.; Hsieh, C.J.; Srivastava, M.B. Genattack: Practical black-box attacks with gradient-free optimization. In Proceedings of the Genetic and Evolutionary Computation Conference, Prague, Czech Republic, 13–17 July 2019; pp. 1111–1119. [Google Scholar]
  31. Deng, L. The mnist database of handwritten digit images for machine learning research. IEEE Signal Process. Mag. 2012, 29, 141–142. [Google Scholar] [CrossRef]
  32. Krizhevsky, A.; Nair, V.; Hinton, G. CIFAR-10; Canadian Institute for Advanced Research: Toronto, ON, Canada, 2009. [Google Scholar]
  33. Gao, X.; Tan, Y.A.; Jiang, H.; Zhang, Q.; Kuang, X. Boosting targeted black-box attacks via ensemble substitute training and linear augmentation. Appl. Sci. 2019, 9, 2286. [Google Scholar] [CrossRef] [Green Version]
  34. Tabacof, P.; Tavares, J.; Valle, E. Adversarial images for variational autoencoders. arXiv 2016, arXiv:1612.00155. [Google Scholar]
  35. Kurakin, A.; Goodfellow, I.; Bengio, S. Adversarial machine learning at scale. arXiv 2016, arXiv:1611.01236. [Google Scholar]
  36. Gu, S.; Rigazio, L. Towards deep neural network architectures robust to adversarial examples. arXiv 2014, arXiv:1412.5068. [Google Scholar]
  37. Siddique, A.; Browne, W.N.; Grimshaw, G.M. Lateralized learning for robustness against adversarial attacks in a visual classification system. In Proceedings of the 2020 Genetic and Evolutionary Computation Conference, Cancún, Mexico, 8–12 July 2020; pp. 395–403. [Google Scholar]
  38. Huq, A.; Pervin, M. Adversarial Attacks and Defense on Textual Data: A Review. arXiv 2020, arXiv:2005.14108. [Google Scholar]
  39. Zhang, J.; Sang, J.; Zhao, X.; Huang, X.; Sun, Y.; Hu, Y. Adversarial Privacy-preserving Filter. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 1423–1431. [Google Scholar]
  40. Wang, Y.; Wang, K.; Zhu, Z.; Wang, F.Y. Adversarial attacks on Faster R-CNN object detector. Neurocomputing 2020, 382, 87–95. [Google Scholar] [CrossRef]
  41. Li, Y.; Zhu, Z.; Zhou, Y.; Xia, Y.; Shen, W.; Fishman, E.K.; Yuille, A.L. Volumetric Medical Image Segmentation: A 3D Deep Coarse-to-Fine Framework and Its Adversarial Examples. In Deep Learning and Convolutional Neural Networks for Medical Imaging and Clinical Informatics; Springer: Cham, Switzerland, 2019; pp. 69–91. [Google Scholar]
  42. Zhang, W.E.; Sheng, Q.Z.; Alhazmi, A.; Li, C. Adversarial attacks on deep-learning models in natural language processing: A survey. ACM Trans. Intell. Syst. Technol. 2020, 11, 1–41. [Google Scholar]
  43. Yu, Y.; Lee, H.J.; Kim, B.C.; Kim, J.U.; Ro, Y.M. Investigating Vulnerability to Adversarial Examples on Multimodal Data Fusion in Deep Learning. arXiv 2020, arXiv:2005.10987. [Google Scholar]
  44. Raval, N.; Verma, M. One word at a time: Adversarial attacks on retrieval models. arXiv 2020, arXiv:2008.02197. [Google Scholar]
  45. Levine, A.; Feizi, S. (De) Randomized Smoothing for Certifiable Defense against Patch Attacks. arXiv 2020, arXiv:2002.10733. [Google Scholar]
  46. Wang, H.; Wang, G.; Li, Y.; Zhang, D.; Lin, L. Transferable, Controllable, and Inconspicuous Adversarial Attacks on Person Re-identification With Deep Mis-Ranking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 342–351. [Google Scholar]
  47. Fawaz, H.I.; Forestier, G.; Weber, J.; Idoumghar, L.; Muller, P.A. Adversarial attacks on deep neural networks for time series classification. In Proceedings of the 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019; pp. 1–8. [Google Scholar]
  48. Yang, Z.; Zhao, Y.; Yan, W. Adversarial Vulnerability in Doppler-based Human Activity Recognition. In Proceedings of the 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, UK, 19–24 July 2020; pp. 1–7. [Google Scholar]
  49. Dong, Y.; Su, H.; Wu, B.; Li, Z.; Liu, W.; Zhang, T.; Zhu, J. Efficient decision-based black-box adversarial attacks on face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 7714–7722. [Google Scholar]
  50. Hafemann, L.G.; Sabourin, R.; Oliveira, L.S. Characterizing and evaluating adversarial examples for Offline Handwritten Signature Verification. IEEE Trans. Inf. Forensics Secur. 2019, 14, 2153–2166. [Google Scholar] [CrossRef] [Green Version]
  51. García, J.; Majadas, R.; Fernández, F. Learning adversarial attack policies through multi-objective reinforcement learning. Eng. Appl. Artif. Intell. 2020, 96, 104021. [Google Scholar] [CrossRef]
  52. Zahoor, S.; Lali, I.U.; Khan, M.A.; Javed, K.; Mehmood, W. Breast cancer detection and classification using traditional computer vision techniques: A comprehensive review. Curr. Med. Imaging 2020, 16, 1187–1200. [Google Scholar] [CrossRef]
  53. Patrício, D.I.; Rieder, R. Computer vision and artificial intelligence in precision agriculture for grain crops: A systematic review. Comput. Electron. Agric. 2018, 153, 69–81. [Google Scholar] [CrossRef] [Green Version]
  54. Saman, G.; Gohar, N.; Noor, S.; Shahnaz, A.; Idress, S.; Jehan, N.; Rashid, R.; Khattak, S.S. Automatic detection and severity classification of diabetic retinopathy. Multimed. Tools Appl. 2020, 79, 31803–31817. [Google Scholar] [CrossRef]
  55. Cheng, Y.; Juefei-Xu, F.; Guo, Q.; Fu, H.; Xie, X.; Lin, S.W.; Lin, W.; Liu, Y. Adversarial Exposure Attack on Diabetic Retinopathy Imagery. arXiv 2020, arXiv:2009.09231. [Google Scholar]
  56. Hirano, H.; Minagi, A.; Takemoto, K. Universal adversarial attacks on deep neural networks for medical image classification. BMC Med. Imaging 2021, 21, 1–13. [Google Scholar] [CrossRef]
  57. Kang, X.; Song, B.; Du, X.; Guizani, M. Adversarial Attacks for Image Segmentation on Multiple Lightweight Models. IEEE Access 2020, 8, 31359–31370. [Google Scholar] [CrossRef]
  58. Pineda, L.; Basu, S.; Romero, A.; Calandra, R.; Drozdzal, M. Active MR k-space sampling with reinforcement learning. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Cham, Switzerland, 2020; pp. 23–33. [Google Scholar]
  59. Chen, C.; Qin, C.; Qiu, H.; Ouyang, C.; Wang, S.; Chen, L.; Tarroni, G.; Bai, W.; Rueckert, D. Realistic adversarial data augmentation for MR image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Cham, Switzerland, 2020; pp. 667–677. [Google Scholar]
  60. Liu, S.; Setio, A.A.A.; Ghesu, F.C.; Gibson, E.; Grbic, S.; Georgescu, B.; Comaniciu, D. No Surprises: Training Robust Lung Nodule Detection for Low-Dose CT Scans by Augmenting with Adversarial Attacks. arXiv 2020, arXiv:2003.03824. [Google Scholar] [CrossRef] [PubMed]
  61. Paul, R.; Schabath, M.; Gillies, R.; Hall, L.; Goldgof, D. Mitigating adversarial attacks on medical image understanding systems. In Proceedings of the 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI), Iowa City, IA, USA, 3–7 April 2020; pp. 1517–1521. [Google Scholar]
  62. Ding, Y.; Wu, G.; Chen, D.; Zhang, N.; Gong, L.; Cao, M.; Qin, Z. DeepEDN: A Deep Learning-based Image Encryption and Decryption Network for Internet of Medical Things. arXiv 2020, arXiv:2004.05523. [Google Scholar] [CrossRef]
  63. Anand, D.; Tank, D.; Tibrewal, H.; Sethi, A. Self-Supervision vs. Transfer Learning: Robust Biomedical Image Analysis Against Adversarial Attacks. In Proceedings of the 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI), Iowa City, IA, USA, 3–7 April 2020; pp. 1159–1163. [Google Scholar]
  64. Sharma, Y.; Chen, P.Y. Attacking the Madry Defense Model with L1-based Adversarial Examples. arXiv 2017, arXiv:1710.10733. [Google Scholar]
  65. Liu, N.; Du, M.; Guo, R.; Liu, H.; Hu, X. Adversarial Machine Learning: An Interpretation Perspective. ACM SIGKDD Explor. Newsl. 2021, 23, 86–99. [Google Scholar] [CrossRef]
  66. Agarwal, A.; Singh, R.; Vatsa, M.; Ratha, N.K. Image transformation based defense against adversarial perturbation on deep learning models. IEEE Trans. Dependable Comput. Secur. 2020. [Google Scholar] [CrossRef]
  67. Huang, X.; Kroening, D.; Ruan, W.; Sharp, J.; Sun, Y.; Thamo, E.; Wu, M.; Yi, X. A survey of safety and trustworthiness of deep neural networks: Verification, testing, adversarial attack and defence, and interpretability. Comput. Sci. Rev. 2020, 37, 100270. [Google Scholar] [CrossRef]
  68. Meng, D.; Chen, H. Magnet: A two-pronged defense against adversarial examples. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, Dallas, TX, USA, 30 October–3 November 2017; pp. 135–147. [Google Scholar]
  69. Bai, Y.; Feng, Y.; Wang, Y.; Dai, T.; Xia, S.T.; Jiang, Y. Hilbert-based generative defense for adversarial examples. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 4784–4793. [Google Scholar]
  70. McCoyd, M.; Wagner, D. Background class defense against adversarial examples. In Proceedings of the 2018 IEEE Security and Privacy Workshops (SPW), San Francisco, CA, USA, 24 May 2018; pp. 96–102. [Google Scholar]
  71. Kabilan, V.M.; Morris, B.; Nguyen, H.P.; Nguyen, A. Vectordefense: Vectorization as a defense to adversarial examples. In Soft Computing for Biomedical Applications and Related Topics; Springer: Cham, Switzerland, 2018; pp. 19–35. [Google Scholar]
  72. Athalye, A.; Carlini, N.; Wagner, D. Obfuscated gradients give a false sense of security: Circumventing defenses to adversarial examples. arXiv 2018, arXiv:1802.00420. [Google Scholar]
  73. Tripathi, A.M.; Mishra, A. Fuzzy Unique Image Transformation: Defense against Adversarial Attacks on Deep COVID-19 Models. arXiv 2020, arXiv:2009.04004. [Google Scholar]
  74. Xu, W.; Evans, D.; Qi, Y. Feature squeezing: Detecting adversarial examples in deep neural networks. arXiv 2017, arXiv:1704.01155. [Google Scholar]
  75. Liu, C.; Ye, D. Defend Against Adversarial Samples by Using Perceptual Hash. Comput. Mater. Contin. 2020, 62, 1365–1386. [Google Scholar] [CrossRef]
  76. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  77. Zheng, H.; Zhang, Z.; Gu, J.; Lee, H.; Prakash, A. Efficient adversarial training with transferable adversarial examples. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1181–1190. [Google Scholar]
  78. Moosavi-Dezfooli, S.; Fawzi, A.; Frossard, P.; Deepfool. A simple and accurate method to fool deep neural networks. In Proceedings of the CVPR, Boston, MA, USA, 8–10 June 2015; pp. 2574–2582. [Google Scholar]
  79. Ojala, T.; Pietikäinen, M.; Harwood, D. A comparative study of texture measures with classification based on featured distributions. Pattern Recognit. 1996, 29, 51–59. [Google Scholar] [CrossRef]
  80. Dalal, N.; Triggs, B. Histograms of Oriented Gradients for Human Detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005. [Google Scholar]
  81. Costa, A.F.; Humpire-Mamani, G.; Traina, A.J.M. An Efficient Algorithm for Fractal Analysis of Textures. In Proceedings of the 2012 25th SIBGRAPI Conference on Graphics, Patterns and Images, Ouro Preto, Brazil, 22–25 August 2012. [Google Scholar]
  82. Otsu, N. A Threshold Selection Method from Gray-Level Histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Block diagram of the proposed system.
Figure 1. Block diagram of the proposed system.
Sensors 21 03922 g001
Figure 2. Addition of FGSM Attacks. The first row shows the original images, while the second row represents the FGSM attacked images that mislead the model.
Figure 2. Addition of FGSM Attacks. The first row shows the original images, while the second row represents the FGSM attacked images that mislead the model.
Sensors 21 03922 g002
Figure 3. Addition of FGSM attacks. The first row shows the original images, while the second row represents the SN attacked images.
Figure 3. Addition of FGSM attacks. The first row shows the original images, while the second row represents the SN attacked images.
Sensors 21 03922 g003
Figure 4. Addition of DF attacks. The first row shows the original images, while the second row indicates the DF attacked images.
Figure 4. Addition of DF attacks. The first row shows the original images, while the second row indicates the DF attacked images.
Sensors 21 03922 g004
Figure 5. Illustration of the adversarial training process.
Figure 5. Illustration of the adversarial training process.
Sensors 21 03922 g005
Figure 6. ROC curve (left) and fusion scatter plot (right).
Figure 6. ROC curve (left) and fusion scatter plot (right).
Sensors 21 03922 g006
Table 1. Comparison of recent related works with their datasets.
Table 1. Comparison of recent related works with their datasets.
ReferenceMethodologyDatasetEvaluation MeasuresResults
[54]Morphological operationDiaretDBSVM classifierMild severity DR detection and classification
[55]Bracketed Exposure Fusion (BEF) and Convolution Bracketed Exposure Fusion (CBEF) AttacksEyepacsComponent-wise multiplicative fusion and element-wise convolutionalDR detection
[56]Iterative algorithms for universal perturbations attacks (UPA)Multiple datasetsClassification of targeted and non-targeted UPA attacks80% Accuracy
[57]Local perturbation and universal attacks,CityscapesNoise function and Gradient of pixelsImage Segmentation
[58]Reinforcement learning, Markov Decision ProcessMRI single-coil knee datasetMSE, NMSE, SSIM and PSNRMRI phase-encoding sampling
[64]Adversarial training by modelling intensity inhomogeneitiesAutomated Cardiac Diagnosis Challenge (ACDC)Low-shot learning, learning from limited populationSemantic features for cardiac image segmentation
[60]Projected gradient descent (PGD), adverse synthetic nodule and adverse perturbationCT dataFalse positive reduction rateLung nodule detection and prediction of lung by false positive reduction (FPR)
[61]Fast Gradient Sign Method (FGSM) and one-pixel attacksNational Lung Screening Trial (NLST) datasetEnsemble-based classificationMalignancy prediction of lung nodules. 1-pixel attack with 82.27% and 81.43%
[62]Cycle-generative adversarial network (Cycle-GAN)Chest X-ray data setX-ray dataset through ROI (region of interest)Encrypting and decrypting the medical image through DeepEDN
[63]Self-supervised transfer learning combined with adversarial trainingChest X-rays, and segmentation of MRI images.MRI segmentation using two PGD, and fast gradient single methodPneumonia classification of x-ray images and MRI segmentation
[65]Untargeted vs Targeted Attack, One-Shot vs Iterative AttackFashion-MNIST datasetFeature-level interpretation and model-level interpretationDefensive graph-based models, causal models generated
[66]Discrete Wavelet Transform and Discrete Sine TransformObject database (validation set of ImageNet) and face recognition (MBGC)SVM ClassifierDefense through which adversarial perturbation can be neutralized
[67]Dimensionality reduction, a characterization of the adversarial region,Multiple datasetCombining input discretization with adversarial trainingActivation transformations for the best and robust defense against these attacks
[68]MagNet with RandomizationAdversarial examples (AEs) on a manifold and normal examples.MagNet DNN classifier3% higher than simple MagNet.
[69]Hilbert-based Generative pixel CNN Hilbert-based PixelDefend (HPD)Adversarial examples (AEs)Ensemble of Hilbert curve with different orientations.PixelDefend mapping pixels from 2-D to 1-D.
[70]Crafts attacks, Background class image classification trainingEMNIST DatasetWeak or small adversarial attacks samples basedConstructing background images between the key classes and artificially expanding the background data
[71]Protrace vectorization algorithmsMNIST handwritten digits datasetIn high-dimensional color image space, simple image tracing may not yield compact and interpretable elements.the vector images are resolution-independent, one could rasterize them back into much smaller-sized images.
[72]Obfuscated Gradients, iterative optimization-based attacks,ICLR 2018False sensitive securityPrevent gradient descent-based attacks) for perceived robustness
[64]Mary EAD elastic-net attack with L∞MNIST digits datasetlocal first-order information, Minimum distortionEAD is able to outperform PGD in transferring in the targeted case.
[73]Fuzzy Unique Image Transformation (FUIT)Chest x-ray and CT image datasetthat downsamples the image pixels into an interval.Diagnosis of COVID-19 through DNN model.
[74]Feature SqueezingMNIST, CIFAR-10, ImageNetJoint detection with multiple squeezers, adversarial adaptationColor depth reduction, median smoothing. non local smoothing
[75]Perceptual hashCIFAR-10JSMA gradient-based attack, One Pixel Attack is an evolutionary-based attackWhite-box attack success rate 36.3%, and in black-box attack 72.8%
Table 2. Testing of Original Training Network with Attack Images.
Table 2. Testing of Original Training Network with Attack Images.
Original Class LabelAttacks AppliedPredicted Label After AttackAccuracy with Class DR1 (%)Accuracy with class DR2 (%)Accuracy with Class DR3 (%)
DR1FGSMDR2093.016.99
DR2FGSMDR181.71018.29
DR3FGSMDR2091.098.91
DR1SNDR212.2787.980
DR2SNDR310.0910.8279.09
DR3SNDR189.8210.180
DR1DFDR201000
DR2DFDR18217.590.41
DR3DFDR2099.750.41
Table 3. Testing of Adversarial Training AT1 Network with Attack Images.
Table 3. Testing of Adversarial Training AT1 Network with Attack Images.
Original Class LabelAttacks AppliedPredicted Label After AttackAccuracy with Class DR1 (%)Accuracy with Class DR2 (%)Accuracy with Class DR3 (%)
DR1FGSMDR188.86011.14
DR2FGSMDR22072.077.93
DR3FGSMDR310.018.9481.05
DR1SNDR24057.232.77
DR2SNDR170030
DR3SNDR30040
DR1DFDR3040.9758.50
DR2DFDR189.790.2110.0
DR3DFDR3035.3264.51
Table 4. Testing of Adversarial Training AT2 Network with Attack Images.
Table 4. Testing of Adversarial Training AT2 Network with Attack Images.
Original Class LabelAttacks AppliedPredicted Label After AttackAccuracy with Class DR1 (%)Accuracy with Class DR2 (%)Accuracy with Class DR3 (%)
DR1FGSMDR32.7120.376.98
DR2FGSMDR2062.3737.26
DR3FGSMDR181.344.713.96
DR1SNDR198.021.980
DR2SNDR20.0288.6611.32
DR3SNDR305.9894.07
DR1DFDR214.6271.813.58
DR2DFDR3010.1589.95
DR3DFDR182.751.815.45
Table 5. Testing of Adversarial Training AT3 Network with Attack Images.
Table 5. Testing of Adversarial Training AT3 Network with Attack Images.
Original Class LabelAttacks AppliedPredicted Label After AttackAccuracy with Class DR1 (%)Accuracy with Class DR2 (%)Accuracy with Class DR3 (%)
DR1FGSMDR168.0520.9720.09
DR2FGSMDR329.9110.9150.09
DR3FGSMDR33.9035.3274.51
DR1SNDR339.91060.09
DR2SNDR29.8790.020.01
DR3SNDR3045.3254.51
DR1DFDR182.9717.020
DR2DFDR20.0499.960
DR3DFDR300100
Table 6. Testing of Mixed Adversarial Training MAT Network with Attack Images.
Table 6. Testing of Mixed Adversarial Training MAT Network with Attack Images.
Original Class LabelAttacks AppliedPredicted Label After AttackAccuracy with Class DR1 (%)Accuracy with Class DR2 (%)Accuracy with Class DR3 (%)
DR1FGSMDR199.830.110
DR2FGSMDR223.5274.941.54
DR3FGSMDR32.890.0297.09
DR1SNDR194.464.830.71
DR2SNDR20991
DR3SNDR317.9082.09
DR1DFDR196.5103.29
DR2DFDR201000
DR3DFDR300.0199.99
Table 7. Summary of proposed defense model.
Table 7. Summary of proposed defense model.
Training DatasetTesting DatasetCorrect Label Prediction %
Original DatasetOriginal dataset100%
Original DatasetFGSM Attacked Dataset0%
Original DatasetSN Attacked Dataset10%
Original DatasetDF Attacked Dataset0%
Adversarial Training (AT1)Original+ FGSM62%
Adversarial Training (AT2)Original+SN52%
Adversarial Training (AT3)Original+DF66%
Adversarial Training (Mixed Data MAT)Original+FGSM+SN+DF92%
Table 8. Accuracy obtained using feature fusion approaches on different models.
Table 8. Accuracy obtained using feature fusion approaches on different models.
ModelSVMKNN (Cubic)Ensemble
DarkNet-5380.9%79.6%90.3%
HOG+SFTA+LBP82.3%84.1%85.5%
Proposed Model99.9%99.5%99.9%
Table 9. Results obtained using feature fusion approaches on different target classes.
Table 9. Results obtained using feature fusion approaches on different target classes.
ClassNo of InstancesAccuracy (%)PrecisionRecallF1-Score
DR1254399.941.01.01.0
DR2250999.951.01.01.0
DR3159199.981.01.01.0
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lal, S.; Rehman, S.U.; Shah, J.H.; Meraj, T.; Rauf, H.T.; Damaševičius, R.; Mohammed, M.A.; Abdulkareem, K.H. Adversarial Attack and Defence through Adversarial Training and Feature Fusion for Diabetic Retinopathy Recognition. Sensors 2021, 21, 3922. https://0-doi-org.brum.beds.ac.uk/10.3390/s21113922

AMA Style

Lal S, Rehman SU, Shah JH, Meraj T, Rauf HT, Damaševičius R, Mohammed MA, Abdulkareem KH. Adversarial Attack and Defence through Adversarial Training and Feature Fusion for Diabetic Retinopathy Recognition. Sensors. 2021; 21(11):3922. https://0-doi-org.brum.beds.ac.uk/10.3390/s21113922

Chicago/Turabian Style

Lal, Sheeba, Saeed Ur Rehman, Jamal Hussain Shah, Talha Meraj, Hafiz Tayyab Rauf, Robertas Damaševičius, Mazin Abed Mohammed, and Karrar Hameed Abdulkareem. 2021. "Adversarial Attack and Defence through Adversarial Training and Feature Fusion for Diabetic Retinopathy Recognition" Sensors 21, no. 11: 3922. https://0-doi-org.brum.beds.ac.uk/10.3390/s21113922

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop