Next Article in Journal
‘All We Have to Decide Is What to Do with the Time That Is Given to Us’ a Photovoice Study on Physical Activity in Nursing Homes
Next Article in Special Issue
Effect of Smartphone Usage on Neck Muscle Endurance, Hand Grip and Pinch Strength among Healthy College Students: A Cross-Sectional Study
Previous Article in Journal
An Evaluation of Two Systems for the Management of the Microbiological Quality of Water in Dental Unit Waterlines: Hygowater® and IGN Calbénium®
Previous Article in Special Issue
Determinants and Prediction of Injury Severities in Multi-Vehicle-Involved Crashes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Skin Cancer Detection: A Review Using Deep Learning Techniques

1
Government Associate College for Women Mari Sargodha, Sargodha 40100, Pakistan
2
Department of Computer Science and Information Technology, University of Sargodha, Sargodha 40100, Pakistan
3
Electrical Engineering Department, College of Engineering, Najran University Saudi Arabia, Najran 61441, Saudi Arabia
4
Department of Computer Science, Wah Campus, Comsats University, Wah Cantt 47040, Pakistan
5
Department of Computer Science, School of Systems and Technology, University of Management and Technology, Lahore 54782, Pakistan
6
Department of Computer Science, COMSATS University Islamabad, Islamabad 440000, Pakistan
7
Department of Internal Medicine, Faculty of Medicine, Najran University, Najran 61441, Saudi Arabia
8
Department of Dermatology, Najran University Hospital, Najran 61441, Saudi Arabia
9
Department of Internal Medicine, Faculty of Medicine, University of Tabuk, Tabuk 71491, Saudi Arabia
10
Department of Medicinal Chemistry, Pharmacy School, Najran University, Najran 61441, Saudi Arabia
*
Author to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2021, 18(10), 5479; https://0-doi-org.brum.beds.ac.uk/10.3390/ijerph18105479
Submission received: 16 February 2021 / Revised: 26 April 2021 / Accepted: 13 May 2021 / Published: 20 May 2021

Abstract

:
Skin cancer is one of the most dangerous forms of cancer. Skin cancer is caused by un-repaired deoxyribonucleic acid (DNA) in skin cells, which generate genetic defects or mutations on the skin. Skin cancer tends to gradually spread over other body parts, so it is more curable in initial stages, which is why it is best detected at early stages. The increasing rate of skin cancer cases, high mortality rate, and expensive medical treatment require that its symptoms be diagnosed early. Considering the seriousness of these issues, researchers have developed various early detection techniques for skin cancer. Lesion parameters such as symmetry, color, size, shape, etc. are used to detect skin cancer and to distinguish benign skin cancer from melanoma. This paper presents a detailed systematic review of deep learning techniques for the early detection of skin cancer. Research papers published in well-reputed journals, relevant to the topic of skin cancer diagnosis, were analyzed. Research findings are presented in tools, graphs, tables, techniques, and frameworks for better understanding.

1. Introduction

Skin cancer is one of the most active types of cancer in the present decade [1]. As the skin is the body’s largest organ, the point of considering skin cancer as the most common type of cancer among humans is understandable [2]. It is generally classified into two major categories: melanoma and nonmelanoma skin cancer [3]. Melanoma is a hazardous, rare, and deadly type of skin cancer. According to statistics from the American Cancer Society, melanoma skin cancer cases are only 1% of total cases, but they result in a higher death rate [4]. Melanoma develops in cells called melanocytes. It starts when healthy melanocytes begin to grow out of control, creating a cancerous tumor. It can affect any area of the human body. It usually appears on the areas exposed to sun rays, such as on the hands, face, neck, lips, etc. Melanoma type of cancers can only be cured if diagnosed early; otherwise, they spread to other body parts and lead to the victim’s painful death [5]. There as various types of melanoma skin cancer such as nodular melanoma, superficial spreading melanoma, acral lentiginous, and lentigo maligna [3]. The majority of cancer cases lie under the umbrella of nonmelanoma categories, such as basal cell carcinoma (BCC), squamous cell carcinoma (SCC), and sebaceous gland carcinoma (SGC). BCC, SGC, and SCC are formed in the middle and upper layers of the epidermis, respectively. These cancer cells have a low tendency of spreading to other body parts. Nonmelanoma cancers are easily treated as compared with melanoma cancers.
Therefore, the critical factor in skin cancer treatment is early diagnosis [6]. Doctors ordinarily use the biopsy method for skin cancer detection. This procedure removes a sample from a suspected skin lesion for medical examination to determine whether it is cancerous or not. This process is painful, slow, and time-consuming. Computer-based technology provides a comfortable, less expensive, and speedy diagnosis of skin cancer symptoms. In order to examine the skin cancer symptoms, whether they represent melanoma or nonmelanoma, multiple techniques, noninvasive in nature, are proposed. The general procedure followed in skin cancer detection is acquiring the image, preprocessing, segmenting the acquired preprocessed image, extracting the desired feature, and classifying it, represented in Figure 1.
Deep learning has revolutionized the entire landscape of machine learning during recent decades. It is considered the most sophisticated machine learning subfield concerned with artificial neural network algorithms. These algorithms are inspired by the function and structure of the human brain. Deep learning techniques are implemented in a broad range of areas such as speech recognition [7], pattern recognition [8], and bioinformatics [9]. As compared with other classical approaches of machine learning, deep learning systems have achieved impressive results in these applications. Various deep learning approaches have been used for computer-based skin cancer detection in recent years. In this paper, we thoroughly discuss and analyze skin cancer detection techniques based on deep learning. This paper focuses on the presentation of a comprehensive, systematic literature review of classical approaches of deep learning, such as artificial neural networks (ANN), convolutional neural networks (CNN), Kohonen self-organizing neural networks (KNN), and generative adversarial neural networks (GAN) for skin cancer detection.
A significant amount of research has been performed on this topic. Thus, it is vital to accumulate and analyze the studies, classify them, and summarize the available research findings. To conduct a valuable systematic review of skin cancer detection techniques using deep neural network-based classification, we built search strings to gather relevant information. We kept our search focused on publications of well-reputed journals and conferences. We established multi-stage selection criteria and an assessment procedure, and on the basis of the devised search, 51 relevant research papers were selected. These papers were thoroughly evaluated and analyzed from different aspects. We are greatly encouraged by the trends in skin cancer detection systems, but still, there is space for further improvement in present diagnostic techniques.
This paper is subdivided into four main sections. Section 2 describes the research methodology for performing the effective analysis of deep learning techniques for skin cancer (SC) detection. It contains a description of the review domain, search strings, search criteria, the sources of information, the information extraction framework, and selection criteria. Selected research papers are evaluated, and a detailed survey of SC detection techniques is presented in Section 3. Section 4 summarizes the whole study and presents a brief conclusion.

2. Research Methodology

The purpose of performing this systematic literature review was to select and categorize the best available approaches to skin cancer detection using neural networks (NNs). Systematic literature reviews collect and analyze existing studies according to predefined evaluation criteria. Such reviews help to determine what is already known in the concerned domain of study [10].
All data collected from primary sources are organized and analyzed. Once systematic literature is completed, it provides a more sensible, logical, and robust answer to the underlying question of the research [11].
The population of studies considered in the current systematic literature review consisted of research papers relevant to SC detection based on deep neural network (DNN) techniques.

2.1. Research Framework

Defining the review framework was the first step in this systematic review. It consisted of an overall plan being followed in the systematic literature review. The plan consisted of three layers: a planning layer, a data selection and evaluation layer, and a results-generation and conclusion layer.

2.1.1. Research Questions

For conducting an effective systematic literature review on a topic, it is necessary to formulate research questions. The research questions formulated for the current systematic research were as follows:
Question No. 1: What are the major deep learning techniques for skin cancer detection?
Question No. 2: What are the main characteristics of datasets available for skin cancer?

2.1.2. Search Strategy

A systematic and well-planned search is very important for collecting useful material from the searched data of the desired domain. In this step, a thorough search was conducted to extract meaningful and relevant information from the mass of data. We created an automated search mechanism for filtering out the desired domain’s data from all sources. Research papers, case studies, American Cancer Society reports, and reference lists of related publications were examined in detail. Websites containing information regarding skin cancer, the dangers of skin cancer, the reasons for skin cancer, and NN techniques of skin cancer detection were all carefully searched. For extraction of the desired and relevant data, we conducted our search according to the following parameters.
  • Search keywords/search term identification based on research questions
  • Words related to the search keywords
  • Search string formulation using logical operators between search words
The keywords related to deep learning techniques for skin cancer detection were selected. Subsequently, the search was extended to synonyms for these keywords.
Furthermore, the search was carried out using logical operators ’AND’ and ‘OR’ between keywords. The keywords used to search information relevant to skin cancer are listed in Table 1.

2.1.3. Resources of Search

We conducted our initial search on well-reputed search engines such as IEEE Xplore, ACM, Springer as well as Google Scholar to extract information relevant to NN techniques for skin cancer detection. Basic research material related to the underlying topic was filtered out in the primary search. The selected research papers and conference proceedings were further analyzed according to evaluation criteria.

2.1.4. Initial Selection Criteria

The initial selection of research papers/conference papers was based on certain specified parameters such as the language of the paper, the year of the paper, and the relevance of the topic within the desired domain. Only research papers written in the English language were included in this research. Our review paper focused on research published between 2011 and 2021. Selected papers had to be relevant to the search terms described in the search strategy.

2.2. Selection and Evaluation Procedure

Using the initial search criteria, the search extracted 1483 research papers and conference reports. From the papers that were identified, we selected 95 papers that had a title considered relevant to our study. Subsequently, the abstracts of those selected papers were examined more closely for their relevance, which led to reducing their number to 64 research papers. The research papers successfully passing abstract-based selection were studied in detail. The quality of those research papers was fully examined, and 51 research papers were selected for final review. In this finalized selection, the percentage of IEEE publications was 25%, Google Scholar’s selection percentage was 16%, 10% papers were selected from ACM DL, 29% from springer, and 20% from Science Direct. The search results are represented in Table 2.
A thorough study of the full text of the selected research papers sought answers to certain quality control questions. The current systematic research asked the following quality assessment questions.
  • Did the selected study cover all aspects of this review’s topic?
  • Was the quality of the selected paper verified?
  • Does the selected study adequately answer the research questions?
The first quality assessment question focused on the thorough coverage of deep learning techniques for skin cancer detection. The quality of a selected paper was verified by the reputation of the journal in which it was published and by its citations. The third question ensured that the research answered the research questions mentioned in Section 2. Only the most relevant research papers to our domain of study were extracted. These papers had to satisfy those above research questions to qualify for selection. Research papers that failed to adequately answer the research or quality control questions and papers with text that was not related to our study topic were excluded.
Each question had Boolean ’yes/no’ responses. Each ‘yes’ was assigned a value Y = 1 and each ‘no’ was assigned a value N = 0. The first quality control question evaluated the topic coverage of the 51 selected research papers and resulted in a value of 77%, which was quite satisfactory. The second research question verified the quality of the selected papers, which resulted in improvement of quality. It generated an 82% result, which was satisfactory. The third question was a very important question in order to answer the review’s main research questions. It generated a 79% result, which was the indicator of the adequacy of the studies to answer the research questions posed by the review. The overall results of answers to these quality questions seemed healthy.

3. Deep Learning Techniques for Skin Cancer Detection

Deep neural networks play a significant role in skin cancer detection. They consist of a set of interconnected nodes. Their structure is similar to the human brain in terms of neuronal interconnectedness. Their nodes work cooperatively to solve particular problems. Neural networks are trained for certain tasks; subsequently, the networks work as experts in the domains in which they were trained. In our study, neural networks were trained to classify images and to distinguish between various types of skin cancer. Different types of skin lesion from International Skin Imaging Collaboration (ISIC) dataset are presented in Figure 2. We searched for different techniques of learning, such as ANN, CNN, KNN, and GAN for skin cancer detection systems. Research related to each of these deep neural networks is discussed in detail in this section.

3.1. Artificial Neural Network (ANN)-Based Skin Cancer Detection Techniques

An artificial neural network is a nonlinear and statistical prediction method. Its structure is borrowed from the biological structure of the human brain. An ANN consists of three layers of neurons. The first layer is known as the input layer; these input neurons transfer data to the second/intermediate layer of neurons. The intermediate layers are referred to as hidden layers. In a typical ANN, there can be several hidden layers. Intermediate neurons send data to the third layer of output neurons. Computations are learned at each layer using backpropagation, which is used for learning the complex associations/relationships between input and output layers. It is similar to a neural network. Currently, in computer science, the term neural network and artificial neural network are used interchangeably. The basic structure of an ANN network is presented in Figure 3.
ANN is used for the classification of extracted features in skin cancer detection systems. Input images are classified as melanoma or nonmelanoma after successful training/classification of the training set. The number of hidden layers in an ANN depends on the number of input images. The input/first layer of the ANN process connects with the hidden layer by the input dataset. The dataset can be labeled or unlabeled, which can be processed accordingly using a supervised or unsupervised learning mechanism. A neural network uses backpropagation or feed-forward architecture to learn weights present at each network connection/link. Both architectures use a different pattern for the underlying dataset. Feed-forward-architecture-based neural networks transfer data only in one direction. Data flows only from the input to the output layer.
Xie et al. [14] proposed a skin lesion classification system that classified lesions into two main classes: benign and malignant. The proposed system worked in three phases. In the initial phase, a self-generating NN was used to extract lesions from images. In the second phase, features such as tumor border, texture, and color details were extracted. The system extracted a total of 57 features, including 7 novel features related to lesion borders descriptions. Principal component analysis (PCA) was used to reduce the dimensionality of the features, which led to the selection of the optimal set of features. Finally, in the last phase, lesions were classified using a NN ensemble model. Ensemble NN improves classification performance by combining backpropagation (BP) NN and fuzzy neural networks. Furthermore, the proposed system classification results were compared with other classifiers, such as SVM, KNN, random forest, Adaboot, etc. With a 91.11% accuracy, the proposed model achieved at least 7.5% higher performance in terms of sensitivity than the other classifiers.
Masood et al. [15] proposed an ANN-based automated skin cancer diagnostic system. The performance of three ANN’s learning algorithms such as Levenberg–Marquardt (LM) [16], resilient backpropagation (RP) [17], scaled conjugate gradient (SCG) [18], was also investigated by this paper. Comparison of performance showed that the LM algorithm achieved the highest specificity score (95.1%) and remained efficient at the classification of benign lesions, while the SCG learning algorithm produced better results if the number of epochs was increased, scoring a 92.6% sensitivity value. A mole classification system for the early diagnosis of melanoma skin cancer was proposed [19]. The proposed system extracted features according to the ABCD rule of lesions. ABCD refers to asymmetry of a mole’s form, borders of mole, color, and diameter of mole. Assessment of a mole’s asymmetry and borders were extracted using the Mumford–Shah algorithm and Harris Stephen algorithm, respectively. Normal moles are composed of black, cinnamon, or brown color, so moles with colors other than those three were considered melanoma in the proposed system. Melanoma moles commonly have a diameter value greater than 6 mm, so that value was used as the threshold value of diameter for melanoma detection. The proposed system used a backpropagation feed-forward ANN to classify moles into three classes, such as common mole, uncommon mole, or melanoma mole, with 97.51% accuracy.
An automated skin cancer diagnostic system based on backpropagation ANN was proposed [20], represented in Figure 4. This system employed a 2D-wavelet transform technique for feature extraction. The proposed ANN model classified the input images into two classes, such as cancerous or noncancerous. Another ANN-based skin cancer diagnostic system was proposed by Choudhari and Biday [21]. Images were segmented with a maximum entropy thresholding measure. A gray-level co-occurrence matrix (GLCM) was used to extract unique features of skin lesions. Finally, a feed-forward ANN classified the input images into either a malignant or benign stage of skin cancer, achieving an accuracy level of 86.66%.
Aswin et al. [22] described a new method for skin cancer detection based on a genetic algorithm (GA) and ANN algorithms. Images were preprocessed for hair removal with medical imaging software named Dull-Rozar and region of interest (ROI) and were extracted with the Otsu thresholding method. Furthermore, the GLCM technique was employed to extract unique features of the segmented images. Subsequently, a hybrid ANN and GA classifier was used for the classification of lesion images into cancerous and noncancerous classes. The proposed system achieved an overall accuracy score of 88%. Comprehensive details of the various skin cancer detection systems based on ANN are listed in Table 3 below.

3.2. Convolutional Neural Network (CNN)-Based Skin Cancer Detection Techniques

A convolution neural network is an essential type of deep neural network, which is effectively being used in computer vision. It is used for classifying images, assembling a group of input images, and performing image recognition. CNN is a fantastic tool for collecting and learning global data as well as local data by gathering more straightforward features such as curves and edges to produce complex features such as shapes and corners [28]. CNN’s hidden layers consist of convolution layers, nonlinear pooling layers, and fully connected layers [29]. CNN can contain multiple convolution layers that are followed by several fully connected layers. Three major types of layers involved in making CNN are convolution layers, pooling layers, and full-connected layers [30]. The basic architecture of a CNN is presented in Figure 5.
CNN-based automated deep learning algorithms have achieved remarkable performance in the detection, segmentation, and classification operations of medical imaging [31]. Lequan et al. [32] proposed a very deep CNN for melanoma detection. A fully convolutional residual network (FCRN) having 16 residual blocks was used in the segmentation process to improve performance. The proposed technique used an average of both SVM and softmax classifier for classification. It showed 85.5% accuracy in melanoma classification with segmentation and 82.8% without segmentation. DeVries and Ramachandram [33] proposed a multi-scale CNN using an inception v3 deep neural network that was trained on an ImageNet dataset. For skin cancer classification, the pre-trained inception v3 was further fined-tuned on two resolution scales of input lesion images: coarse-scale and finer scale. The coarse-scale was used to capture shape characteristics as well as overall contextual information of lesions. In contrast, the finer scale gathered textual detail of lesion for differentiation between various types of skin lesions.
Mahbod et al. [34] proposed a technique to extract deep features from various well-established and pre-trained deep CNNs for skin lesions classification. Pretrained AlexNet, ResNet-18 and VGG16 were used as deep-feature generators, then a multi-class SVM classifier was trained on these generated features. Finally, the classifier results were fused to perform classification. The proposed system was evaluated on the ISIC 2017 dataset and showed 97.55% and 83.83% area under the curve (AUC) performance for seborrheic keratosis (SK) and melanoma classification. A deep CNN architecture based on pre-trained ResNet-152 was proposed to classify 12 different kinds of skin lesions [35]. Initially, it was trained on 3797 lesion images; however, later, 29-times augmentation was applied based on lighting positions and scale transformations. The proposed technique provided an AUC value of 0.99 for the classification of hemangioma lesion, pyogenic granuloma (PG) lesion, and intraepithelial carcinoma (IC) skin lesions.
A technique for the classification of four different types of skin lesion images was proposed by Dorj et al. [36]. A pre-trained deep CNN named AlexNet was used for feature extraction, after which error-correcting output coding SVM worked as a classifier. The proposed system produced the highest scores of the average sensitivity, specificity, and accuracy for SCC, actinic keratosis (AK), and BCC: 95.1%, 98.9%, and 94.17%, respectively. Kalouche [37] proposed a pre-trained deep CNN architecture VGG-16 with a final three fine-tuned layers and five convolutional blocks. The proposed VCG-16 model is represented in Figure 6. VCG-16 models showed 78% accuracy for the classification of lesion images as melanoma skin cancer. A deep CNN-based system was proposed to detect the borders of skin lesions in images. The deep learning model was trained on 1200 normal skin images and 400 images of skin lesions. The proposed system classified the input images into two main classes, normal skin image and lesion image, with 86.67% accuracy. A comprehensive list of skin cancer detection systems using CNN classifiers is presented in Table 4.

3.3. Kohonen Self-Organizing Neural Network (KNN)-Based Skin Cancer Detection Techniques

The Kohonen self-organizing map is a very famous type of deep neural network. CNNs are trained on the basis of unsupervised learning, which means that a KNN does not require any developer’s intervention in the learning process as well as requiring little information about the attributes of the input data. A KNN generally consists of two layers. In the 2-D plane, the first layer is called an input layer, while another is named a competitive layer. Both of these layers are fully connected, and every connection is from the first to second layer dimension. A KNN can be used for data clustering without knowing the relationships between input data members. It is also known as a self-organizing map. KNNs do not contain an output layer; every node in the competitive layer also acts as the output node itself.
A KNN basically works as a dimensionality reducer. It can reduce the high dimensional data into a low dimension, such as a two-dimensional plane. Thus, it provides discrete types of representation of the input dataset. KNNs are different from other types of NN in terms of learning strategy because it uses competitive learning rather than the learning based on error correction found in BPN or feed-forward learning. A KNN preserves the topological structure of the input data space during mapping dimensionality from high to low. Preservation refers to the preservation of relative distance between data points in space. Data points that are closer in input data space are mapped closer to each other in this scheme; far points are mapped far from each other as well as, according to the relative distance present among them. Consequently, a KNN is the best tool for high dimensional data. Another important feature provided by a KNN is its generalization ability. The network has the ability to recognize and organize unknown input data. The architecture of a KNN is shown in Figure 7. A KKN’s main quality is its ability to map complex relationships of data points in which even nonlinear relations exist between data points. Due to these benefits, nowadays, KNNs are being used in skin cancer detection systems.
Lenhardt et al. [59] proposed a KNN-based skin cancer detection system. The proposed system processed synchronous fluorescence spectra of melanoma, nevus, and normal skin samples for neural network training. A fluorescence spectrophotometer was used to measure the fluorescence spectra of the samples, whereas samples were collected from human patients immediately after surgical resection. The dimensionality of measured spectra was reduced with the PCA technique. Both KNN and ANN were trained, and their performance for melanoma detection was compared. On the test dataset, the classification error of KNN was 2–3%, while the classification error for ANN lay in the range of 3% to 4%.
A combination of self-organizing NN and radial basis function (RBF) neural network was proposed to diagnose three different types of skin cancer, such as BCC, melanoma, and SCC [60]. The proposed system extracted color, GLCM, and morphological features of lesion images, after which the classification model used those features as input. Furthermore, the classification performance of the proposed system was compared with k-nearest neighbor, ANN, and naïve-Bayes classifiers. The proposed system achieved 93.150685% accuracy while k-nearest neighbor showed 71.232877%, ANN showed 63.013699%, and naïve Bayes showed 56.164384% accuracy scores.
Another KNN-based automated skin cancer diagnostic system was proposed by Sajid et al. [61]. The proposed system employed a median filter as a noise removal technique. Then filtered images were segmented with a statistical region growing and merging technique. In this system, a collection of textual and statistical features was used. Statistical features were extracted from lesion images, whereas textual features were extracted from a curvelet domain. Finally, the proposed system classified the input images into cancerous or noncancerous with 98.3% accuracy. In this work, other classifiers such as SVM, BPN, and 3-layer NN were also implemented, and their performance was compared with the proposed system’s classification performance. SVM produced 91.1% accuracy, BPN showed 90.4% accuracy, 3-layer NN showed 90.5%, whereas the proposed system achieved the highest accuracy of 98.3% for skin cancer diagnosis. Details on the KNN-based skin cancer diagnostic systems is presented in Table 5.

3.4. Generative Adversarial Network (GAN)-Based Skin Cancer Detection Techniques

A generative adversarial neural network is a powerful class of DNN that is inspired by zero-sum game theory [62]. GANs are based on the idea that two neural networks, such as a generator and a discriminator, compete with each other to analyze and capture the variance in a database. The generator module uses the data distribution to produce fake data samples and tries to misguide the discriminator module. On the other hand, the discriminator module aims to distinguish between real and fake data samples [63]. In the training phase, both of these neural networks repeat these steps, and their performance improves after each competition. The ability to generate fake samples that are similar to a real sample using the same data distribution, such as photorealistic images, is the major power of a GAN network. It can also solve a major problem in deep learning: the insufficient training examples problem. Research scholars have been implementing various types of GANs, such as Vanilla GAN, condition GAN (CGAN), deep convolutional GAN (DCGAN), super-resolution GAN (SRGAN), and Laplacian Pyramid GAN (LPGAN). Nowadays, GANs are successfully being used in skin cancer diagnostic systems. The architecture of a GAN is shown in Figure 8.
Rashid et al. [7] proposed a GAN-based skin lesion classification system. The proposed system performed augmentation on a training set of images with realistic-looking skin lesion images generated via GAN. A deconvolutional network was used as the generator module, while the discriminator module used CNN as a classifier. The CNN learned to classify seven different categories of skin lesions. Results of the proposed system were compared with ResNet-50 and DenseNet. ResNet-50 produced 79.2% accuracy, DenseNet showed 81.5% accuracy, whereas the proposed approach achieved the highest accuracy of 86.1% for skin lesion classification. Deep learning methods provide sufficient accuracy but require pure, unbalanced, and large training datasets. To overcome these limitations, Bisla et al. [8] proposed a deep learning approach for data purification and GAN for data augmentation. The proposed system used decoupled deep convolutional GANs for data generation. A pre-trained ResNet-50 model was further refined with a purified and augmented dataset and was used to classify dermoscopic images into three categories: melanoma, SK, and nevus. The proposed system outperformed the baseline ResNet-50 model for skin lesion classification and achieved 86.1% accuracy.
A novel data augmentation method for a skin lesion on the basis of self-attention progressive GAN (PGAN) was proposed. Moreover, the generative model was enhanced with the stabilization technique. The proposed system achieved 70.1% accuracy as compared with 67.3% accuracy produced by a non-augmented system. A list of GAN-based skin cancer detection systems with their diagnosed skin cancer type, classifier, dataset, and the obtained result is presented in Table 6.

4. Datasets

Several computer-based systems for skin cancer diagnosis have been proposed. Evaluating their diagnostic performance and validating predicted results requires a solid and reliable collection of dermoscopic images. Various skin cancer datasets have lacked size and diversity other than for images of nevi or melanoma lesions. Training of artificial neural networks for skin lesion classification is hampered by the small size of the datasets and a lack of diverse data. Although patients commonly suffer from a variety of non-melanocytic lesions, past research for automated skin cancer diagnosis primarily focused on diagnosing melanocytic lesions, resulting in a limited number of diagnoses in the available datasets [66]. Therefore, the availability of a standard, reliable dataset of dermoscopic images is very crucial. Real-world datasets for the evaluation of proposed skin cancer detection techniques are discussed in this section. Table 7 summarizes the important details of these datasets.

4.1. HAM10000

There is a human-against-machine dataset with 10,000 training images that is referred to as HAM10000 [66]. It is the latest publicly available skin lesions dataset, and it overcomes the problem of the lack of diversity. The final dataset of HAM10000 contains 10,015 dermoscopic images, collected from two sources: Cliff Rosendahl’s skin cancer practice in Queensland, Australia, and the Dermatology Department of the Medical University of Vienna, Austria. This collection has taken twenty years to compile. Before widespread use of digital cameras, photographic prints of lesions were deposited and stored at the Dermatology Department of the Medical University of Vienna, Austria. These photographic prints were digitalized with the help of Nikon-Coolscan-5000-ED scanner, manufactured by Nikon corporation Japan and converted into 8-bit color JPEG images having 300 DPI quality. The images were then manually cropped and saved at 800 × 600 pixels resolution at 72 DPI.
Several acquisition functions and cleaning methods were applied to the images and a semi-automatic workflow was developed using a neural network to attain diversity. The resulting dataset contains 327 images of AK, 514 images of basal cell carcinomas, 1099 images of benign keratoses, 115 images of dermatofibromas, 1113 images of melanocytic nevi, 6705 images of melanomas, and 142 images of vascular skin lesions.

4.2. PH2

The dermoscopic images in the PH² dataset were collected at the Dermatology Center of Pedro Hispano Hospital, Portugal [68]. These images were obtained using a Tuebinger-Mole-Analyzer system under the same conditions and magnification rate of 20×. PH2 dataset contains 8-bit RGB color images having 768 × 560 pixels resolution. The dataset contains 200 dermoscopic images, divided into 80 images of common nevi, 80 images of atypical nevi, and 40 images of melanoma skin cancers. This dataset contains medical annotation of the lesion images, such as medical segmentation of pigmented skin lesions, histological and clinical diagnosis, and evaluation of various dermoscopic criteria. The assessment was performed according to dermoscopic criteria of streaks, colors, regression areas, pigment network, and blue-whitish veil globules.

4.3. ISIC Archive

The ISIC archive [69] is a collection of various skin lesions datasets. The ISIC dataset [70] was originally released by the International Skin Imaging Collaboration at the International Symposium on Biomedical Imaging (ISBI) challenge 2016, named as ISIC2016. The ISIC2016 archive is divided into two parts: training and testing. The training subset of ISIC contains 900 images, while the testing subset contains 379 dermoscopic images. It includes images of two classes: malignant melanomas and benign nevi. Approximately 30.3% of the dataset’s images are of melanoma lesions and the remaining images belong to the benign nevi class. ISIC increases the number of images in its archive every year and has established a design challenge for the development of a system for skin cancer automated diagnosis.
In the ISIC2017 dataset, there were three categories of images: melanomas, seborrheic-keratoses (SK), and benign nevi. The dataset contains 2000 training images, 150 validation images, and 600 images for testing. The training dataset contains 374 images of melanomas, 254 SK images, and 1372 images of benign nevi. The validation dataset contains 30 melanoma images, 42 SK images, and 78 benign nevus images. The test dataset includes 117 melanoma images, 90 SK images, and 393 benign nevus images. ISIC2018 contains 12,594 training images, 100 validation images, and 1000 test images. The ISIC2019 dataset includes 25,331 images of eight different categories of skin lesions, such as melanoma, melanocytic-nevus, BCC, AK, benign keratosis, dermatofibroma, vascular lesion, and SCC. It contains 8239 images in the test dataset and an additional outlier class that was not included in the training dataset. The new proposed skin cancer diagnostic systems must be able to identify these images. The ISIC2019 dataset also includes metadata for images, such as sex, age, and area of the patient.

4.4. Derm Quest

The publicly available DermQuest dataset [71] contained 22,082 dermoscopic images. Among all dermoscopic datasets, only the DermQuest dataset contained lesion tags for skin lesions. There were 134 lesion tags for all images in the dataset. The DermQuest dataset redirected to Derm101 in 2018. However, this dataset was deactivated recently on 31 December 2019.

4.5. DermIS

The Dermoscopic dataset Dermatology Information System is commonly known as DermIS [72]. This dataset was built through cooperation between the Department of Dermatology of the University of Erlangen and the Department of Clinical Social Medicine of the University of Heidelberg. It contains 6588 images. This dataset has recently been divided into two parts: a dermatology online image atlas (DOIA) and a pediatric dermatology online image atlas (PeDOIA). The DOIA includes 3000 lesion images covering approximately 600 dermatological diagnoses. It provides dermoscopic images complete with differential and provisional diagnoses, case reports, and other information on nearly all types of skin diseases.

4.6. AtlasDerm

The Atlas of Dermoscopy dataset is commonly referred to as AtlasDerm [73]. It is a unique and well-organized combination of a book and images on CD-ROM with sample examples for training. It was originally designed as a tool to help physicians in the diagnosis of skin lesions and the recognition of dermoscopic criteria related to melanoma. The AtlasDerm dataset considers various cases of skin lesions, with corresponding dermoscopic images for every case. It contains 5 images of AK, 42 images of BCC, 70 images of benign keratosis, 20 images of dermatofibroma, 275 images of melanocytic nevus, 582 images of melanoma, and 30 images of vascular skin lesions.

4.7. Dermnet

The Dermnet Skin Disease Atlas dataset is commonly referred to as Dermnet [74]. It was built in 1998 by Dr. Thomas Habif in Portsmouth, New Hampshire. It consists of more than 23,000 dermoscopic images. This database contains images of 643 different types of skin diseases. These diseases are biologically organized into a two-level taxonomy. The bottom level contains more than 600 skin diseases in fine granularity. The top-level taxonomy contains 23 different classes of skin diseases, such as connective tissue disease, benign tumors, eczema, melanomas, moles, nevi, etc.

5. Open Research Challenges

5.1. Extensive Training

One of the major challenges in neural network-based skin cancer detection techniques is the extensive training that is required. In other words, to successfully analyze and interpret the features from dermoscopic images, the system must undergo detailed training, which is a time-consuming process and demands extremely powerful hardware.

5.2. Variation in Lesion Sizes

Another challenge is the variation in the sizes of lesions. A group of Italian and Austrian researchers collected many benign and cancerous melanoma lesion images in the 1990s [73]. The diagnostic accuracy of the identification of the lesions was as high as 95% to 96% [75]. However, the diagnostic process, with earlier stage and smaller lesions of 1mm or 2mm in size, was much more difficult and error-prone.

5.3. Images of Light Skinned People in Standard Datasets

Existing standard dermoscopic datasets contain images of light-skinned people, mostly from Europe, Australia, and the United States. For accurate skin cancer detection in dark-skinned people, a neural network must learn to account for skin color [76]. However, doing so is possible only if the neural network observes enough images of dark-skinned people during the process of training. Therefore, datasets having sufficient lesion images of dark-skinned and light-skinned people is necessary for increasing the accuracy of skin cancer detection systems.

5.4. Small Interclass Variation in Skin Cancer Images

Unlike the other types of images, medical images have very small interclass variation; that is, the difference between melanoma and nonmelanoma skin cancer lesion images has much less variation than, say, the variation between images of cats and dogs. It is also very difficult to differentiate between a birthmark and a melanoma. The lesions of some disease are so similar that it is extremely hard to distinguish them. This limited variation makes the task of image analysis and classification very complex [32].

5.5. Unbalanced Skin Cancer Datasets

Real-world datasets used for skin cancer diagnosis are highly unbalanced. Unbalanced datasets contain a very different number of images for each type of skin cancer. For example, they contain hundreds of images of common skin cancer types but only a few images for the uncommon types, making it difficult to draw generalizations from the visual features of the dermoscopic images [12].

5.6. Lack of Availability of Powerful Hardware

Powerful hardware resources with high graphical processing unit (GPU) power are required for the NN software to be able to extract the unique features of a lesion’s image, which is critical for achieving better skin cancer detection. The lack of availability of high computing power is a major challenge in deep learning-based skin cancer detection training.

5.7. Lack of Availability of Age-Wise Division of Images In Standard Datasets

Various types of skin cancers such as Merkel cell cancer, BCC, and SCC typically appear after the age of 65 years [77]. Existing standard dermoscopic datasets contain images of young people. However, for an accurate diagnosis of skin cancer in elderly patients, it is necessary that neural networks observe enough images of people aged more than 50 years.

5.8. Use of Various Optimization Techniques

Preprocessing and detection of lesion edges are very crucial steps in the automated detection of skin cancer. Various optimization algorithms such as artificial the bee colony algorithm [78], ant colony optimization [79], social spider optimization [80], and particle swarm optimization [81] can be explored to increase the performance of automated skin cancer diagnostic systems.

5.9. Analysis of Genetic and Environmental Factors

Researchers have identified various genetic risk factors for melanoma, such as fair skin, light colored eyes, red hair, a large number of moles on the body, and a family history of skin cancer. When these genetic risk factors are combined with environmental risks such as high ultraviolet light exposure, the chances of developing skin cancer become very high [82]. These factors can be combined with existing deep learning approaches for better performance.

6. Conclusion and Future Work

This systematic review paper has discussed various neural network techniques for skin cancer detection and classification. All of these techniques are noninvasive. Skin cancer detection requires multiple stages, such as preprocessing and image segmentation, followed by feature extraction and classification. This review focused on ANNs, CNNs, KNNs, and RBFNs for classification of lesion images. Each algorithm has its advantages and disadvantages. Proper selection of the classification technique is the core point for best results. However, CNN gives better results than other types of a neural networks when classifying image data because it is more closely related to computer vision than others.
Most of the research related to skin cancer detection focuses on whether a given lesion image is cancerous. However, when a patient asks if a particular skin cancer symptom appears on any part of their body, the current research cannot provide an answer. Thus far, the research has focused on the narrow problem of classification of the signal image. Future research can include full-body photography to seek the answer to the question that typically arises. Autonomous full-body photography will automate and speed up the image acquisition phase.
The idea of auto-organization has recently emerged within the area of deep learning. Auto-organization refers to the process of unsupervised learning, which aims to identify features and to discover relations or patterns in the image samples of the dataset. Under the umbrella of convolutional neural networks, auto-organization techniques increase the level of features representation that is retrieved by expert systems [47]. Currently, auto-organization is a model that is still in research and development. However, its study can improve the accuracy of image processing systems in the future, particularly in the area of medical imaging, where the smallest details of features are extremely crucial for the correct diagnosis of disease.

Author Contributions

M.D. and S.A., M.I. and H.U.K. developed the idea and collected the data. M.R. and A.R.M. analyzed the data and wrote the manuscript. S.A.A. prepared the figures. A.H.M.S., M.O.A. and M.H.M. reviewed the data and manuscript as well. M.O.A. and M.H.M. were involved in the analysis of the images, datasets, labelling, and data. All authors have read and agreed to the published version of the manuscript.

Funding

This research study has not received any research funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable as it is a review article, no experiment has been performed by using any data.

Acknowledgments

The authors acknowledge support from the Deanship of Scientific Research, Najran University, Kingdom of Saudi Arabia.

Conflicts of Interest

Authors have no conflicts of interest.

References

  1. Ashraf, R.; Afzal, S.; Rehman, A.U.; Gul, S.; Baber, J.; Bakhtyar, M.; Mehmood, I.; Song, O.Y.; Maqsood, M. Region-of-Interest Based Transfer Learning Assisted Framework for Skin Cancer Detection. IEEE Access 2020, 8, 147858–147871. [Google Scholar] [CrossRef]
  2. Byrd, A.L.; Belkaid, Y.; Segre, J.A. The Human Skin Microbiome. Nat. Rev. Microbiol. 2018, 16, 143–155. [Google Scholar] [CrossRef]
  3. Elgamal, M. Automatic Skin Cancer Images Classification. IJACSA 2013, 4. [Google Scholar] [CrossRef] [Green Version]
  4. Key Statistics for Melanoma Skin Cancer. Am. Cancer Soc. Available online: https://www.cancer.org/content/dam/CRC/PDF/Public/8823.00.pdf (accessed on 8 February 2021).
  5. Khan, M.Q.; Hussain, A.; Rehman, S.U.; Khan, U.; Maqsood, M.; Mehmood, K.; Khan, M.A. Classification of Melanoma and Nevus in Digital Images for Diagnosis of Skin Cancer. IEEE Access 2019, 7, 90132–90144. [Google Scholar] [CrossRef]
  6. Premier Surgical Staff. What Is the Difference between Melanoma And non-Melanoma Skin Cancer? PSS. Available online: https://www.premiersurgical.com/01/whats-the-difference-between-melanoma-and-non-melanoma-skin-cancer/ (accessed on 6 February 2021).
  7. Rashid, H.; Tanveer, M.A.; Aqeel Khan, H. Skin Lesion Classification Using GAN Based Data Augmentation. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 916–919. [Google Scholar] [CrossRef]
  8. Bisla, D.; Choromanska, A.; Stein, J.A.; Polsky, D.; Berman, R. Towards Automated Melanoma Detection with Deep Learning: Data Purification and Augmentation. arXiv 2019, arXiv:1902.06061. Available online: http://arxiv.org/abs/1902.06061 (accessed on 10 February 2021).
  9. Farag, A.; Lu, L.; Roth, H.R.; Liu, J.; Turkbey, E.; Summers, R.M. A Bottom-Up Approach for Pancreas Segmentation Using Cascaded Superpixels and (Deep) Image Patch Labeling. IEEE Trans. Image Process. 2017, 26, 386–399. [Google Scholar] [CrossRef] [Green Version]
  10. Schlosser, R.W. The Role of Systematic Reviews in Evidence-Based Practice, Research and Development. Focus 2006, 15, 1–4. Available online: https://ktdrr.org/ktlibrary/articles_pubs/ncddrwork/focus/focus15 (accessed on 2 February 2021).
  11. Mallett, R.; Hagen-Zanker, J.; Slater, R.; Duvendack, M. The Benefits and Challenges of Using Systematic Reviews in International Development Research. J. Dev. Eff. 2012, 4, 445–455. [Google Scholar] [CrossRef]
  12. Milton, M.A.A. Automated Skin Lesion Classification Using Ensemble of Deep Neural Networks in ISIC 2018: Skin Lesion Analysis Towards Melanoma Detection Challenge. arXiv 2019, arXiv:1901.10802. Available online: http://arxiv.org/abs/1901.10802 (accessed on 22 January 2021).
  13. Aqib, M.; Mehmood, R.; Albeshri, A.; Alzahrani, A. Disaster Management in Smart Cities by Forecasting Traffic Plan Using Deep Learning and GPUs. In Smart Societies, Infrastructure, Technologies and Applications; Mehmood, R., Bhaduri, B., Katib, I., Chlamtac, I., Eds.; Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering; Springer International Publishing: Cham, Switzerland, 2018; Volume 224, pp. 139–154. [Google Scholar] [CrossRef]
  14. Xie, F.; Fan, H.; Li, Y.; Jiang, Z.; Meng, R.; Bovik, A. Melanoma Classification on Dermoscopy Images Using a Neural Network Ensemble Model. IEEE Trans. Med. Imaging 2017, 36, 849–858. [Google Scholar] [CrossRef]
  15. Masood, A.; Al-Jumaily, A.A.; Adnan, T. Development of Automated Diagnostic System for Skin Cancer: Performance Analysis of Neural Network Learning Algorithms for Classification. In Artificial Neural Networks and Machine Learning–ICANN 2014; Wermter, S., Weber, C., Duch, W., Honkela, T., Koprinkova-Hristova, P., Magg, S., Palm, G., Villa, A.E.P., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2014; Volume 8681, pp. 837–844. [Google Scholar] [CrossRef]
  16. Transtrum, M.K.; Sethna, J.P. Improvements to the Levenberg-Marquardt Algorithm for Nonlinear Least-Squares Minimization. arXiv 2012, arXiv:1201.5885. Available online: http://arxiv.org/abs/1201.5885 (accessed on 24 January 2021).
  17. Al-Naima, F.M.; Al-Timemy, A.H. Resilient Back Propagation Algorithm for Breast Biopsy Classification Based on Artificial Neural Networks. In Computational Intelligence and Modern Heuristics; Ali, A.-D., Ed.; InTech: Shanghai, China, 2010; Available online: https://www.intechopen.com/books/computational-intelligence-and-modern-heuristics/resilient-back-propagation-algorithm-for-breast-biopsy-classification-based-on-artificial-neural-net (accessed on 20 January 2021). [CrossRef] [Green Version]
  18. Møller, M.F. A Scaled Conjugate Gradient Algorithm for Fast Supervised Learning. Neural Netw. 1993, 6, 525–533. [Google Scholar] [CrossRef]
  19. Cueva, W.F.; Munoz, F.; Vasquez, G.; Delgado, G. Detection of Skin Cancer ”Melanoma” through Computer Vision. In Proceedings of the 2017 IEEE XXIV International Conference on Electronics, Electrical Engineering and Computing (INTERCON), Cusco, Peru, 15–18 August 2017; pp. 1–4. [Google Scholar] [CrossRef]
  20. Jaleel, J.A.; Salim, S.; Aswin, R. Artificial Neural Network Based Detection of Skin Cancer. Int. J. Adv. Res. Electr. Electron. Instrum. Eng. 2012, 1, 200–205. [Google Scholar]
  21. Choudhari, S.; Biday, S. Artificial Neural Network for SkinCancer Detection. IJETTCS 2014, 3, 147–153. [Google Scholar]
  22. Aswin, R.B.; Jaleel, J.A.; Salim, S. Hybrid Genetic Algorithm: Artificial Neural Network Classifier for Skin Cancer Detection. In Proceedings of the 2014 International Conference on Control, Instrumentation, Communication and Computational Technologies (ICCICCT), Kanyakumari, India, 10–11 July 2014; pp. 1304–1309. [Google Scholar] [CrossRef]
  23. Kanimozhi, T.; Murthi, D.A. Computer-Aided Melanoma Skin Cancer Detection Using Artificial Neural Network Classifier. J. Sel. Areas Microelectron. 2016, 8, 35–42. [Google Scholar]
  24. Ercal, F.; Chawla, A.; Stoecker, W.V.; Hsi-Chieh, L.; Moss, R.H. Neural Network Diagnosis of Malignant Melanoma from Color Images. IEEE Trans. Biomed. Eng. 1994, 41, 837–845. [Google Scholar] [CrossRef]
  25. Mahmoud, K.A.; Al-Jumaily, A.; Takruri, M. The Automatic Identification of Melanoma by Wavelet and Curvelet Analysis: Study Based on Neural Network Classification. In Proceedings of the 2011 11th International Conference on Hybrid Intelligent Systems (HIS), Melacca, Malaysia, 5–8 December 2011; pp. 680–685. [Google Scholar] [CrossRef]
  26. Jaleel, J.A.; Salim, S.; Aswin, R.B. Computer Aided Detection of Skin Cancer. In Proceedings of the 2013 International Conference on Circuits, Power and Computing Technologies (ICCPCT), Nagercoil, India, 20–21 March 2013; pp. 1137–1142. [Google Scholar] [CrossRef]
  27. Bayot, A.R.; Ann, L.; Niño, M.S.; Santiago, A.D. Malignancy Detection of Candidate for Basal Cell Carcinoma Using Image Processing and Artificial Neural Network. Dlsu Eng. E J. 2007, 1, 70–79. [Google Scholar]
  28. Rehman, M.; Khan, S.H.; Danish Rizvi, S.M.; Abbas, Z.; Zafar, A. Classification of Skin Lesion by Interference of Segmentation and Convolotion Neural Network. In Proceedings of the 2018 2nd International Conference on Engineering Innovation (ICEI), Bangkok, Thailand, 5–6 July 2018; pp. 81–85. [Google Scholar] [CrossRef]
  29. Harley, A.W. An Interactive Node-Link Visualization of Convolutional Neural Networks. In Advances in Visual Computing; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2015; Volume 9474, pp. 867–877. [Google Scholar] [CrossRef]
  30. Zeiler, M.D.; Fergus, R. Visualizing and Understanding Convolutional Networks. arXiv 2013, arXiv:1311.2901. Available online: http://arxiv.org/abs/1311.2901 (accessed on 15 February 2021).
  31. Nasr-Esfahani, E.; Samavi, S.; Karimi, N.; Soroushmehr, S.M.R.; Jafari, M.H.; Ward, K.; Najarian, K. Melanoma Detection by Analysis of Clinical Images Using Convolutional Neural Network. In Proceedings of the 2016 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Orlando, FL, USA, 16–20 August 2016; p. 1373. [Google Scholar] [CrossRef]
  32. Yu, L.; Chen, H.; Dou, Q.; Qin, J.; Heng, P.-A. Automated Melanoma Recognition in Dermoscopy Images via Very Deep Residual Networks. IEEE Trans. Med. Imaging 2017, 36, 994–1004. [Google Scholar] [CrossRef]
  33. DeVries, T.; Ramachandram, D. Skin Lesion Classification Using Deep Multi-Scale Convolutional Neural Networks. arXiv 2017, arXiv:1703.01402. Available online: http://arxiv.org/abs/1703.01402 (accessed on 13 February 2021).
  34. Mahbod, A.; Schaefer, G.; Wang, C.; Ecker, R.; Ellinge, I. Skin Lesion Classification Using Hybrid Deep Neural Networks. In Proceedings of the ICASSP 2019–2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 1229–1233. [Google Scholar] [CrossRef] [Green Version]
  35. Mendes, D.B.; da Silva, N.C. Skin Lesions Classification Using Convolutional Neural Networks in Clinical Images. arXiv 2018, arXiv:1812.02316. Available online: http://arxiv.org/abs/1812.02316 (accessed on 25 January 2021).
  36. Dorj, U.-O.; Lee, K.-K.; Choi, J.-Y.; Lee, M. The Skin Cancer Classification Using Deep Convolutional Neural Network. Multimed. Tools Appl. 2018, 77, 9909–9924. [Google Scholar] [CrossRef]
  37. Kalouche, S. Vision-Based Classification of Skin Cancer Using Deep Learning. 2016. Available online: https://www.semanticscholar.org/paper/Vision-Based-Classification-of-Skin-Cancer-using-Kalouche/b57ba909756462d812dc20fca157b3972bc1f533 (accessed on 10 January 2021).
  38. Ali, A.A.; Al-Marzouqi, H. Melanoma Detection Using Regular Convolutional Neural Networks. In Proceedings of the 2017 International Conference on Electrical and Computing Technologies and Applications (ICECTA), Ras Al Khaimah, United Arab Emirates, 21–23 November 2017; pp. 1–5. [Google Scholar] [CrossRef]
  39. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-Level Classification of Skin Cancer with Deep Neural Networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef] [PubMed]
  40. Shoieb, D.A.; Aly, W.M.; Youssef, S.M. Basal Cell Carcinoma Detection in Full-Field OCT Images Using Convolutional Neural Networks. JOIG 2016, 4, 122–129. [Google Scholar] [CrossRef] [Green Version]
  41. Harangi, B.; Baran, A.; Hajdu, A. Classification Of Skin Lesions Using An Ensemble Of Deep Neural Networks. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; pp. 2575–2578. [Google Scholar] [CrossRef]
  42. Mandache, D.; Dalimier, E.; Durkin, J.R.; Boceara, C.; Olivo-Marin, J.-C.; Meas-Yedid, V. Basal Cell Carcinoma Detection in Full Field OCT Images Using Convolutional Neural Networks. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 784–787. [Google Scholar] [CrossRef] [Green Version]
  43. Sabouri, P.; GholamHosseini, H. Lesion Border Detection Using Deep Learning. In Proceedings of the 2016 IEEE Congress on Evolutionary Computation (CEC), Vancouver, BC, Canada, 24–29 July 2016; pp. 1416–1421. [Google Scholar] [CrossRef]
  44. Hasan, M.; Barman, S.D.; Islam, S.; Reza, A.W. Skin Cancer Detection Using Convolutional Neural Network. In Proceedings of the 2019 5th International Conference on Computing and Artificial Intelligence-ICCAI 19; Association for Computing Machinery: New York, NY, USA, 2019; pp. 254–258. [Google Scholar] [CrossRef]
  45. Rezvantalab, A.; Safigholi, H.; Karimijeshni, S. Dermatologist Level Dermoscopy Skin Cancer Classification Using Different Deep Learning Convolutional Neural Networks Algorithms. arXiv 2018, arXiv:1810.10348. Available online: http://arxiv.org/abs/1810.10348 (accessed on 5 January 2021).
  46. Nida, N.; Irtaza, A.; Javed, A.; Yousaf, M.H.; Mahmood, M.T. Melanoma Lesion Detection and Segmentation Using Deep Region Based Convolutional Neural Network and Fuzzy C-Means Clustering. Int. J. Med. Inform. 2019, 124, 37–48. [Google Scholar] [CrossRef]
  47. Mahecha, M.S.S.; Parra, O.J.S.; Velandia, J.B. Design of a System for Melanoma Detection Through the Processing of Clinical Images Using Artificial Neural Networks. In Challenges and Opportunities in the Digital Era; Al-Sharhan, S.A., Simintiras, A.C., Dwivedi, Y.K., Janssen, M., Mäntymäki, M., Tahat, L., Moughrabi, I., Ali, T.M., Rana, N.P., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2018; Volume 11195, pp. 605–616. [Google Scholar] [CrossRef]
  48. Attia, M.; Hossny, M.; Nahavandi, S.; Yazdabadi, A. Skin Melanoma Segmentation Using Recurrent and Convolutional Neural Networks. In Proceedings of the 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI 2017), Melbourne, Australia, 18–21 April 2017; pp. 292–296. [Google Scholar] [CrossRef]
  49. Albahar, M.A. Skin Lesion Classification Using Convolutional Neural Network With Novel Regularizer. IEEE Access 2019, 7, 38306–38313. [Google Scholar] [CrossRef]
  50. Namozov, A.; Ergashev, D.; Cho, Y.I. Adaptive Activation Functions for Skin Lesion Classification Using Deep Neural Networks. In Proceedings of the 2018 Joint 10th International Conference on Soft Computing and Intelligent Systems (SCIS) and 19th International Symposium on Advanced Intelligent Systems (ISIS), Toyama, Japan, 5–8 December 2018; pp. 232–235. [Google Scholar] [CrossRef]
  51. Singh, V.; Nwogu, I. Analyzing Skin Lesions in Dermoscopy Images Using Convolutional Neural Networks. In Proceedings of the 2018 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Miyazaki, Japan, 7–10 October 2018; pp. 4035–4040. [Google Scholar] [CrossRef]
  52. Liao, H.; Li, Y.; Luo, J. Skin Disease Classification versus Skin Lesion Characterization: Achieving Robust Diagnosis Using Multi-Label Deep Neural Networks. In Proceedings of the 2016 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 355–360. [Google Scholar] [CrossRef] [Green Version]
  53. Yoshida, T.; Celebi, M.E.; Schaefer, G.; Iyatomi, H. Simple and Effective Pre-Processing for Automated Melanoma Discrimination Based on Cytological Findings. In Proceedings of the 2016 IEEE International Conference on Big Data (Big Data), Washington, DC, USA, 5–8 December 2016; pp. 3439–3442. [Google Scholar] [CrossRef]
  54. Kawahara, J.; BenTaieb, A.; Hamarneh, G. Deep Features to Classify Skin Lesions. In Proceedings of the 2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI), Prague, Czech Republic, 13–16 April 2016; pp. 1397–1400. [Google Scholar] [CrossRef]
  55. Ho, C.; Calderon-Delgado, M.; Chan, C.; Lin, M.; Tjiu, J.; Huang, S.; Chen, H.H. Detecting Mouse Squamous Cell Carcinoma from Submicron Full-field Optical Coherence Tomography Images by Deep Learning. J. Biophotonics 2021, 14. [Google Scholar] [CrossRef]
  56. Sagar, A.; Dheeba, J. Convolutional Neural Networks for Classifying Melanoma Images. bioRxiv 2020. Available online: https://www.biorxiv.org/content/10.1101/2020.05.22.110973v2 (accessed on 17 May 2021). [CrossRef]
  57. Jojoa Acosta, M.F.; Caballero Tovar, L.Y.; Garcia-Zapirain, M.B.; Percybrooks, W.S. Melanoma Diagnosis Using Deep Learning Techniques on Dermatoscopic Images. BMC Med. Imaging 2021, 21, 6. [Google Scholar] [CrossRef]
  58. Burguillo, J.C.; Dorronsoro, B. Using Complex Network Topologies and Self-Organizing Maps for Time Series Prediction. In Nostradamus 2013: Prediction, Modeling and Analysis of Complex Systems; Zelinka, I., Chen, G., Rössler, O.E., Snasel, V., Abraham, A., Eds.; Advances in Intelligent Systems and Computing; Springer International Publishing: Berlin/Heidelberg, Germany, 2013; Volume 210, pp. 323–332. [Google Scholar] [CrossRef]
  59. Lenhardt, L.; Zeković, I.; Dramićanin, T.; Dramićanin, M.D. Artificial Neural Networks for Processing Fluorescence Spectroscopy Data in Skin Cancer Diagnostics. Phys. Scr. 2013, T157, 014057. [Google Scholar] [CrossRef]
  60. Mengistu, A.D.; Alemayehu, D.M. Computer Vision for Skin Cancer Diagnosis and Recognition Using RBF and SOM. Int. J. Image Process. 2015, 9, 311–319. [Google Scholar]
  61. Sajid, P.M.; Rajesh, D.A. Performance Evaluation of Classifiers for Automatic Early Detection of Skin Cancer. J. Adv. Res. Dyn. Control. Syst. 2018, 10, 454–461. [Google Scholar]
  62. Gonog, L.; Zhou, Y. A Review: Generative Adversarial Networks. In Proceedings of the 2019 14th IEEE Conference on Industrial Electronics and Applications (ICIEA), Xi’an, China, 20 June 2019; pp. 505–510. [Google Scholar] [CrossRef]
  63. Goodfellow, I. NIPS 2016 Tutorial: Generative Adversarial Networks. arXiv 2017, arXiv:1701.00160. Available online: http://arxiv.org/abs/1701.00160 (accessed on 25 January 2021).
  64. Bissoto, A.; Perez, F.; Valle, E.; Avila, S. Skin Lesion Synthesis with Generative Adversarial Networks. arXiv 2018, arXiv:1902.03253. [Google Scholar] [CrossRef] [Green Version]
  65. Ali, I.S.; Mohamed, M.F.; Mahdy, Y.B. Data Augmentation for Skin Lesion Using Self-Attention Based Progressive Generative Adversarial Network. arXiv 2019, arXiv:1910.11960. Available online: http://arxiv.org/abs/1910.11960 (accessed on 22 January 2021).
  66. Tschandl, P.; Rosendahl, C.; Kittler, H. The HAM10000 Dataset, a Large Collection of Multi-Source Dermatoscopic Images of Common Pigmented Skin Lesions. Sci. Data. 2018, 5, 180161. [Google Scholar] [CrossRef] [PubMed]
  67. Shoieb, D.A.; Youssef, S.M.; Aly, W.M. Computer-Aided Model for Skin Diagnosis Using Deep Learning. J. Image Graph. 2016, 4, 116–121. [Google Scholar] [CrossRef] [Green Version]
  68. Mendonca, T.; Ferreira, P.M.; Marques, J.S.; Marcal, A.R.S.; Rozeira, J. PH2-A Dermoscopic Image Database for Research and Benchmarking. In Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Osaka, Japan, 3–7 July 2013; pp. 5437–5440. [Google Scholar] [CrossRef]
  69. ISIC Archive. Available online: https://isic-archive.com/ (accessed on 1 January 2021).
  70. Codella, N.C.F.; Gutman, D.; Celebi, M.E.; Helba, B.; Marchetti, M.A.; Dusza, S.W.; Kalloo, A.; Liopyris, K.; Mishra, N.; Kittler, H.; et al. Skin Lesion Analysis toward Melanoma Detection: A Challenge at the 2017 International Symposium on Biomedical Imaging (ISBI), Hosted by the International Skin Imaging Collaboration (ISIC). In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 168–172. [Google Scholar] [CrossRef] [Green Version]
  71. Boer, A.; Nischal, K. Www.Derm101.Com: A Growing Online Resource for Learning Dermatology and Dermatopathology. Indian J. Derm. Venereol. Leprol. 2007, 73, 138. [Google Scholar] [CrossRef]
  72. DermIS. Available online: https://www.dermis.net/dermisroot/en/ (accessed on 5 January 2021).
  73. Argenziano, G.; Soyer, H.P.; Giorgi, V.D.; Piccolo, D.; Carli, P.; Delfino, M.; Ferrari, A.; Hofmann-Wellenhof, R.; Massi, D.; Mazzocchetti, G.; et al. Interactiveatlas of Dermoscopy; EDRA Medical Publishing and New Media: Milan, Italy, 2000. [Google Scholar]
  74. Dermnet. Available online: http://www.dermnet.com/dermatology-pictures-skin-disease-pictures/ (accessed on 8 January 2021).
  75. Stanley, R.J.; Stoecker, W.V.; Moss, R.H.; Gupta, K.; Jella, P. Color and Structural Features for Automatic Skin Lesion Discrimination in Dermoscopy Images. In Proceedings of the 6th World Congress on Melanoma, Vancouver, BC, Canada, 6–10 September 2005. [Google Scholar]
  76. Brinker, T.J.; Hekler, A.; Utikal, J.S.; Grabe, N.; Schadendorf, D.; Klode, J.; Berking, C.; Steeb, T.; Enk, A.H.; von Kalle, C. Skin Cancer Classification Using Convolutional Neural Networks: Systematic Review. J. Med. Internet Res. 2018, 20, e11936. [Google Scholar] [CrossRef]
  77. Garcovich, S.; Colloca, G.; Sollena, P.; Andrea, B.; Balducci, L.; Cho, W.C.; Bernabei, R.; Peris, K. Skin Cancer Epidemics in the Elderly as An Emerging Issue in Geriatric Oncology. Aging Dis. 2017, 8, 643–661. [Google Scholar] [CrossRef] [Green Version]
  78. Öztürk, Ş.; Ahmad, R.; Akhtar, N. Variants of Artificial Bee Colony Algorithm and Its Applications in Medical Image Processing. Appl. Soft Comput. 2020, 97, 106799. [Google Scholar] [CrossRef]
  79. Fahad, L.G.; Tahir, S.F.; Shahzad, W.; Hassan, M.; Alquhayz, H.; Hassan, R. Ant Colony Optimization-Based Streaming Feature Selection: An Application to the Medical Image Diagnosis. Sci. Program. 2020, 2020, 1–10. [Google Scholar] [CrossRef]
  80. Husodo, A.Y.; Jati, G.; Octavian, A.; Jatmiko, W. Enhanced Social Spider Optimization Algorithm for Increasing Performance of Multiple Pursuer Drones in Neutralizing Attacks From Multiple Evader Drones. IEEE Access 2020, 8, 22145–22161. [Google Scholar] [CrossRef]
  81. Berlin, S.J.; John, M. Particle Swarm Optimization with Deep Learning for Human Action Recognition. Multimed. Tools Appl. 2020, 79, 17349–17371. [Google Scholar] [CrossRef]
  82. Sturm, R.A. Skin Colour and Skin Cancer -MC1R, the Genetic Link. Melanoma Res. 2002, 12, 405–416. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The process of skin cancer detection. ANN = Artificial neural network; CNN = Convolutional neural network; KNN = Kohonen self-organizing neural network; GAN = Generative adversarial neural network.
Figure 1. The process of skin cancer detection. ANN = Artificial neural network; CNN = Convolutional neural network; KNN = Kohonen self-organizing neural network; GAN = Generative adversarial neural network.
Ijerph 18 05479 g001
Figure 2. Skin disease categories from International Skin Imaging Collaboration (ISIC) dataset [12].
Figure 2. Skin disease categories from International Skin Imaging Collaboration (ISIC) dataset [12].
Ijerph 18 05479 g002
Figure 3. Basic ANN structure [13].
Figure 3. Basic ANN structure [13].
Ijerph 18 05479 g003
Figure 4. Skin cancer detection using ANN [19].
Figure 4. Skin cancer detection using ANN [19].
Ijerph 18 05479 g004
Figure 5. Basic CNN Architecture [9].
Figure 5. Basic CNN Architecture [9].
Ijerph 18 05479 g005
Figure 6. Skin cancer diagnosis using CNN [37].
Figure 6. Skin cancer diagnosis using CNN [37].
Ijerph 18 05479 g006
Figure 7. Basic KNN structure [58], BMU= Best matching unit.
Figure 7. Basic KNN structure [58], BMU= Best matching unit.
Ijerph 18 05479 g007
Figure 8. GAN architecture [64].
Figure 8. GAN architecture [64].
Ijerph 18 05479 g008
Table 1. Search terms.
Table 1. Search terms.
Search TermSet of Keywords
Skin *Skin cancer, skin diseases, skin treatment
Cancer *Cancer disease, cancer types, cancer diagnosis, cancer treatment
Deep *Deep learning, deep neural networks
Neural *Neural network, neural networking
Network *Neural network, neural networking
Melano *Networking, network types
NonMelano *Melanoma skin cancer, melanoma death rate, melanoma treatment, melanoma diagnosis, melanoma causes, melanoma symptoms
Basal *Basal cell carcinoma, basal cell carcinoma skin cancer, basal cell carcinoma diagnosis, basal cell carcinoma causes, basal cell carcinoma symptoms
Squamous *Squamous cell carcinoma, squamous cell carcinoma skin cancer, squamous cell carcinoma diagnosis, squamous cell carcinoma causes, squamous cell carcinoma symptoms
Artificial *Artificial neural network, artificial neural networking,
Back *Backpropagation neural network
Conv *Convolutional neural network
* = All words that start with the string written before asterisk *.
Table 2. Search results.
Table 2. Search results.
Sr. NoResourceInitial SearchTitle-Based
Selection
Abstract-Based
Selection
Full Paper-Based
Selection
1IEEE Xplore123211513
2Google Scholar 45129118
3ACM DL3271995
4Springer235111715
5Science Direct347151210
Total1483956451
Table 3. A comparative analysis of skin cancer detection using ANN-based approaches.
Table 3. A comparative analysis of skin cancer detection using ANN-based approaches.
RefSkin Cancer
Diagnoses
Classifier and Training
Algorithm
DatasetDescriptionResults (%)
[23]MelanomaANN with backpropagation algorithm31 dermoscopic imagesABCD parameters for feature extraction,Accuracy (96.9)
[20]Melanoma/Non- melanomaANN with backpropagation algorithm90 dermoscopic imagesmaximum entropy for thresholding, and gray- level co-occurrence matrix for features extractionAccuracy (86.66)
[19]Cancerous/non- cancerousANN with backpropagation algorithm31 dermoscopic images 2D-wavelet transform for feature extraction and thresholding for segmentationNil
[24]Malignant
/benign
Feed-forward ANN with the backpropagation training algorithm326 lesion
images
Color and shape characteristics of the tumor were used as discriminant features for classificationAccuracy (80)
[25]Malignant/non-MalignantBackpropagation neural network as NN classifier448 mixed-type imagesROI and SRM for segmentationAccuracy (70.4)
[21]Cancerous/noncancerousANN with backpropagation algorithm30 cancerous/noncancerous imagesRGB color features and GLCM techniques for feature extractionAccuracy (86.66)
[18]Common mole/non-common mole/melanomaFeed-forward BPNN200 dermoscopic imagesFeatures extracted according to ABCD ruleAccuracy (97.51)
[26]Cancerous/noncancerousArtificial neural network with backpropagation algorithm50 dermoscopic imagesGLCM technique for feature extractionAccuracy (88)
[27]BCC/non-BCCANN180 skin lesion images Histogram equalization for contrast enhancementReliability (93.33)
[14]Melanoma/Non-melanomaANN with Levenberg–Marquardt (LM), resilient backpropagation (RBP), and scaled conjugate gradient (GCG) learning algorithms135 lesion
images
Combination of multiple classifiers to avoid the misclassificationAccuracy (SCG:91.9, LM: 95.1, RBP:88.1)
[13]Malignant/benignANN meta-ensemble model consisting of BPN and fuzzy neural networkCaucasian race and xanthous-race datasetsSelf-generating neural network was used for
lesion extraction
Accuracy (94.17)
Sensitivity (95), specificity (93.75)
ANN = Artificial neural network, NN = Neural network. ROI = Region of interest, SRM = Statistical region merging, GLCM = Gray level co-occurrence matrix, BPNN = Backpropagation neural network.
Table 4. A comparative analysis of skin cancer detection using CNN-based approaches.
Table 4. A comparative analysis of skin cancer detection using CNN-based approaches.
RefSkin Cancer DiagnosesClassifier and Training
Algorithm
DatasetDescriptionResults (%)
[38]Benign/malignantLightNet (deep learning framework), used for classificationISIC 2016 datasetFewer parameters and well suited for mobile applicationsAccuracy (81.6), sensitivity (14.9), specificity (98)
[31]Melanoma/benignCNN classifier170 skin lesion imagesTwo convolving layers in CNNAccuracy (81), sensitivity (81), specificity (80)
[36]BCC/SCC/melanoma/AKSVM with deep CNN3753 dermoscopic images Pertained to deep CNN and AlexNet for features extractionAccuracy (SCC: 95.1, AK: 98.9, BCC: 94.17)
[39]Melanoma /benign
Keratinocyte carcinomas/benign SK
Deep CNNISIC-Dermoscopic ArchiveExpert-level performance against 21 certified dermatologistsAccuracy (72.1)
[35]Malignant melanoma and BC carcinomaCNN with Res-Net 152 architectureThe first dataset has 170 images the second dataset contains 1300 images Augmentor Python library for augmentation.AUC (melanoma: 96, BCC: 91)
[40]Melanoma/nonmelanomaSVM-trained, with CNN, extracted featuresDermIS dataset and DermQuest dataA median filter for noise removal and CNN for feature extractionAccuracy (93.75)
[41]Malignant melanoma/nevus/SKCNN as single neural-net architectureISIC 2017 datasetCNN ensemble of AlexNet, VGGNet, and GoogleNetfor classificationAverage AUC:9 84.8), average accuracy (83.8)
[42]BCC/nonBCCCNN40 FF-OCT imagesTrained CNN, consisted of 10 layers for features extractionAccuracy (95.93), sensitivity (95.2), specificity (96.54)
[43]Cancerous/noncancerousCNN1730 skin lesion and background imagesFocused on edge detectionAccuracy (86.67)
[37]Benign/melanomaVGG-16 and CNNISIC datasetDataset was trained on three separate learning modelsAccuracy (78)
[44]Benign/malignantCNNISIC databaseABCD symptomatic checklist for feature extractionAccuracy (89.5)
[45]Melanoma/benign keratosis/ melanocytic nevi/BCC/AK/IC/atypical nevi/dermatofibroma/vascular lesionsDeep CNN architecture (DenseNet 201, Inception v3, ResNet 152 and
InceptionResNet v2)
HAM10000 and PH2 datasetDeep learning models outperformed highly trained dermatologists in overall mean results by at least 11%ROC AUC
(DenseNet 201: 98.79–98.16, Inception v3:
98.60–97.80,
ResNet 152: 98.61–98.04,
InceptionResNet v2: 98.20–96.10)
[46]Lipoma/fibroma/sclerosis/melanomaDeep region-based CNN
and fuzzy C means clustering
ISIC datasetCombination of the region-based CNN and fuzzy C-means ensured more accuracy in disease detectionAccuracy (94.8) sensitivity (97.81) specificity (94.17) F1_score (95.89)
[47]Malignant/benign6-layers deep CNNMED-NODE and ISIC datasetsIllumination factor in images affected performance of the systemAccuracy (77.50)
[48]Melanoma/non melanomaHybrid of fully CNN with autoencoder and decoder and RNNISIC datasetProposed model outperformed state-of-art SegNet, FCN, and ExB architectureAccuracy (98) Jaccard index (93), sensitivity (95), specificity (94)
[49]Benign/malignant2-layer CNN with a novel regularizerISIC datasetProposed regularization technique controlled complexity by adding a penalty on the dispersion value of classifier’s weight matrix Accuracy (97.49) AUC (98), sensitivity (94.3), specificity (93.6)
[34]Malignant melanoma/SKSVM classification with features extracted with pretrained deep models named AlexNet, ResNet-18, and VGG16ISIC datasetSVM scores were mapped to
probabilities with logistic regression function for evaluation
Average AUC (90.69)
[12]Melanoma/BCC/melanocytic nevus/Bowen’s disease/AK/benign keratosis/vascular lesion/dermatofibromaInceptionResNetV2, PNASNet-5-Large,
InceptionV4, and
SENet154
ISIC datasetA trained image-net model was used to initialize network parameters and fine-tuning Validation Score (76)
[50]melanoma/BCC/melanocytic nevus/AK/benign keratosis/vascular lesion/dermatofibromaCNN model with
LeNet approach
ISIC datasetThe adaptive piecewise linear activation function was used to increase system performanceAccuracy (95.86)
[51]Benign/malignantDeep CNNISIC datasetData augmentation was performed for data balancingAccuracy (80.3), precision (81), AUC (69)
[52]Compound nevus/malignant melanomaCNNAtlasDerm, Derma, Dermnet, Danderm, DermIS and DermQuest datasetsBVLC-AlexNet model, pretrained from ImageNet dataset was used for fine-tuningMean average precision (70)
[33]Melanoma/SKDeep multi-scale CNNISIC datasetThe proposed model used Inception-v3 model, which was trained on the ImageNet.Accuracy (90.3), AUC (94.3)
[53]Benign/malignantCNN with 5-fold cross-validation1760 dermoscopic images Images were preprocessed on the basis of melanoma cytological findings Accuracy (84.7), sensitivity (80.9),
specificity (88.1)
[32]Benign/malignantA very deep residual CNN and FCRNISIC 2016 databaseFCRN incorporated with a multi-scale contextual information integration technique was proposed for accurate lesions segmentationAccuracy (94.9), sensitivity (91.1), specificity (95.7), Jaccard index (82.9), dice coefficient (89.7)
[54]AK/melanocytic nevus/BCC/SK/SCCCNN1300 skin lesion imagesMean subtraction for each image, pooled multi-scale feature extraction process and pooling in augmented-feature spaceAccuracy (81.8)
[55]BCC/non-BCCPruned ResNet18297 FF-OCT imagesK-fold cross-validation was applied to measure the performance of the proposed systemAccuracy (80)
[56]Melanoma/non melanomaResNet-50 with deep transfer learning3600 lesion images from the ISIC datasetThe proposed model showed better performance than o InceptionV3, Densenet169, Inception ResNetV2, and MobilenetAccuracy (93.5), precision (94)
recall (77), F1_ score (85)
[57]Benign/malignantRegion-based CNN with ResNet1522742 dermoscopic images from ISIC datasetRegion of interest was extracted by mask and region-based CNN, then ResNet152 is used for classification.Accuracy (90.4), sensitivity (82),
specificity (92.5)
CNN = Convolutional neural network; ISIC = International skin imaging collaboration; SVM = Support vector machine; BCC = Basal cell carcinoma; SCC = Squamous cell carcinoma; AK = Actinic keratosis; IC = Intraepithelial carcinoma; HAM10000 = Human-against-machine dataset with 10,000 images; BVLC = Berkeley Vision and Learning Center; SK= Seborrheic keratosis; FCRN = Fully convolutional residual network; FF-OCT = Full field optical coherence tomography; FCN = Fully convolutional network.
Table 5. A comparative analysis of skin cancer detection using KNN-based approaches.
Table 5. A comparative analysis of skin cancer detection using KNN-based approaches.
RefSkin Cancer
Diagnoses
Classifier and Training AlgorithmDatasetDescriptionResults (%)
[59]Melanoma/nevus/normal skinSOM and feed-forward NN50 skin lesion imagesPCA for decreasing spectra’s dimensionalityAccuracy (96–98)
[60]BCC, SCC, and melanomaSOM and RBFDermQuest and Dermnet datasets15 features consisting of GCM morphological and color features were extractedAccuracy (93.15)
[61]Cancerous/noncancerousModified KNN500 lesion imagesAutomated Otsu method of thresholding for segmentationAccuracy (98.3)
SOM = Self organizing map; PCA = Principal component analysis; GCM = Generalized co-occurrence matrices; RBF = Radial Basis Function; KNN = Kohonen self-organizing neural network.
Table 6. A comparative analysis of skin cancer detection using GAN-based approaches.
Table 6. A comparative analysis of skin cancer detection using GAN-based approaches.
RefSkin Cancer DiagnosesClassifier and Training AlgorithmDatasetDescriptionResults (%)
[7]AK/BCC/benign keratosis/dermatofibroma/melanoma/melanocytic nevus/vascular lesionGANISIC 2018The proposed system used deconvolutional network and CNN as generator and discriminator moduleAccuracy (86.1)
[8]Melanoma/nevus/SKDeep convolutional GANISIC 2017, ISIC 2018, PH2Decoupled deep convolutional GANs for data augmentationROC AUC (91.5), accuracy (86.1)
[65]BCC/vascular/pigmented benign keratosis/pigmented Bowen’s/nevus/dermatofibromaSelf-attention-based PGANISIC 2018A generative model was enhanced with a stabilization techniqueAccuracy (70.1)
GAN = Generative adversarial neural network, PGAN = Progressive generative adversarial network, ROC AUC= Area under the receiver operating characteristic curve.
Table 7. Skin Cancer Datasets.
Table 7. Skin Cancer Datasets.
Sr. NoName of DatasetYear of ReleaseNo. of ImagesReference Used
1HAM10000201810,015[45]
2PH22013200[45]
3ISIC archive201625,331[12,33,34,37,44,46,47,48,49,51,53]
4DermQuest199922,082[52,63,67]
5DermIS 6588[52,63]
6AtlasDerm20001024[52]
7Dermnet199823,000[52,59]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Dildar, M.; Akram, S.; Irfan, M.; Khan, H.U.; Ramzan, M.; Mahmood, A.R.; Alsaiari, S.A.; Saeed, A.H.M.; Alraddadi, M.O.; Mahnashi, M.H. Skin Cancer Detection: A Review Using Deep Learning Techniques. Int. J. Environ. Res. Public Health 2021, 18, 5479. https://0-doi-org.brum.beds.ac.uk/10.3390/ijerph18105479

AMA Style

Dildar M, Akram S, Irfan M, Khan HU, Ramzan M, Mahmood AR, Alsaiari SA, Saeed AHM, Alraddadi MO, Mahnashi MH. Skin Cancer Detection: A Review Using Deep Learning Techniques. International Journal of Environmental Research and Public Health. 2021; 18(10):5479. https://0-doi-org.brum.beds.ac.uk/10.3390/ijerph18105479

Chicago/Turabian Style

Dildar, Mehwish, Shumaila Akram, Muhammad Irfan, Hikmat Ullah Khan, Muhammad Ramzan, Abdur Rehman Mahmood, Soliman Ayed Alsaiari, Abdul Hakeem M Saeed, Mohammed Olaythah Alraddadi, and Mater Hussen Mahnashi. 2021. "Skin Cancer Detection: A Review Using Deep Learning Techniques" International Journal of Environmental Research and Public Health 18, no. 10: 5479. https://0-doi-org.brum.beds.ac.uk/10.3390/ijerph18105479

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop