Selection
A thorough study of the full text of the selected research papers sought answers to certain quality control questions. The current systematic research asked the following quality assessment questions.
The first quality assessment question focused on the thorough coverage of deep learning techniques for skin cancer detection. The quality of a selected paper was verified by the reputation of the journal in which it was published and by its citations. The third question ensured that the research answered the research questions mentioned in Section 2 . Only the most relevant research papers to our domain of study were extracted. These papers had to satisfy those above research questions to qualify for selection. Research papers that failed to adequately answer the research or quality control questions and papers with text that was not related to our study topic were excluded.
Each question had Boolean ’yes/no’ responses. Each ‘yes’ was assigned a value Y = 1 and each ‘no’ was assigned a value N = 0. The first quality control question evaluated the topic coverage of the 51 selected research papers and resulted in a value of 77%, which was quite satisfactory. The second research question verified the quality of the selected papers, which resulted in improvement of quality. It generated an 82% result, which was satisfactory. The third question was a very important question in order to answer the review’s main research questions. It generated a 79% result, which was the indicator of the adequacy of the studies to answer the research questions posed by the review. The overall results of answers to these quality questions seemed healthy.
Deep neural networks play a significant role in skin cancer detection. They consist of a set of interconnected nodes. Their structure is similar to the human brain in terms of neuronal interconnectedness. Their nodes work cooperatively to solve particular problems. Neural networks are trained for certain tasks; subsequently, the networks work as experts in the domains in which they were trained. In our study, neural networks were trained to classify images and to distinguish between various types of skin cancer. Different types of skin lesion from International Skin Imaging Collaboration (ISIC) dataset are presented in Figure 2 . We searched for different techniques of learning, such as ANN, CNN, KNN, and GAN for skin cancer detection systems. Research related to each of these deep neural networks is discussed in detail in this section.
Skin disease categories from International Skin Imaging Collaboration (ISIC) dataset [ 12 ].
An artificial neural network is a nonlinear and statistical prediction method. Its structure is borrowed from the biological structure of the human brain. An ANN consists of three layers of neurons. The first layer is known as the input layer; these input neurons transfer data to the second/intermediate layer of neurons. The intermediate layers are referred to as hidden layers. In a typical ANN, there can be several hidden layers. Intermediate neurons send data to the third layer of output neurons. Computations are learned at each layer using backpropagation, which is used for learning the complex associations/relationships between input and output layers. It is similar to a neural network. Currently, in computer science, the term neural network and artificial neural network are used interchangeably. The basic structure of an ANN network is presented in Figure 3 .
Basic ANN structure [ 13 ].
ANN is used for the classification of extracted features in skin cancer detection systems. Input images are classified as melanoma or nonmelanoma after successful training/classification of the training set. The number of hidden layers in an ANN depends on the number of input images. The input/first layer of the ANN process connects with the hidden layer by the input dataset. The dataset can be labeled or unlabeled, which can be processed accordingly using a supervised or unsupervised learning mechanism. A neural network uses backpropagation or feed-forward architecture to learn weights present at each network connection/link. Both architectures use a different pattern for the underlying dataset. Feed-forward-architecture-based neural networks transfer data only in one direction. Data flows only from the input to the output layer.
Xie et al. [ 14 ] proposed a skin lesion classification system that classified lesions into two main classes: benign and malignant. The proposed system worked in three phases. In the initial phase, a self-generating NN was used to extract lesions from images. In the second phase, features such as tumor border, texture, and color details were extracted. The system extracted a total of 57 features, including 7 novel features related to lesion borders descriptions. Principal component analysis (PCA) was used to reduce the dimensionality of the features, which led to the selection of the optimal set of features. Finally, in the last phase, lesions were classified using a NN ensemble model. Ensemble NN improves classification performance by combining backpropagation (BP) NN and fuzzy neural networks. Furthermore, the proposed system classification results were compared with other classifiers, such as SVM, KNN, random forest, Adaboot, etc. With a 91.11% accuracy, the proposed model achieved at least 7.5% higher performance in terms of sensitivity than the other classifiers.
Masood et al. [ 15 ] proposed an ANN-based automated skin cancer diagnostic system. The performance of three ANN’s learning algorithms such as Levenberg–Marquardt (LM) [ 16 ], resilient backpropagation (RP) [ 17 ], scaled conjugate gradient (SCG) [ 18 ], was also investigated by this paper. Comparison of performance showed that the LM algorithm achieved the highest specificity score (95.1%) and remained efficient at the classification of benign lesions, while the SCG learning algorithm produced better results if the number of epochs was increased, scoring a 92.6% sensitivity value. A mole classification system for the early diagnosis of melanoma skin cancer was proposed [ 19 ]. The proposed system extracted features according to the ABCD rule of lesions. ABCD refers to asymmetry of a mole’s form, borders of mole, color, and diameter of mole. Assessment of a mole’s asymmetry and borders were extracted using the Mumford–Shah algorithm and Harris Stephen algorithm, respectively. Normal moles are composed of black, cinnamon, or brown color, so moles with colors other than those three were considered melanoma in the proposed system. Melanoma moles commonly have a diameter value greater than 6 mm, so that value was used as the threshold value of diameter for melanoma detection. The proposed system used a backpropagation feed-forward ANN to classify moles into three classes, such as common mole, uncommon mole, or melanoma mole, with 97.51% accuracy.
An automated skin cancer diagnostic system based on backpropagation ANN was proposed [ 20 ], represented in Figure 4 . This system employed a 2D-wavelet transform technique for feature extraction. The proposed ANN model classified the input images into two classes, such as cancerous or noncancerous. Another ANN-based skin cancer diagnostic system was proposed by Choudhari and Biday [ 21 ]. Images were segmented with a maximum entropy thresholding measure. A gray-level co-occurrence matrix (GLCM) was used to extract unique features of skin lesions. Finally, a feed-forward ANN classified the input images into either a malignant or benign stage of skin cancer, achieving an accuracy level of 86.66%.
Skin cancer detection using ANN [ 19 ].
Aswin et al. [ 22 ] described a new method for skin cancer detection based on a genetic algorithm (GA) and ANN algorithms. Images were preprocessed for hair removal with medical imaging software named Dull-Rozar and region of interest (ROI) and were extracted with the Otsu thresholding method. Furthermore, the GLCM technique was employed to extract unique features of the segmented images. Subsequently, a hybrid ANN and GA classifier was used for the classification of lesion images into cancerous and noncancerous classes. The proposed system achieved an overall accuracy score of 88%. Comprehensive details of the various skin cancer detection systems based on ANN are listed in Table 3 below.
A comparative analysis of skin cancer detection using ANN-based approaches.
Ref | Skin Cancer Diagnoses | Classifier and Training Algorithm | Dataset | Description | Results (%) |
---|---|---|---|---|---|
[ ] | Melanoma | ANN with backpropagation algorithm | 31 dermoscopic images | ABCD parameters for feature extraction, | Accuracy (96.9) |
[ ] | Melanoma/Non- melanoma | ANN with backpropagation algorithm | 90 dermoscopic images | maximum entropy for thresholding, and gray- level co-occurrence matrix for features extraction | Accuracy (86.66) |
[ ] | Cancerous/non- cancerous | ANN with backpropagation algorithm | 31 dermoscopic images | 2D-wavelet transform for feature extraction and thresholding for segmentation | Nil |
[ ] | Malignant /benign | Feed-forward ANN with the backpropagation training algorithm | 326 lesion images | Color and shape characteristics of the tumor were used as discriminant features for classification | Accuracy (80) |
[ ] | Malignant/non-Malignant | Backpropagation neural network as NN classifier | 448 mixed-type images | ROI and SRM for segmentation | Accuracy (70.4) |
[ ] | Cancerous/noncancerous | ANN with backpropagation algorithm | 30 cancerous/noncancerous images | RGB color features and GLCM techniques for feature extraction | Accuracy (86.66) |
[ ] | Common mole/non-common mole/melanoma | Feed-forward BPNN | 200 dermoscopic images | Features extracted according to ABCD rule | Accuracy (97.51) |
[ ] | Cancerous/noncancerous | Artificial neural network with backpropagation algorithm | 50 dermoscopic images | GLCM technique for feature extraction | Accuracy (88) |
[ ] | BCC/non-BCC | ANN | 180 skin lesion images | Histogram equalization for contrast enhancement | Reliability (93.33) |
[ ] | Melanoma/Non-melanoma | ANN with Levenberg–Marquardt (LM), resilient backpropagation (RBP), and scaled conjugate gradient (GCG) learning algorithms | 135 lesion images | Combination of multiple classifiers to avoid the misclassification | Accuracy (SCG:91.9, LM: 95.1, RBP:88.1) |
[ ] | Malignant/benign | ANN meta-ensemble model consisting of BPN and fuzzy neural network | Caucasian race and xanthous-race datasets | Self-generating neural network was used for lesion extraction | Accuracy (94.17) Sensitivity (95), specificity (93.75) |
ANN = Artificial neural network, NN = Neural network. ROI = Region of interest, SRM = Statistical region merging, GLCM = Gray level co-occurrence matrix, BPNN = Backpropagation neural network.
A convolution neural network is an essential type of deep neural network, which is effectively being used in computer vision. It is used for classifying images, assembling a group of input images, and performing image recognition. CNN is a fantastic tool for collecting and learning global data as well as local data by gathering more straightforward features such as curves and edges to produce complex features such as shapes and corners [ 28 ]. CNN’s hidden layers consist of convolution layers, nonlinear pooling layers, and fully connected layers [ 29 ]. CNN can contain multiple convolution layers that are followed by several fully connected layers. Three major types of layers involved in making CNN are convolution layers, pooling layers, and full-connected layers [ 30 ]. The basic architecture of a CNN is presented in Figure 5 .
Basic CNN Architecture [ 9 ].
CNN-based automated deep learning algorithms have achieved remarkable performance in the detection, segmentation, and classification operations of medical imaging [ 31 ]. Lequan et al. [ 32 ] proposed a very deep CNN for melanoma detection. A fully convolutional residual network (FCRN) having 16 residual blocks was used in the segmentation process to improve performance. The proposed technique used an average of both SVM and softmax classifier for classification. It showed 85.5% accuracy in melanoma classification with segmentation and 82.8% without segmentation. DeVries and Ramachandram [ 33 ] proposed a multi-scale CNN using an inception v3 deep neural network that was trained on an ImageNet dataset. For skin cancer classification, the pre-trained inception v3 was further fined-tuned on two resolution scales of input lesion images: coarse-scale and finer scale. The coarse-scale was used to capture shape characteristics as well as overall contextual information of lesions. In contrast, the finer scale gathered textual detail of lesion for differentiation between various types of skin lesions.
Mahbod et al. [ 34 ] proposed a technique to extract deep features from various well-established and pre-trained deep CNNs for skin lesions classification. Pretrained AlexNet, ResNet-18 and VGG16 were used as deep-feature generators, then a multi-class SVM classifier was trained on these generated features. Finally, the classifier results were fused to perform classification. The proposed system was evaluated on the ISIC 2017 dataset and showed 97.55% and 83.83% area under the curve (AUC) performance for seborrheic keratosis (SK) and melanoma classification. A deep CNN architecture based on pre-trained ResNet-152 was proposed to classify 12 different kinds of skin lesions [ 35 ]. Initially, it was trained on 3797 lesion images; however, later, 29-times augmentation was applied based on lighting positions and scale transformations. The proposed technique provided an AUC value of 0.99 for the classification of hemangioma lesion, pyogenic granuloma (PG) lesion, and intraepithelial carcinoma (IC) skin lesions.
A technique for the classification of four different types of skin lesion images was proposed by Dorj et al. [ 36 ]. A pre-trained deep CNN named AlexNet was used for feature extraction, after which error-correcting output coding SVM worked as a classifier. The proposed system produced the highest scores of the average sensitivity, specificity, and accuracy for SCC, actinic keratosis (AK), and BCC: 95.1%, 98.9%, and 94.17%, respectively. Kalouche [ 37 ] proposed a pre-trained deep CNN architecture VGG-16 with a final three fine-tuned layers and five convolutional blocks. The proposed VCG-16 model is represented in Figure 6 . VCG-16 models showed 78% accuracy for the classification of lesion images as melanoma skin cancer. A deep CNN-based system was proposed to detect the borders of skin lesions in images. The deep learning model was trained on 1200 normal skin images and 400 images of skin lesions. The proposed system classified the input images into two main classes, normal skin image and lesion image, with 86.67% accuracy. A comprehensive list of skin cancer detection systems using CNN classifiers is presented in Table 4 .
Skin cancer diagnosis using CNN [ 37 ].
A comparative analysis of skin cancer detection using CNN-based approaches.
Ref | Skin Cancer Diagnoses | Classifier and Training Algorithm | Dataset | Description | Results (%) |
---|---|---|---|---|---|
[ ] | Benign/malignant | LightNet (deep learning framework), used for classification | ISIC 2016 dataset | Fewer parameters and well suited for mobile applications | Accuracy (81.6), sensitivity (14.9), specificity (98) |
[ ] | Melanoma/benign | CNN classifier | 170 skin lesion images | Two convolving layers in CNN | Accuracy (81), sensitivity (81), specificity (80) |
[ ] | BCC/SCC/melanoma/AK | SVM with deep CNN | 3753 dermoscopic images | Pertained to deep CNN and AlexNet for features extraction | Accuracy (SCC: 95.1, AK: 98.9, BCC: 94.17) |
[ ] | Melanoma /benign Keratinocyte carcinomas/benign SK | Deep CNN | ISIC-Dermoscopic Archive | Expert-level performance against 21 certified dermatologists | Accuracy (72.1) |
[ ] | Malignant melanoma and BC carcinoma | CNN with Res-Net 152 architecture | The first dataset has 170 images the second dataset contains 1300 images | Augmentor Python library for augmentation. | AUC (melanoma: 96, BCC: 91) |
[ ] | Melanoma/nonmelanoma | SVM-trained, with CNN, extracted features | DermIS dataset and DermQuest data | A median filter for noise removal and CNN for feature extraction | Accuracy (93.75) |
[ ] | Malignant melanoma/nevus/SK | CNN as single neural-net architecture | ISIC 2017 dataset | CNN ensemble of AlexNet, VGGNet, and GoogleNetfor classification | Average AUC:9 84.8), average accuracy (83.8) |
[ ] | BCC/nonBCC | CNN | 40 FF-OCT images | Trained CNN, consisted of 10 layers for features extraction | Accuracy (95.93), sensitivity (95.2), specificity (96.54) |
[ ] | Cancerous/noncancerous | CNN | 1730 skin lesion and background images | Focused on edge detection | Accuracy (86.67) |
[ ] | Benign/melanoma | VGG-16 and CNN | ISIC dataset | Dataset was trained on three separate learning models | Accuracy (78) |
[ ] | Benign/malignant | CNN | ISIC database | ABCD symptomatic checklist for feature extraction | Accuracy (89.5) |
[ ] | Melanoma/benign keratosis/ melanocytic nevi/BCC/AK/IC/atypical nevi/dermatofibroma/vascular lesions | Deep CNN architecture (DenseNet 201, Inception v3, ResNet 152 and InceptionResNet v2) | HAM10000 and PH2 dataset | Deep learning models outperformed highly trained dermatologists in overall mean results by at least 11% | ROC AUC (DenseNet 201: 98.79–98.16, Inception v3: 98.60–97.80, ResNet 152: 98.61–98.04, InceptionResNet v2: 98.20–96.10) |
[ ] | Lipoma/fibroma/sclerosis/melanoma | Deep region-based CNN and fuzzy C means clustering | ISIC dataset | Combination of the region-based CNN and fuzzy C-means ensured more accuracy in disease detection | Accuracy (94.8) sensitivity (97.81) specificity (94.17) F1_score (95.89) |
[ ] | Malignant/benign | 6-layers deep CNN | MED-NODE and ISIC datasets | Illumination factor in images affected performance of the system | Accuracy (77.50) |
[ ] | Melanoma/non melanoma | Hybrid of fully CNN with autoencoder and decoder and RNN | ISIC dataset | Proposed model outperformed state-of-art SegNet, FCN, and ExB architecture | Accuracy (98) Jaccard index (93), sensitivity (95), specificity (94) |
[ ] | Benign/malignant | 2-layer CNN with a novel regularizer | ISIC dataset | Proposed regularization technique controlled complexity by adding a penalty on the dispersion value of classifier’s weight matrix | Accuracy (97.49) AUC (98), sensitivity (94.3), specificity (93.6) |
[ ] | Malignant melanoma/SK | SVM classification with features extracted with pretrained deep models named AlexNet, ResNet-18, and VGG16 | ISIC dataset | SVM scores were mapped to probabilities with logistic regression function for evaluation | Average AUC (90.69) |
[ ] | Melanoma/BCC/melanocytic nevus/Bowen’s disease/AK/benign keratosis/vascular lesion/dermatofibroma | InceptionResNetV2, PNASNet-5-Large, InceptionV4, and SENet154 | ISIC dataset | A trained image-net model was used to initialize network parameters and fine-tuning | Validation Score (76) |
[ ] | melanoma/BCC/melanocytic nevus/AK/benign keratosis/vascular lesion/dermatofibroma | CNN model with LeNet approach | ISIC dataset | The adaptive piecewise linear activation function was used to increase system performance | Accuracy (95.86) |
[ ] | Benign/malignant | Deep CNN | ISIC dataset | Data augmentation was performed for data balancing | Accuracy (80.3), precision (81), AUC (69) |
[ ] | Compound nevus/malignant melanoma | CNN | AtlasDerm, Derma, Dermnet, Danderm, DermIS and DermQuest datasets | BVLC-AlexNet model, pretrained from ImageNet dataset was used for fine-tuning | Mean average precision (70) |
[ ] | Melanoma/SK | Deep multi-scale CNN | ISIC dataset | The proposed model used Inception-v3 model, which was trained on the ImageNet. | Accuracy (90.3), AUC (94.3) |
[ ] | Benign/malignant | CNN with 5-fold cross-validation | 1760 dermoscopic images | Images were preprocessed on the basis of melanoma cytological findings | Accuracy (84.7), sensitivity (80.9), specificity (88.1) |
[ ] | Benign/malignant | A very deep residual CNN and FCRN | ISIC 2016 database | FCRN incorporated with a multi-scale contextual information integration technique was proposed for accurate lesions segmentation | Accuracy (94.9), sensitivity (91.1), specificity (95.7), Jaccard index (82.9), dice coefficient (89.7) |
[ ] | AK/melanocytic nevus/BCC/SK/SCC | CNN | 1300 skin lesion images | Mean subtraction for each image, pooled multi-scale feature extraction process and pooling in augmented-feature space | Accuracy (81.8) |
[ ] | BCC/non-BCC | Pruned ResNet18 | 297 FF-OCT images | K-fold cross-validation was applied to measure the performance of the proposed system | Accuracy (80) |
[ ] | Melanoma/non melanoma | ResNet-50 with deep transfer learning | 3600 lesion images from the ISIC dataset | The proposed model showed better performance than o InceptionV3, Densenet169, Inception ResNetV2, and Mobilenet | Accuracy (93.5), precision (94) recall (77), F1_ score (85) |
[ ] | Benign/malignant | Region-based CNN with ResNet152 | 2742 dermoscopic images from ISIC dataset | Region of interest was extracted by mask and region-based CNN, then ResNet152 is used for classification. | Accuracy (90.4), sensitivity (82), specificity (92.5) |
CNN = Convolutional neural network; ISIC = International skin imaging collaboration; SVM = Support vector machine; BCC = Basal cell carcinoma; SCC = Squamous cell carcinoma; AK = Actinic keratosis; IC = Intraepithelial carcinoma; HAM10000 = Human-against-machine dataset with 10,000 images; BVLC = Berkeley Vision and Learning Center; SK= Seborrheic keratosis; FCRN = Fully convolutional residual network; FF-OCT = Full field optical coherence tomography; FCN = Fully convolutional network.
The Kohonen self-organizing map is a very famous type of deep neural network. CNNs are trained on the basis of unsupervised learning, which means that a KNN does not require any developer’s intervention in the learning process as well as requiring little information about the attributes of the input data. A KNN generally consists of two layers. In the 2-D plane, the first layer is called an input layer, while another is named a competitive layer. Both of these layers are fully connected, and every connection is from the first to second layer dimension. A KNN can be used for data clustering without knowing the relationships between input data members. It is also known as a self-organizing map. KNNs do not contain an output layer; every node in the competitive layer also acts as the output node itself.
A KNN basically works as a dimensionality reducer. It can reduce the high dimensional data into a low dimension, such as a two-dimensional plane. Thus, it provides discrete types of representation of the input dataset. KNNs are different from other types of NN in terms of learning strategy because it uses competitive learning rather than the learning based on error correction found in BPN or feed-forward learning. A KNN preserves the topological structure of the input data space during mapping dimensionality from high to low. Preservation refers to the preservation of relative distance between data points in space. Data points that are closer in input data space are mapped closer to each other in this scheme; far points are mapped far from each other as well as, according to the relative distance present among them. Consequently, a KNN is the best tool for high dimensional data. Another important feature provided by a KNN is its generalization ability. The network has the ability to recognize and organize unknown input data. The architecture of a KNN is shown in Figure 7 . A KKN’s main quality is its ability to map complex relationships of data points in which even nonlinear relations exist between data points. Due to these benefits, nowadays, KNNs are being used in skin cancer detection systems.
Basic KNN structure [ 58 ], BMU= Best matching unit.
Lenhardt et al. [ 59 ] proposed a KNN-based skin cancer detection system. The proposed system processed synchronous fluorescence spectra of melanoma, nevus, and normal skin samples for neural network training. A fluorescence spectrophotometer was used to measure the fluorescence spectra of the samples, whereas samples were collected from human patients immediately after surgical resection. The dimensionality of measured spectra was reduced with the PCA technique. Both KNN and ANN were trained, and their performance for melanoma detection was compared. On the test dataset, the classification error of KNN was 2–3%, while the classification error for ANN lay in the range of 3% to 4%.
A combination of self-organizing NN and radial basis function (RBF) neural network was proposed to diagnose three different types of skin cancer, such as BCC, melanoma, and SCC [ 60 ]. The proposed system extracted color, GLCM, and morphological features of lesion images, after which the classification model used those features as input. Furthermore, the classification performance of the proposed system was compared with k -nearest neighbor, ANN, and naïve-Bayes classifiers. The proposed system achieved 93.150685% accuracy while k -nearest neighbor showed 71.232877%, ANN showed 63.013699%, and naïve Bayes showed 56.164384% accuracy scores.
Another KNN-based automated skin cancer diagnostic system was proposed by Sajid et al. [ 61 ]. The proposed system employed a median filter as a noise removal technique. Then filtered images were segmented with a statistical region growing and merging technique. In this system, a collection of textual and statistical features was used. Statistical features were extracted from lesion images, whereas textual features were extracted from a curvelet domain. Finally, the proposed system classified the input images into cancerous or noncancerous with 98.3% accuracy. In this work, other classifiers such as SVM, BPN, and 3-layer NN were also implemented, and their performance was compared with the proposed system’s classification performance. SVM produced 91.1% accuracy, BPN showed 90.4% accuracy, 3-layer NN showed 90.5%, whereas the proposed system achieved the highest accuracy of 98.3% for skin cancer diagnosis. Details on the KNN-based skin cancer diagnostic systems is presented in Table 5 .
A comparative analysis of skin cancer detection using KNN-based approaches.
Ref | Skin Cancer Diagnoses | Classifier and Training Algorithm | Dataset | Description | Results (%) |
---|---|---|---|---|---|
[ ] | Melanoma/nevus/normal skin | SOM and feed-forward NN | 50 skin lesion images | PCA for decreasing spectra’s dimensionality | Accuracy (96–98) |
[ ] | BCC, SCC, and melanoma | SOM and RBF | DermQuest and Dermnet datasets | 15 features consisting of GCM morphological and color features were extracted | Accuracy (93.15) |
[ ] | Cancerous/noncancerous | Modified KNN | 500 lesion images | Automated Otsu method of thresholding for segmentation | Accuracy (98.3) |
SOM = Self organizing map; PCA = Principal component analysis; GCM = Generalized co-occurrence matrices; RBF = Radial Basis Function; KNN = Kohonen self-organizing neural network.
A generative adversarial neural network is a powerful class of DNN that is inspired by zero-sum game theory [ 62 ]. GANs are based on the idea that two neural networks, such as a generator and a discriminator, compete with each other to analyze and capture the variance in a database. The generator module uses the data distribution to produce fake data samples and tries to misguide the discriminator module. On the other hand, the discriminator module aims to distinguish between real and fake data samples [ 63 ]. In the training phase, both of these neural networks repeat these steps, and their performance improves after each competition. The ability to generate fake samples that are similar to a real sample using the same data distribution, such as photorealistic images, is the major power of a GAN network. It can also solve a major problem in deep learning: the insufficient training examples problem. Research scholars have been implementing various types of GANs, such as Vanilla GAN, condition GAN (CGAN), deep convolutional GAN (DCGAN), super-resolution GAN (SRGAN), and Laplacian Pyramid GAN (LPGAN). Nowadays, GANs are successfully being used in skin cancer diagnostic systems. The architecture of a GAN is shown in Figure 8 .
GAN architecture [ 64 ].
Rashid et al. [ 7 ] proposed a GAN-based skin lesion classification system. The proposed system performed augmentation on a training set of images with realistic-looking skin lesion images generated via GAN. A deconvolutional network was used as the generator module, while the discriminator module used CNN as a classifier. The CNN learned to classify seven different categories of skin lesions. Results of the proposed system were compared with ResNet-50 and DenseNet. ResNet-50 produced 79.2% accuracy, DenseNet showed 81.5% accuracy, whereas the proposed approach achieved the highest accuracy of 86.1% for skin lesion classification. Deep learning methods provide sufficient accuracy but require pure, unbalanced, and large training datasets. To overcome these limitations, Bisla et al. [ 8 ] proposed a deep learning approach for data purification and GAN for data augmentation. The proposed system used decoupled deep convolutional GANs for data generation. A pre-trained ResNet-50 model was further refined with a purified and augmented dataset and was used to classify dermoscopic images into three categories: melanoma, SK, and nevus. The proposed system outperformed the baseline ResNet-50 model for skin lesion classification and achieved 86.1% accuracy.
A novel data augmentation method for a skin lesion on the basis of self-attention progressive GAN (PGAN) was proposed. Moreover, the generative model was enhanced with the stabilization technique. The proposed system achieved 70.1% accuracy as compared with 67.3% accuracy produced by a non-augmented system. A list of GAN-based skin cancer detection systems with their diagnosed skin cancer type, classifier, dataset, and the obtained result is presented in Table 6 .
A comparative analysis of skin cancer detection using GAN-based approaches.
Ref | Skin Cancer Diagnoses | Classifier and Training Algorithm | Dataset | Description | Results (%) |
---|---|---|---|---|---|
[ ] | AK/BCC/benign keratosis/dermatofibroma/melanoma/melanocytic nevus/vascular lesion | GAN | ISIC 2018 | The proposed system used deconvolutional network and CNN as generator and discriminator module | Accuracy (86.1) |
[ ] | Melanoma/nevus/SK | Deep convolutional GAN | ISIC 2017, ISIC 2018, PH | Decoupled deep convolutional GANs for data augmentation | ROC AUC (91.5), accuracy (86.1) |
[ ] | BCC/vascular/pigmented benign keratosis/pigmented Bowen’s/nevus/dermatofibroma | Self-attention-based PGAN | ISIC 2018 | A generative model was enhanced with a stabilization technique | Accuracy (70.1) |
GAN = Generative adversarial neural network, PGAN = Progressive generative adversarial network, ROC AUC= Area under the receiver operating characteristic curve.
Several computer-based systems for skin cancer diagnosis have been proposed. Evaluating their diagnostic performance and validating predicted results requires a solid and reliable collection of dermoscopic images. Various skin cancer datasets have lacked size and diversity other than for images of nevi or melanoma lesions. Training of artificial neural networks for skin lesion classification is hampered by the small size of the datasets and a lack of diverse data. Although patients commonly suffer from a variety of non-melanocytic lesions, past research for automated skin cancer diagnosis primarily focused on diagnosing melanocytic lesions, resulting in a limited number of diagnoses in the available datasets [ 66 ]. Therefore, the availability of a standard, reliable dataset of dermoscopic images is very crucial. Real-world datasets for the evaluation of proposed skin cancer detection techniques are discussed in this section. Table 7 summarizes the important details of these datasets.
Skin Cancer Datasets.
Sr. No | Name of Dataset | Year of Release | No. of Images | Reference Used |
---|---|---|---|---|
1 | HAM10000 | 2018 | 10,015 | [ ] |
2 | PH | 2013 | 200 | [ ] |
3 | ISIC archive | 2016 | 25,331 | [ , , , , , , , , , , ] |
4 | DermQuest | 1999 | 22,082 | [ , , ] |
5 | DermIS | 6588 | [ , ] | |
6 | AtlasDerm | 2000 | 1024 | [ ] |
7 | Dermnet | 1998 | 23,000 | [ , ] |
There is a human-against-machine dataset with 10,000 training images that is referred to as HAM10000 [ 66 ]. It is the latest publicly available skin lesions dataset, and it overcomes the problem of the lack of diversity. The final dataset of HAM10000 contains 10,015 dermoscopic images, collected from two sources: Cliff Rosendahl’s skin cancer practice in Queensland, Australia, and the Dermatology Department of the Medical University of Vienna, Austria. This collection has taken twenty years to compile. Before widespread use of digital cameras, photographic prints of lesions were deposited and stored at the Dermatology Department of the Medical University of Vienna, Austria. These photographic prints were digitalized with the help of Nikon-Coolscan-5000-ED scanner, manufactured by Nikon corporation Japan and converted into 8-bit color JPEG images having 300 DPI quality. The images were then manually cropped and saved at 800 × 600 pixels resolution at 72 DPI.
Several acquisition functions and cleaning methods were applied to the images and a semi-automatic workflow was developed using a neural network to attain diversity. The resulting dataset contains 327 images of AK, 514 images of basal cell carcinomas, 1099 images of benign keratoses, 115 images of dermatofibromas, 1113 images of melanocytic nevi, 6705 images of melanomas, and 142 images of vascular skin lesions.
The dermoscopic images in the PH² dataset were collected at the Dermatology Center of Pedro Hispano Hospital, Portugal [ 68 ]. These images were obtained using a Tuebinger-Mole-Analyzer system under the same conditions and magnification rate of 20×. PH2 dataset contains 8-bit RGB color images having 768 × 560 pixels resolution. The dataset contains 200 dermoscopic images, divided into 80 images of common nevi, 80 images of atypical nevi, and 40 images of melanoma skin cancers. This dataset contains medical annotation of the lesion images, such as medical segmentation of pigmented skin lesions, histological and clinical diagnosis, and evaluation of various dermoscopic criteria. The assessment was performed according to dermoscopic criteria of streaks, colors, regression areas, pigment network, and blue-whitish veil globules.
The ISIC archive [ 69 ] is a collection of various skin lesions datasets. The ISIC dataset [ 70 ] was originally released by the International Skin Imaging Collaboration at the International Symposium on Biomedical Imaging (ISBI) challenge 2016, named as ISIC2016. The ISIC2016 archive is divided into two parts: training and testing. The training subset of ISIC contains 900 images, while the testing subset contains 379 dermoscopic images. It includes images of two classes: malignant melanomas and benign nevi. Approximately 30.3% of the dataset’s images are of melanoma lesions and the remaining images belong to the benign nevi class. ISIC increases the number of images in its archive every year and has established a design challenge for the development of a system for skin cancer automated diagnosis.
In the ISIC2017 dataset, there were three categories of images: melanomas, seborrheic-keratoses (SK), and benign nevi. The dataset contains 2000 training images, 150 validation images, and 600 images for testing. The training dataset contains 374 images of melanomas, 254 SK images, and 1372 images of benign nevi. The validation dataset contains 30 melanoma images, 42 SK images, and 78 benign nevus images. The test dataset includes 117 melanoma images, 90 SK images, and 393 benign nevus images. ISIC2018 contains 12,594 training images, 100 validation images, and 1000 test images. The ISIC2019 dataset includes 25,331 images of eight different categories of skin lesions, such as melanoma, melanocytic-nevus, BCC, AK, benign keratosis, dermatofibroma, vascular lesion, and SCC. It contains 8239 images in the test dataset and an additional outlier class that was not included in the training dataset. The new proposed skin cancer diagnostic systems must be able to identify these images. The ISIC2019 dataset also includes metadata for images, such as sex, age, and area of the patient.
The publicly available DermQuest dataset [ 71 ] contained 22,082 dermoscopic images. Among all dermoscopic datasets, only the DermQuest dataset contained lesion tags for skin lesions. There were 134 lesion tags for all images in the dataset. The DermQuest dataset redirected to Derm101 in 2018. However, this dataset was deactivated recently on 31 December 2019.
The Dermoscopic dataset Dermatology Information System is commonly known as DermIS [ 72 ]. This dataset was built through cooperation between the Department of Dermatology of the University of Erlangen and the Department of Clinical Social Medicine of the University of Heidelberg. It contains 6588 images. This dataset has recently been divided into two parts: a dermatology online image atlas (DOIA) and a pediatric dermatology online image atlas (PeDOIA). The DOIA includes 3000 lesion images covering approximately 600 dermatological diagnoses. It provides dermoscopic images complete with differential and provisional diagnoses, case reports, and other information on nearly all types of skin diseases.
The Atlas of Dermoscopy dataset is commonly referred to as AtlasDerm [ 73 ]. It is a unique and well-organized combination of a book and images on CD-ROM with sample examples for training. It was originally designed as a tool to help physicians in the diagnosis of skin lesions and the recognition of dermoscopic criteria related to melanoma. The AtlasDerm dataset considers various cases of skin lesions, with corresponding dermoscopic images for every case. It contains 5 images of AK, 42 images of BCC, 70 images of benign keratosis, 20 images of dermatofibroma, 275 images of melanocytic nevus, 582 images of melanoma, and 30 images of vascular skin lesions.
The Dermnet Skin Disease Atlas dataset is commonly referred to as Dermnet [ 74 ]. It was built in 1998 by Dr. Thomas Habif in Portsmouth, New Hampshire. It consists of more than 23,000 dermoscopic images. This database contains images of 643 different types of skin diseases. These diseases are biologically organized into a two-level taxonomy. The bottom level contains more than 600 skin diseases in fine granularity. The top-level taxonomy contains 23 different classes of skin diseases, such as connective tissue disease, benign tumors, eczema, melanomas, moles, nevi, etc.
5.1. extensive training.
One of the major challenges in neural network-based skin cancer detection techniques is the extensive training that is required. In other words, to successfully analyze and interpret the features from dermoscopic images, the system must undergo detailed training, which is a time-consuming process and demands extremely powerful hardware.
Another challenge is the variation in the sizes of lesions. A group of Italian and Austrian researchers collected many benign and cancerous melanoma lesion images in the 1990s [ 73 ]. The diagnostic accuracy of the identification of the lesions was as high as 95% to 96% [ 75 ]. However, the diagnostic process, with earlier stage and smaller lesions of 1mm or 2mm in size, was much more difficult and error-prone.
Existing standard dermoscopic datasets contain images of light-skinned people, mostly from Europe, Australia, and the United States. For accurate skin cancer detection in dark-skinned people, a neural network must learn to account for skin color [ 76 ]. However, doing so is possible only if the neural network observes enough images of dark-skinned people during the process of training. Therefore, datasets having sufficient lesion images of dark-skinned and light-skinned people is necessary for increasing the accuracy of skin cancer detection systems.
Unlike the other types of images, medical images have very small interclass variation; that is, the difference between melanoma and nonmelanoma skin cancer lesion images has much less variation than, say, the variation between images of cats and dogs. It is also very difficult to differentiate between a birthmark and a melanoma. The lesions of some disease are so similar that it is extremely hard to distinguish them. This limited variation makes the task of image analysis and classification very complex [ 32 ].
Real-world datasets used for skin cancer diagnosis are highly unbalanced. Unbalanced datasets contain a very different number of images for each type of skin cancer. For example, they contain hundreds of images of common skin cancer types but only a few images for the uncommon types, making it difficult to draw generalizations from the visual features of the dermoscopic images [ 12 ].
Powerful hardware resources with high graphical processing unit (GPU) power are required for the NN software to be able to extract the unique features of a lesion’s image, which is critical for achieving better skin cancer detection. The lack of availability of high computing power is a major challenge in deep learning-based skin cancer detection training.
Various types of skin cancers such as Merkel cell cancer, BCC, and SCC typically appear after the age of 65 years [ 77 ]. Existing standard dermoscopic datasets contain images of young people. However, for an accurate diagnosis of skin cancer in elderly patients, it is necessary that neural networks observe enough images of people aged more than 50 years.
Preprocessing and detection of lesion edges are very crucial steps in the automated detection of skin cancer. Various optimization algorithms such as artificial the bee colony algorithm [ 78 ], ant colony optimization [ 79 ], social spider optimization [ 80 ], and particle swarm optimization [ 81 ] can be explored to increase the performance of automated skin cancer diagnostic systems.
Researchers have identified various genetic risk factors for melanoma, such as fair skin, light colored eyes, red hair, a large number of moles on the body, and a family history of skin cancer. When these genetic risk factors are combined with environmental risks such as high ultraviolet light exposure, the chances of developing skin cancer become very high [ 82 ]. These factors can be combined with existing deep learning approaches for better performance.
This systematic review paper has discussed various neural network techniques for skin cancer detection and classification. All of these techniques are noninvasive. Skin cancer detection requires multiple stages, such as preprocessing and image segmentation, followed by feature extraction and classification. This review focused on ANNs, CNNs, KNNs, and RBFNs for classification of lesion images. Each algorithm has its advantages and disadvantages. Proper selection of the classification technique is the core point for best results. However, CNN gives better results than other types of a neural networks when classifying image data because it is more closely related to computer vision than others.
Most of the research related to skin cancer detection focuses on whether a given lesion image is cancerous. However, when a patient asks if a particular skin cancer symptom appears on any part of their body, the current research cannot provide an answer. Thus far, the research has focused on the narrow problem of classification of the signal image. Future research can include full-body photography to seek the answer to the question that typically arises. Autonomous full-body photography will automate and speed up the image acquisition phase.
The idea of auto-organization has recently emerged within the area of deep learning. Auto-organization refers to the process of unsupervised learning, which aims to identify features and to discover relations or patterns in the image samples of the dataset. Under the umbrella of convolutional neural networks, auto-organization techniques increase the level of features representation that is retrieved by expert systems [ 47 ]. Currently, auto-organization is a model that is still in research and development. However, its study can improve the accuracy of image processing systems in the future, particularly in the area of medical imaging, where the smallest details of features are extremely crucial for the correct diagnosis of disease.
The authors acknowledge support from the Deanship of Scientific Research, Najran University, Kingdom of Saudi Arabia.
M.D. and S.A., M.I. and H.U.K. developed the idea and collected the data. M.R. and A.R.M. analyzed the data and wrote the manuscript. S.A.A. prepared the figures. A.H.M.S., M.O.A. and M.H.M. reviewed the data and manuscript as well. M.O.A. and M.H.M. were involved in the analysis of the images, datasets, labelling, and data. All authors have read and agreed to the published version of the manuscript.
This research study has not received any research funding.
Not applicable.
Data availability statement, conflicts of interest.
Authors have no conflicts of interest.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
To ensure we keep this website safe, please can you confirm you are a human by ticking the box below.
If you are unable to complete the above request please contact us using the below link, providing a screenshot of your experience.
https://ioppublishing.org/contacts/
A not-for-profit organization, IEEE is the world's largest technical professional organization dedicated to advancing technology for the benefit of humanity. © Copyright 2024 IEEE - All rights reserved. Use of this web site signifies your agreement to the terms and conditions.
Advertisement
Skin cancer poses a significant health hazard, necessitating the utilization of advanced diagnostic methodologies to facilitate timely detection, owing to its escalating prevalence in recent years. This paper proposes a novel approach to tackle the issue by introducing a method for detecting skin cancer that uses MixNets to enhance diagnosis and leverages mobile network-based transfer learning. Skin cancer has diverse forms, each distinguishable by its structural attributes, morphological characteristics, texture, and coloration. The pressing demand for accurate and efficient diagnostic instruments has spurred the investigation of novel techniques. The present study utilizes the ISIC dataset, comprising a validation set of 660 images and a training set of 2637 images. Moreover, the research employs a combination of MixNets and mobile network-based transfer learning as its chosen approach. Transfer learning is a technique that leverages preexisting models to enhance the diagnostic capabilities of the proposed system. Integrating MobileNet and MixNets allows for utilizing their respective functionalities, resulting in a dual-model methodology that enhances the comprehensiveness of skin cancer diagnosis. The results demonstrate impressive performance metrics, with MobileNet and MixNets models, and the proposed approach achieves an outstanding accuracy rate of 99.58%. The above findings underscore the efficacy of the dual-model method in effectively discerning between benign and malignant skin lesions. Moreover, the present study aims to examine the potential integration of emerging technologies to enhance the accuracy and practicality of diagnostics within real-world healthcare settings.
This is a preview of subscription content, log in via an institution to check access.
Subscribe and save.
Price includes VAT (Russian Federation)
Instant access to the full article PDF.
Rent this article via DeepDyve
Institutional subscriptions
Explore related subjects.
Artificial Intelligence
The dataset is available on reasonable request.
International Skin Imaging Collaboration Laboratory database
Machine Learning
Magnetic Resonance Imaging
Visual Geometry Group
Human Against Machine
Residual Network
Principal Component Analysis
Receiver Operating Characteristic
True Negative/False Negative
Ultraviolet
Support Vector Machine
Convolutional Neural Network
Internet of Things
Geometric Active Contour
Exploratory Data Analysis
Rectified Linear Unit
False Positive/True Positive
Area Under the ROC Curve
Anand V, Gupta S, Altameem A, Nayak SR, Poonia RC, Saudagar AKJ (2022) An enhanced transfer learning based classification for diagnosis of skin cancer. Diagnostics 12(7):1628. https://doi.org/10.3390/diagnostics12071628
Article Google Scholar
Al-Rasheed A, Ksibi A, Ayadi M, Alzahrani AIA, Zakariah M, Ali Hakami N (2022) “An ensemble of transfer learning models for the prediction of skin cancers with conditional generative adversarial networks.” Diagnostics 12(12):3145. https://doi.org/10.3390/diagnostics12123145
Khandizod S, Patil T, Dode A, Banale V, Prof CD, Bawankar (2022) “Deep Learning based skin cancer classifier using MobileNet.” Int J Res Appl Sci Eng Technol 10(5):629–633. https://doi.org/10.22214/ijraset.2022.42260
Ghazal TM, Hussain S, Khan MF, Khan MA, Said RAT, Ahmad M (2022) Detection of benign and malignant tumours in skin empowered with transfer learning. Comput Intell Neurosci 2022:1–9. https://doi.org/10.1155/2022/4826892
K. V. Reddy and L. R. Parvathy, “An Innovative Analysis of predicting Melanoma Skin Cancer using MobileNet and Convolutional Neural Network Algorithm,” in 2022 2nd International Conference on Technological Advancements in Computational Sciences (ICTACS), IEEE, Oct. 2022, pp. 91–95. https://doi.org/10.1109/ICTACS56270.2022.9988569 .
M. Castro-Fernandez, A. Hernandez, H. Fabelo, F. J. Balea-Fernandez, S. Ortega, and G. M. Callico, “Towards Skin Cancer Self-Monitoring through an Optimized MobileNet with Coordinate Attention,” in 2022 25th Euromicro Conference on Digital System Design (DSD), IEEE, Aug. 2022, pp. 607–614. https://doi.org/10.1109/DSD57027.2022.00087 .
Fraiwan M, Faouri E (2022) “On the automatic detection and classification of skin cancer using deep transfer learning.” Sensors 22(13):4963. https://doi.org/10.3390/s22134963
Bassel A, Abdulkareem AB, Alyasseri ZAA, Sani NS, Mohammed HJ (2022) “Automatic malignant and benign skin cancer classification using a hybrid deep learning approach.” Diagnostics 12(10):2472. https://doi.org/10.3390/diagnostics12102472
Alam TM et al (2022) “An efficient deep learning-based skin cancer classifier for an imbalanced dataset.” Diagnostics 12(9):2115. https://doi.org/10.3390/diagnostics12092115
Lu X, Firoozeh Abolhasani Zadeh YA (2022) “Deep learning-based classification for melanoma detection using xceptionnet.” J Healthc Eng. 2022:1–10
Google Scholar
Kousis I, Perikos I, Hatzilygeroudis I, Virvou M (2022) “Deep learning methods for accurate skin cancer recognition and mobile application.” Electronics (Basel) 11(9):1294. https://doi.org/10.3390/electronics11091294
Viknesh CK, Kumar PN, Seetharaman R, Anitha D (2023) “Detection and classification of melanoma skin cancer using image processing technique.” Diagnostics 13(21):3313. https://doi.org/10.3390/diagnostics13213313
Mazhar T et al (2023) “The role of machine learning and deep learning approaches for the detection of skin cancer.” Healthcare 11(3):415. https://doi.org/10.3390/healthcare11030415
Yaqoob MM, Alsulami M, Khan MA, Alsadie D, Saudagar AKJ, AlKhathami M (2023) “Federated machine learning for skin lesion diagnosis: an asynchronous and weighted approach.” Diagnostics 13(11):1964. https://doi.org/10.3390/diagnostics13111964
Bakheet S, Alsubai S, El-Nagar A, Alqahtani A (2023) “A multi-feature fusion framework for automatic skin cancer diagnostics.” Diagnostics 13(8):1474. https://doi.org/10.3390/diagnostics13081474
Z. E. Diame, M. N. Al-Berry, M. A.-M. Salem, and M. Roushdy, “Deep Learning Architectures For Aided Melanoma Skin Disease Recognition: A Review,” in 2021 International Mobile, Intelligent, and Ubiquitous Computing Conference (MIUCC), IEEE, May 2021, pp. 324–329. https://doi.org/10.1109/MIUCC52538.2021.9447615 .
Imran A, Nasir A, Bilal M, Sun G, Alzahrani A, Almuhaimeed A (2022) “Skin cancer detection using combined decision of deep learners.” IEEE Access 10:118198–118212. https://doi.org/10.1109/ACCESS.2022.3220329
Melarkode N, Srinivasan K, Qaisar SM, Plawiak P (2023) “AI-powered diagnosis of skin cancer: a contemporary review, open challenges and future research directions.” Cancers (Basel) 15(4):1183. https://doi.org/10.3390/cancers15041183
N. Kumar and T. Sandhan, “Alternating Sequential and Residual Networks for Skin Cancer Detection from Biomedical Images,” in 2023 National Conference on Communications (NCC), IEEE, Feb. 2023, pp. 1–5. https://doi.org/10.1109/NCC56989.2023.10068074 .
Salih O, Duffy KJ (2023) “Optimization convolutional neural network for automatic skin lesion diagnosis using a genetic algorithm.” Appl Sci 13(5):3248. https://doi.org/10.3390/app13053248
Cai H, Brinti Hussin N, Lan H, Li H (2023) “A skin cancer detector based on transfer learning and feature fusion.” Curr Bioinform 18(6):517–526. https://doi.org/10.2174/1574893618666230403115540
Rashid J et al (2022) “Skin cancer disease detection using transfer learning technique.” Appl Sci 12(11):5714. https://doi.org/10.3390/app12115714
Hussien MA, Alasadi AHH (2023) “A review of skin cancer detection: traditional and deep learning-based techniques.” J Univ Babylon Pure Appl Sci 31(2):253–262. https://doi.org/10.29196/jubpas.v31i2.4682
Ali MS, Miah MS, Haque J, Rahman MM, Islam MK (2021) “An enhanced technique of skin cancer classification using deep convolutional neural network with transfer learning models.” Mach Learn Appl 5:100036. https://doi.org/10.1016/j.mlwa.2021.100036
Toğaçar M, Cömert Z, Ergen B (2021) “Intelligent skin cancer detection applying autoencoder, MobileNetV2 and spiking neural networks.” Chaos Solitons Fractals 144:110714. https://doi.org/10.1016/j.chaos.2021.110714
Article MathSciNet Google Scholar
Shinde RK et al (2022) “Squeeze-mnet: precise skin cancer detection model for low computing IoT devices using transfer learning.” Cancers (Basel) 15(1):12. https://doi.org/10.3390/cancers15010012
Jain S, Singhania U, Tripathy B, Nasr EA, Aboudaif MK, Kamrani AK (2021) “Deep learning-based transfer learning for classification of skin cancer.” Sensors 21(23):8142. https://doi.org/10.3390/s21238142
Khan IU et al (2021) “Remote diagnosis and triaging model for skin cancer using efficientnet and extreme gradient boosting.” Complexity 2021:1–13. https://doi.org/10.1155/2021/5591614
Shehzad K et al (2023) “A deep-ensemble-learning-based approach for skin cancer diagnosis.” Electronics (Basel) 12(6):1342. https://doi.org/10.3390/electronics12061342
Srinivasu PN, SivaSai JG, Ijaz MF, Bhoi AK, Kim W, Kang JJ (2021) “Classification of skin disease using deep learning neural networks with MobileNet V2 and LSTM.” Sensors 21(8):2852. https://doi.org/10.3390/s21082852
Dildar M et al (2021) “Skin cancer detection: a review using deep learning techniques.” Int J Environ Res Public Health 18(10):5479. https://doi.org/10.3390/ijerph18105479
Mridha K, Uddin MdM, Shin J, Khadka S, Mridha MF (2023) “An interpretable skin cancer classification using optimized convolutional neural network for a smart healthcare system.” IEEE Access 11:41003–41018. https://doi.org/10.1109/ACCESS.2023.3269694
A. H. Jui, S. Sharnami, and A. Islam, “A CNN Based Approach to Classify Skin Cancers using Transfer Learning,” in 2022 25th International Conference on Computer and Information Technology (ICCIT), IEEE, Dec. 2022, pp. 1063–1068. https://doi.org/10.1109/ICCIT57492.2022.10055838 .
T. Guergueb and M. A. Akhloufi, “Skin Cancer Detection using Ensemble Learning and Grouping of Deep Models,” in International Conference on Content-based Multimedia Indexing, New York, NY, USA: ACM, Sep. 2022, pp. 121–125. https://doi.org/10.1145/3549555.3549584 .
Gong X, Xiao Y (2021) “A skin cancer detection interactive application based on CNN and NLP.” J Phys Conf Ser 2078(1):012036. https://doi.org/10.1088/1742-6596/2078/1/012036
Olayah F, Senan EM, Ahmed IA, Awaji B (2023) “AI techniques of dermoscopy image analysis for the early detection of skin lesions based on combined cnn features.” Diagnostics 13(7):1314. https://doi.org/10.3390/diagnostics13071314
Download references
The authors would like to acknowledge the Researchers Supporting Project number (RSP2024R206), King Saud University, Riyadh, Saudi Arabia.
Funding for this study was received from the Researchers Supporting Project number (RSP2024R206), King Saud University, Riyadh, Saudi Arabia.
Authors and affiliations.
Department of Computer Sciences and Engineering, College of Applied Science, King Saud University, 11543, Riyadh, Saudi Arabia
Mohammed Zakariah
Department of Software Engineering, College of Computer and Information Sciences, King Saud University, 11345, Riyadh, Saudi Arabia
Muna Al-Razgan
Department of Information Systems, College of Computer and Information Sciences, King Saud University, 11543, Riyadh, Saudi Arabia
Taha Alfakih
You can also search for this author in PubMed Google Scholar
Correspondence to Mohammed Zakariah .
Conflict of interest.
The authors declare there is no conflict of interest.
Not applicable.
Publisher's note.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
Reprints and permissions
Zakariah, M., Al-Razgan, M. & Alfakih, T. Skin cancer detection with MobileNet-based transfer learning and MixNets for enhanced diagnosis. Neural Comput & Applic (2024). https://doi.org/10.1007/s00521-024-10227-w
Download citation
Received : 22 January 2024
Accepted : 12 July 2024
Published : 28 August 2024
DOI : https://doi.org/10.1007/s00521-024-10227-w
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.
Provided by the Springer Nature SharedIt content-sharing initiative
IMAGES
VIDEO
COMMENTS
In order to analyze current research findings that have been suggested for skin disease detection and classification using conventional machine learning methods, deep learning methods, and hybrid methodologies, this systematic literature review work set three main objectives: (1) to identify the commonly available datasets that could be ...
A machine learning approach for skin disease detection ...
Skin lesions affect millions of people worldwide. They can be easily recognized based on their typically abnormal texture and color but are difficult to diagnose due to similar symptoms among certain types of lesions. The motivation for this study is to collate and analyze machine learning (ML) applications in skin lesion research, with the goal of encouraging the development of automated ...
This paper presents a comparative analysis of 5 different. machine learning algorithms random forest, naive Bayes, logistic regression, kernel SVM and CNN. All these. algorithms are implemented on ...
Skin is the human body's exterior integument. Human skin pigmentation varies from person to person, and skin types include dry, oily, and mixed. The human skin's diversity offers bacteria and other microbes with a diverse home. Melanocytes in the human skin create melanin, which can absorb harmful UV radiation from the sun, causing skin damage and cancer. In most third-world societies, the ...
Skin disease detection using deep learning
3 Proposed Work. The proposed technique is a valuable tool for studying people's skin diseases and predicting skin disease. A hybrid architecture of image processing and machine learning techniques is used in this proposed framework to predict disease types with promising accuracy in a short period of time.
Objective: Skin diseases constitute a widespread health concern, and the application of machine learning and deep learning algorithms has been instrumental in improving diagnostic accuracy and treatment effectiveness. This paper aims to provide a comprehensive review of the existing research on the utilization of machine learning and deep learning in the field of skin disease diagnosis, with a ...
In skin cancer research, image processing, machine learning, CNN, and DL have all ... Koc K.O. Detection of skin diseases from dermoscopy image using the combination of convolutional neural network and one-versus-all. ... Ijaz M.F., Bhoi A.K., Kim W., Kang J.J. Classification of skin disease using deep learning neural networks with MobileNet V2 ...
An automatic detection method for skin disorders identifies skin diseases at high efficiency inside a short period of time. If skin disorders are found early, life is spared from recurring skin diseases such as skin cancer. These chapter contains the design criteria of a protection system for hybridized skin diseases.
Kotian AL, Deepa K (2017) Detection and classification of skin diseases by image analysis using MATLAB. Int J Emerg Res Manag Technol 6(5):779-784. Google Scholar Sundaramurthy S, Saravanabhavan C, Kshirsagar P (2020) Prediction and Classification of rheumatoid arthritis using ensemble machine learning approaches.
Skin is the largest organ of the body which provides protection, regulates the body fluids and temperature, and enables sense of the external environment. 1 Skin diseases are the most common cause of all human illnesses which affects almost 900 million people in the world at any time. 2 According to the global burden of disease project, skin ...
A Method Of Skin Disease Detection Using Image ...
Classification of Skin Disease Using Deep Learning Neural ...
Among the most avoidable diseases on the planet is epidermal concerns. At any rate being normal, its research is quite difficult due to its intricate colors, hair is present, hiding. Early diagnosis of skin diseases is critical to successful treatment. The skills and experience of such expert specialist are used to determine the procedure for identifying and treating skin damage. The ...
The main objective of this research is to develop an application based on Deep learning, Computer vision and cloud computing that detects the different kinds of skin diseases caused by different types of viruses, Bacteria, Fungus and Environment. This study has also developed and integrated a recommendation system, which recommends the medicines and care taking process for a particular disease ...
The paper proposed a skin disease detection tool based on image processing, machine learning and deep learning techniques. The proposed tools are non-invasive, easy-touse and accurate to identify ...
Skin Cancer Detection: A Review Using Deep Learning ...
Skin Disease Classification System Based on Machine Learning Technique: A Survey. Saja Salim Mohammed1 and Jamal Mustafa Al-Tuwaijari1. IOP Conference Series: Materials Science and Engineering, Volume 1076, 2nd International Scientific Conference of Engineering Sciences (ISCES 2020) 16th-17th December 2020, Diyala, Iraq Citation Saja Salim ...
The world population is growing very fast and the lifestyle of human beings is changing with time and place. So, there is a need for disease management which includes disease diagnosis, its detection and classification, cure and lastly for future disease prevention. The outermost protective layer of a human body is the skin. Skin not only impacts a person's health but also psychologically ...
These datasets play crucial roles in advancing research and development across various fields, from dermatology to machine learning and pharmaceuticals. In summary, the availability of public datasets specifically curated for melanoma identification has played a pivotal role in facilitating the development and evaluation of DL models aimed at ...
The application of deep learning methods to diagnose diseases has become a new research topic in the medical field. In the field of medicine, skin disease is one of the most common diseases, and its visual representation is more prominent compared with the other types of diseases. Accordingly, the use of deep learning methods for skin disease image recognition is of great significance and has ...
Skin diseases are the 4th common cause of skin burden worldwide. Robust and Automated system have been developed to lessen this burden and to help the patients to conduct the early assessment of the skin lesion. Mostly this system available in the literature only provide skin cancer classification. Treatments for skin are more effective and ...
Skin is an extraordinary human structure. As a result of inherited traits and environmental variables, skin conditions are the most prevalent worldwide. People frequently neglect the effects of skin diseases in their initial stages. It commonly experienced both well-known and rare diseases. Identifying skin diseases and their kinds in the medical field is a very difficult process. It can be ...
Skin cancer poses a significant health hazard, necessitating the utilization of advanced diagnostic methodologies to facilitate timely detection, owing to its escalating prevalence in recent years. This paper proposes a novel approach to tackle the issue by introducing a method for detecting skin cancer that uses MixNets to enhance diagnosis and leverages mobile network-based transfer learning ...