1. Introduction
Skin acts as the most important and massive part of the human body, covering about 20 square feet. The skin plays the role of regulating temperature, allowing the sense of touch, feeling hot and cold, and protecting the inner body from ultraviolet rays [1]. Skin accounts for 15% weight of the whole body, with a surface area of about [2]. Skin consists of 3 main layers. skin [3]. In skin cancer the rare growing of the skin cell has become uncontrolled [4].In daily routine, some skin cells die, and new cells come on their place [5]. Skin cancer has become common these days. According to the report of cancer statistics estimation in the US in 2021, the new skin cancer (melanoma) estimated score has reached 34,920, where 19,320 are male and 15,600 are female. The death rate estimation is 12,410, with 5570 females and 6840 males [6]. The support of computer-aided diagnosis can motivate dermatologists to develop real-time skin cancer identification algorithms. One of the most essential steps for the analysis of the problem is to extract and select the most prominent and promising features. After that, the designed algorithm must be able to provide better measures than the previous one. The limitation in existing approaches acts as a motivation for the presented work, as semantic segmentation is required to extract the exact boundaries of the lesion and deep features, and their selection is required for more accurate classification. The main contributions of the presented work are:
▪. Skin lesions are segmented using the proposed segmentation model, in which features are drawn out through a pre-trained Mobilenetv2 model, which acts as a base of DeepLabv3+ for boundary extraction. The model is attained on the chosen hyperparameters that provide more accurate segmentation results.
▪. A classification framework is designed in which features are taken through a pre-trained DenseNet-201 model and optimal features are picked using SMA. These optimal features are passed to the machine learning classifier along with labels to perform classification.
2. Related Work
Timely and accurate skin lesion recognition and classification [7,8] is a very important task [9,10,11,12]. In the skin lesion analysis segmentation [13,14], it is the most important and second step, coming after pre-processing [15,16]. It divides the image into parts, with these parts being called segments [17]. A hybrid model is proposed which is the combination of k-means with a level set [18,19]. A method is defined in which, segments the input image using k-mean clustering [20,21]. An initial contour edges Chan–Vese model is applied with a genetic algorithm for the recognition of skin lesion boundaries [22,23]. The researcher proposed new pyramid pooling for lesion segmentation [24,25]. A system with Mask-R-CNN is proposed [26,27]. A dense framework is utilised for improvement [28,29]. Segmentation is performed using an adaptive dual attention module [30,31]. An algorithm using Bezier curves used for global optimization [32,33]. The segmentation is performed to accurately discover the lesion using deep learning-based methods, i.e., DeepLab V3+ and Mask R-CNN [34,35,36]. The encoder is joined with DeepLabV3 and decoder [37,38] for lesion segmentation. Deconvolutional coating are utilised to change the volume of input and output [39,40]. Hierarchical supervision is used to refine the prediction mask [41,42]. To segment, the image fuzzy clustering is utilised [43]. Researchers utilised colour features to partition the image [44,45,46]. The CNN classification with the novel regularising method proposed provided an accuracy of 0.974 [47]. The ensembles for melanoma classification, are utilised [48,49]. The ARL-CNN classification model is used for effectiveness [50].
3. Proposed Methodology
We developed novel segmentation and classification models. In the proposed segmentation model, a pre-trained Mobilenetv2 model and DeepLabv3+ are utilised. In the proposed classification framework, features are investigated using DesnseNet-201 and novel features are extracted with SMA for multi-classification of skin lesions as presented in Figure 1.
3.1. Segmentation of Skin Lesion
In the proposed segmentation model, features are extracted using Mobilenetv2 [51]. Features obtained from pre-trained Mobilenetv2 are input to the DeepLabv3+ network. DeepLabv3+ [52] is an enhanced version of atrous spatial pyramid pooling, with the addition of image-level features and batch normalization. Atrous convolutional in the last few blocks of the backbone to control the feature map size. The atrous spatial pyramid pooling is added on the peak of taken features that classify every pixel corresponding to their classes. The proposed framework is joined with Mobilenetv2 and Deeplabv3+, which contains 186 layers, in which 01 input, 67 convolutions, 59 batch-norm, 40 flip ReLU, 13 addition, 02 2D-crops, 02 depth concatenation, 01 softmax, and 01 classification layers are included, as illustrated in Figure 2.
The parameters of segmentation are mentioned in Table 1.
In Table 1, the parameters are concluded after long experimentation, in which 32 batch-size, 100 epochs, 0.0001 rate of learning with sgdm optimizer solver provide good segmentation results.
3.2. Classification of Skin Lesions
The proposed classification model consists of three phases, including features extraction using DenseNet-201, optimal features selection using a slime mould algorithm, and pictorial classification, as in Figure 3.
3.2.1. Features Extraction and Selection
Pre-trained DenseNe-t201 [53,54] model is used to obtain the feature, taken from a fully-connected FC-1000 layer measuring , and is input into the slime mould algorithm (SMA) [55]. SMA is an optimization technique used for the best feature selection. SMA [56] is naturally established within slime mould oscillation. Thus, SMA is influenced by the actions of morphological alterations and slime mould. The individual swarms are categorized into three groups. Some of them are picked at the origin, through a proportional number, to be resurrected and carry out their exploration. Some of them pursue their investigation built on their current position and the remaining would be direct towards the foremost candidate. The selected SMA parameters are described in Table 2.
Table 2 depicts the selected parameters of SMA which are utilised for the selection of optimum features, in which the total number of 5 neighbours, 0.2 hold-out validation ratio, the total number of 100 solutions, and maximum 100 iterations are included. The convergence curve in terms of fitness is obtained using SMA, as revealed in Figure 4.
The above graph shows the outcomes of the best feature selection on the PH2 dataset using the SMA algorithm. The SMA’s mathematical model is discussed below:
(1)
where the in Equation (3), linearly decreasing towards one to zero. The iu presents the current iteration. The stands for the current highest accuracy position and it describes the current position, defines slime Mould location, and present randomly selected two individuals from swarms, defines the weight of slime mould’s, and h is shown in Equation (2):h = tanh |V (j) − eF|(2)
where the V (j) shows the fitness of where j 1, 2, 3, … n, and eF define the finest fitness in iterations. The is defined as bellow:(3)
where the calculation of d is shown in Equation (4)(4)
where ) shows the maximum iteration.The is presented as follows:
(5)
where aF stands for best fitness, cF stands for worst fitness, cond. describes that V (j) categorizes the population in the initial half, g defines a random value between [0, 1] interval. Index defines sorted values of fitness and computed, as described in Equation (6).index = Sort (V)(6)
The uncertainty is described in Equation (5), simulated using r. The log decreased the change rate of numerical values; therefore, not too many changes occur in frequency.
The slime mould changes its search pattern, conforming to the nature of the worth of the food. When the mass is greater, food concentration becomes sufficient, and the mass should decrease when the food concentration becomes poor, as presented in Equation (7).
(7)
where the ub and lb define the upper and lower boundaries, and g defines the random value of 0 and 1. Figure 5 depicts the optimization process of the feature vector.Table 3 depicts the selected feature vector dimensions after applying SMA.
Table 3 shows the best-selected features number on PH2, MED-NODE, HAM10000, and ISIC 2019 datasets in both the training and testing phases.
3.2.2. Classification Using Selected Classifiers
The classifier takes the value of numerous features to make a prediction and consists of the number of parameters that it should learn from training data. The learned classifier shows the correspondence between the labels in training data and features [57]. In the proposed methodology, by using optimal features, the three classifiers have been utilised to differentiate the skin lesions into relevant classes.
The cubic kernel SVM [58] and its chosen parameters are stated in Table 4.
For classification purposes, the weighted KNN [59] and fine KNN [60] selected parameters are presented in Table 5.
4. Experimental Discussion/Setup
The achievement of the proposed segmentation approach is estimated on four public datasets ISIC 2016 [61], 2017 [62], 2018 [63], and PH2 [64,65]. The four public datasets ISIC 2019 [66,67], HAM10000 [66], PH2 [64], and MED-NODE [68,69] were utilised to estimate the performance of the proposed classification framework after augmentation. MATLAB 2020b is utilised as an implementation tool, using Intel core i5 6th Generation hardware on Windows 10.
4.1. Experiment#1: Segmentation
The proposed segmentation approach performance is computed based on global accuracy, mean Accuracy, meanIoU, weightedIoU, and mean BF score using ISIC 2016,17,18, and PH2 datasets, as shown in Table 6.
Table 6 depicts the proposed segmentation results, in which we achieved a global accuracy of 0.97481, 0.97297, 0.98642, 0.95914 on ISIC 2016, 2017, 2018, and PH2, respectively. The proposed framework segmentation outcomes using benchmark ISIC 2016 and PH2 datasets are stated in Figure 6 and Figure 7.
The achieved results are also compared to existing research work, as presented in Table 7.
On the 2016 challenge dataset, the existing technique provides a maximum of 96.2% accuracy using GA-based optimization [22]. On the 2017 segmentation challenging dataset, FC-DPN provides 95.14% accuracy but some lesions are not segmented accurately due to blurry and low-contrast images [74]. The w-net model provides 97.39% accuracy of segmentation, though an improvement is required in the deep learning framework to increase the segmentation results [77]. Antialiasing convolution model is utilised for skin lesion segmentation, providing 95% prediction scores. The segmentation scores might be increased using the improved features optimization approach [70].
The proposed method in this article consists of Mobilenetv2 and DeepLabv3+, which detects lesion boundaries more accurately, with an accuracy of 97.48%, 97.29%, 98.64% and 95.91% on challenge 2016, 17, 18 and PH2, respectively, making it far more efficient compared to the existing work.
4.2. Experiment#2: Skin Lesions Classification
In the classification experiment, features are computed using pre-trained DenseNet-201 and selected optimum features by SMA that are supplied to the classifiers on 5-fold cross-validation. The graphical depiction of the proposed classification results is expressed in Figure 8, Figure 9, Figure 10 and Figure 11. The classification results are described in Table 8.
As given in Table 8, the classification of the MED-NODE dataset was performed using three classifiers: cubic SVM, weighted KNN, and fine KNN, with an overall accuracy of 97.32%, 97.62%, and 99.33%, respectively. All the classifiers were trained using cross-validation 5 folds dataset distribution. The classification outcomes on the PH2 are mentioned in Table 9.
In Table 9, cubic SVM achieved an accuracy of 97.87%. The results of the weighted KNN and fine KNN classifiers are 98.09% and 98.88%, respectively. The classification outcomes on the HAM10000 are mentioned in Table 10.
Table 11 shows the outcomes of the cubic SVM, weighted KNN, and fine KNN, which obtained an overall accuracy of 90.65%, 86.90%, and 92.01%, respectively. The classification outcomes of the ISIC 2019 are depicted in Table 11.
Table 11 shows the outcomes of the cubic SVM, weighted KNN, and fine KNN with an overall accuracy of 89.99%, 90.22%, and 91.7%. The result shows that fine KNN performs best among the three classifiers. The classification results comparison is mentioned in Table 12.
In Table 12, the classification results with existing methods using ISIC 2019, HAM10000, PH2, and MED-NODE datasets are shown. On dataset ISIC 2019, deep learning and an entropy-based approach provided 91% accuracy; however, there is still room to improve the model for better accuracy [83]. On the HAM-10000 dataset, the accuracy rate achieved is 89.8% based on deep features extraction and selection approach [89]. On the dataset PH2, 97.5% accuracy is achieved using a combination of deep and texture features. The classification results might be increased using shape and colour features [91]. On MED-NODE dataset, the accuracy is 97.70% using transfer learning model [69].
In this research, however, features from the selected layers of the pre-trained models and the best features are selected using an SMA model that provides an accuracy of 91.7% on ISIC 2019, 92.01% on HAM10000, 98.88% on PH2, and 99.33% on MED-NODE datasets. The experimental outcomes show that the achieved outcomes are finer compared to the newest works in this domain.
5. Conclusions
Skin lesions’ detection is a complex job due to resemblances among the classes of skin lesions. To overcome the existing challenge, novel deep learning models are designed for skin lesion analysis. To perform semantic segmentation, the deep features are taken through the pre-trained model Mobilenetv2, which are then passed to the DeepLabv3+ for the extraction of the exact border of the lesion. The proposed segmentation approach is evaluated based on Mean Accuracy, Global Accuracy, BF Score, Weighted IoU, and Mean IoU on ISIC 2016, 2017, 2018, and PH2 datasets, which provide a global accuracy of 0.97481, 0.97297, 0.98642, and 0.95914, respectively.
In the proposed classification model, deep features are taken using DenseNet-201, and select optimal features by SMA, which are then evaluated on the MED-NODE, PH2, HAM-10000, and ISIC 2019 benchmark datasets, providing an accuracy of 99.33%, 98.88%, 92.01%, and 91.7, respectively. The achieved outcomes of segmentation and classification are far better compared to existing techniques.
Conceptualization, M.Z. and J.A.; Methodology, M.S., G.A.M. and S.K.; Formal analysis, M.A.A. All authors have read and agreed to the published version of the manuscript.
Not applicable.
The authors declare no conflict of interest.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Figure 1. Skin lesions’ recognition. Segmentation using Mobilenetv2 and DeepLabv3+; classification based on DenseNet-201 and SMA with SVM, KNN classifiers.
Figure 6. Visualisation results (a) input image of ISIC 2016 (b) segmented output of ISIC 2016 (c) input image of ISIC 2017 (d) segmented output of ISIC 2017.
Figure 7. Visualisation results (a) input image of ISIC 2018 (b) segmented output of ISIC 2018 (c) input image of PH2 (d) segmented output of PH2.
Figure 8. Classification results on MED-NODE (a) ROC curve of fine KNN (b) confusion matrix of fine KNN.
Figure 9. Classification results on PH2 (a) ROC curve of fine KNN (b) confusion matrix of fine KNN.
Figure 10. Classification results on HAM10000 (a) ROC curve of fine KNN (b) confusion matrix of fine KNN.
Figure 11. Classification results on ISIC 2019 (a) ROC Curve of fine KNN (b) confusion matrix of fine KNN.
Hyper parameters of the proposed framework.
Batch-size | 32 |
Training epochs | 100 |
Rate of learning | 0.0001 |
Optimizer solver | Sgdm |
Selected parameters of SMA.
Features selection model | SMA |
No. of K Nearest Neighbour | Opts.k = 5 |
Validation data ratio | Ho = 0.2 |
No. of solutions | Opts. N = 10 |
Maximum iterations | Opts. T = 100 |
Selected numbers of features in each dataset.
Datasets | Total Features | No. of Selected Features |
---|---|---|
PH2 |
|
N × 50 |
MED-NODE | N × 30 | |
HAM10000 | N × 227 | |
ISIC 2019 | N × 251 |
Parameters of SVM.
Model | Cubic SVM |
---|---|
Function of Kernel | Cubic |
Scale of Kernel | Automatic |
Level of box constraint | 01 |
Multiclass method | One-vs-One |
Data Standardization | True |
Parameters of KNN classifier.
Model | Fine KNN | Weighted KNN |
---|---|---|
No. of neighbours | 1 | 10 |
Metric of distance | Euclidean | Euclidean |
Distance weight | Equal | Square inverse |
Data Standardization | True | True |
Skin lesion segmentation performance on ISIC-2016, 2017, 2018, and PH2 datasets.
Dataset | Global Accuracy | Mean Accuracy | Mean IoU | Weighted IoU | Mean BF Score |
---|---|---|---|---|---|
ISIC 2016 | 0.97481 | 0.96253 | 0.93960 | 0.95082 | 0.88649 |
ISIC 2017 | 0.97297 | 0.96841 | 0.94483 | 0.94724 | 0.84741 |
ISIC 2018 | 0.98642 | 0.91472 | 0.88139 | 0.97390 | 0.78364 |
PH2 | 0.95914 | 0.96005 | 0.90477 | 0.92299 | 0.82448 |
Comparison of the existing research works on similar datasets.
Ref# | Year | Datasets | Accuracy |
---|---|---|---|
[ |
2022 | ISIC 2016 | 96% |
[ |
2021 | 95.4% | |
[ |
2021 | 96.2% | |
[ |
2020 | 93.8% | |
[ |
2020 | 95.24% | |
[ |
2019 | 95.78% | |
Proposed Method | 97.48% | ||
[ |
2022 | ISIC 2017 | 95% |
[ |
2021 | 92.6% | |
[ |
2020 | 93.8% | |
[ |
2020 | 95.14% | |
[ |
2020 | 94.58 | |
[ |
2018 | 94.03% | |
Proposed Method | 97.29% | ||
[ |
2022 | ISIC 2018 | 97.39 |
[ |
2021 | 96.95% | |
[ |
2021 | 95.0% | |
[ |
2020 | 94.7% | |
[ |
2019 | 96.23 | |
[ |
2018 | 96.80% | |
Proposed Method | 98.64% | ||
[ |
2022 | PH2 | 95.14 |
[ |
2021 | 94.3% | |
[ |
2021 | 94.6% | |
[ |
2020 | 94.9% | |
[ |
2020 | 93.2% | |
[ |
2019 | 93.1% | |
Proposed Method | 95.91% |
Proposed classification model outcomes on the MED-NODE dataset.
Classifiers | Classes | Accuracy | Precision | Recall | F1 Score | Overall |
|
---|---|---|---|---|---|---|---|
Melanoma (M) | Nevus (N) | ||||||
Cubic SVM | ✓ | 97.32% | 0.98 | 0.97 | 0.98 | 97.87% | |
✓ | 97.32% | 0.96 | 0.97 | 0.97 | |||
Weighted KNN | ✓ | 97.62% | 0.99 | 0.97 | 0.98 | 97.62% | |
✓ | 97.62% | 0.96 | 0.98 | 0.97 | |||
Fine KNN | ✓ | 99.33% | 0.99 | 0.99 | 0.99 | 99.33% | |
✓ | 99.33% | 0.99 | 0.99 | 0.99 |
Proposed classification model results on PH2.
Classifiers | Classes | Accuracy | Precision | Recall | F1 Score | Overall |
||
---|---|---|---|---|---|---|---|---|
AN | CN | M | ||||||
Cubic SVM | ✓ | 97.95% | 0.97 | 0.97 | 0.97 | 97.87% | ||
✓ | 98.06% | 0.97 | 0.97 | 0.97 | ||||
✓ | 99.74% | 1.00 | 1.00 | 1.00 | ||||
Weighted KNN | ✓ | 98.17% | 0.97 | 0.98 | 0.97 | 98.09% | ||
✓ | 98.43% | 0.98 | 0.97 | 0.98 | ||||
✓ | 99.59% | 1.00 | 0.99 | 0.99 | ||||
Fine KNN | ✓ | 98.99% | 0.98 | 0.99 | 0.98 | 98.88% | ||
✓ | 98.92% | 0.99 | 0.98 | 0.98 | ||||
✓ | 99.85% | 1.00 | 1.00 | 1.00 |
Proposed classification model results on HAM10000.
Classifiers | Classes | Accuracy | Precision | Recall | F1 Score | Overall |
||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
AK | BCC | BK | D | M | N | VL | ||||||
Cubic SVM | ✓ | 97.18% | 0.92 | 0.89 | 0.90 | 90.65% | ||||||
✓ | 97.35% | 0.92 | 0.90 | 0.91 | ||||||||
✓ | 95.67% | 0.84 | 0.86 | 0.85 | ||||||||
✓ | 98.93% | 0.97 | 0.96 | 0.96 | ||||||||
✓ | 95.57% | 0.84 | 0.85 | 0.84 | ||||||||
✓ | 96.91% | 0.87 | 0.91 | 0.89 | ||||||||
✓ | 99.68% | 0.99 | 0.99 | 0.99 | ||||||||
Weighted KNN | ✓ | 95.89% | 0.94 | 0.80 | 0.87 | 86.96% | ||||||
✓ | 96.49% | 0.89 | 0.86 | 0.88 | ||||||||
✓ | 94.55% | 0.71 | 0.88 | 0.79 | ||||||||
✓ | 97.22% | 0.99 | 0.84 | 0.91 | ||||||||
✓ | 94.73% | 0.80 | 0.82 | 0.81 | ||||||||
✓ | 95.67% | 0.77 | 0.91 | 0.83 | ||||||||
✓ | 99.38% | 0.98 | 0.98 | 0.98 | ||||||||
Fine KNN | ✓ | 97.67% | 0.95 | 0.89 | 0.92 | 92.01% | ||||||
✓ | 98.02% | 0.94 | 0.92 | 0.93 | ||||||||
✓ | 96.74% | 0.87 | 0.90 | 0.88 | ||||||||
✓ | 98.36% | 0.98 | 0.91 | 0.95 | ||||||||
✓ | 96.82% | 0.89 | 0.88 | 0.89 | ||||||||
✓ | 96.69% | 0.81 | 0.96 | 0.87 | ||||||||
✓ | 99.7% | 0.99 | 0.99 | 0.99 |
Classification results on ISIC 2019 dataset.
Classifiers | Classes | Accuracy | Precision | Recall | F1 Score | Overall |
|||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
AK | BCC | BK | D | M | MN | SCC | VL | ||||||
Cubic SVM | ✓ | 97.58% | 0.88 | 0.88 | 0.88 | 89.99% | |||||||
✓ | 96.81% | 0.89 | 0.88 | 0.88 | |||||||||
✓ | 95.81% | 0.84 | 0.86 | 0.85 | |||||||||
✓ | 98.76% | 0.96 | 0.94 | 0.95 | |||||||||
✓ | 96.23% | 0.83 | 0.88 | 0.85 | |||||||||
✓ | 97.77% | 0.90 | 0.90 | 0.90 | |||||||||
✓ | 97.37% | 0.91 | 0.87 | 0.89 | |||||||||
✓ | 97.77% | 0.91 | 0.87 | 0.89 | |||||||||
Weighted KNN | ✓ | 96.65% | 0.86 | 0.88 | 0.87 | 90.22% | |||||||
✓ | 96.44% | 0.83 | 0.91 | 0.87 | |||||||||
✓ | 98.17% | 0.98 | 0.88 | 0.93 | |||||||||
✓ | 96.4% | 0.82 | 0.90 | 0.86 | |||||||||
✓ | 97.93% | 0.90 | 0.91 | 0.91 | |||||||||
✓ | 97.67% | 0.94 | 0.87 | 0.90 | |||||||||
✓ | 99.41% | 0.99 | 0.97 | 0.98 | |||||||||
✓ | 97.77% | 0.91 | 0.87 | 0.89 | |||||||||
Fine KNN | ✓ | 98.13% | 0.93 | 0.89 | 0.91 | 91.7% | |||||||
✓ | 97.16% | 0.88 | 0.91 | 0.89 | |||||||||
✓ | 96.95% | 0.88 | 0.90 | 0.89 | |||||||||
✓ | 98.66% | 0.98 | 0.92 | 0.95 | |||||||||
✓ | 96.82% | 0.85 | 0.91 | 0.88 | |||||||||
✓ | 97.97% | 0.89 | 0.93 | 0.91 | |||||||||
✓ | 98.14% | 0.95 | 0.89 | 0.92 | |||||||||
✓ | 99.57% | 0.99 | 0.98 | 0.98 |
Comparisons of proposed classification method outcomes with existing approaches.
Ref# | Year | Datasets | Accuracy |
---|---|---|---|
[ |
2019 | ISIC 2019 | 91% |
[ |
2020 | 84.79% | |
[ |
2020 | 82.5% | |
Proposed | 91.7% | ||
[ |
2022 | HAM10000 | 80% |
[ |
2021 | 85.50% | |
[ |
2021 | 88.50% | |
[ |
2019 | 89.8% | |
Proposed | 92.01% | ||
[ |
2022 | PH2 | 94.97% |
[ |
2021 | 97.5% | |
[ |
2020 | 96.9% | |
[ |
2020 | 85.7% | |
[ |
2020 | 94.0 | |
Proposed | 98.88% | ||
[ |
2022 | MED-NODE | 92% |
[ |
2021 | 97% | |
[ |
2020 | 83.33% | |
[ |
2019 | 97.70% | |
Proposed | 99.33% |
References
1. Hameed, N.; Ruskin, A.; Hassan, K.A.; Hossain, M.A. A comprehensive survey on image-based computer aided diagnosis systems for skin cancer. Proceedings of the 2016 10th International Conference on Software, Knowledge, Information Management & Applications (SKIMA); Chengdu, China, 15–17 December 2016.
2. Lai-Cheong, J.E.; McGrath, J.A. Structure and function of skin, hair and nails. Medicine; 2013; 41, pp. 317-320. [DOI: https://dx.doi.org/10.1016/j.mpmed.2013.04.017]
3. Gordon, R. Skin cancer: An overview of epidemiology and risk factors. Semin. Oncol. Nurs.; 2013; 29, pp. 160-169. [DOI: https://dx.doi.org/10.1016/j.soncn.2013.06.002] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/23958214]
4. Javed, R.; Rahim, M.S.M.; Saba, T.; Rehman, A. A comparative study of features selection for skin lesion detection from dermoscopic images. Netw. Model. Anal. Heal. Inform. Bioinform.; 2019; 9, 4.
5. Zhang, N.; Cai, Y.-X.; Wang, Y.-Y.; Tian, Y.-T.; Wang, X.-L.; Badami, B. Skin cancer diagnosis based on optimized convolutional neural network. Artif. Intell. Med.; 2020; 102, 101756. [DOI: https://dx.doi.org/10.1016/j.artmed.2019.101756] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/31980095]
6. Key Statistics for Melanoma Skin Cancer. Available online: https://www.cancer.org/cancer/melanoma-skin-cancer/about/key-statistics.html (accessed on 12 October 2021).
7. Shetty, B.; Fernandes, R.; Rodrigues, A.P.; Chengoden, R.; Bhattacharya, S.; Lakshmanna, K. Skin lesion classification of dermoscopic images using machine learning and convolutional neural network. Sci. Rep.; 2022; 12, 18134. [DOI: https://dx.doi.org/10.1038/s41598-022-22644-9]
8. Liang, Y.; Sun, L.; Ser, W.; Lin, F.; Thng, S.T.G.; Chen, Q.; Lin, Z. Classification of non-tumorous skin pigmentation disorders using voting based probabilistic linear discriminant analysis. Comput. Biol. Med.; 2018; 99, pp. 123-132. [DOI: https://dx.doi.org/10.1016/j.compbiomed.2018.05.026]
9. Amin, J.; Sharif, M.; Raza, M.; Saba, T.; Rehman, A. Brain tumor classification: Feature fusion. Proceedings of the 2019 International Conference on Computer and Information Sciences (ICCIS); Sakaka, Saudi Arabia, 3–4 April 2019.
10. Amin, J.; Sharif, M.; Yasmin, M.; Saba, T.; Raza, M. Use of machine intelligence to conduct analysis of human brain data for detection of abnormalities in its cognitive functions. Multimed. Tools Appl.; 2020; 79, pp. 10955-10973. [DOI: https://dx.doi.org/10.1007/s11042-019-7324-y]
11. Amin, J.; Almas-Anjum, M.; Sharif, M.; Kadry, S.; Nam, Y. Fruits and vegetable diseases recognition using convolutional neural networks. Comput. Mater. Contin.; 2021; 70, pp. 619-635. [DOI: https://dx.doi.org/10.32604/cmc.2022.018562]
12. Saleem, S.; Amin, J.; Sharif, M.; Anjum, M.A.; Iqbal, M.; Wang, S.-H. A deep network designed for segmentation and classification of leukemia using fusion of the transfer learning models. Complex Intell. Syst.; 2022; 8, pp. 3105-3120. [DOI: https://dx.doi.org/10.1007/s40747-021-00473-z]
13. Wei, Z.; Shi, F.; Song, H.; Ji, W.; Han, G. Attentive boundary aware network for multi-scale skin lesion segmentation with adversarial training. Multimed. Tools Appl.; 2020; 79, pp. 27115-27136. [DOI: https://dx.doi.org/10.1007/s11042-020-09334-2]
14. Wei, Z.; Song, H.; Li, L.; Chen, Q.; Han, G. Attention-based DenseUnet network with adversarial training for skin lesion segmentation. IEEE Access; 2019; 7, pp. 136616-136629. [DOI: https://dx.doi.org/10.1109/ACCESS.2019.2940794]
15. Sharif, M.I.; Li, J.P.; Amin, J.; Sharif, A. An improved framework for brain tumor analysis using MRI based on YOLOv2 and convolutional neural network. Complex Intell. Syst.; 2021; 7, pp. 2023-2036. [DOI: https://dx.doi.org/10.1007/s40747-021-00310-3]
16. Sharif, M.; Amin, J.; Raza, M.; Yasmin, M.; Satapathy, S.C. An integrated design of particle swarm optimization (PSO) with fusion of features for detection of brain tumor. Pattern Recognit. Lett.; 2020; 129, pp. 150-157. [DOI: https://dx.doi.org/10.1016/j.patrec.2019.11.017]
17. Saba, T.; Mohamed, A.S.; El-Affendi, M.; Amin, J.; Sharif, M. Brain tumor detection using fusion of hand crafted and deep learning features. Cogn. Syst. Res.; 2020; 59, pp. 221-230. [DOI: https://dx.doi.org/10.1016/j.cogsys.2019.09.007]
18. Hwang, Y.N.; Seo, M.J.; Kim, S.M. A segmentation of melanocytic skin lesions in dermoscopic and standard images using a hybrid two-stage approach. BioMed Res. Int.; 2021; 2021, 5562801. [DOI: https://dx.doi.org/10.1155/2021/5562801]
19. Amin, J.; Sharif, M.; Gul, N.; Yasmin, M.; Shad, S.A. Brain tumor classification based on DWT fusion of MRI sequences using convolutional neural network. Pattern Recognit. Lett.; 2020; 129, pp. 115-122. [DOI: https://dx.doi.org/10.1016/j.patrec.2019.11.016]
20. Garg, S.; Jindal, B. Skin lesion segmentation using k-mean and optimized fire fly algorithm. Multimed. Tools Appl.; 2021; 80, pp. 7397-7410. [DOI: https://dx.doi.org/10.1007/s11042-020-10064-8]
21. Amin, J.; Sharif, A.; Gul, N.; Anjum, M.A.; Nisar, M.W.; Azam, F.; Bukhari, S.A.C. Integrated design of deep features fusion for localization and classification of skin cancer. Pattern Recognit. Lett.; 2020; 131, pp. 63-70. [DOI: https://dx.doi.org/10.1016/j.patrec.2019.11.042]
22. Ashour, A.S.; Nagieb, R.M.; El-Khobby, H.A.; Abd-Elnaby, M.M.; Dey, N. Genetic algorithm-based initial contour optimization for skin lesion border detection. Multimed. Tools Appl.; 2021; 80, pp. 2583-2597. [DOI: https://dx.doi.org/10.1007/s11042-020-09792-8]
23. Amin, J.; Sharif, M.; Gul, N.; Raza, M.; Anjum, M.A.; Nisar, M.W.; Bukhari, S.A.C. Brain tumor detection by using stacked autoencoders in deep learning. J. Med. Syst.; 2020; 44, 32. [DOI: https://dx.doi.org/10.1007/s10916-019-1483-2]
24. Kaur, R.; GholamHosseini, H.; Sinha, R.; Lindén, M. Automatic lesion segmentation using atrous convolutional deep neural networks in dermoscopic skin cancer images. BMC Med. Imaging; 2022; 22, 103. [DOI: https://dx.doi.org/10.1186/s12880-022-00829-y]
25. Sharif, M.; Amin, J.; Raza, M.; Anjum, M.A.; Afzal, H.; Shad, S.A. Brain tumor detection based on extreme learning. Neural Comput. Appl.; 2020; 32, pp. 15975-15987. [DOI: https://dx.doi.org/10.1007/s00521-019-04679-8]
26. Bagheri, F.; Tarokh, M.J.; Ziaratban, M. Skin lesion segmentation from dermoscopic images by using Mask R-CNN, Retina-Deeplab, and graph-based methods. Biomed. Signal Process. Control.; 2021; 67, 102533. [DOI: https://dx.doi.org/10.1016/j.bspc.2021.102533]
27. Amin, J.; Sharif, M.; Rehman, A.; Raza, M.; Mufti, M.R. Diabetic retinopathy detection and classification using hybrid feature set. Microsc. Res. Tech.; 2018; 81, pp. 990-996. [DOI: https://dx.doi.org/10.1002/jemt.23063] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/30447130]
28. Qamar, S.; Ahmad, P.; Shen, L. Dense encoder-decoder–based architecture for skin lesion segmentation. Cogn. Comput.; 2021; 13, pp. 583-594. [DOI: https://dx.doi.org/10.1007/s12559-020-09805-6]
29. Amin, J.; Sharif, M.; Anjum, M.A.; Raza, M.; Bukhari, S.A.C. Convolutional neural network with batch normalization for glioma and stroke lesion detection using MRI. Cogn. Syst. Res.; 2020; 59, pp. 304-311. [DOI: https://dx.doi.org/10.1016/j.cogsys.2019.10.002]
30. Wu, H.; Pan, J.; Li, Z.; Wen, Z.; Qin, J. Automated skin lesion segmentation via an adaptive dual attention module. IEEE Trans. Med. Imaging; 2020; 40, pp. 357-370. [DOI: https://dx.doi.org/10.1109/TMI.2020.3027341]
31. Muhammad, N.; Sharif, M.; Amin, J.; Mehboob, R.; Gilani, S.A.; Bibi, N.; Javed, H.; Ahmed, N. Neurochemical alterations in sudden unexplained perinatal deaths—A review. Front. Pediatr.; 2018; 6, 6. [DOI: https://dx.doi.org/10.3389/fped.2018.00006] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/29423392]
32. Abdulhamid, I.A.M.; Sahiner, A.; Rahebi, J. New auxiliary function with properties in nonsmooth global optimization for melanoma skin cancer segmentation. BioMed Res. Int.; 2020; 2020, 5345923. [DOI: https://dx.doi.org/10.1155/2020/5345923]
33. Sharif, M.; Amin, J.; Nisar, M.W.; Anjum, M.A.; Muhammad, N.; Shad, S.A. A unified patch based method for brain tumor detection using features fusion. Cogn. Syst. Res.; 2020; 59, pp. 273-286. [DOI: https://dx.doi.org/10.1016/j.cogsys.2019.10.001]
34. Goyal, M.; Oakley, A.; Bansal, P.; Dancey, D.; Yap, M.H. Skin lesion segmentation in dermoscopic images with ensemble deep learning methods. IEEE Access; 2019; 8, pp. 4171-4181. [DOI: https://dx.doi.org/10.1109/ACCESS.2019.2960504]
35. Sharif, M.; Amin, J.; Siddiqa, A.; Khan, H.U.; Malik, M.S.A.; Anjum, M.A.; Kadry, S. Recognition of different types of leukocytes using YOLOv2 and optimized bag-of-features. IEEE Access; 2020; 8, pp. 167448-167459. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.3021660]
36. Anjum, M.A.; Amin, J.; Sharif, M.; Khan, H.U.; Malik, M.S.A.; Kadry, S. Deep semantic segmentation and multi-class skin lesion classification based on convolutional neural network. IEEE Access; 2020; 8, pp. 129668-129678. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.3009276]
37. Lameski, J.; Jovanov, A.; Zdravevski, E.; Lameski, P.; Gievska, S. Skin lesion segmentation with deep learning. Proceedings of the IEEE EUROCON 2019-18th International Conference on Smart Technologies; Novi Sad, Serbia, 1–4 July 2019.
38. Sharif, M.; Amin, J.; Yasmin, M.; Rehman, A. Efficient hybrid approach to segment and classify exudates for DR prediction. Multimed. Tools Appl.; 2020; 79, pp. 11107-11123. [DOI: https://dx.doi.org/10.1007/s11042-018-6901-9]
39. Amin, J.; Sharif, M.; Gul, E.; Nayak, R.S. 3D-semantic segmentation and classification of stomach infections using uncertainty aware deep neural networks. Complex Intell. Syst.; 2022; 8, pp. 3041-3057. [DOI: https://dx.doi.org/10.1007/s40747-021-00328-7]
40. Amin, J.; Anjum, M.A.; Sharif, M.; Saba, T.; Tariq, U. An intelligence design for detection and classification of COVID-19 using fusion of classical and convolutional neural network and improved microscopic features selection approach. Microsc. Res. Tech.; 2021; 84, pp. 2254-2267. [DOI: https://dx.doi.org/10.1002/jemt.23779]
41. Li, H.; He, X.; Zhou, F.; Yu, Z.; Ni, D.; Chen, S.; Wang, T.; Lei, B. Dense deconvolutional network for skin lesion segmentation. IEEE J. Biomed. Health Inform.; 2018; 23, pp. 527-537. [DOI: https://dx.doi.org/10.1109/JBHI.2018.2859898]
42. Amin, J.; Sharif, M.; Anjum, M.A.; Khan, H.U.; Malik, M.S.A.; Kadry, S. An integrated design for classification and localization of diabetic foot ulcer based on CNN and YOLOv2-DFU models. IEEE Access; 2020; 8, pp. 228586-228597. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.3045732]
43. Amin, J.; Sharif, M.; Yasmin, M. Segmentation and classification of lung cancer: A review. Immunol. Endocr. Metab. Agents Med. Chem.; 2016; 16, pp. 82-99. [DOI: https://dx.doi.org/10.2174/187152221602161221215304]
44. Jaisakthi, S.A.; Chandrabose, A.; Mirunalini, P. Automatic skin lesion segmentation using semi-supervised learning technique. arXiv; 2017; arXiv: 1703.04301
45. Lynn, N.C.; Kyu, Z.M. Segmentation and classification of skin cancer melanoma from skin lesion images. Proceedings of the 2017 18th International Conference on Parallel and Distributed Computing, Applications and Technologies (PDCAT); Taipei, Taiwan, 18–20 December 2017.
46. Amin, J.; Anjum, M.A.; Sharif, M.; Kadry, S.; Nam, Y.; Wang, S. Convolutional Bi-LSTM based human gait recognition using video sequences. Comput. Mater. Contin.; 2021; 68, pp. 2693-2709. [DOI: https://dx.doi.org/10.32604/cmc.2021.016871]
47. Albahar, M.A. Skin lesion classification using convolutional neural network with novel regularizer. IEEE Access; 2019; 7, pp. 38306-38313. [DOI: https://dx.doi.org/10.1109/ACCESS.2019.2906241]
48. Milton, M.A.A. Automated skin lesion classification using ensemble of deep neural networks in ISIC 2018: Skin lesion analysis towards melanoma detection challenge. arXiv; 2019; arXiv: 1901.10802
49. Linsky, T.W.; Vergara, R.; Codina, N.; Nelson, J.W.; Walker, M.J.; Su, W.; Barnes, C.O.; Hsiang, T.-Y.; Esser-Nobis, K.; Yu, K. De novo design of potent and resilient hACE2 decoys to neutralize SARS-CoV-2. Science; 2020; 370, pp. 1208-1214. [DOI: https://dx.doi.org/10.1126/science.abe0075] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33154107]
50. Zhang, J.; Xie, Y.; Xia, Y.; Shen, C. Attention residual learning for skin lesion classification. IEEE Trans. Med. Imaging; 2019; 38, pp. 2092-2103. [DOI: https://dx.doi.org/10.1109/TMI.2019.2893944]
51. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. Mobilenetv2: Inverted residuals and linear bottlenecks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; Salt Lake City, UT, USA, 18–22 June 2018.
52. Chen, L.-C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. Proceedings of the European Conference on Computer Vision (ECCV); Munich, Germany, 8–14 September 2018.
53. Nalini, M.; Radhika, K. Comparative analysis of deep network models through transfer learning. Proceedings of the 2020 Fourth International Conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud)(I-SMAC); Palladam, India, 7–9 October 2020.
54. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; Honolulu, HI, USA, 21–26 July 2017.
55. Li, S.; Chen, H.; Wang, M.; Heidari, A.A.; Mirjalili, S. Slime mould algorithm: A new method for stochastic optimization. Future Gener. Comput. Syst.; 2020; 111, pp. 300-323. [DOI: https://dx.doi.org/10.1016/j.future.2020.03.055]
56. Wazery, Y.M.; Saber, E.; Houssein, E.H.; Ali, A.A.; Amer, E. An efficient slime mould algorithm combined with k-nearest neighbor for medical classification tasks. IEEE Access; 2021; 9, pp. 113666-113682. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3105485]
57. Pereira, F.; Mitchell, T.; Botvinick, M. Machine learning classifiers and fMRI: A tutorial overview. Neuroimage; 2009; 45, pp. S199-S209. [DOI: https://dx.doi.org/10.1016/j.neuroimage.2008.11.007]
58. Singh, S.; Kumar, R. Histopathological image analysis for breast cancer detection using cubic SVM. Proceedings of the 2020 7th International Conference on Signal Processing and Integrated Networks (SPIN); Noida, India, 27–28 February 2020.
59. Ali, A.; Alrubei, M.; Hassan, L.; Al-Ja’afari, M.; Abdulwahed, S. Diabetes classification based on KNN. IIUM Eng. J.; 2020; 21, pp. 175-181. [DOI: https://dx.doi.org/10.31436/iiumej.v21i1.1206]
60. Subbarao, M.V.; Samundiswary, P. Performance analysis of modulation recognition in multipath fading channels using pattern recognition classifiers. Wirel. Pers. Commun.; 2020; 115, pp. 129-151. [DOI: https://dx.doi.org/10.1007/s11277-020-07564-z]
61. Gutman, D.; Codella, N.C.; Celebi, E.; Helba, B.; Marchetti, M.; Mishra, N.; Halpern, A. Skin lesion analysis toward melanoma detection: A challenge at the international symposium on biomedical imaging (ISBI) 2016, hosted by the international skin imaging collaboration (ISIC). arXiv; 2016; arXiv: 1605.01397
62. Codella, N.C.; Gutman, D.; Celebi, M.E.; Helba, B.; Marchetti, M.A.; Dusza, S.W.; Kalloo, A.; Liopyris, K.; Mishra, N.; Kittler, H. Skin lesion analysis toward melanoma detection: A challenge at the 2017 international symposium on biomedical imaging (isbi), hosted by the international skin imaging collaboration (isic). Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018); Washington, DC, USA, 4–7 April 2018.
63. Codella, N.; Rotemberg, V.; Tschandl, P.; Celebi, M.E.; Dusza, S.; Gutman, D.; Helba, B.; Kalloo, A.; Liopyris, K.; Marchetti, M. Skin lesion analysis toward melanoma detection 2018: A challenge hosted by the international skin imaging collaboration (isic). arXiv; 2019; arXiv: 1902.03368
64. Mendonça, T.; Ferreira, P.M.; Marques, J.S.; Marcal, A.R.; Rozeira, J. PH 2-A dermoscopic image database for research and benchmarking. Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC); Osaka, Japan, 3–7 July 2013.
65. Ünver, H.M.; Ayan, E. Skin lesion segmentation in dermoscopic images with combination of YOLO and grabcut algorithm. Diagnostics; 2019; 9, 72. [DOI: https://dx.doi.org/10.3390/diagnostics9030072]
66. Tschandl, P.; Rosendahl, C.; Kittler, H. The HAM10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions. Sci. Data; 2018; 5, 180161. [DOI: https://dx.doi.org/10.1038/sdata.2018.161] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/30106392]
67. Combalia, M.; Codella, N.C.; Rotemberg, V.; Helba, B.; Vilaplana, V.; Reiter, O.; Carrera, C.; Barreiro, A.; Halpern, A.C.; Puig, S. Bcn20000: Dermoscopic lesions in the wild. arXiv; 2019; arXiv: 1908.02288
68. Giotis, I.; Molders, N.; Land, S.; Biehl, M.; Jonkman, M.F.; Petkov, N. MED-NODE: A computer-assisted melanoma diagnosis system using non-dermoscopic images. Expert Syst. Appl.; 2015; 42, pp. 6578-6585. [DOI: https://dx.doi.org/10.1016/j.eswa.2015.04.034]
69. Hosny, K.M.; Kassem, M.A.; Foaud, M.M. Classification of skin lesions using transfer learning and augmentation with Alex-net. PloS ONE; 2019; 14, e0217293. [DOI: https://dx.doi.org/10.1371/journal.pone.0217293] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/31112591]
70. Le, P.T.; Chang, C.-C.; Li, Y.-H.; Hsu, Y.-C.; Wang, J.-C. Antialiasing Attention Spatial Convolution Model for Skin Lesion Segmentation with Applications in the Medical IoT. Wirel. Commun. Mob. Comput.; 2022; 2022, 1278515. [DOI: https://dx.doi.org/10.1155/2022/1278515]
71. Tong, X.; Wei, J.; Sun, B.; Su, S.; Zuo, Z.; Wu, P. ASCU-Net: Attention gate, spatial and channel attention u-net for skin lesion segmentation. Diagnostics; 2021; 11, 501. [DOI: https://dx.doi.org/10.3390/diagnostics11030501] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33809048]
72. Xie, F.; Yang, J.; Jiang, Z.; Zheng, Y.; Wang, Y. Skin lesion segmentation using high-resolution convolutional neural network. Comput. Methods Programs Biomed.; 2020; 186, 105241. [DOI: https://dx.doi.org/10.1016/j.cmpb.2019.105241]
73. Bi, L.; Kim, J.; Ahn, E.; Kumar, A.; Feng, D.; Fulham, M. Step-wise integration of deep class-specific learning for dermoscopic image segmentation. Pattern Recognit.; 2019; 85, pp. 78-89. [DOI: https://dx.doi.org/10.1016/j.patcog.2018.08.001]
74. Shan, P.; Wang, Y.; Fu, C.; Song, W.; Chen, J. Automatic skin lesion segmentation based on FC-DPN. Comput. Biol. Med.; 2020; 123, 103762. [DOI: https://dx.doi.org/10.1016/j.compbiomed.2020.103762]
75. Kaymak, R.; Kaymak, C.; Ucar, A. Skin lesion segmentation using fully convolutional networks: A comparative experimental study. Expert Syst. Appl.; 2020; 161, 113742. [DOI: https://dx.doi.org/10.1016/j.eswa.2020.113742]
76. Al-Masni, M.A.; Al-Antari, M.A.; Choi, M.-T.; Han, S.-M.; Kim, T.-S. Skin lesion segmentation in dermoscopy images via deep full resolution convolutional networks. Comput. Methods Programs Biomed.; 2018; 162, pp. 221-231. [DOI: https://dx.doi.org/10.1016/j.cmpb.2018.05.027] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/29903489]
77. Khouloud, S.; Ahlem, M.; Fadel, T.; Amel, S. W-net and inception residual network for skin lesion segmentation and classification. Appl. Intell.; 2022; 52, pp. 3976-3994. [DOI: https://dx.doi.org/10.1007/s10489-021-02652-4]
78. Arora, R.; Raman, B.; Nayyar, K.; Awasthi, R. Automated skin lesion segmentation using attention-based deep convolutional neural network. Biomed. Signal Process. Control.; 2021; 65, 102358. [DOI: https://dx.doi.org/10.1016/j.bspc.2020.102358]
79. Bi, L.; Feng, D.; Kim, J. Improving automatic skin lesion segmentation using adversarial learning based data augmentation. arXiv; 2018; arXiv: 1807.08392
80. Jahanifar, M.; Tajeddin, N.Z.; Koohbanani, N.A.; Gooya, A.; Rajpoot, N. Segmentation of skin lesions and their attributes using multi-scale convolutional neural networks and domain specific augmentations. arXiv; 2018; arXiv: 1809.10243
81. Chen, P.; Huang, S.; Yue, Q. Skin Lesion Segmentation Using Recurrent Attentional Convolutional Networks. IEEE Access; 2022; 10, pp. 94007-94018. [DOI: https://dx.doi.org/10.1109/ACCESS.2022.3204280]
82. Zhang, G.; Shen, X.; Chen, S.; Liang, L.; Luo, Y.; Yu, J.; Lu, J. DSM: A deep supervised multi-scale network learning for skin cancer segmentation. IEEE Access; 2019; 7, pp. 140936-140945. [DOI: https://dx.doi.org/10.1109/ACCESS.2019.2943628]
83. Pacheco, A.G.; Ali, A.-R.; Trappenberg, T. Skin cancer detection based on deep learning and entropy to detect outlier samples. arXiv; 2019; arXiv: 1909.04525
84. Setiawan, A.W. Effect of Color Enhancement on Early Detection of Skin Cancer using Convolutional Neural Network. Proceedings of the 2020 IEEE International Conference on Informatics, IoT, and Enabling Technologies (ICIoT); Doha, Qatar, 2–5 February 2020.
85. Pacheco, A.G.; Sastry, C.S.; Trappenberg, T.; Oore, S.; Krohling, R.A. On out-of-distribution detection algorithms with deep neural skin cancer classifiers. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops; Seattle, WA, USA, 14–19 June 2020.
86. Saarela, M.; Geogieva, L. Robustness, Stability, and Fidelity of Explanations for a Deep Skin Cancer Classification Model. Appl. Sci.; 2022; 12, 9545. [DOI: https://dx.doi.org/10.3390/app12199545]
87. Afza, F.; Sharif, M.; Mittal, M.; Khan, M.A.; Hemanth, D.J. A hierarchical three-step superpixels and deep learning framework for skin lesion classification. Methods; 2022; 202, pp. 88-102. [DOI: https://dx.doi.org/10.1016/j.ymeth.2021.02.013]
88. Khan, M.A.; Akram, T.; Zhang, Y.-D.; Sharif, M. Attributes based skin lesion detection and recognition: A mask RCNN and transfer learning-based deep learning framework. Pattern Recognit. Lett.; 2021; 143, pp. 58-66. [DOI: https://dx.doi.org/10.1016/j.patrec.2020.12.015]
89. Khan, M.A.; Javed, M.Y.; Sharif, M.; Saba, T.; Rehman, A. Multi-model deep neural network based features extraction and optimal selection approach for skin lesion classification. Proceedings of the 2019 International Conference on Computer and Information Sciences (ICCIS); Sakaka, Saudi Arabia, 3–4 April 2019.
90. Hosny, K.M.; Kassem, M.A. Refined residual deep convolutional network for skin lesion classification. J. Digit. Imaging; 2022; 35, pp. 258-280. [DOI: https://dx.doi.org/10.1007/s10278-021-00552-0]
91. Alizadeh, S.M.; Mahloojifar, A. Automatic skin cancer detection in dermoscopy images by combining convolutional neural networks and texture features. Int. J. Imaging Syst. Technol.; 2021; 31, pp. 695-707. [DOI: https://dx.doi.org/10.1002/ima.22490]
92. Kumar, M.; Alshehri, M.; AlGhamdi, R.; Sharma, P.; Deep, V. A de-ann inspired skin cancer detection approach using fuzzy c-means clustering. Mob. Netw. Appl.; 2020; 25, pp. 1319-1329. [DOI: https://dx.doi.org/10.1007/s11036-020-01550-2]
93. Arora, G.; Dubey, A.K.; Jaffery, Z.A.; Rocha, A. Bag of feature and support vector machine based early diagnosis of skin cancer. Neural Comput. Appl.; 2020; 34, pp. 8385-8392. [DOI: https://dx.doi.org/10.1007/s00521-020-05212-y]
94. Xie, Y.; Zhang, J.; Xia, Y.; Shen, C. A mutual bootstrapping model for automated skin lesion segmentation and classification. IEEE Trans. Med. Imaging; 2020; 39, pp. 2482-2493. [DOI: https://dx.doi.org/10.1109/TMI.2020.2972964]
95. Gerges, F.; Shih, F.Y. A Convolutional Deep Neural Network Approach for Skin Cancer Detection Using Skin Lesion Images. Int. J. Electr. Comput. Eng.; 2021; 15, pp. 475-478.
96. Mukherjee, S.; Adhikari, A.; Roy, M. Malignant melanoma detection using multi layer preceptron with visually imperceptible features and PCA components from MED-NODE dataset. Int. J. Med. Eng. Inform.; 2020; 12, pp. 151-168. [DOI: https://dx.doi.org/10.1504/IJMEI.2020.106899]
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
The development of abnormal cell growth is caused by different pathological alterations and some genetic disorders. This alteration in skin cells is very dangerous and life-threatening, and its timely identification is very essential for better treatment and safe cure. Therefore, in the present article, an approach is proposed for skin lesions’ segmentation and classification. So, in the proposed segmentation framework, pre-trained Mobilenetv2 is utilised in the act of the back pillar of the DeepLabv3+ model and trained on the optimum parameters that provide significant improvement for infected skin lesions’ segmentation. The multi-classification of the skin lesions is carried out through feature extraction from pre-trained DesneNet201 with N × 1000 dimension, out of which informative features are picked from the Slim Mould Algorithm (SMA) and input to SVM and KNN classifiers. The proposed method provided a mean ROC of 0.95 ± 0.03 on MED-Node, 0.97 ± 0.04 on PH2, 0.98 ± 0.02 on HAM-10000, and 0.97 ± 0.00 on ISIC-2019 datasets.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details



1 Department of Computer Science, COMSATS University Islamabad, Wah Campus, Wah Cantt 47040, Pakistan
2 Department of Computer Science, University of Wah, Wah Cantt 47040, Pakistan
3 National University of Technology (NUTECH), Islamabad 44000, Pakistan
4 Department of Computer Science, Shah Abdul Latif University, Khairpur 66111, Pakistan
5 Department of Applied Data Science, Noroff University College, 4612 Kristiansand, Norway; Artificial Intelligence Research Center (AIRC), Ajman University, Ajman P.O. Box 346, United Arab Emirates; Department of Electrical and Computer Engineering, Lebanese American University, Byblos P.O. Box 13-5053, Lebanon