Academia.eduAcademia.edu

Classification of Skin Cancer with Deep Transfer Learning Method

Computer Science

Skin cancer is a serious health hazard for human society. This disease is developed when the pigments that produce skin color become cancerous. Dermatologists face difficulties in diagnosing skin cancer since many skin cancer colors seem identical. As a result, early diagnosis of lesions (the foundation of skin cancer) is very crucial and beneficial in totally curing skin cancer patients. Significant progress has been made in creating automated methods with the development of artificial intelligence (AI) technologies to aid dermatologists in the identification of skin cancer. The widespread acceptance of AI-powered technologies has enabled the use of a massive collection of photos of lesions and benign sores authorized by histology. This research compares six alternative transfer learning networks (deep networks) for skin cancer classification using the International Skin Imaging Collaboration (ISIC) dataset. DenseNet, Xception, InceptionResNetV2, ResNet50, and MobileNet were the tr...

ISSN,e-ISSN: 2548-1304 Volume:IDAP-2022, pp:202-210, 2022 https://doi.org/ 10.53070/bbd.1172782 Research Paper Journal of Computer Science https://dergipark.org.tr/en/pub/bbd Classification of Skin Cancer with Deep Transfer Learning Method Doaa Khalid Abdulridha Al-Saedi*1 1 , Serkan Savaş2 Department of Electronics and Computer Engineering, Çankırı Karatekin University, Çankırı, Türkiye Department of Computer Engineering, Kırıkkale University, Kırıkkale, Türkiye 2 ([email protected], [email protected]) Received:Sep.08,2022 Accepted:Sep.16,2022 Published:Oct.10,2022 Abstract— Skin cancer is a serious health hazard for human society. This disease is developed when the pigments that produce skin color become cancerous. Dermatologists face difficulties in diagnosing skin cancer since many skin cancer colors seem identical. As a result, early diagnosis of lesions (the foundation of skin cancer) is very crucial and beneficial in totally curing skin cancer patients. Significant progress has been made in creating automated methods with the development of artificial intelligence (AI) technologies to aid dermatologists in the identification of skin cancer. The widespread acceptance of AI-powered technologies has enabled the use of a massive collection of photos of lesions and benign sores authorized by histology. This research compares six alternative transfer learning networks (deep networks) for skin cancer classification using the International Skin Imaging Collaboration (ISIC) dataset. DenseNet, Xception, InceptionResNetV2, ResNet50, and MobileNet were the transfer learning networks employed in the investigation which were successful in different studies recently. To compensate for the imbalance in the ISIC dataset, the photos of classes with low frequencies are augmented. The results show that augmentation is appropriate for the classification success, with high classification accuracies and F-scores with decreased false negatives. With an accuracy rate of 98.35%, modified DenseNet121 was the most successful model against the rest of the transfer learning nets utilized in the study. Keywords : Skin cancer, deep learning, ISIC, transfer learning, DenseNet. 1. Introduction Skin cancer may be divided into two primary categories: melanoma and non-melanoma. Squamous cell carcinoma and basal cell carcinoma are the most prevalent non-melanoma tumors. The 17th most prevalent cancer globally is cutaneous melanoma. It is the 13th and 15th most prevalent cancer, respectively, in men and women (WCRF, 2022). Over the past several decades, there has been an increase in the incidence of both non-melanoma and melanoma skin cancers. Currently, 132,000 cases of melanoma and 2 to 3 million cases of non-melanoma skin cancer are reported annually worldwide (WHO, 2017). In 2020, there were more than 150,000 brand-new instances of cutaneous melanoma. Table 1 displays the overall incidence and rates of melanoma skin cancer worldwide in 2020 (WCRF, 2022). Table 1. Global melanoma skin cancer incidence and rates in 2020 (WCRF, 2022) No Country Number ASR/ 100,000 1 Australia 16,171 36.6 2 New Zealand 2,801 31.6 3 Denmark 2,886 29.7 4 The Netherlands 8,310 27.0 5 Norway 2,567 26.4 6 Sweden 4,266 23.3 202 7 Switzerland 3,357 21.6 8 Germany 31,468 20.5 9 Slovenia 735 19.7 2,090 19.5 324,635 3.4 10 Finland World According to the Table 1, Australia and New Zealand were the countries with the highest rates of cutaneous melanoma in 2020. Modern medical research is attempting to aid dermatologists in their diagnosis without the need for specialized or expensive equipment due to the rise in skin cancer globally. Early identification and prevention of skin cancer are the most effective ways to control it. Skin patches or growths that are new or changing, especially those that appear strange, should be investigated. Any new lesions or increasing changes in the appearance of an existing lesion (size, shape, or color) should be assessed by a physician. Deep learning (DL) technology has allowed for the classification of skin cancer into seven diagnostic groups, including melanocytic nevi, melanoma, benign keratosis-like lesions, basal cell carcinoma, actinic keratosis, vascular lesions, and dermatofibroma. A dermatologist who specializes in the detection of skin cancer often follows a predetermined process that starts with a visual examination of the worrisome lesion, is followed by a dermoscopy, and ends with a biopsy (Haenssle et al., 2018). When compared to relying just on a visual diagnosis today, the efficacy of anticipating a result using artificial intelligence (AI) and deep learning (DL) in diagnostics grows tremendously (Ayoub et al., 2021). Deep convolutional neural networks (DCNNs) process dermoscopic images to identify skin lesions, including all skin cancer lesions, whereas convolutional neural networks (CNNs) may be used for feature selection and object categorization. DNNs are effective for classifying medical images, but they need a lot of training data. High-performance GPUs are used to train a network of DNNs using largescale datasets (Savaş et al., 2019, 2022). DL systems enabled by GPUs have shown superior skin cancer detection than people (Nugroho et al., 2019). From this point of view, in this study, it is aimed to classify skin cancer by transfer learning-based deep learning method. In this way, it is aimed to assist clinics in diagnosis and treatment procedures with early detection of skin cancer. Some studies have been carried out in the literature to diagnose skin cancer using machine learning (ML) and DL algorithms for this reason. These studies are summarized in the second section. In the third section of the research, the material and method used in this study are explained. In the fourth section, the experimental results are explained. In the fifth section, the contributions of the research and its comparative discussion with other researches were realized. 2. Related Works A brand-new method of diagnosing skin cancer based on metaheuristics and DL is proposed by Qiao et al., (2022). In their method, the skin dermoscopy images are first trained using a modified AlexNet that has already been trained using batch normalization layers, and the remaining few layers are then handled by an Extreme Learning Machine (ELM). The Fractional-order Red Fox Optimization (FORFO) Algorithm, a recently updated metaheuristic, is used to increase the effectiveness of the ELM network. The proposed approach has a 97.14% overall accuracy. On the other hand, to examine polarization speckle images obtained from the main categories of malignant and benign lesions, Wang et al., (2021) used DL and classical ML algorithms. A collection of 122 cancerous and 196 benign skin lesion speckle images was improved using patch cropping for DL, which was an advantageous strategy given the patterns' statistical homogeneity. In the easier classification job of distinguishing malignant melanoma from benign melanoma, the ML technique achieved high 90% accuracy. However, in the general classification test of malignant and benign tumors, ResNet, their selected DL architecture, got the best result of 82% diagnosis accuracy. Another novel model based on the auto-encoder, spiking, and CNN is proposed by Toğaçar et al., (2021). The dataset used is an open-access dataset called the International Skin Imaging Collaboration (ISIC) skin cancer, which included 1800 benign and 1497 malignant tumor images. The dataset is reconstructed using the auto-encoder model in the proposed method. The MobileNetV2 model, which comprises of residual blocks and spiking networks, was used to train and classify the original dataset and structured dataset. The study's classification success rate was 95.27%. A pipeline technique for melanoma diagnosis utilizing dermoscopy images was investigated by Fu et al., (2022). After picture preprocessing, the region of interest (RoI) is segmented using a technique based on the Kernel Fuzzy C-means algorithm. The segmented area's key attributes were then optimally retrieved and chosen utilizing 203 a new optimized approach. An ideal classification approach based on multi-layer perceptron is offered for the final diagnosis. The feature selection and classification are optimized using a newly created Red Fox Optimization (DRFO) algorithm. The accuracy of 90.5% indicated that the outcome was trustworthy. Asymmetry, Border, Color, and Diameter (ABCD) criteria are used to diagnose skin cancer automatically in the research of Senan & Jadhav, (2021). The PH2 standard dataset consisting three classes of skin diseases (Atypical Nevi, Melanoma, and Common Nevus) was utilized to test the proposed system. The proposed mechanism consists of two stages. The Gaussian filter method is used to enhance the photos and remove undesirable pixels in the first stage, which uses preprocessing to improve image quality. The contour approach was used to extract the RoI from dermoscopy images. Morphology is also being considered for improving the quality of skin lesions. The second stage involves implementing ABCD rules to extract relevant characteristics and the total accuracy of the study is 84%. Tumpa & Kabir, (2021) aimed to create a neural network that can accurately detect and classify Melanoma in their study. Their procedure began with dermoscopic images being preprocessed to remove hairs using the Maximum Gradient Intensity technique, as well as image enhancement. To separate skin lesions from the images, a segmentation technique based on the Otsu Thresholding algorithm is used. The segmented images are then utilized to calculate many features such as ABCD, GLCM, and LBP, which used to train a neural network. On the combined dataset of ISIC archive and the PH2 dermoscopic image database, the network achieved an accuracy of 97.7%. Ali et al. (2021) introduced a DCNN model for the precise categorization of benign and malignant skin lesions. They first apply a filter or kernel to reduce noise and artefacts, after which they normalize the input images and extract features that help with effective classification, they then increase the number of images by using data augmentation, which increases classification rate accuracy, and finally they add more images. The DCNN model is contrasted with AlexNet, ResNet, VGG-16, DenseNet, MobileNet, and other transfer learning techniques. The HAM10000 dataset was used to test the model, and the results showed that it had the greatest training and testing accuracy, respectively, of 93.16% and 91.93%. Sharpening and smoothening filters, as well as enhancing techniques, are employed in the preprocessing step to remove noise in the study of Dabhi et al., (2021). After these processes, Otsu segmentation was employed to detect skin cancer in its early stages. Finally, a back propagated based artificial neural network (BP-ANN) was created for classification of skin cancer with the spatially grey level dependency matrix (SGLD) characteristics to archive the system's optimum efficiency in this study. As a result, the research work may be used to efficiently classify Benign and Melanoma skin cancers, with a total accuracy of 99.13%. Hosny et al. (2019), used transfer learning and a pre-trained DL network (AlexNet) in their study. They classified three distinct lesions in the study with using fine-tuning. In the study, data augmentation techniques were also used to balance the number of dataset. The accuracy of the model on the PH2 dataset was 98.61%. Researchers also used other metrics such as sensitivity (98.33%), specificity (98.93%), and precision (97.73%). Another study using transfer learning is proposed by Demir et al., (2019). Their dataset consists of 3297 images in total and they used ResNet-101 and Inception-v3 DL architectures for the classification. Accuracy rates of 84.09% and 87.42% are get in ResNet-101 and Inception-v3 architectures, respectively. Transfer learning has developed out of DL research nowadays. Among the models that have demonstrated their efficiency in various image classifications, it is crucial to identify the dominant model or models to be employed in biomedical image processing and classification operations. It is preferable to develop general models that may be applied to other research and employed in this instance rather than subject and situation-specific models. It's crucial to find models that produce consistent and trustworthy findings across a variety of imaging modalities, especially given the sparse amount of images in the area of medicine and the complexity of image capture and processing (Savaş, 2022). For this reason, pre-trained deep architectures were used in this work to carry out the classification procedure. 3. Material and Method 3.1. Dataset: The dataset of the study is made up of 64,000 images by The International Skin Imaging Collaboration of cancerous and noncancerous skin diseases (ISIC, 2022).The dataset was splitted into three sets such as training set with 50,000 images, 10,000 images for validation, and 4000 images for testing. All of the images were put in order based on how they were classified with ISIC, and each subset had the same number of images, except for melanomas and moles, whose images are slightly more common. The dataset consists of two categories benign and malignent. Figure 1 represents a sample of the dataset. 204 \ Figure 1. Sample of dataset 3.2. Transfer Learning Nets: Transfer learning is a type of ML in which a model built for one task is used for different ones. It's usually used when researcher doesn't have enough data to train on. However, the problem with the data can be fixed by adding more data or data augmentation. The main reason for need of transfer learning is that Melanoma and Benign lesions look a lot alike, making it hard to tell them apart and put them in different groups. Transfer learning is also better at classifying lesions that look the same, which makes it the first choice. Transfer learning nets are trained on large datasets, and then their model weights are saved. For the application for a different dataset, the last few layers of pre-trained nets are changed in transfer learning. In this paper, DenseNet, Xception, InceptionResNetV2, ResNet50, and MobileNet are the models selected for comparison of the results. But in this case, we do not only used the saved weights, but also retrained the models on our dataset so that the network layers could tell the difference between the two types of lesions more accurately. We used six transfer learning nets to train the models on the skin lesion dataset and then looked at their predictions. For the evaluation of the performances, accuracy, loss, and confusion matrices results were plotted. Then, the accuracy of all of these learning nets were compared to found the one that could find all of the lesions with the most accuracy. 3.2.1. InceptionResNet: Simple Inception modules are not employed in this net, but rather the residual form of Inception Nets (Szegedy et al., 2016). In this network, each Inception block is followed by a filter-expansion layer (1x1 convolution without activation), which is used to increase the dimensionality of the filter bank before the addition to correspond to the depth of the input. The Inception-v4 network's computational cost is the same as that of the Inception-ResNet-v2 network.. The batch normalization employed in Inception-ResNetv2 is different from that used in non-residual Inception versions. In that it is applied only on top of the standard layers and not the summations (Szegedy et al., 2016). 3.2.2. Xception: An addition to the Inception architecture is the Xception architecture. Convolutions that can be separated based on depth are used in place of the conventional Inception modules. It does not divide the input data; instead, it maps the spatial correlations for each output channel independently. The next step is 1𝑥1 depth wise convolution, which allows the Xception net to record cross-channel correlation. It substantially excels Inception V3 on larger data while just somewhat outperforming it on smaller data (Chollet, 2016). 3.2.3. MobileNetV2: Similar to the Xception net, this net also uses depth-wise separable connections. Every input channel is subjected to a single filter as part of the depth wise convolution for MobileNets. The pointwise convolution employing a 1x1 convolution then combines the results of the depth-wise convolution. A typical convolution layer filters and blends inputs in one step to produce a new set of outputs. The depth-wise separable convolution separates this into two layers: a layer for filtering and a layer for combining. This factorization leads to a large reduction in computation and model size. Applications for mobile and embedded vision benefit especially from MobileNet. In comparison to other systems, it is simpler and has fewer parameters (Howard et al., 2017). 3.2.4. ResNet50: These utilise skip connections and are the deeper CNNs. These residual blocks significantly reduce total parameters while also resolving gradient deterioration. Two simple design tenets are followed by the ResNet (Residual Networks) architecture. In addition, when the size of the feature map is cut in half, the number of filters doubles. Layers have the same number of filters for the same output map size. Batch normalization is performed 205 before the ReLU activation function and after each convolution layer. The shortcut is utilized if the input and output are the same size. The projection shortcut is applied as the dimensions expand (He et al., 2015). 3.2.5. DenseNet121: In this net, because all levels, including those in the same dense block and transition layers, spread their weights over several inputs, deeper layers can utilize characteristics that were collected earlier. 120 Convolutions and 4 AvgPool make up DenseNet-121 (He et al., 2015). 3.2.6. EfficientNet: EfficientNet uses a technique known as compound coefficient to quickly and simply scale up models. Compound scaling, as opposed to using a random scaling scheme, equally scales each dimension using a predetermined set of scaling factors (Tan & Le, 2019). 3.3. Proposed Methodology: We describe how skin lesions are classified in this section. The proposed approach is shown diagrammatically in Figure 2. In order to tackle more effectively computer vision challenges, researchers have constructed increasingly sophisticated CNNs in recent years. In this study, we used TensorFlow, a DL framework built by Google, to train six DCNNs to categorize skin cancer images. Figure 2. Block diagram of the study 3.3.1. Preprocessing: In this step, the dataset were converted to 224 x 224 x 3 dimensions to be used in the pre-trained model, and also implementing the preprocess input function for every transfer learning algorithms. This function apply image scaling. The next step in this phase contains feature extraction. Pre-trained models were used to extract features through transfer learning. This entails identifying key elements in an image and extrapolating information from them. To create a model, several CNNs are piled on top of one another. 3.3.2. Implementation of Transfer Learning Models: In this step, every kind of pre-trained models was implemented separately without the output layer from the original model. We added three layers in the end of each model, the first and the second layers are same in their components. Every layer of these consists of 128 neuron with ReLU activation function. The output layer consists of 2 classes which is equal to the number of classes (benign, Malignent). 4. Experimental Results and Discussion In this section, we discuss our research and contrast the trained models' diagnostic acuity with that of a recognized panel of dermatologists. In this study, the metrics were used are precision, recall (sensitivity), F1-score, specificity, and accuracy to determine the evaluation and outcomes of trained models as mentioned in the five equations below. In this case, binary classification is used to comprehend the model. The class for positive points 206 is positive, while the class for negative points is negative. The models' efficacy was calculated using the following equations, and they were then compared. 𝑇𝑃 Precision = (𝑇𝑃 + 𝐹𝑃) F1-Score = (2 × 𝑃𝑅𝐸 × 𝑅𝐸𝐶) Recall (Sensitivity) = TP (2) (TP + FN) (3) (𝑃𝑅𝐸 + 𝑅𝐸𝐶) Specificity = Accuracy = (1) 𝑇𝑁 (4) (TN + FP) (TP + TN ) (5) (𝑇𝑃 + 𝐹𝑃 + 𝑇𝑁 + 𝐹𝑁) TP stands for the fraction of correctly categorized positive instances, TN for correctly labelled negative cases, FP for wrongly labelled positive cases, and FN for incorrectly labelled negative cases in the equations. Additionally, the F1-score result, which is the harmonic mean of accuracy and recall, is presented since the distribution of sample sizes among databases is significantly imbalanced. Table 2 provides a summary of the comparison of six models by presenting average values for variables such as accuracy, precision, F1-score, recall (sensitivity), and specificity. The precision demonstrates a classifier's ability to avoid classifying a negative sample as positive. The F1-score peaks close to 1 and vice versa. This demonstrates that, as compared to previous models, DenseNet121 has produced overall better results for the categorization of dermoscopy skin cancer images. Table 2. Results comparison between proposed models Accuracy (%) Precision (%) F1-Score (%) ResNet50 98.1 97.8 98.1 Recall (Sensitivity) (%) 98.45 Xception 99.1 98 99 99.6 98.6 MobilNetV2 99.4 99.4 99.4 99.5 99.4 InceptionResNetV2 98.45 98.6 98.4 98.2 98.7 EfficientNet 98.5 98.7 98.5 98.2 98.75 DenseNet121 99.6 99.7 99.5 99.4 99.7 Model Specificity (%) 97.8 As shown in the above Table 2, the DenseNet121 is the best model for skin cancer detection with 99.6% accuracy, 99.7% precision, 99.5% F1-score, 99.4% recall (sensitivity), and 99.7% specificity. Table 3 represents a comparison between the proposed DenseNet121 with state of arts. It is appear that the proposed model overcome all other models. Table 3. Comparison between the proposed model and the state of arts Study Skin Cancer Diagnosis Classifier and training algorithm Dataset Results A. A. Ali & AlMarzouqi, (2017) Harangi et al., (2018) Benign/malignant ISIC 2016 dataset ISIC 2017 dataset Acc: 81.6% Malignant melanoma/nevus/SK LightNet (deep learning framework), CNN as single neural-net architecture Kalouche, (2016) Benign/melanoma VGG-16 and CNN ISIC dataset Acc: 78% Hasan et al., (2019) Benign/malignant CNN ISIC dataset Acc: 89.5% 207 Acc: 83.8% (Rezvantalab et al., 2018) Lipoma/fibroma/ sclerosis/melanoma Deep region-based CNN and fuzzy C means clustering ISIC dataset Namozov et al., (2018) nevus/AK/benign, keratosis/vascular lesion/dermatofibrom Benign/malignant CNN model with keratosis/vascular LeNet approach Modified DenseNet121 ISIC dataset Proposed work ISIC dataset Acc:94.8% Sensitivity: 97.81% Specificity: 94.17% F1_score: 95.89% Acc: 95.86% Acc: 99.6% Precision: 99.7% F1_Score: 99.5% Recall: 99.4% Specificity: 99.7% Figure 3 represents the progress of accuracy with increasing the number of epochs and represents the relation between the training loss and the validation loss. The final loss is very small which reaches to near of zero. Figure 3.Training and validation relations with accuracy and loss As can be seen in Figure 3, the training accuracy of the proposed model was above 99%, while the validation rate reached a high success rate of 98.35%. Loss is another important performance metric. A loss value close to zero indicates that the performance of the study is high. As seen in Figure 3, the training Loss value decreased below a very small value of 0.025, especially after the 10th epoch. The validation Loss value was generally around 0.075 in the study, which can be considered as very low. 5. Conclusion and Future Work In this study, skin cancer classification was made using pre-trained models. Skin cancer is one of the most common types of cancer in the world, and if not detected early, it can lead to fatal outcomes. For this reason, the need to support clinical processes has arisen with the state-of-art technology studies. In this study, the suggested DenseNet121 model outperforms existing transfer learning models in terms of classification accuracy. The suggested technique has the capacity to categorize benign and malignant skin lesions by substituting a sigmoid for the output activation layer for binary classification. The best accuracy rate that can be achieved with the best model is 99.6% in the study. In addition, the validation accuracy rate of this model has reached a high rate of 98.35%. In the next stages of this study, research will be carried out to obtain similar success rates in other datasets by using different feature selection techniques together with different pre-trained models. This research will be carried out by expanding the dataset and the results will be announced. 208 References Ali, A. A., & Al-Marzouqi, H. (2017). Melanoma detection using regular convolutional neural networks. 2017 International Conference on Electrical and Computing Technologies and Applications, ICECTA 2017, 2018January, 1–5. https://doi.org/10.1109/ICECTA.2017.8252041 Ali, M. S., Miah, M. S., Haque, J., Rahman, M. M., & Islam, M. K. (2021). An enhanced technique of skin cancer classification using deep convolutional neural network with transfer learning models. Machine Learning with Applications, 5, 100036. https://doi.org/10.1016/J.MLWA.2021.100036 Ayoub, A., Mahboob, K., Javed, A. R., Rizwan, M., Gadekallu, T. R., Abidi, M. H., & Alkahtani, M. (2021). Classification and categorization of COVID-19 outbreak in Pakistan. Computers, Materials and Continua, 69(1), 1253–1269. https://doi.org/10.32604/CMC.2021.015655 Chollet, F. (2016). Xception: Deep Learning with Depthwise Separable Convolutions. Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, 2017-January, 1800–1807. https://doi.org/10.48550/arxiv.1610.02357 Dabhi, V. M., Kashyap, S. S., Nithin, G., Vamshi, A. C., & Krishna, G. A. (2021). Detection and Classification of Skin Cancer using Back Propagated Artificial Neural Networks. JES - Journal of Engineering Sciences, 12(06), 686–693. Demir, A., Yilmaz, F., & Kose, O. (2019). Early detection of skin cancer using deep learning architectures: Resnet101 and inception-v3. TIPTEKNO 2019 - Tip Teknolojileri Kongresi, 2019-January. https://doi.org/10.1109/TIPTEKNO47231.2019.8972045 Fu, Z., An, J., Yang, Q., Yuan, H., Sun, Y., & Ebrahimian, H. (2022). Skin cancer detection using Kernel Fuzzy C-means and Developed Red Fox Optimization algorithm. Biomedical Signal Processing and Control, 71, 103160. https://doi.org/10.1016/J.BSPC.2021.103160 Haenssle, H. A., Fink, C., Schneiderbauer, R., Toberer, F., Buhl, T., Blum, A., Kalloo, A., ben Hadj Hassen, A., Thomas, L., Enk, A., Uhlmann, L., Alt, C., Arenbergerova, M., Bakos, R., Baltzer, A., Bertlich, I., Blum, A., Bokor-Billmann, T., Bowling, J., … Zalaudek, I. (2018). Man against machine: diagnostic performance of a deep learning convolutional neural network for dermoscopic melanoma recognition in comparison to 58 dermatologists. Annals of Oncology, 29(8), 1836–1842. https://doi.org/10.1093/ANNONC/MDY166 Harangi, B., Baran, A., & Hajdu, A. (2018). Classification of Skin Lesions Using An Ensemble of Deep Neural Networks. Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBS, 2018-July, 2575–2578. https://doi.org/10.1109/EMBC.2018.8512800 Hasan, M., Barman, S. das, Islam, S., & Reza, A. W. (2019). Skin cancer detection using convolutional neural network. ACM International Conference Proceeding Series, 254–258. https://doi.org/10.1145/3330482.3330525 He, K., Zhang, X., Ren, S., & Sun, J. (2015). Deep Residual Learning for Image Recognition. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2016-December, 770–778. https://doi.org/10.48550/arxiv.1512.03385 Hosny, K. M., Kassem, M. A., & Foaud, M. M. (2019). Skin Cancer Classification using Deep Learning and Transfer Learning. 2018 9th Cairo International Biomedical Engineering Conference, CIBEC 2018 Proceedings, 90–93. https://doi.org/10.1109/CIBEC.2018.8641762 Howard, A. G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., Andreetto, M., & Adam, H. (2017). MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. Arxiv. https://doi.org/10.48550/arxiv.1704.04861 ISIC. (2022). ISIC Archive. The International Skin Imaging Collaboration. https://www.isicarchive.com/#!/topWithHeader/wideContentTop/main Kalouche, S. (2016). Vision-Based Classification of Skin Cancer using Deep Learning. Semantic Scholar. https://www.semanticscholar.org/paper/Vision-Based-Classification-of-Skin-Cancer-usingKalouche/b57ba909756462d812dc20fca157b3972bc1f533 Namozov, A., Ergashev, D., & Cho, Y. I. (2018). Adaptive activation functions for skin lesion classification using deep neural networks. Proceedings - 2018 Joint 10th International Conference on Soft Computing and Intelligent Systems and 19th International Symposium on Advanced Intelligent Systems, SCIS-ISIS 2018, 232– 235. https://doi.org/10.1109/SCIS-ISIS.2018.00048 Nugroho, A. A., Slamet, I., & Sugiyanto. (2019). Skins cancer identification system of HAMl0000 skin cancer dataset using convolutional neural network. AIP Conference Proceedings, 2202(1), 020039. https://doi.org/10.1063/1.5141652 Qiao, L., Xue, Y., Tang, W., & Jimenez, G. (2022). Skin cancer diagnosis based on a hybrid AlexNet/extreme learning machine optimized by Fractional-order Red Fox Optimization algorithm. Proceedings of the Institution of Mechanical Engineers. Part H, Journal of Engineering in Medicine. https://doi.org/10.1177/09544119221075941 209 Rezvantalab, A., Safigholi, H., & Karimijeshni, S. (2018). Dermatologist level dermoscopy skin cancer classification using different deep learning convolutional neural networks algorithms. ArXiv Preprint ArXiv:1810.10348. Savaş, S. (2022). Detecting the Stages of Alzheimer’s Disease with Pre-trained Deep Learning Architectures. Arabian Journal for Science and Engineering, 47(2), 2201–2218. https://doi.org/10.1007/s13369-021-061313 Savaş, S., Topaloğlu, N., Kazcı, Ö., & Koşar, P. N. (2019). Performance Comparison of Carotid Artery Intima Media Thickness Classification by Deep Learning Methods. International Congress on Human-Computer Interaction, Optimization and Robotic Applications Proceedings, 4(5), 125–131. https://doi.org/10.36287/setsci.4.5.025 Savaş, S., Topaloğlu, N., Kazcı, Ö., & Koşar, P. N. (2022). Comparison of Deep Learning Models in Carotid Artery Intima-Media Thickness Ultrasound Images: CAIMTUSNet. Bilişim Teknolojileri Dergisi, 15(1), 1– 12. Senan, E. M., & Jadhav, M. E. (2021). Analysis of dermoscopy images by using ABCD rule for early detection of skin cancer. Global Transitions Proceedings, 2(1), 1–7. https://doi.org/10.1016/J.GLTP.2021.01.001 Szegedy, C., Ioffe, S., Vanhoucke, V., & Alemi, A. A. (2016). Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. 31st AAAI Conference on Artificial Intelligence, AAAI 2017, 4278–4284. https://doi.org/10.48550/arxiv.1602.07261 Tan, M., & Le, Q. v. (2019). EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. 36th International Conference on Machine Learning, ICML 2019, 2019-June, 10691–10700. https://doi.org/10.48550/arxiv.1905.11946 Toğaçar, M., Cömert, Z., & Ergen, B. (2021). Intelligent skin cancer detection applying autoencoder, MobileNetV2 and spiking neural networks. Chaos, Solitons & Fractals, 144, 110714. https://doi.org/10.1016/J.CHAOS.2021.110714 Tumpa, P. P., & Kabir, M. A. (2021). An artificial neural network based detection and classification of melanoma skin cancer using hybrid texture features. Sensors International, 2, 100128. https://doi.org/10.1016/J.SINTL.2021.100128 Wang, Y., Louie, D. C., Cai, J., Tchvialeva, L., Lui, H., Jane Wang, Z., & Lee, T. K. (2021). Deep learning enhances polarization speckle for in vivo skin cancer detection. Optics & Laser Technology, 140, 107006. https://doi.org/10.1016/J.OPTLASTEC.2021.107006 WCRF. (2022). Skin cancer statistics. World Cancer Research Fund International. https://www.wcrf.org/cancertrends/skin-cancer-statistics/ WHO. (2017). Radiation: Ultraviolet (UV) radiation and skin cancer. World Health Organization. https://www.who.int/news-room/questions-and-answers/item/radiation-ultraviolet-(uv)-radiation-and-skincancer 210