Many skin lesion segmentation algorithms have been proposed by researchers in the past. The algorithms are based on extraction of features for each image from an image dataset and accordingly creating a segmentation model. The model is then used for segmentation of query images and the accuracy and dice score of segmentation is analysed. Some of the related works are as follow.
The DullRazor algorithm by (Lee,1997) has been one of the first solution to the hair removal problem. Using thresholds and morphological processes, this program eliminates hairs from the lesion. (Ali,2022) offered a study on skin cancer picture hair removal and lesion segmentation. They stated that hair removal is critical because the CNN model would find correlations between the noise and the aim (classification of skin cancer). If the noise in the image is not removed, the CNN must learn to ignore it using gradient descent and a big image dataset. In another study (Alom,2020), a Nabla N-Network has been proposed with enhanced fusion units in decoder of the segmentation network. The segmented features are then fed into an Inception Recurrent Residual Convolutional Neural Network (IRRCNN) for final image classification. This study achieved an accuracy of 87% for ISIC-2018 dataset. Anjum et al. first used a YOLOv2 model based on Open Neural networks and SqueezeNet Model for localization of skin lesions.A pixel classification layer has been used for computing the overlapping regions between segmented and ground truth images. Finally, the classification is achieved using Res-Net-18 and Ant colony optimization with Global Accuracy of 0.93, 0.95 on ISBI 2017, and ISBI 2018 respectively(Anjum,2020).Using Mask R_CNN architecture, Jojoa Acosta et al. first clipped the lesions from the skin picture. After that, they classified the clipped lesions using the pre-trained ResNet152 architecture. The ISIC 2017 dataset produced a 90.4% accuracy from this model(Acosta, 2021).In a different study, Malibari et al. proposed a deep model that used deep convolutional neural networks for skin classification and detection. They used the UNET architecture for segmentation when pre-trained Squeezenet architecture-based whale optimization was applied for the classification step. They were able to classify melanoma with 99% accuracy using the ISIC 2019 dataset [10] (Malibari, 2022). A deep convolutional network was employed in a hybrid structure by Jayapriya and Jacob. The lesions were segmented using fully trained pre-convolutional networks, and a method was put forth to extract features from the lesions and classify them using SVM. Experimental studies found that the accuracy of this hybrid technique was 88.92% for ISIC 2016 and 85.3% for ISIC 2017 (Jayapriya, 2019).
Authors in(Monika,2020)proffered a method for the identification and classification of distinct forms of skin cancer by using Multi-class Support Vector Machine (MSVM). The authors used the ISIC 2019 challenge datasetand were able to achieve an accuracy of about 96.25%. In (Vidya2020), authorsused different machine learning approaches, such as SVM, KNN, and Naive Bayes classifier, to categorise skin lesions between benign and melanoma. Using SVM classifiers, a classification accuracy of 97.8% and an area under the curve of 0.94 were attained. Additionally, by using KNN, the Sensitivity and Specificity were both 86.2% and 85%, respectively.Another research work in (Vijayalakshmi,2019) offered a method for removing hair, shading, and glares from images during the pre-processing stage after that segmentation and feature extraction are performed. They trained their model on the back propagation technique (feed-forward neural network), SVM, and CNN in the final step of their process. The models were amalgamated (combined) utilizing image processing tools after classification, yielding an accuracy of 85% on the ISIC dataset.
In (Kumar,2022), images are pre-processed with a Gaussian filter and Region of Interest (ROI) extraction to weed out noise and mine important parts. Integrating U-Net and RP-Net allows for segmentation, and the output from both models is combined using the Jaccard similarity-based fusion model. In order to increase the effectiveness of the detection process, the data augmentation is processed. SqueezeNet, which was trained using the suggested Aquila Whale Optimization (AWO) method, is then used to identify skin cancer. The newly created AWO method was created by fusing the Aquila Optimizer (AO) and Whale Optimization Algorithm (WOA). With the greatest testing accuracy of 92.5%, sensitivity of 92.1%, and specificity of 91.7%, the created AWO-based SqueezeNet exceeded.(Araújo,2022) proposed a U-Net model along with Link-Net for segmentation of skin cancer by using three datasets viz. pH2, ISIC 2018 and DermiS. The proposed model has been compared in terms of Dice Coefficient and obtained an average Dice of 0.923 for PH2 dataset, Dice value of 0.893 for ISIC 2018, and Dice value of 0.879 forDermISdataset.In (Ries,2022), authors proposed U-Net Model for segmentation of skin cancer images and an InsiNet Model for the classification of melanoma images. The model has been used over ISIC and HAM 10K data set of skin cancer and is compared with similar techniques in terms of accuracy. The results reveal that the created InSiNet architecture surpasses the other techniques, with 94.59%, 91.89%, and 90.54% accuracy in the ISIC 2018, 2019, and 2020 datasets, respectively. (Akyel,2022) proposed a LinkNet-B7 model based on the EfficientNetB7 encoder. Here ISIC and pH2 datasets have been used. The Efficient-Net model used is based on Mobile Inverted Bottleneck Convolution. Also this encoder is then fused with Res-Net Model for final segmentation. This model’s training accuracy for noise removal and lesion segmentation was calculated to be 95.72% and 97.80%, respectively. A fully convolution encoder decoder network (FCEDN) with hyper-parameter optimization is suggested in (Muhakud,2022) for the segmentation of dermatoscopy images. The unique Exponential Neighbourhood Grey Wolf Optimization (EN-GWO) algorithm is used to optimise the hyper-parameters of FCEDN rather than manually setting them, with an emphasis on the right balance between exploration and exploitation.The EN-GWO has been validated by the authors against four variations of GWO, GA, and PSO based hyper-parameter optimization approaches using the International Skin Imaging Collaboration (ISIC) 2016 and ISIC 2017 datasets. The proposed model can segment skin cancer images for the ISIC 2016 and ISIC 2017 datasets with a Jaccard coefficient of 96.41%, 86.85%, Dice coefficient of 98.48%, 87.23%, and accuracy of 98.32%, 95.25%, respectively. Authors in(Goyal,2017) proposed an improved fully convolutional network with the help of transfer learning and a hybrid loss function to perform multi-class semantic segmentation. A custom hybrid loss function has been designed for multi-class segmentation of skin lesion images. The results showed that the two-tier level transfer learning FCN-8s achieved the best overall result, with a Dice score of 78.5% in a naevus category, 65.3% in melanoma, and 55.7% in seborrhoeic keratosis in multi-class segmentation, and an accuracy of 84.62% in melanoma recognition in lesion diagnosis. In another study (Dimša,2021), authors used three types of U-Nets namely, Base U-Net, U-Net + + and MultiResU-Net models for the segmentation of skin lesion images. In U-Net + + skip connections are redesigned so that feature maps undergo a dense convolution block. In Multi-ResU-Net, 3x3 and 7x7 convolutions are carried in parallel to 5x5 convolutions and the output of all three types of convolutions are thereby concatenated. In terms of Dice co-efficient, MultiResU-Net outperformed the rest two types of U-Net Models.
Dong et al. (Dong,2021) introduced FAC-Net, a deep learning model. They attained a 91.19% dice coefficient using the ISIC 2018 dataset. Unlike the U-Net architecture, FAC-Net employs upsampling techniques. Among the various modified LinkNet architectures, one notable example is D-LinkNet. D-LinkNet utilizes Res-Net34 as the encoder and incorporates a middle block containing dilated convolutional layers. In comparison to LinkNet34 (Zhou,2018), D-LinkNet achieved a 2% higher accuracy rate.Another related study by Xiong et al. (Xiong2021) presented Dp-LinkNet, which shares similarities with D-LinkNet but utilizes a different center block. They were able to achieve a 0.9% higher accuracy rate than D-LinkNet, albeit with an increase in training time. Malik et al. (Malik,2022) proposed a signet-based model for lesion segmentation. To address hair noise, they applied the Dullrazor algorithm. However, the study revealed that Dullrazor was not adequate for handling thin hairs. Despite this, they achieved a dice accuracy of 88.43%, demonstrating that hair removal enhances accuracy.Hasan et al. (Hasan2020) in 2020 introduce the Dermoscopic Skin Network (DSNet), an autonomous semantic segmentation network for skin lesion segmentation, along with a new loss function that combines an intersection over-union and a binary cross-entropy. In the studies they did, their proposed loss function produced greater true positive rates and played a significant impact in semantic segmentation.Meanwhile, Bagheri et al. (Bagheri2021) proposed a mask R-CNN-based model in another study, achieving a notable 89.83% dice accuracy on the PH2 dataset.While Al-masni et al. (Al-Masni,2018) suggest a revolutionary segmentation methodology in 2020. In order to increase pixel-wise segmentation performance, they employed Full resolution Convolutional Networks (FrCN), which learn the full resolution characteristics of each individual pixel of the input data without the need for pre- or post-processing steps. In the datasets that were tested, they obtained high evaluation metrics values.