Next Article in Journal
A Novel Paradigm for Non-Invasive Prenatal Genetic Screening: Trophoblast Retrieval and Isolation from the Cervix (TRIC)
Previous Article in Journal
Diagnostic Value of Dynamic 18F-Fluorodeoxyglucose Positron Emission Tomography-Computed Tomography (18F-FDG PET-CT) in Cervical Lymph Node Metastasis of Nasopharyngeal Cancer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Assist-Dermo: A Lightweight Separable Vision Transformer Model for Multiclass Skin Lesion Classification

1
College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh 11432, Saudi Arabia
2
Department of Computer Science, Quaid-i-Azam University, Islamabad 44000, Pakistan
3
Department of Electrical Engineering, Benha Faculty of Engineering, Benha University, Qalubia, Benha 13518, Egypt
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(15), 2531; https://doi.org/10.3390/diagnostics13152531
Submission received: 23 June 2023 / Revised: 22 July 2023 / Accepted: 26 July 2023 / Published: 29 July 2023
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
A dermatologist-like automatic classification system is developed in this paper to recognize nine different classes of pigmented skin lesions (PSLs), using a separable vision transformer (SVT) technique to assist clinical experts in early skin cancer detection. In the past, researchers have developed a few systems to recognize nine classes of PSLs. However, they often require enormous computations to achieve high performance, which is burdensome to deploy on resource-constrained devices. In this paper, a new approach to designing SVT architecture is developed based on SqueezeNet and depthwise separable CNN models. The primary goal is to find a deep learning architecture with few parameters that has comparable accuracy to state-of-the-art (SOTA) architectures. This paper modifies the SqueezeNet design for improved runtime performance by utilizing depthwise separable convolutions rather than simple conventional units. To develop this Assist-Dermo system, a data augmentation technique is applied to control the PSL imbalance problem. Next, a pre-processing step is integrated to select the most dominant region and then enhance the lesion patterns in a perceptual-oriented color space. Afterwards, the Assist-Dermo system is designed to improve efficacy and performance with several layers and multiple filter sizes but fewer filters and parameters. For the training and evaluation of Assist-Dermo models, a set of PSL images is collected from different online data sources such as Ph2, ISBI-2017, HAM10000, and ISIC to recognize nine classes of PSLs. On the chosen dataset, it achieves an accuracy (ACC) of 95.6%, a sensitivity (SE) of 96.7%, a specificity (SP) of 95%, and an area under the curve (AUC) of 0.95. The experimental results show that the suggested Assist-Dermo technique outperformed SOTA algorithms when recognizing nine classes of PSLs. The Assist-Dermo system performed better than other competitive systems and can support dermatologists in the diagnosis of a wide variety of PSLs through dermoscopy. The Assist-Dermo model code is freely available on GitHub for the scientific community.

1. Introduction

Skin cancer is becoming more widespread in the Western world, with significant ramifications for both general skincare and the availability of dermatological treatments. Day after day, about 99,780 individuals in the United States are identified with melanoma or skin cancer. Among them, two or more are likely to die per hour. Skin cancer affects more individuals in the United States each year than all other cancers combined [1]. Europe accounts for 9% of the global population yet bears 25% of the worldwide cancer cases. If tumors are recognized and treated early, cancer mortality can be considerably decreased. Thus, it is crucial to devote research resources to implementing systems for primary cancer recognition. Late-stage melanomas have a poor prognosis, like other malignancies. However, an early-identified melanoma can be treated by a simple resection. Early diagnosis and treatment can preserve the lives of far more than 95% of patients. These numbers are why so much work has been invested in evolving novel imaging approaches that improve the presentation of skin lesions and automated methods to detect melanomas. Human interpreting is laborious and susceptible to faults [2]. Artificial intelligence techniques offer great assistance in skin blemish identification.
Due to the weak contrast of skin carious lesions, the wide intra-class variance of skin cancers, and the high level of likeness among cancerous and non-cancerous lesions, the presence of numerous defects in the image makes automated pigmented skin lesions identification difficult. Often, there is a relatively weak contrast between healthy and affected skin. Additional factors comprise differences in skin color, skin abnormalities, the occurrence of defects (such as hairs, dark spots, foams, rulers, and date markers), uneven lighting and outer black circles, the lesion’s precise position, but more crucially, the lesion’s individual changes regarding color, roughness, form, size, and arrangement within the frame of the image. When developing robust lesion separation algorithms, researchers should consider these aspects. The majority of these factors can have a minimal impact on lesion segmentation with the proper preprocessing steps [3,4]. Nine different forms of pigmented skin lesions are represented visually in Figure 1.
Recently, researchers worldwide have successfully utilized deep learning for different visual tasks [5,6]. Moreover, deep learning in dermatology is mainly used for lesion-specific diagnostic processes, identifying suspicious lesions among several benign lesions, and monitoring lesion progression over time. Skin cancer categorization has benefited significantly from growing study attention since it is amenable to visual shape identification. Research articles demonstrate that DL-based dermoscopic classifiers can enhance skin cancer diagnosis or are better than human specialists [7].

1.1. Research Highlights

  • We used cross validation to test our model and the results showed that it worked much better than the methods already being used.
  • This study proposes a new classification model (Assist-Dermo) to recognize multiclass PSLs.
  • A new preprocessing step is integrated into the perceptual-oriented color space to enhance contrast and adjust the light.
  • The proposed Assist-Dermo model is constructed with many layers and various filter sizes but fewer filters, and these parameters are selected by using lightweight SqueezeNet on a depthwise separable CNN.
  • It is assessed using experimental findings from the many datasets that were gathered with regards to sensitivity, specificity, and other metrics.
  • The Assist-Dermo can reduce overfitting because the dense connection better protects against the overfitting problem, especially when learning from small amounts of data.

1.2. Research Outline

The remainder of the article is divided into the following sections: a survey of the latest research studies regarding automatic recognition of skin tumors in Section 2. The applied methodology including employed datasets the developed DL architecture is illustrated in Section 3. The findings of the suggested skin lesions DL model along with analysis of these results are demonstrated in Section 4. Lastly, conclusions and recommendations for further research are offered in Section 5.

2. Review of Related Research

Classification of skin images using older methods [3,4] required arduous preprocessing, segmentation, and feature extraction procedures. Researchers from all over the world have recently been using deep learning to solve a variety of visual problems with success [5,6]. In order to exploit the merits of both machine learning (ML) and deep learning methods some researchers [8,9,10,11,12] presented an integration between outmoded ML approaches and DL algorithms for segmentation or categorization of skin cancer lesions.
Researchers of [8] described a technique for segmenting cutaneous melanoma that combines fuzzy k-means (FKM) clustering and faster region-based CNN (FRCNN). In order to attain a fixed-length feature vector, FRCNN was used. The FKM was employed to partition the potentially cancerous zone of skin to variable size segments and borders. Three common databases—ISBI-2016, ISIC-2017, and PH2—were used for performance evaluation. Regarding the ISIC-2016, ISIC-2017, and PH2 databases, their proposed technique attained an average accurateness of 95.40%, 93.1%, and 95.6%, respectively. In [9], the authors proposed a hybrid Inception adaptive neuro-fuzzy (ANF) model for classifying dermoscopic photos into different seven labels. All images were first preprocessed using hat filter and segmented with Grabcut technique. Then the Inception was used to automatically attain discriminative features, which were fed to the ANF classifier. Their method achieved an accuracy of 97.91%, specificity (SP) of 98.70%, and sensitivity (SE) of 93.40% on the ISIC-2018 dataset. Meanwhile, in [10], the authors suggested a hybrid ResNet-SVM framework for efficient binary classification of skin lesions. Furthermore, the issue of a shortage of annotated datasets was resolved via transfer learning (TL) together with data augmentation. Among the three employed datasets, ISIC-2017, HAM10000, and ISBI-2016, the ISIC-2017 dataset showed 99.19% accuracy for the suggested approach. Similarly, in [11], the authors classified three skin lesions using four dissimilar CNN-based architectures, specifically, AlexNet, VGG, ResNet-18, and ResNet-101 for feature extraction and SVM, RF, and MLP for classification. To provide a sole categorization for the input skin lacerations, several classification findings were combined. The results on ISIC 2017 showed an AUROC of 87.3% for PSL categorization and 95.5% regarding seborrheic keratosis categorization.
In [12], the authors suggested using a combination of CNN and local descriptor encoding techniques to classify skin lesions. Using ResNet50 and ResNet101, the lesion features were recovered from the pictures. Then, using the ResNet features that were extracted, a fisher vector (FV) was employed to create a general image representation. Finally, classification was performed using an SVM and a Chi-squared kernel. ISBI 2016 dataset was used to assess the efficacy. These findings demonstrated that their method outperformed other methods. Particle swarm optimization (PSO) was employed by [13] for segmenting skin lesions utilizing a variety of techniques, including the “Firefly Algorithm (FA), spiral research activity, probability distributions, crossover, and mutation”. K-means was applied to enhance lesion segmentation. The creation of CNN made use of the hybrid learning PSO (HLPSO). Melanoma and nevus lesions could be distinguished using the classification system. In [14], the authors developed a hybrid classifier for skin cancer employing both ECOC-SVM and DCNN. A collection of colored images of different skin cancers were gathered from the internet. For features extraction, a pre-trained AlexNet model was employed. Then, ECOC-SVM was used for skin cancer differentiation. More than 3750 images comprised of four skin cancer classes were employed for assessing efficacy. Their findings exhibited an average accurateness of 93.35%.
However, the authors in [15] provided an “intelligent Region of Interest (ROI)” algorithm to distinguish tumors from nevus malignancies with the aid of a transfer learning technique. A more effective variation of the k-mean technique was utilized to take out the ROIs from photos facilitating the recognition of distinctive traits, since only photos including malignancy cells were used for training. Furthermore, transfer learning along with data augmentation were utilized for extraction of ROI from the “DermIS” and “DermQuest” databases. The accuracy values from the suggested system for DermIS and DermQuest were 97.9% and 97.4%, respectively. Additionally, a CNN architecture that was trained utilizing ROI pigmented skin lesions was proposed by [16]. An AUC of 0.96 was achieved by their method. In [17], the authors presented a CNN-based six-class skin lesion classification technique. The authors took out 5846 medical photos of PSL from 3551 cases. A training dataset of 4732 images, 2885 patients with bounding box annotations and a test set of 666 images from 666 cases. A FRCNN developed architecture attained an accuracy of 86.2% while the derma experts showed an accuracy of 79.5% on the test set.
To accurately identify the greatest architectures for binary classification of skin cancer in more than 24,000 images, Ref. [18] applied three CNN models (Inception-V3, VGG19, and ResNet). The ISIC-19 and ISIC-2020 datasets were utilized to assess the classification accuracy. The Inception-V3 outperformed the other tested architectures with an accuracy of 86.90%. In [19], a DL architecture based on cloud infrastructure was utilized to implement different models for accurate skin cancer identification. SqueezeNet, DenseNet, and Inception-v3 pretrained models had ROC AUCs around 0.99, surpassing the ResNet model.
In order to increase performance and efficiency, Refs. [20,21] implemented a CNN model with many layers, various filter sizes, and a lower number of filters and settings for skin cancer categorization. For efficacy evaluation, Ref. [20] utilized the ISIC2017, ISIC2018, and ISIC2019 sets. Their method helps derma experts to categorize skin lesions, since it achieved 94% precision, 93% SE, 91% SP, and 0.964 AUC in ISIC2017 for PSL identification and categorization. While in [21], the authors used an idScore dataset that mixed dermoscopic images with clinical and histological information for classifying atypical moles. They achieved an AUC of 90.3%, SE of 86.5%, and SP of 73.6%. In [22], they compared the performance of CNN-based binary melanoma classifier trained with dermoscopic images to CNN as input data joint with an ANN for patients’ metadata. The ISIC-2019 dataset was used for assessing the performance of both models. The CNN+ANN model achieved an accuracy of 92.34%, thus outperforming the CNN model, which only attained 73.69%. However, the authors did not discuss the low accuracy results of their CNN model compared to the most recent research.
For three well-known architectures such as Inception-V3, VGG16, and VGG19, the authors in [23] created models using the transfer learning insight. The ISIC-17 dataset was used for training these models, which includes 604 test photos and 2487 test set of seven dissimilar classes of PSL. Their results showed accuracy of 74% for Inception-V3, 77% for VGG16, and 76% for VGG19. The authors of [24] described the development of a computerized skin lesion categorization algorithm employing dermoscopic photos and CNN as the base classifier. The model’s highest level of accuracy was 88.6%.
A hybrid-CNN [25] was developed based on three different feature extracting modules. These steps were combined to produce lesion feature vectors with better depth. Their system achieved an AUC of 0.96, 0.95, and 0.97, when tested against the ISIC2016, ISIC2017, and ISIC2018 datasets, respectively. In [26], the authors presented the classification of four different forms of skin cancer using the SCDNet, a vgg16-based framework. A classification accuracy of 96.91% was attained by SCDNet. Additionally, the proposed method’s accuracy is contrasted with that of four cutting-edge pre-trained classifiers.
For precise skin lesion subdivision, a “fully convolutional residual network (FCRN)” was built in [27]. Experimental results showed the suggested architecture had improved significantly in terms of performance, placing it first among 25 teams for classification and second among 28 teams for segmentation. The authors in [28] employed a set of data including 129,450 medical pictures for CNN training. The most common malignancies are represented in the first case, while the worst skin cancer is shown in the second case. In both missions, CNN performed as well as all tested specialists, demonstrating that AI is proficient of categorizing skin malignancies with a level of efficacy analogous to derma experts. Using a unique regularization technique, Ref. [29] proposed a different forecasting scheme which divides skin lacerations into “benign or malignant” lesions. Their new model achieved a 97.49% average rightness evaluated on the ISIC-2018. However, their new regularization method is not usable for feature extraction. In addition, choosing the appropriate parameters is a tedious and time-consuming process.
The authors of [30] reported that the average accuracy attained by derma experts for skin cancer identification is between 62% to 80% and they proposed the use of four ensemble and five CNN models for multiple class skin lesions discrimination. The HAM10000 dataset was utilized to assess the performance after applying preprocessing. Their results indicated a 93.2% maximum accuracy for ResNeXt101 model while ensemble model achieved 92.83% accuracy. Meanwhile, Ref. [31] separated photos into benign and cancerous categories. Their models were trained and tested using the open ISIC2020 database. According to the ISIC-2020 database, melanoma is categorized as cancerous. The effectiveness of three pre-trained architectures was then reported with a classification accuracy of 98.39%. Deep learning-based methods are employed by a popular yearly competition created by the “International Skin Imaging Collaboration Project (ISIC)” [32,33].
The authors in [34] used transfer learning and a pre-trained deep learning architecture to classify skin lesions. Through substituting the last layer with a softmax aiming to divide lesions into three classes, “transfer learning” is applied to AlexNet as well as soft-adjustment and data enlargement. Utilizing the ph2 database, their proposed architecture was trained and tested. The efficacy of their architecture obtained values of 98.61% accurateness, 98.33% sensitivity, and 98.93% specificity. Whereas, in [35], the authors described the creation of an ensemble of deep CNNs to further improve the efficiency of each CNN while identifying dermoscopy photographs into three cancerous levels: “melanoma, nevus, and seborrheic keratosis”. They were unable to receive sufficient training on an appropriate quantity of annotated photographs. They examine many fusion-based aggregating techniques and choose the most effective one for this issue. Regarding the three categories categorization mission, the average AUROC was 0.891.
The authors in [36] suggested an architecture based on weighted mean ensemble learning to categorize seven different kinds of skin infections. As the foundation of the ensemble, they used five DL architectures: ResNeXt, SeResNeXt, ResNet, Xception, and DenseNet. About 18,730 dermoscopy photographs from the HAM10000 and ISIC-2019 datasets at the same time with class balance, noise elimination, and data enlargement techniques were utilized for the training and validation of the evaluated models. The weighted mean ensemble scored a 94% recall rate, whereas the basic mean ensemble model attained 93%. The impact was revealed via the grid search technique. In order to train a DCNN, authors of [37] employed 4867 clinical photos from 1842 patients who had been diagnosed with skin tumors as a dataset. The DCNN model achieved an accuracy of malignant/benign classification of 92.4%. In [38], the authors introduced a bilinear CNN strategy that made use of transfer learning, a soft-adjustment step, and data augmentation to enhance classification performance while lowering the computing cost. Over the HAM10000 dataset, several imitations were run. According to their results, a bilinear method using the ResNet50 and VGG16 models improved accurateness by 2.7% above the up-to-date architectures. For example, the suggested method needed 238.6 min to train and averaged 93.21% accuracy. Whereas in [39], the authors proposed a DCNN (GoogLeNet Inception-v3) architecture to categorize derma-microscopy images into seven categories comprising the merit of binary support, showing that these categories can be unified into only normal/cancerous.
Also, in [40], the authors trained a DL architecture with 220,680 photographs of 174 illnesses and assessed it. Their system provided multi-class categorization among 134 illnesses, provided accurate malignancy prediction, recommended initial treatment alternatives, and enhanced the performance of medical experts. AUC for detecting malignancy was 0.928 and 0.937. In [41], the authors matched deep learning’s efficiency to that of derma specialists in categorizing histopathologic melanoma photos. For categorization of histological melanoma photos, CNN moderately outperformed eleven histopathologists despite having access to less image data, obtaining a mean sensitivity of 76%, specificity of 60%, and accuracy of 68%. Table 1 presented a comparison of deep learning approaches for classifying skin cancer.

3. Materials and Methods

Since skin melanoma ranks among the highly common cancers worldwide, precise and non-invasive diagnosis based on dermoscopic images has become critical and promising. This paper introduces the Assist-Dermo system for classifying nine classes of PSLs through an advanced deep learning approach by integrating preprocessing and data augmentation techniques. The overall architecture of the Assist-Dermo system is presented in Figure 2. The classification model is trained on the selected dataset. An enhanced lightweight version of SqueezeNet architecture that adopted a depthwise-separable CNN model was developed. The key aim of this article is to identify nine malignance types using images of skin lesions such as “actinic keratosis (AK), basal cell carcinoma (BCC), dermatofibroma (DF), melanoma (MEL), nevus (NV), pigmented benign keratosis (PBK), seborrheic keratosis (SK), squamous cell carcinoma (SCC), vascular lesion (VASL)”. The proposed Assist-Dermo was tested on the 24,000 images dataset. This section illustrates the different stages of the proposed Assist-Dermo PSLs classification system.

3.1. Acquisition and Preparation of Dataset

This section explores the recently utilized datasets for skin cancer classification, focusing on multiclass. The International Skin Imaging Collaboration (ISIC) offered various skin lesion datasets starting in 2016. These datasets were collected from a fair trial of affected participants who had undergone skin malignancy testing at various organizations, utilizing a range of derma-microscopy procedures over all anatomical sites (apart from the mucous membrane and nails).
The HAM10000 (“Human Against Machine with 10,000 Training Images”) database, also known as the ISIC2018 dataset [32,33], consists of seven PSL classes. It was separated into training and test databases, each of which includes 10,015 and 1512 images, respectively. The ISIC-2019 (BCN 20000) [42] comprises 25,331 photos grouped into eight PSL classes: “Actinic keratosis (AK), basal cell carcinoma (BCC), dermatofibroma (DF), melanoma (MEL), nevus (NV), pigmented benign keratosis (PBK), seborrheic keratosis (SK), and squamous cell carcinoma (SCC)”. The ISIC2020 [43] database contains 33,126 dermoscopy pictures categorized into nine well-known PSL classes as well as an unidentified picture class. More than 2000 patients provided the dermoscopy photos for these eight PSLs. The same eight PSL classes of ISIC 2019 in addition to “vascular lesion (VASC)” make up the nine PSL groups of photos.
Table 2 shows these datasets, indicating the number of selected images as well as the number of PSL classes. Table 3 represents the initial collection of PSLs with respect to multiclass. Those PSL images were used by the data augmentation technique to eradicate the problem of data imbalance as mentioned in Table 4. The visual result of data augmentation techniques is shown in Figure 3.
We discovered that the number of photos included within the various classes of PSLs in the selected dataset varied significantly. For example, the NV class has many more samples than the other classes. Additionally, the DF and VASC categories have fewer samples. We require enough balanced data to successfully train a DL-based model. Data balancing is performed by purposefully generating the required samples to prevent biased sampling during the DL model’s training. In addition, the unbalanced data may cause the model training to continue favoring classes with many examples. Therefore, to balance our dataset we used data augmentation, which comprises nine classes to increase our dataset by more than 24,000 and balance it for each class. Because affine transformations like rotation and shearing proved to have a detrimental effect on performance, we decided to include this type of augmentation in our investigation. Additional augmentation processes include adding blur with a probability of 25%, adding random Gaussian noise, altering brightness and contrast, and randomly flipping an image horizontally. During the training process, these data augmentation techniques were used. Moreover, in the selected datasets, there were different sizes of the 24-bit RGB PSL images ranged from (512 × 718) to (2848 × 4288). To address the variable-sized pictures issue, all dermoscopic images were reduced to a consistent size (512 × 512 × 3) by an image resize technique.

3.2. Preprocessing

Color space conversion is a crucial step that even experts in medicine use to separate non-melanoma skin lesions from healthy skin to simplify the segmentation process. The chosen color space must therefore be compatible with how people perceive color. Additionally, it was discovered that the selection of color space had a significant impact on the final image classification stage. Dermoscopy images have been used in the literature in the “RGB (Red, Green, Blue), YUV, HSV (Hue, Saturation, Value), CIE L*a*b*, and CIE L*u*v*” color spaces. Although the color spaces “RGB, YCbCr, and YUV” are frequently used in raw data and coding standards, human perception is incompatible with them. CIE color standards, in contrast, are computationally more complex but perceptually more homogeneous. The color appearance model CIECAM02 [20] offered the best features of previous color models. “Brightness Q, lightness J, colorfulness M, chroma C, saturation s, and hue h” are the six dimensions of color appearance included in the CIECAM02 color form scheme.
For picture segmentation, it has been demonstrated that the CIECAM02 (JCh) color space performs better than even RGB [28] and HSV [25] and is capable of reproducing human visual experience in colors. Therefore, both the preprocessing and skin lesion segmentation processes in this research used the JCh color space. The normalized ranges of the three planes are determined and are defined as (h: [0, 360]/15, C: [0, 1]/0.43, and J: [0, 1]/0.43) in the JCh color space transition step. To emphasize the significance of hue and reduce computational expenses, the 15 and 0.43 values are derived by empirical investigation. J*C*h* stands for the upgraded CIECAM02 (JCh) components. The “J*C*h*” perceptually-based color standard is developed.
To enhance the contrast and adjust the light illumination, we employed a technique developed in [46] but in perceptual-oriented (JCh) color space. We have selected this technique because it supports both local and global contrast of pixels without disturbing lesions patterns. The local improvement in PSL images is then created using an h-plane of JCh color space by maintaining the local contrast. To produce a finding that reflects a compromise between global and local contrasts, a contrast-brightness-based fusion algorithm is finally used. This technique enhances the visual quality while maintaining the authenticity of the image. The three-color channels are sometimes treated separately in color picture enhancement techniques, which alter the hue of the source photos. Hue preservation is a crucial strategy for color enhancement algorithms because these methods frequently produce unnatural-looking photos. The concept of hue in the JCh scheme was employed in our technique. Equation (1) determines the hue of an individual RGB colored point, p, with r, g, and b values:
h p = 0                           i f   r = g = b θ                                             i f   b g 2 π θ                       i f   b > g
where, the θ parameter is calculated as:
θ = a r c o s 1 / 2 ( r g + r b ) ( r g 2 + r b g b ) 1 / 2
Both the single pixel p1 and the other single pixel p2 have values of (r1, g1, b1) that are [0, L − 1]3. If and only if there are a(.) values R such that (r1, g1, b1) = a (r2, g2, b2) +d3 = T, where d3 = (1, 1, 1)T, L is the range of pixel values, and for 8-bit pictures, L = 256, then 3 have the same hue. The claim is a fundamental principle of hue preservation. The analysis of hue-preserving-based enhancement techniques shows that these techniques are superior to the conventional channel-by-channel improvements or techniques that just improve the intensity channel. In our approach for global contrast improvement, the range of pixel values of a colored photo Xc is first translated to the entire range [0, 1].
X c = ( J 1 ) X c X m i n X m a x X m i n
where L = 256 for 8-bit photos, and Xmin and Xmax are the image’s smallest and highest intensity values for its three-color channels, respectively. Using Equation (4), the rigid color photo Xc is transformed into the associated intensity photo J.
J = 0.299 × X c + 0.587 × X G + 0.114 × X B
where the three-color channels of the stretched color image X are XR, XG, and XB. To obtain the appropriate increased intensity picture GI, a new contrast improvement algorithm is then employed for J. Finally, we create the final improved color image Gc using the hue preservation enhancement framework.
G c k = X c k G I k I k                                                                                                                         i f   G I k J k 1 J 1 G I k J 1 I ( k )   ( X c k J k + G I k )                                           i f   G I k J ( k ) > 1
where k is the index of the pixels in each color channel, and c represents the corresponding R, G, and B color channel. With this equation, we have examined and demonstrated the effectiveness of color preservation. Figure 4 shows the contrast enhancement results in a perceptual-oriented color space.

3.3. Architecture of SqueezeNet-Light

The primary goal is to find a CNN architecture with few parameters that has competitive accuracy. For improved runtime performance, this paper modifies the SqueezeNet design into Squeeze-Light by utilizing depthwise separable convolutions (SepConv) rather than simple conventional ones. Overall steps of this architecture are mentioned in Algorithm 1. To categorize PSLs into nine types, our improved SqueezeNet-Light classifier was used. The structure of the Squeeze-Light is described in Figure 5a.
The fire module is one of SqueezeNet’s main components. The fire module uses several strategies, including reducing the quantity of 3 × 3 filters and substituting 1 × 1 filters. These two methods have both been used to reduce the number of variables. The expand layer comes after the squeeze layer, and it subjects the output of the squeeze layer to two distinct convolution procedures using 1 × 1 and 3 × 3 kernels. Like the squeezing layer, each convolution operation is preceded by batch normalization and followed by the GELU procedure. In our work, the 3 × 3 convolutions in the extend layer E3 are replaced with depthwise separable convolutions (SepConv), which considerably compress the fire module of SqueezeNet. Section 3.3.2 provides a detailed explanation of depthwise separable convolutions. This smaller fire module will thereafter be figured out as a “spark” module (Figure 4). The number of input channels is shown by C in the diagram. The number of the squeeze layer’s channels is S. The 1 × 1 and 3 × 3 extension layers’ channel counts are E1 and E3, respectively. The number of channels in the output depends only on the expansion layers.
We write the batch normalization and depthwise separable convolution layers with the abbreviations BN and SepConv, respectively (Figure 4a). In addition, SepFire denotes the substitution of SepConv for Conv in the fire layer, while SepFire + BN denotes the addition of the BN layer after the SepConv layer in the Fire module (Figure 5b).
Algorithm 1: ShuffleNet-Light Architecture for Features Extraction and Classification of PSLs
Input:Input Tensor ( X ), 2-D of (256 × 256 × 3) PSLs training dataset.
Output: Obtained and Classified feature map x = ( x 1 , x 2 , , x n ) augmented 2-D image
Main Process:
Step 1. Define number of stages = 4
Step 2. Iterate for Each Stage
(a) 
“Depthwise-CNN is applied to tensor x by kernel size of (3 × 3), which includes a number of filters; branch normaliza-tion, the ReLU activation function, Pointwise-CNN by kernel size of (1 × 1), branch normalization, and the GELU” activation function is applied.
(b) 
“Pointwise-CNN is applied to tensor x by kernel size of (1 × 1), which includes a number of filters, branch normalization, ReLU activation function, Pointwise-CNN by kernel size of (1 × 1), branch normalization, GELU” activation function is applied.
Step 3. Fscale = Squeeze and Excitation (SE) block contains expansion (1 × 1 × 3) layers.
Step 4. Fcat(i) = concatenation (# features-maps)
Step 5. channel = shuffle (x)
[End Step 2]
Step 6. Model Construction
(a) 
Define global-average-pooling layer
(b) 
Define fully-connected (FC) Layer and applied GELU function.
Step 7. Afterward, the feature map x = ( x 1 , x 2 , , x n ) generated, which is recognized by Softmax function.
Step 8. Test samples y i t   are predicted to the class label using the decision function of the below equation. y i t = t = 0 M 1 f t ( x i )

3.3.1. SqueezeNet Neural Network

Recently, increasing the accuracy and number of network layers has been the primary research focus for deep CNNs. Existing deep CNNs demand a lot of hardware and have slow prediction performance because of their vast volume and number of layers. However, there are nine classification categories for the problem of classifying PSLs. Therefore, the network does not require a lot of layers and hyperparameters. As a result, a deep learning-based PSL identification system is particularly challenging to employ in actual scenarios. With fewer network layers, deep learning can produce superior classification accuracy, whereas using too many layers can overfit the model and produce low accuracy. A tiny DL network with high-level accuracy and a compact network design must be created as a result.
SqueezeNet is a sort of extremely small deep CNN that attains similar accuracy as AlexNet on ImageNet while using only one-fiftieth the number of parameters. Compared to standard deep learning networks, SqueezeNet is more widely deployable in a variety of scenarios, has faster training and testing speeds, and has minimal hardware configuration needs. SqueezeNet uses a more compact and effective framework, particularly in the following three areas:
(1)
To decrease network parameters, a smaller 1 × 1 convolution kernel is used in place of the 3 × 3 convolution (Conv) kernel;
(2)
The parameters and model volume are reduced using SqueezeNet’s developed fire layer;
(3)
A thinner (a smaller number of) pooling layer resulted; only three maxpooling levels and a global pooling-layer are present in SqueezeNet.
As a result, SqueezeNet has a greater output layer with improved classification accuracy and information retention. Figure 4 shows the SqueezeNet network structure diagram as well as the layout of the fire layer. Conv and Concat are the abbreviations for convolution and concatenation, respectively. Additionally, we changed the data’s input format to width (W), height (H), and channel (N). Figure 4a depicts the overall layout of SqueezeNet, which is made up of numerous fire layers. To prevent overfitting, the SqueezeNet network includes global pooling and dropout layers next to the Fire 9 unit before classifying with a Softmax classifier. Because there are few pooling layers in the architecture and they are positioned backward, a complete connection layer does not have to be added to the network, which greatly minimizes the volume of SqueezeNet and eases parameter modification. The fire layer’s precise layout is depicted in Figure 4b, and it utilizes a 1 × 1 convolution layer for linear projection. Then, to lower the model size, speed up the detection, and cap the number of input channels, the network is enlarged using a combination of 1 × 1 and 3 × 3 convolutions.

3.3.2. Depthwise Separable Convolution

Even though SqueezeNet shrinks the size of the model and speeds up detection by fire layers, the efficacy of the conventional layer in SqueezeNet is constrained. Consequently, the model’s accuracy is not different from that of CNN. Hence, SqueezeNet network optimization is crucial. In this paper, aiming to enhance the SqueezeNet architecture, we substitute the conventional convolutions with a “depth-wise separable convolution”. The use of depthwise separable convolution allows for fewer training weight factors and fewer floating-point tasks, leading to a lighter, faster, and more accurate model [47].
Standard convolution operates by simultaneously handling the input channel and the convolution window to extract various characteristics in accordance with various convolution kernels. In depthwise separable convolution, two independent tasks are carried out. The first convolution in space task is carried out separately on every input channel while using a single 1-dimensional kernel to ensure an equal number of input and output channels (DConv). The channels computed in the preceding stage are then projected onto a new channel space using point-based convolution with a 1′1 kernel (PConv), as particularly presented in Figure 4c.
When biases are not considered, the number of input and output channels for a (N’N) standard conventional layer is C1 and C2, respectively. The standard convolutional layer needs the following number of parameters: C2 × C1 × N × N. However, SepConv simply needs C1 + N + C2 + C1 + 1 factors to produce an identical result. The SepConv layer’s parameters have been drastically cut back. Considering the larger convolution kernel, the SepConv parameters are considerably lower than those of regular convolutions. Hence, the impact on architecture optimization is greater. Batch normalization is employed after the SepConv or SepFire layers because it (1) increases the network’s training speed while stabilizing the distribution of input data for each layer, (2) simplifies the model parameter adjustment method to enhance the stability of architecture learning, and (3) lessens the issue of gradient disappearance, which, to some extent, regularizes the neural network.
The classical convolution is represented by Equation (6). While the depthwise separable convolution is mathematically represented by the Equations (7)–(9):
S t d C o n v ( θ , x ) ( i , j ) = h , w , c H , W , C θ h , w , c . x ( i + h ,   j + w , c )
D C o n v ( θ , x ) ( i , j ) = h , w H , W θ h , w   x ( i + h ,   j + w )
P C o n v ( θ , x ) ( i , j ) = c C θ c   . x ( i ,   j ,   c )
S e p C o n v θ p , θ d , x i , j = P C o n v i , j ( θ p , D C o n v ( i , j ) ( θ d , x ) )

3.3.3. Network Parameter Optimization

The values of a function’s parameters to reach a minimum can be either approximated or algebraically derived through a closed-form solution. Iterative methods like gradient descent are frequently the only option because cost functions in machine learning depend on a vast number of variables, and there is almost never a feasible way to obtain a closed-form solution for the minimum. When utilizing gradient descent [47], the parameter values must first be initialized so that optimization can begin. The value of the cost function is then reduced by iteratively changing the parameter values. In each iteration, parameter values are modified in the opposite direction as the cost gradient, which lowers the cost.
Even though adaptive techniques generalize [48] less effectively than SGD for many models, like convolutional neural networks (CNNs), they are typically employed as the default method because of their stability in challenging situations, such as the SqueezeNet model. To simultaneously achieve three goals—training stability, good generalization (like SGD), and speedy convergence—we propose AdaBelief (like adaptive techniques). AdaBelief’s fundamental concept is to adjust the step size in accordance with one’s “belief” in the current gradient’s direction. We mistrust the present observation and move slowly if the observed gradient considerably deviates from the prediction; if the observed gradient is near the prediction, we trust it and move rapidly. This is achieved by considering the “exponential moving average (EMA)” of the noisy gradient as the forecast of the gradient at the following time step. Through extensive testing, we assess AdaBelief and prove that it beats rival methods for image classification and language modeling with brisk convergence and superb accuracy. The mathematical formulas are as follows:
Y p r e d i c t = m o d e l W ( x ) × h
w = w α f ( Y , Y p r e d i c t ) w

4. Experimental Results

In this section we illustrate the setup for all experiments along with a brief explanation for the employed performance metrices. In addition, we present the results of the conducted experiments as well as analysis for these results.

4.1. Experimental Setup

A processor from Intel, the Core-i7 7700HQ, with a frequency of 2.8 GHz and memory of 16 GB, along with a GPU from the NVIDIA GeForce series, the GTX-1050 Ti, with committed 4 GB of memory, are utilized as the experimentation hardware. We put the suggested deep Light-SqueezeNet approach for PSL multi-labelling into practice using the Python-based Keras environment, a high-level neural network API. The mini-batch size was fixed at 16, and the number of epochs was set to 40. In addition, we utilized adaptive belief (AdaBelief) to train the proposed model with a learning rate of 0.001, and the learning policy is set to “step” with a gamma of 0.5. Also, we employed the AdaBelief optimization technique with the following configurations: default settings of 0.9, 0.999, and 1 × 10−8 for the exponential decay, the moment estimates, and epsilon, as shown in Table 5.
The datasets include photos of variable sizes. Therefore, we resized the images to the customary size of 512 × 512. In experiments, the data are divided into two groups: the training set and the test set, and a 10-fold cross-validation test is applied. The test set is utilized for model evaluation and prediction. To ensure fair comparisons, the same settings are used for all models.

4.2. Model Evaluation Metrics

In this study, classifications into nine categories were made using a particular dataset. We evaluated the results through applying “confusion matrix”, accurateness, precision, recall, F1Score, and the “Matthews Correlation Coefficient (MCC)”. The confusion matrix lists various variations of anticipated and real class values. The “true positive (TP)” and “true negative (TN)” values represent, respectively, the properly categorized benign and malignant cases. The wrongly classified benign and malignant cases are indicated, respectively, by the “false positive (FP)” and “false negative (FN)” values. The model’s accuracy, recall, and F-measure were calculated using the macro-average method for each of the nine categorization classes. MCC is a popular metric that is applied even if the classes have drastically diverse sizes because it is a balanced measure. Following is how these metrics are calculated:
A c c u r a c y   A C C = T P + T N T P + T N + F P + F N
P r e c i s i o n   P R = T P T P + F P
R e c a l l R C = T P T P + F N
F 1 s c o r e = 2 × P R × R C P R + R C
M C C = T P × T N ( F P × F N ) ( T P + F P ) ( T P + F N ) ( T N + F P ) ( T N + F N )
We carried out the trials utilizing several potent CNN models, namely VGG16, AlexNet, InceptionV3, GoogleNet, Xception, MobileNet, SqueezeNet, and SqueezeNet-Light techniques for recognizing multiclass scenarios. In addition, we have also performed SOTA comparisons. We utilized the validation and testing splits to assess the effectiveness of these procedures. The primary distinction between the two splits is that, while the testing data were derived from separate sources, the validation split was produced using the same sources as the training data. We review the two, three, five, seven, and nine-class scenarios. In all experiments, all performance measures are computed using a 10-fold cross-validation technique split by the prevalence of the diagnosis category. The means of the metrics are calculated and provided in the remaining sections of the study to evaluate the models’ efficacy.

4.3. Network Settings

In addition to accuracy, the loss function plays a crucial role in image categorization during the transfer learning (TL) training phase. An optimizer is applied to the weights and learning rate of the network. This step is performed to cut down on loss functions while training in the DL layers. In this paper, the weight and acceleration time inside the model layers of each pre-trained TL model were changed by applying various optimizers such as ADAM, SGD, Adadelta, AdaBelief, and RMSprop. The optimizer in DL algorithms manages weight and bias throughout the network fitting process. This is the main emphasis of this optimizer. We choose the best hyperparameters for the layer feature maps, filter size, activation function, pool size, dropout, and fine-tuning. Then, the optimization step was performed to attain the best parameters. This clarifies that each optimizer has a specific function in various deep-learning applications.
In contrast, AdaBelief optimization is used to determine the best set of hyperparameters to use when assessing the entire deep learning framework. A total of 12 hyperparameters are used in this work. We are trying to examine hyperparameters and fine-tuning layers, which include freezing the top or lower network layer. The automated hyperparameters and fine-tuning steps are going through various stages of evaluation to find the best set of hyperparameters and layers for the recognition of nine classes of PSLs. When using the AdaBelief method to evaluate the hyperparameter and fine tune the pre-trained layers, a default value for the hyperparameter and fine-tuning layers must be set. After completing automatic hyperparameter tuning and automated fine tuning against all of the TL models by using different optimizations, the second stage involves assessing the performance of different transfer learning (TL) models. The final section examines how well TL performs when some layers are frozen throughout the automated hyperparameter tweaking procedure. Each TL model was optimized using hyperparameters and fine-tuning steps.

4.4. Computational Cost

According to the complexity of computations, SOTA models and the suggested Light-SqueezeNet system were also compared. According to Table 6, the suggested DL architecture required a total processing time of about 184.5 s. Where the total processing times were 246.2, 230.1, 217.4, 211.8, 207.5, 195.7, and 193.4 s for the VGG16, AlexNet, InceptionV3, GoogleNet, Xception, MobileNet, and SqueezeNet, respectively. Accordingly, our suggested Light-SqueezeNet technique took less time to identify several PSL classes, which is essential in a setting where computational performance is crucial. This demonstrates how effective the suggested idea is in relation to the current paradigm.
It has been shown that SqueezeNet-Light may improve network detection performance and is better at finding nine classes of PSLs in images than traditional convolutional layers. Therefore, on several datasets, the suggested approach outperformed SOTA systems. The parameter size was reduced in part thanks to this separable transfer network. Table 7 compares the number of parameters in the convolutional layers of the VGG16, AlexNet, InceptionV3, GoogleNet, Xception, MobileNet, SqueezeNet, and the proposed SqueezeNet-Light CNN. The findings showed that the suggested model significantly reduced the number of parameters on the convoluted layer. According to the experiment’s findings, reducing parameters did not produce degenerate models but rather more network generalization. In conclusion, the suggested SqueezeNet-Light performs better than certain other conventional models, especially when using large datasets, and lays the framework for its usage in PSL analysis systems thanks to its faster running speed.

4.5. Visual Feature Representation

SqueezeNet-Light provides an efficient and effective model for the classification of multiple classes of PSLs on a vast dataset. The SqueezeNet-Light is used as an efficacious feature extractor, which is exhibited in Figure 6. Figure 6 shows a visual comparison of how the proposed (SqueezeNet-Light) model and the classic (SqueezeNet) model affect the feature reactions of PSLs. The images in the first left column are input PSL images, along with their feature reactions in the correct columns of Figure 6. It is noticeably observed that the patterns of PSL using the original SqueezeNet are unclear, including noise and vital details that are missing (Figure 6a). Although, in Figure 6b images, using our SqueezeNet, the PSL’s crucial texture details are easily noticed. Accordingly, our suggested model feature extractor effectively extracted the crucial textural information needed to distinguish distinct PSL lesions. As a result, our proposed model can handle the pattern classification problem of PSLs with a better response while ignoring artifacts like hairs.

4.6. Performance of Proposed System

The dropout value (the percentage of deleted layers at the training phase) and initial learning rate were selected at random. To avoid its unique and accidental nature and to decrease the influence of manual parameter modification, the network was taught to choose the best SqueezeNet architecture. The dropout rate ranged from 0.2 to 0.6. The SqueezeNet model contained batch normalization layers, allowing us to start training at a higher learning rate. As a result, their initial learning rates were between 102 and 104. Under batch sizes of 32, 64, and 128, the initial learning rate and dropout values are chosen at random, and the network is trained using the Adam optimization technique. For training under different training batches, each network contained 30 unique initial sequence parameters. The 10,000 training sessions were allowed for each network with combined parameters. The training was over before ten training sessions, when the verification set loss remained constant. Statistical analysis was used to calculate the accuracy (ACC), specificity (SP), sensitivity (SE), precision (PR), recall (RL), F1-score, and MCC values. These metrics were employed for the performance assessment of the proposed system and to compare it to formerly built, pre-trained transfer learning algorithms. The different experiments measured the accuracy in the convolutional layers of the VGG16, AlexNet, InceptionV3, GoogleNet, Xception, MobileNet, SqueezeNet, and SqueezeNet-Light models. This section contains an exhaustive overview of the many investigations conducted to assess the efficacy of the purported SqueezeNet-Light model. In addition, we used the area under the receiver operating curve (AUC) to show the effectiveness of the training and validation datasets with a 10-fold cross-validation test. Figure 7, Figure 8 and Figure 9 reveal the best plot loss, accuracy, AUC, and recall on the train and validation sets with data augmentation, run for 40 epochs, for the proposed SqueezeNet-Light model.
Experiment 1: This experiment showed a 10-fold cross-validation testing methodology to compare the outcomes in terms of the confusion matrix when the pre-processing step is used in different color spaces. Figure 10a–c shows the results of the proposed model with preprocessing in different color spaces like CIECAM, CIELab, and HSV. Figure 10d–f shows CIECAM, CIELab, and HSV color spaces without a preprocessing step to increase contrast and adjust brightness. As shown in these confusion matrices, the color space CIECAM is perfect for feature extraction and classification tasks.
Experiment 2: We used 10-fold cross-validation testing to compare the statistical results to different TL algorithms, such as VGG16, AlexNet, InceptionV3, GoogleNet, Xception, MobileNet, SqueezeNet, and SqueezeNet-Light models. Table 8, Table 9 and Table 10 show the classification results of the DL models that had already been trained based on different batch sizes like 16, 32, and 64. The performance of the developed SqueezeNet-Light system has been the same with all different batch sizes (16, 32, and 64). But the effect of batch size was observed in terms of the number of parameters and computational time. However, the classification results remain the same for other pre-trained TL algorithms. The developed approach yielded excellent results: SE of 94%, SP of 96%, ACC of 95.6%, PR of 94.12%, F1-score of 95.2, and MCC of 96.7, as well as a low training error (0.76) in identifying multiclass PSLs.
Experiment 3: In this experiment, the impact of various optimization techniques on classification results was examined. We have also utilized different optimizers to build an efficient SqueezeNet-Light model. Adaptive algorithms like Adam have a good convergence speed, while algorithms like stochastic gradient descent (SGD) have better generalization (better response to new data). To combine both optimizer qualities, AdaBelief was developed in the past as an optimizer to control the loss function. We believe that AdaBelief can take care of regions with “large gradient, small curvature” cases, while Adam does not handle them. In the same 10-fold of training data, Table 11 compares the optimizers with the weighted ones. The AdaBelief optimizer, with learning rate, weight decay, and momentum set at 1 × 10−5, 1 × 10−8, and 0.9, respectively, is compared to the different optimization methods. When training the model with the AdaBelief optimizer with momentum, the number of epochs is set to 40. This is because previous experiments have shown that the value of the loss function curve keeps going down after the 30th epoch and that convergence is not reached at the end of the 30th epoch. The epoch number is fixed to 40 using all optimizers for the sole purpose of this study to provide a fair comparison. Table 10 provides a summary of the numerical outcomes. When the AdaBelief optimizer is used, the SE value is seen to increase considerably to 94%. As a result, the experiments’ optimization method of choice is the AdaBelief optimizer.
Experiment 4: In this experiment, we have also evaluated different loss functions. The findings demonstrate that the weighted-cross entropy loss function can improve classification performance by correcting class imbalance. Cross entropy loss was 82.2, 97.8, 96.9, 76, 79, and 82.2, whereas weighted-cross entropy loss was provided at 94, 96, 95.6, 94.12, 95.2, and 96.7 for SE, SP, ACC, PR, F1-score, and MCC metrics, respectively. As a result, we have used the weighted-cross entropy loss function. Table 12 demonstrates those results.
Experiment 5: In experiment 5, we have also evaluated the proposed SqueezeNet-Light model in terms of computational cost based on different benchmarks. Compared to other CNN and TL-based architectures, the SqueezeNet-Light model is an effective method for classifying PSLs. The original SqueezeNet network has more parameters and FLOPs compared to the proposed SqueezeNet-Light model (shown in Table 13). As a result, the proposed enhanced architecture has fewer parameters and converges faster than its baselines. Table 12 shows that the SqueezeNet-Light model has FLOPs of 68.3 MFLOPs, a model Size of 9.3 MB, and a GPU Speed of 0.7 S. As a result, the SqueezeNet-Light model created a new and improved architecture with computational efficiency, as detailed in Section 3.
Experiment 6: In this experiment, we compared the execution of the proposed SqueezeNet-Light model on CPU, GPU, and TPU with respect to batch size and computational speed. Layer-by-layer examination of the CNN implementation on the CPU, TPU, and GPU is necessary in practice [43]. The SqueezeNet-Light network should be built with each job being a multiple instruction, single data (MISD) task to maximize its performance in TPU. Prioritizing the neural network’s tasks is necessary when building a network. In actuality, the GPU provides more programming simplicity and flexibility for small programs. GPUs are better suited to small batch quantities of data because of the execution pattern in wraps and scheduling on straightforward on-stream multi-processors. The GPU performs well for large datasets and network models by maximizing memory reuse. Fully linked neural networks have lesser weight reuse, which causes increased memory traffic as the model size increases. The GPU can be utilized for applications that require memory because of its memory bandwidth. When processing big neural networks, GPUs outperform CPUs due to their added parallelism capacity. For fully-linked neural networks, the GPU performs better than the CPU, while the TPU shines when dealing with large batch sizes.
Whereas in the case of TPU, the array structure has been utilized, which works better on the SqueezeNet-Light architecture with large batches to offer high throughput during training. Large batches of data are necessary to properly utilize the matrix multiply units in the systolic array of the TPU. As the batch size increases, the architecture speeds up. Due to the networks’ ability to reuse space for big batch sizes and intricate CNNs, TPU is the best. The performance of the proposed SqueezeNet-Light model’s CPU, TPU, and GPU benchmarks in terms of batch size is shown in Table 14.
These experiments showed that SqueezeNet-Light produces classification rates for nine classes of PSLs that are superior compared to those of standard models. The improved SqueezeNet-Light-based classification accuracy of PSLs was 95.6%. As a result, the use of SepConv instead of classical convolution layers in the SqueezeNet model provided the best architecture for developing a lightweight Assist-Dermo system. In this paper, the speed and the number of parameters of the model are the key performance measures in addition to classification accuracy.

4.7. Comparisons with SOTA

Comparisons were made between traditional models such as Salama-ResNet-SVM [10], Ashraf-FRCNN [15], Naeem-VGG16 [26], Hosny-AlexNet [34], Fujisawa-DCNN [37], Harangi-Inception [39], Hekler-CNN [41], SqueezeNet, and the proposed SqueezeNet-Light model. As the identification and classification models, the SqueezeNet-Light-based models with the highest verification set under various configurations were chosen. The training set based on 10-fold cross validation was used to train the deep learning models for comparison, and the verification set adjustment parameters were used to choose the best model. The performance of the classifier with regards to accuracy, recall rate, precision rate, and F1-score of several approaches that rely on the unbalanced data set was quantifiably explored by evaluating each model with the test set.
We calculated the typical test time (forward propagation time), the number of parameters, and the model size 1000 times for each image in the test set. Based on the detailed indicators, our aim is to choose the model with the best F1 score, accuracy, and detection speed. Table 14 provides information on the sensitivity, specificity, precision, accuracy, and F1-score of each model on the test set. The computational cost, total parameters, model sizes, and average prediction time for the 1000 test images are shown in Table 15. In this table, we demonstrate that SqueezeNet is faster at runtime but is more accurate thanks to the additional batch normalization layers. Although the performance is significantly slower after the batch normalizing layer is added, precision increases. SqueezeNet-Light contains an additional deep separable convolution layer in addition to the added batch normalizing layer. SqueezeNet-light outperforms SqueezeNet in terms of accuracy, precision rate, and F1 score, among other metrics. The pace of detection and classification has also been enhanced. Additionally, SqueezeNet-light has far fewer model parameters than SqueezeNet and AlexNet, which makes it more usable and feasible for implementation in a realistic context.
When compared to other SOTA models, SqueezeNet-light shows an accuracy that is higher than that of the original SqueezeNet and other SOTA TL-based CNN models. The proposed SqueezeNet-Light’s accuracy rating (95.6%) is higher than classical SqueezeNet (87.6%) and even Inception (85%). While it has a considerably shorter prediction time and a smaller model size than all other SOTA models. Those statistical results are mentioned in Table 15 and Figure 11. Additionally, the SqueezeNet-Light model reaches the model’s smallest volume. SqueezeNet-Light is therefore superior to other models in terms of accuracy, prediction speed, and model size for the detection of PSL lesions.
In this study, dermoscopic pictures are used to differentiate between various classes of pigmented skin lesions (PSLs) using a variety of datasets. A visual example of the results attained using our proposed SqueezeNet-Light classifier is displayed in Figure 12 to recognize nine classes of PSLs. The testing splits for other classes were also gathered from various sources. We assessed various pre-trained TL CNN architectures such as VGG16, AlexNet, InceptionV3, GoogleNet, Xception, MobileNet, and SqueezeNet techniques to discriminate among various skin cancer types in various scenarios. We employed data augmentation to avoid the issue that CNN architectures need a lot of labeled data to train on. The findings obtained demonstrated that our proposed model beats CNN architectures. In three-class, five-class, seven-class, and nine-class scenarios, our suggested SqueezeNet-Light architecture had an accuracy of 95.6% in the recognition of multiclass PLS classes. Computational performance of different SOTA architectures are mentioned in Table 16. According to this table, our proposed model achieved low computational burden as compared to other architectures.

5. Discussion

There is an urgent need to address this worldwide public health concern, since skin cancer incidence rates have been rising over the past few decades. Pretrained-based deep-learning CNNs are now being used for the classification of skin cancer due to their outstanding effectiveness in classifying medical images. Although numerous studies have been conducted in the past to classify skin cancer, none of them have been successful in expanding their research to include nine different types of skin cancer. In this study, we classified skin cancer more accurately than dermatologists and existing deep learning techniques. On the different datasets, the performance of the proposed SqueezeNet-Light is examined to identify which approach is most effective for classifying skin cancer. We conducted a significant study to identify the ideal configuration of hyperparameters. Comparing SqueezeNet-Light to earlier proposed deep learning models, the performance and computational efficiency have significantly improved. Therefore, we suggest using SqueezeNet-Light to classify skin cancer. We also concluded that deep learning models trained with the ideal hyperparameter setup cannot perform as well as other pre-trained CNN models. While these methods are frequently used to increase classification job accuracy, they also greatly increase the model’s architectural complexity and may not have a major impact on how well deep learning models perform when set up with the best hyperparameters.
The workload of dermatologists can be significantly reduced by creating an automated classification system for skin lesions, which will also lessen subjectivity and human error-related subjectivity in the classification process. There have been a few incidents of inappropriate therapy because of an inaccurate or delayed diagnosis. Misdiagnosis might occasionally result in a greater requirement for surgical intervention and a longer hospital stay because the effects of treatment can take time to manifest. The ability of dermatologists to perform skin examinations as diagnosticians has a significant impact on the detection of skin cancer. However, dermatologists with more than ten years of expertise scarcely top 80% memory, and those with three to five years of experience only reach 62% recall in the skin cancer screening. When competent dermatologists are hard to come by for unskilled practitioners in developing nations, this proposed method can become vital and more valuable. This SqueezeNet-Light model is effective at speeding up the automated classification process for skin lesions and can even be used to apply a class label to a brand-new skin lesion. While these results are reassuring and offer compelling evidence that the deep learning approach can help doctors and healthcare systems, more clinical data on topics like age, gender, race, and family history is still needed for further validation and advancement before the deep models can be evaluated in clinical practice.
Also, our method worked well in the three-class and five-class scenarios, with an overall identification accuracy (ACC) of 94.5%, sensitivity (SE) of 96.7%, specificity (SP) of 95%, and area under the receiver operating curve (AUC) of 0.95. We are attempting to gather additional hospital cases of skin cancer for future research. Moreover, based on the available dermoscopy scans, we intend to define additional skin cancer disease groups. The SqueezeNet-Light technique, on the other hand, aims to deploy more deep CNN architectures. Additionally, combining the advanced features of many architectures is an exciting tactic that can improve performance. A preprocessing step based on contrast enhancement in a perceptual-oriented color space is also needed to further improve the classification results. Additionally, skin lesions will be automatically separated in HSV color space using Grab-Cut with little human involvement. The ABCD (asymmetry, border irregularity, color, and dermoscopic patterns) rule to distinguish malignant melanoma from benign lesions will be automatically applied using image processing techniques. Different pretrained convolutional neural networks (CNNs), including InceptionV3 and MobileNet, are also evaluated to categorize skin lesions as benign or malignant. However, to categorize nine classes of PSLs, we have developed an effective and efficient classifier, Assist-Dermo.
Using a separable vision transformer (SVT) approach, a dermatologist-like automatic classification system is built to classify nine different kinds of pigmented skin lesions (PSLs) to aid clinical professionals in the early diagnosis of skin cancer. The authors have developed a few methods in the past to identify the nine PSL classes. However, they frequently require many calculations to attain good performance, making their deployment on devices with limited resources difficult. In this article, a novel SVT architectural design based on the SqueezeNet, and depthwise-separable CNN models is presented. The primary objective is to identify CNN architectures with minimal parameters and competitive precision. This study modifies the SqueezeNet architecture by employing depthwise separable convolutions rather than the basic conventional ones for enhanced runtime performance. An intuitive comprehension of CNN principles is provided by the display of the results. Plotting the feature maps obtained at successive levels, for instance, enables comparison of the modifications made to the features by the pooling, batch normalization, and activation layers. In the second scenario, the user can further investigate how the network’s performance is impacted by changes to the dataset and hyperparameters. Examples in this sense can include comparing performance when the learning rates of layers are changed or even frozen. Additionally, we think that making this technology more approachable for non-experts would strengthen the collaboration between dermatologists and computer scientists toward the joint effort of improving image-based medical diagnosis.
Table 17 demonstrates the limitations as well as the advantages of our SqueezeNet-Light DL architecture compared to SOTA architectures. Our SqueezeNet-Light architecture outperforms other DL methods in the number of classified PSL classes. Since it classifies the input PSL images into nine classes while others are ranged from two to seven classes without addressing the perceptual-oriented color space, the contrast is enhanced, and there is no control over class imbalance issues. Furthermore, the vast majority of them ignored classifier generalizability and trained and tested models on a single dataset. Moreover, the proposed SqueezeNet-Light model is a lightweight architecture since it has a tiny number of parameters and smaller memory requirements compared to other SOTA systems. Consequently, this leads to a faster processing and classification system.
In the future, we will also compare this proposed SqueezeNet-Light architecture with other recently developed lightweight [49,50] architectures to confirm the generalizability of the network.

6. Conclusions

To aid clinical professionals in the early diagnosis of skin cancer, a derma expert-like automated categorization (Assist-Dermo) system is created in this work to classify several types of PSLs. Using a separable vision transformer (SVT) approach, a derma expert-like automated classification system is built to classify nine different kinds of pigmented skin lesions (PSLs) to aid clinical professionals in the early diagnosis of skin cancer. The authors have developed a few methods in the past to identify the nine PSL classes. However, they frequently require many calculations to attain good performance, making their deployment on devices with limited resources difficult. In this article, a novel SqueezeNet vision transformer (SVT) architectural design based on the SqueezeNet, and depthwise-separable CNN models is presented. The primary objective is to identify CNN architectures with minimal parameters and competitive precision. This study modifies the SqueezeNet architecture by employing depthwise separable convolutions rather than the basic conventional ones for enhanced runtime performance. For the development of this Assist-Dermo system, a data-augmentation approach was used to address the PSLs’ imbalance issue. Next, a pre-processing phase is used to choose the most dominant area, followed by the enhancement of lesion patterns in a color space geared toward perception. To increase efficacy and performance, the Assist-Dermo system is designed with many layers and numerous filter sizes but fewer filters and parameters. For the training and assessment of Assist-Dermo models, a collection of 24,000 photos of PSLs from online data sources such as Ph2, ISBI-2017, HAM10000, and ISIC is used to identify nine classes of PSLs. On the collected dataset, it obtained 95.6 accuracy, 94 sensitivity, 96% specificity, 94.12% precision, and a 95.2 F1-Score. The experimental findings demonstrate that the proposed Assist-Dermo approach outperforms existing algorithms in recognizing nine types of PSLs. The Assist-Dermo system performs better than current state-of-the-art methods and can help dermatologists diagnose a variety of PSLs via dermoscopy images. Also, the proposed Assist-Dermo method performed better than other SOTA methods at classifying different PSLs from dermoscopy images without accurate scarification in terms of speed of prediction and model size.
Computational efficiency shows that the proposed system can be easily deployed in an environment where is a requirement of resource-constrained devices such as mobile devices. However, it is important to note that the lightweight nature of the model should also be tested on the Internet of Things (IoT) devices, which makes it suitable for deployment on a wider scale for the accurate classification of pigmented skin lesions (PSLs). This point of view will be addressed in future works.

Author Contributions

Conceptualization, M.E.A.I., Q.A., Y.D. and U.R.; methodology, M.E.A.I. and Q.A.; software, Q.A.; validation, M.E.A.I., Y.D. and U.R.; resources, M.E.A.I., Q.A., Y.D. and U.R.; data curation, M.E.A.I. and Q.A.; writing—original draft preparation, M.E.A.I., Q.A., Y.D. and U.R.; writing—review and editing, M.E.A.I., Q.A. and U.R.; visualization, M.E.A.I., U.R. and Q.A. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported and funded by the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University (IMSIU) (grant number IMSIU-RP23122).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Datasets used in these experiments are available online. Following are links used to access these datasets such as: [32,33] HAM10000 https://www.kaggle.com/datasets/kmader/skin-cancer-mnist-ham10000 [Accessed on 12 January 2023]; [42] ISIC 2019: https://challenge.isic-archive.com/landing/2019/ [Accessed on 12 January 2023]; [43] ISIC 2020: https://challenge2020.isic-archive.com/ [Accessed on 12 January 2023]; [44] Ph2: https://www.kaggle.com/datasets/synked/ph2-modified [Accessed on 12 January 2023]; [45] ISBI-2017 https://skinclass.de/TestSet.zip [Accessed on 12 January 2023].

Acknowledgments

This work was supported and funded by the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University (IMSIU) (grant number IMSIU-RP23122).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. American Cancer Society. Cancer Facts and Figures. 2022. Available online: https://www.cancer.org/content/dam/cancer-org/research/cancer-facts-and-statistics/annual-cancer-facts-and-figures/2022/2022-cancer-facts-and-figures.pdf (accessed on 12 February 2023).
  2. Ferlay, J.; Colombet, M.; Soerjomataram, I.; Dyba, T.; Randi, G.; Bettio, M.; Gavin, A.; Visser, O.; Bray, F. Cancer incidence and mortality patterns in Europe: Estimates for 40 countries and 25 major cancers in 2018. Eur. J. Cancer 2018, 103, 356–387. [Google Scholar] [CrossRef] [PubMed]
  3. Baig, A.R.; Abbas, Q.; Almakki, R.; Ibrahim, M.E.; AlSuwaidan, L.; Ahmed, A.E. Light-Dermo: A Lightweight Pretrained Convolution Neural Network for the Diagnosis of Multiclass Skin Lesions. Diagnostics 2023, 13, 385. [Google Scholar] [CrossRef] [PubMed]
  4. Alsahafi, Y.S.; Kassem, M.A.; Hosny, K.M. Skin-Net: A novel deep residual network for skin lesions classification using multilevel feature extraction and cross-channel correlation with detection of outlier. J. Big Data 2023, 10, 105. [Google Scholar] [CrossRef]
  5. Menghani, G. Efficient deep learning: A survey on making deep learning models smaller, faster, and better. ACM Comput. Surv. 2023, 55, 1–37. [Google Scholar] [CrossRef]
  6. Abbas, Q.; Ibrahim, M.E.A.; Jaffar, M.A. Video scene analysis: An overview and challenges on deep learning algorithms. Multimed. Tools Appl. 2019, 77, 20415–20453. [Google Scholar] [CrossRef]
  7. Maron, R.C.; Schlager, J.G.; Haggenmüller, S.; von Kalle, C.; Utikal, J.S.; Meier, F.; Gellrich, F.F.; Hobelsberger, S.; Hauschild, A.; French, L.; et al. A benchmark for neural network robustness in skin cancer classification. Eur. J. Cancer 2021, 155, 191–199. [Google Scholar] [CrossRef]
  8. Nadipineni, H. Method to Classify Skin Lesions using Dermoscopic images. arXiv 2020, arXiv:2008.09418. [Google Scholar]
  9. Sikkandar, M.Y.; Alrasheadi, B.A.; Prakash, N.B.; Hemalakshmi, G.R.; Mohanarathinam, A.; Shankar, K. Deep learning based an automated skin lesion segmentation and intelligent classification model. J. Ambient Intell. Humaniz. Comput. 2020, 12, 3245–3255. [Google Scholar] [CrossRef]
  10. Salama, W.M.; Aly, M.H. Deep learning design for benign and malignant classification of skin lesions: A new approach. Multimed. Tools Appl. 2021, 80, 26795–26811. [Google Scholar] [CrossRef]
  11. Mahbod, A.; Schaefer, G.; Ellinger, I.; Ecker, R.; Pitiot, A.; Wang, C. Fusing fine-tuned deep features for skin lesion classification. Comput. Med. Imaging Graph. 2019, 71, 19–29. [Google Scholar] [CrossRef] [Green Version]
  12. Yu, Z.; Jiang, X.; Zhou, F.; Qin, J.; Ni, D.; Chen, S.; Lei, B.; Wang, T. Melanoma recognition in dermoscopy images via aggregated deep convolutional features. IEEE Trans. Biomed. Eng. 2018, 66, 1006–1016. [Google Scholar] [CrossRef] [PubMed]
  13. Tan, T.Y.; Zhang, L.; Lim, C.P. Adaptive melanoma diagnosis using evolving clustering, ensemble and deep neural networks. Knowl. Based Syst. 2020, 187, 104807. [Google Scholar] [CrossRef]
  14. Dorj, U.O.; Lee, K.K.; Choi, J.Y.; Lee, M. The skin cancer classification using deep convolutional neural network. Multimed. Tools Appl. 2018, 77, 9909–9924. [Google Scholar] [CrossRef]
  15. Ashraf, R.; Afzal, S.; Rehman, A.U.; Gul, S.; Baber, J.; Bakhtyar, M.; Mehmood, I.; Song, O.-Y.; Maqsood, M. Region-of-interest based transfer learning assisted framework for skin cancer detection. IEEE Access 2020, 8, 147858–147871. [Google Scholar] [CrossRef]
  16. Abbas, Q.; Celebi, M.E. DermoDeep-A classification of melanoma-nevus skin lesions using multi-feature fusion of visual features and deep neural network. Multimed. Tools Appl. 2019, 78, 23559–23580. [Google Scholar] [CrossRef]
  17. Jinnai, S.; Yamazaki, N.; Hirano, Y.; Sugawara, Y.; Ohe, Y.; Hamamoto, R. The development of a skin cancer classification system for pigmented skin lesions using deep learning. Biomolecules 2020, 10, 1123. [Google Scholar] [CrossRef]
  18. Mijwil, M.M. Skin cancer disease images classification using deep learning solutions. Multimed. Tools Appl. 2021, 80, 26255–26271. [Google Scholar] [CrossRef]
  19. Kadampur, M.A.; Al Riyaee, S. Skin cancer detection: Applying a deep learning based model driven architecture in the cloud for classifying dermal cell images. Inform. Med. Unlocked 2020, 18, 100282. [Google Scholar] [CrossRef]
  20. Iqbal, I.; Younus, M.; Walayat, K.; Kakar, M.U.; Ma, J. Automated multi-class classification of skin lesions through deep convolutional neural network with dermoscopic images. Comput. Med. Imaging Graph. 2021, 88, 101843. [Google Scholar] [CrossRef]
  21. Tognetti, L.; Bonechi, S.; Andreini, P.; Bianchini, M.; Scarselli, F.; Cevenini, G.; Moscarella, E.; Farnetani, F.; Longo, C.; Lallas, A.; et al. A new deep learning approach integrated with clinical data for the dermoscopic differentiation of early melanomas from atypical nevi. J. Dermatol. Sci. 2021, 101, 115–122. [Google Scholar] [CrossRef]
  22. Ningrum, D.N.A.; Yuan, S.-P.; Kung, W.-M.; Wu, C.-C.; Tzeng, I.-S.; Huang, C.-Y.; Li, J.Y.-C.; Wang, Y.-C. Deep Learning Classifier with Patient’s Metadata of Dermoscopic Images in Malignant Melanoma Detection. J. Multidiscip. Healthc. 2021, 14, 877. [Google Scholar] [CrossRef] [PubMed]
  23. Jasil, S.P.; Ulagamuthalvi, V. Deep learning architecture using transfer learning for classification of skin lesions. J. Ambient. Intell. Humaniz. Comput. 2021, 1–8. [Google Scholar] [CrossRef]
  24. Wei, L.; Ding, K.; Hu, H. Automatic Skin Cancer Detection in Dermoscopy Images Based on Ensemble Lightweight Deep Learning Network. IEEE Access 2020, 8, 99633–99647. [Google Scholar] [CrossRef]
  25. Hasan, M.K.; Elahi MT, E.; Alam, M.A.; Jawad, M.T.; Martí, R. DermoExpert: Skin lesion classification using a hybrid convolutional neural network through segmentation, transfer learning, and augmentation. Inform. Med. Unlocked 2022, 28, 100819. [Google Scholar] [CrossRef]
  26. Naeem, A.; Anees, T.; Fiza, M.; Naqvi, R.A.; Lee, S.W. SCDNet: A Deep Learning-Based Framework for the Multiclassification of Skin Cancer Using Dermoscopy Images. Sensors 2022, 22, 5652. [Google Scholar] [CrossRef]
  27. Yu, L.; Chen, H.; Dou, Q.; Qin, J.; Heng, P.A. Automated melanoma recognition in dermoscopy images via very deep residual networks. IEEE Trans. Med. Imaging 2016, 36, 994–1004. [Google Scholar] [CrossRef]
  28. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef]
  29. Albahar, M.A. Skin lesion classification using convolutional neural network with novel regularizer. IEEE Access 2019, 7, 38306–38313. [Google Scholar] [CrossRef]
  30. Chaturvedi, S.S.; Tembhurne, J.V.; Diwan, T. A multi-class skin Cancer classification using deep convolutional neural networks. Multimed. Tools Appl. 2020, 79, 28477–28498. [Google Scholar] [CrossRef]
  31. Singha, S.; Roy, P. Skin Cancer Classification and Comparison of Pretrained Models Performance using Transfer Learning. J. Inf. Syst. Eng. Bus. Intell. 2022, 8, 218–225. [Google Scholar] [CrossRef]
  32. Codella, N.; Rotemberg, V.; Tschandl, P.; Celebi, M.E.; Dusza, S.; Gutman, D.; Halpern, A.; Kalloo, A.; Liopyris, K.; Marchetti, M.; et al. Skin lesion analysis toward melanoma detection 2018: A challenge hosted by the international skin imaging collaboration (isic). arXiv 2019, arXiv:1902.03368. [Google Scholar]
  33. Tschandl, P.; Rosendahl, C.; Kittler, H. The HAM10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions. Sci. Data 2022, 5, 180161. [Google Scholar] [CrossRef] [PubMed]
  34. Hosny, K.M.; Kassem, M.A.; Foaud, M.M. Skin cancer classification using deep learning and transfer learning. In Proceedings of the 2018 9th Cairo International Biomedical Engineering Conference (CIBEC), Cairo, Egypt, 20–22 December 2018; pp. 90–93. [Google Scholar]
  35. Harangi, B. Skin lesion classification with ensembles of deep convolutional neural networks. J. Biomed. Inform. 2018, 86, 25–32. [Google Scholar] [CrossRef] [PubMed]
  36. Rahman, Z.; Hossain, M.S.; Islam, M.R.; Hasan, M.M.; Hridhee, R.A. An approach for multiclass skin lesion classification based on ensemble learning. Inform. Med. Unlocked 2021, 25, 100659. [Google Scholar] [CrossRef]
  37. Fujisawa, Y.; Otomo, Y.; Ogata, Y.; Nakamura, Y.; Fujita, R.; Ishitsuka, Y.; Watanabe, R.; Okiyama, N.; Ohara, K.; Fujimoto, M. Deep-learning-based, computer-aided classifier developed with a small dataset of clinical images surpasses board-certified dermatologists in skin tumour diagnosis. Br. J. Dermatol. 2019, 180, 373–381. [Google Scholar] [CrossRef]
  38. Calderón, C.; Sanchez, K.; Castillo, S.; Arguello, H. BILSK: A bilinear convolutional neural network approach for skin lesion classification. Comput. Methods Programs Biomed. Update 2021, 1, 100036. [Google Scholar] [CrossRef]
  39. Harangi, B.; Baran, A.; Hajdu, A. Assisted deep learning framework for multi-class skin lesion classification considering a binary classification support. Biomed. Signal Process. Control 2020, 62, 102041. [Google Scholar] [CrossRef]
  40. Han, S.S.; Park, I.; Chang, S.E.; Lim, W.; Kim, M.S.; Park, G.H.; Chae, J.B.; Huh, C.H.; Na, J.-I. Augmented intelligence dermatology: Deep neural networks empower medical professionals in diagnosing skin cancer and predicting treatment options for 134 skin disorders. J. Investig. Dermatol. 2020, 140, 1753–1761. [Google Scholar] [CrossRef]
  41. Hekler, A.; Utikal, J.S.; Enk, A.H.; Solass, W.; Schmitt, M.; Klode, J.; Schadendorf, D.; Sondermann, W.; Franklin, C.; Bestvater, F.; et al. Deep learning outperformed 11 pathologists in the classification of histopathological melanoma images. Eur. J. Cancer 2019, 118, 91–96. [Google Scholar] [CrossRef] [Green Version]
  42. Combalia, M.; Codella, N.C.; Rotemberg, V.; Helba, B.; Vilaplana, V.; Reiter, O.; Carrera, C.; Barreiro, A.; Halpern, A.C.; Puig, S.; et al. Bcn20000: Dermoscopic lesions in the wild. arXiv 2019, arXiv:1908.02288. [Google Scholar]
  43. Rotemberg, V.; Kurtansky, N.; Betz-Stablein, B.; Caffery, L.; Chousakos, E.; Codella, N.; Combalia, M.; Dusza, S.; Guitera, P.; Gutman, D.; et al. A patient-centric dataset of images and metadata for identifying melanomas using clinical context. Sci. Data 2019, 8, 34. [Google Scholar] [CrossRef] [PubMed]
  44. Mendonça, T.; Ferreira, P.M.; Marques, J.S.; Marcal, A.R.; Rozeira, J. PH2-A dermoscopic image database for research and benchmarking. In Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Osaka, Japan, 3–7 July 2013; pp. 5437–5440. [Google Scholar]
  45. Maron, R.C.; Weichenthal, M.; Utikal, J.S.; Hekler, A.; Berking, C.; Hauschild, A.; Enk, A.H.; Haferkamp, S.; Klode, J.; Schadendorf, D.; et al. Systematic outperformance of 112 dermatologists in multiclass skin cancer image classification by convolutional neural networks. Eur. J. Cancer 2017, 119, 57–65. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Tian, Q.C.; Cohen, L.D. Global and local contrast adaptive enhancement for non-uniform illumination color images. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Montreal, BC, Canada, 11–17 October 2021; pp. 3023–3030. [Google Scholar]
  47. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  48. Zhuang, J.; Tang, T.; Ding, Y.; Tatikonda, S.C.; Dvornek, N.; Papademetris, X.; Duncan, J. Adabelief optimizer: Adapting stepsizes by the belief in observed gradients. Adv. Neural Inf. Process Syst. 2017, 33, 18795–18806. [Google Scholar]
  49. Zhu, R.; Cui, Y.; Huang, J.; Hou, E.; Zhao, J.; Zhou, Z.; Li, H. YOLOv5s-SA: Light-Weighted and Improved YOLOv5s for Sperm Detection. Diagnostics 2023, 13, 1100. [Google Scholar] [CrossRef]
  50. Li, W.; Zhang, L.; Wu, C.; Cui, Z.; Niu, C. A new lightweight deep neural network for surface scratch detection. Int. J. Adv. Manuf. Technol. 2022, 123, 1999–2015. [Google Scholar] [CrossRef]
Figure 1. A visual example of nine types PSLs, where (a) actinic keratosis (AK), (b) “basal cell carcinoma” (BCC), (c) dermatofibroma (DF), (d) melanoma (MEL), (e) nevus (NV), (f) pigmented benign keratosis (PBK), (g) “seborrheic keratosis” (SK), and (h) “squamous cell carcinoma” (SCC).
Figure 1. A visual example of nine types PSLs, where (a) actinic keratosis (AK), (b) “basal cell carcinoma” (BCC), (c) dermatofibroma (DF), (d) melanoma (MEL), (e) nevus (NV), (f) pigmented benign keratosis (PBK), (g) “seborrheic keratosis” (SK), and (h) “squamous cell carcinoma” (SCC).
Diagnostics 13 02531 g001
Figure 2. A methodical illustration of suggested Assist-Dermo system to recognize nine classes of pigmented skin lesions.
Figure 2. A methodical illustration of suggested Assist-Dermo system to recognize nine classes of pigmented skin lesions.
Diagnostics 13 02531 g002
Figure 3. An example of data augmentation techniques applied on the selected datasets from the ISIC-2019, ISIC-2020, and HAM10000 sources in case of benign and malignant skin lesions.
Figure 3. An example of data augmentation techniques applied on the selected datasets from the ISIC-2019, ISIC-2020, and HAM10000 sources in case of benign and malignant skin lesions.
Diagnostics 13 02531 g003aDiagnostics 13 02531 g003b
Figure 4. An example of preprocessing enhancement step to enhance the contrast of SCC lesion taken from Figure 1, where (ae) represents the original PSLs, and (fj) shows the corresponding contrast enhancement using nonlinear sigmoidal function.
Figure 4. An example of preprocessing enhancement step to enhance the contrast of SCC lesion taken from Figure 1, where (ae) represents the original PSLs, and (fj) shows the corresponding contrast enhancement using nonlinear sigmoidal function.
Diagnostics 13 02531 g004
Figure 5. SqueezeNet-Light architecture (a) SqueezeNet-Light base structure, (b) fire module structure and (c) SepConv module basic structure.
Figure 5. SqueezeNet-Light architecture (a) SqueezeNet-Light base structure, (b) fire module structure and (c) SepConv module basic structure.
Diagnostics 13 02531 g005
Figure 6. (a) Feature representations example of original SequeezeNet model using PSLs image and (b) feature representations example of our SequeezeNet-Light model using PSLs image.
Figure 6. (a) Feature representations example of original SequeezeNet model using PSLs image and (b) feature representations example of our SequeezeNet-Light model using PSLs image.
Diagnostics 13 02531 g006
Figure 7. Plot of loss, accuracy, AUC, and recall on train/validation sets without data augmentations.
Figure 7. Plot of loss, accuracy, AUC, and recall on train/validation sets without data augmentations.
Diagnostics 13 02531 g007
Figure 8. Plot loss, accuracy, AUC, and recall on the train and validation sets with data augmentation.
Figure 8. Plot loss, accuracy, AUC, and recall on the train and validation sets with data augmentation.
Diagnostics 13 02531 g008
Figure 9. Plot loss, accuracy, AUC, and recall on train/validation sets with data augmentation and 40 epochs.
Figure 9. Plot loss, accuracy, AUC, and recall on train/validation sets with data augmentation and 40 epochs.
Diagnostics 13 02531 g009
Figure 10. Confusion matrices: Sub-figures (ac) show the result of proposed model with preprocessing in CIECAM, CIELab, and HSV color spaces. Sub-figures (df) show the result of proposed model without preprocessing step to enhance the contrast and adjust brightness, for the same color spaces.
Figure 10. Confusion matrices: Sub-figures (ac) show the result of proposed model with preprocessing in CIECAM, CIELab, and HSV color spaces. Sub-figures (df) show the result of proposed model without preprocessing step to enhance the contrast and adjust brightness, for the same color spaces.
Diagnostics 13 02531 g010
Figure 11. Performance of SOTA systems when compared to our Assist-Dermo system (a) binary classification (malignant and benign lesions), (b) represents five-classes’ classifications (AK, BCC, DF, Mel, NV), (c) illustrates seven-classes’ classifications (AK, BCC, DF, Mel, NV, PBK, SK), and (d) shows nine-classes’ classifications (AK, BCC, DF, Mel, NV, PBK, SK, SCC, Vasc).
Figure 11. Performance of SOTA systems when compared to our Assist-Dermo system (a) binary classification (malignant and benign lesions), (b) represents five-classes’ classifications (AK, BCC, DF, Mel, NV), (c) illustrates seven-classes’ classifications (AK, BCC, DF, Mel, NV, PBK, SK), and (d) shows nine-classes’ classifications (AK, BCC, DF, Mel, NV, PBK, SK, SCC, Vasc).
Diagnostics 13 02531 g011aDiagnostics 13 02531 g011b
Figure 12. A visual example of proposed SqueezeNet-Light classification, where (a) actinic keratosis, (b) basal cell carcinoma, (c) dermatofibroma, (d) melanoma, (e) nevus, (f) pigmented benign keratosis, (g) seborrheic keratosis, (h) squamous cell carcinoma, and (i) vascular lesion.
Figure 12. A visual example of proposed SqueezeNet-Light classification, where (a) actinic keratosis, (b) basal cell carcinoma, (c) dermatofibroma, (d) melanoma, (e) nevus, (f) pigmented benign keratosis, (g) seborrheic keratosis, (h) squamous cell carcinoma, and (i) vascular lesion.
Diagnostics 13 02531 g012
Table 1. Comparison of DL-based approaches for classifying skin cancer.
Table 1. Comparison of DL-based approaches for classifying skin cancer.
Ref.DiagnosisSegment?ClassificationDL ModelDatasetsResult (%) **Limitations
[9]Proposed a hybrid Inception adaptive-neuro-fuzzy (ANF) model for discriminating dermoscopic photos into different seven labels.Yes7 classesInception-v4ISIC-2018ACC: 97.91%
SE: 93.4%
SP: 98.7%
Classes imbalance problem, evaluated on single dataset, classifies only seven classes, and computationally expensive.
[10]Suggested a hybrid ResNet-SVM framework for efficient binary classification of skin lesions.NoBinaryResNet50,
VGG-16 and SVM
ISIC 2017
ISBI 2016
ACC: 99.19%Classes imbalance problem, binary classification only two classes, and computationally expensive.
[15]Offered an augmented ROI -based CNN system to recognize and separate melanoma from nevus malignancy.YesBinaryCNN + Transfer learningDermIS,
DermQuest
ACC: 97.9%
ACC: 97.4%
Classes imbalance problem, evaluated on single dataset, classifies only two classes, and computationally expensive.
[17]Introduced a six-class cutaneous lesions classification system based on CNN.No6 classesFaster region-based CNN (FRCNN)Private (5846 images)ACC:
Six-classes 86.2%
Two-classes 91.5%
Image processing, handcrafted-based feature extraction approach, which limits the detection accuracy, 6 classes only, and computationally expensive
[19]Suggested a hybrid ResNet-SVM framework for efficient binary classification of skin lesions.NoBinarySqueezeNet, DenseNet, inception v3 and ResNetHAM10000AUC: 0.997Evaluated on one dataset, classifies only two classes, and computationally expensive.
[20]Implemented a CNN model with many layers, various filter sizes, lower number of filters and settings for skin cancer categorization.No3 classesDCNNISIC 2017
ISIC 2018
ISIC 2019
AUC: 0.964Three classes of PSLs and reduced hyper-parameters, so computationally expensive.
[25]Suggested a hybrid-CNN made up of three different feature extracting modules that were combined to produce lesion feature vectors with better depth.Yes7 classesCNNISIC-2016, ISIC-2017, ISIC-2018AUC:
ISIC-2016: 0.96
ISIC-2017: 0.95
ISIC-2018: 0.97
Seven classes only and used only one limited dataset, classifier is not generalized.
[26]Presented the classification of four different forms of skin cancer using the SCDNet, a vgg16-based framework.No4 classesVgg16ISIC 2019ACC: 96.91%Four classes only, many hyper-parameters required, and tested on signal dataset, so classifier is not generalized.
[29]Proposed a new regularization technique for CNN model to binary classify skin lesions.NoBinaryCNNISIC-2018ACC: 97.49%Two classes (benign vs. malignant), required huge parameters, evaluate on single dataset, and computationally expensive.
[31]Examined how well three of the best pretrained DL models classified skin cancer in binary form.NoBinaryResNet, VGG16, MobileNetV2ISIC 2020ACC: 98.39%Two classes (benign vs. malignant), required huge parameters, evaluate on single dataset, and computationally expensive.
[34]Used transfer learning and a pre-trained deep learning architecture to classify three skin lesions.No3 classesAlexNetph2Acc: 98.61%
SE: 98.33%
SP: 98.93%
Three classes only, required huge parameters, evaluate on single dataset, and computationally expensive.
[35]Described the creation of an ensemble of deep CNNs to further improve the efficiency of each CNN while identifying dermoscopy photos into three categoriesNo3 classesGoogLeNet, AlexNet, ResNet, VGGNetISBI 2017AUC: 0.891Single dataset used, three classes of PSLs and reduced hyper-parameters, so computationally expensive.
[36]Suggested an architecture based on weighted mean ensemble learning to categorize seven different kinds of skin infectionsNo7 classesResNeXt, SeResNeXt, DenseNet, Xception, ResNetHAM10000
ISIC 2019
ACC/recall:
avg.: 87%/93%
weight avg.: 88%/94%
Seven classes with two datasets used, three classes of PSLs and reduced hyper-parameters, so computationally expensive.
[37]Employed 4867 clinical photos from 1842 patients who had been diagnosed with skin tumors as a dataset to train a DCNN architecture.NoBinaryDCNNPrivate (4867 images)ACC: 92.4%Two classes (benign vs. malignant), required huge parameters, evaluate on single dataset, and computationally expensive.
[38]a bilinear CNN strategy that made use of transfer learning, a soft-adjustment step, and data augmentation to enhance classification performance while lowering the computing cost.No7 classesResNet50 and VGG16HAM10000ACC: 93.21%Three classes only and used only one limited dataset, classifier is not generalized.
[39]classified dermoscopy images into seven classes comprising the advantage of binary support.No7 classesGoogLeNet and Inception-v3ISIC 2018BMA: 67.7%Seven classes, required huge parameters, evaluate on single dataset, and computationally expensive.
[41]Deep learning’s efficiency to that of derma specialists in categorizing histopathologic melanoma photos.NoBinaryCNNPrivate
(695 lesions)
ACC: 68%
SE: 76%
SP: 60%
Two classes (benign vs. malignant), required huge parameters, evaluate on single dataset, and computationally expensive.
D dermoscopic, C clinical, H Histopathological WSIs. ** ACC: accuracy, SE: sensitivity, SP: Specificity, BMA: balanced multi-class accuracy, AUC: area under curve.
Table 2. Dermoscopic datasets brief description.
Table 2. Dermoscopic datasets brief description.
DatasetRef.ImagesSelected ImagesNumber of Classes *
HAM10000[32,33]10,01510,0157 (“AK, BCC, BKL, DF, NV, MEL, VASC”)
ISIC 2019[42]25,33125,3318 (“MEL, NV, BCC, AKIEC, BKL, DF, VASC, SCC”)
ISIC 2020[43]33,12633,1269 (“MEL, NV, BCC, AKIEC, BKL, DF, VASC, SCC, PBK”)
Ph2[44]2002002(NV, and Mel)
ISBI 2017[45]275027503 (NV, Mel, and SK)
* Nv: Nevus, Mel: melanoma, SK: seborrheic keratosis, BKL: “benign keratosis lesion”, DF: “dermatofibroma”, AK: actinic keratosis, BCC: “basal cell carcinoma”, VASC: “vascular lesion”, SCC: squamous cell carcinoma, and PBK: pigmented benign keratosis.
Table 3. Number of images in each class.
Table 3. Number of images in each class.
Classes *No. of ImagesSize
AK700(512,512,3)
BCC3300(512,512,3)
BKL2600(512,512,3)
DF200(512,512,3)
NV12,000(512,512,3)
MEL4000(512,512,3)
SCC600(512,512,3)
VASC300(512,512,3)
PBK300(512,512,3)
Total24,000(512,512,3)
* AK: Actinic keratosis, BCC: “basal cell carcinoma”, BKL: “benign keratosis lesion”, DF: “dermatofibroma”, NV: nevus, MEL: melanoma, SCC: squamous cell carcinoma, VASC: “vascular lesion”, and PBK: pigmented benign keratosis.
Table 4. Data augmentation techniques used to develop a light-dermo system.
Table 4. Data augmentation techniques used to develop a light-dermo system.
ParametersAngleBrightnessZoomShearModeHorizontalVerticalRescaleNoise
Values30°[0.9, 1.1]0.10.1ConstantFlipFlip1./2550.45
Table 5. Hyper parameters configurations of the developed SqueezeNet-Light architecture.
Table 5. Hyper parameters configurations of the developed SqueezeNet-Light architecture.
Tensor FlowGPULearning RateOptimizerNumber of EpochBatch sizeValidation
2.9.1GeForce GTX 1050 Ti1 × 10−3AdaBelief401610-fold
Table 6. Average processing time on a PSLs dataset by various DL algorithms.
Table 6. Average processing time on a PSLs dataset by various DL algorithms.
MethodPreprocessingFeature
Extraction
TrainingPredictionOverall
VGG1620.5 s14.4 s200.5 s10.8 s246.2 s
AlexNet18.6 s12.2 s190.5 s8.8 s230.1 s
InceptionV316.3 s14.8 s178.5 s7.8 s217.4 s
GoogleNet17.2 s17.3 s170.5 s6.8 s211.8 s
Xception18.1 s15.1 s165.5 s8.8 s207.5 s
MobileNet14.1 s13.3 s160.5 s7.8 s195.7 s
SqueezeNet10.8 s8.3 s168.5 s5.8 s193.4 s
Proposed1.8 s1.9 s165.5 s1.5 s184.5 s
Table 7. The number of parameters in the existing pretrained TL models compared to SqueezeNet-Light model.
Table 7. The number of parameters in the existing pretrained TL models compared to SqueezeNet-Light model.
ModelsImage SizeParametersValidation Accuracy
VGG16512 × 512 × 3
256 × 256 × 3
200 × 200 × 3
14,714,688
14,865,222
14,911,302
79
AlexNet512 × 512 × 3
256 × 256 × 3
200 × 200 × 3
23,587,712
14,911,302
81.3
InceptionV3512 × 512 × 3
256 × 256 × 3
200 × 200 × 3
42,658,176
14,911,302
82.7
GoogleNet512 × 512 × 3
256 × 256 × 3
200 × 200 × 3
14,714,688
14,865,222
14,911,302
83.5
Xception512 × 512 × 3
256 × 256 × 3
200 × 200 × 3
14,714,688
14,865,222
14,911,302
82.4
MobileNet512 × 512 × 3
256 × 256 × 3
200 × 200 × 3
3,228,864
14,865,222
14,911,302
84.3
SqueezeNet512 × 512 × 3
256 × 256 × 3
200 × 200 × 3
7,037,504
14,911,302
87.6
SqueezeNet-Light512 × 512 × 3
256 × 256 × 3
200 × 200 × 3
3,182,412
3,182,412
3,182,412
95.6
Table 8. Results of the proposed system model’s classification using 16 batches of data.
Table 8. Results of the proposed system model’s classification using 16 batches of data.
ModelEpochsSESPACCPRF1-ScoreMCC
VGG1640788079767980
AlexNet40798281.38080.481.1
InceptionV340818082.78282.783.5
GoogleNet40838183.58383.584.5
Xception40828382.48384.385.4
MobileNet408484.284.38485.286.3
SqueezeNet408586.287.68586.187.2
Squeeze-Light40949695.694.1295.296.7
SE: Sensitivity, SP: Specificity, RL: Recall, PR: Precision, ACC: Accuracy.
Table 9. Results of the proposed system model’s classification using 32 batches of data.
Table 9. Results of the proposed system model’s classification using 32 batches of data.
ModelEpochsSESPACCPRF1-ScoreMCC
VGG1640788079767980
AlexNet40798281.18080.081.0
InceptionV340818082.38282.283.4
GoogleNet40838183.68383.384.3
Xception40828382.68384.485.2
MobileNet408484.084.38485.186.1
SqueezeNet408586.187.28586.087.0
Squeeze-Light40949695.694.1295.296.7
SE: Sensitivity, SP: Specificity, RL: Recall, PR: Precision, ACC: Accuracy.
Table 10. Results of the proposed system model’s classification using 64 batches of data.
Table 10. Results of the proposed system model’s classification using 64 batches of data.
ModelEpochsSESPACCPRF1-ScoreMCC
VGG1640788080767980.5
AlexNet40808180.38080.482.3
InceptionV340828281.78282.784.0
GoogleNet40828382.58382.585.0
Xception40848483.48383.386.0
MobileNet408382.285.38484.283.0
SqueezeNet408585.286.68585.186.1
Squeeze-Light40949695.694.1295.296.7
SE: Sensitivity, SP: Specificity, RL: Recall, PR: Precision, ACC: Accuracy.
Table 11. Results of the proposed system model’s classification using various optimizers.
Table 11. Results of the proposed system model’s classification using various optimizers.
OptimizationSESPACCPRF1-ScoreMCC
SGD with Momentum80.281.882.9858082.2
Adam828382.58382.585.0
RMSProp848483.48383.386.0
AdaGrad8382.285.38484.283.0
AdaBelief949695.694.1295.296.7
SE: Sensitivity, SP: Specificity, RL: Recall, PR: Precision, ACC: Accuracy.
Table 12. Proposed system model’s classification results using various loss functions.
Table 12. Proposed system model’s classification results using various loss functions.
Loss FunctionsSESPACCPRF1-ScoreMCC
Cross Entropy Loss82.297.896.9767982.2
Weighted Cross Entropy Loss949695.694.1295.296.7
SE: Sensitivity, SP: Specificity, RL: Recall, PR: Precision, ACC: Accuracy.
Table 13. Computational performance of different architectures.
Table 13. Computational performance of different architectures.
DL ArchitecturesComplexity (MFLOPs)Model Size
(MB)
GPU Speed
(S)
SqueezeNet-Light68.39.30.7
SqueezeNet96.914.51.6
MobileNet95.412.31.3
GoogleNet272.815.22.7
Xception281.816.32.6
InceptionV3554.317.53.0
AlexNet65.914.52.8
VGG16295.812.33.4
MFLOPS: million floating-point operations per second, M: millions, MB: megabytes, S: seconds.
Table 14. Performance of CPU/TPU/GPU Comparisons of the proposed SqueezeNet-Light model.
Table 14. Performance of CPU/TPU/GPU Comparisons of the proposed SqueezeNet-Light model.
Batch SizeEpochsCPU/TPU/GPU (mS)
6440700/500/400
12840750/400/500
25640750/400/500
51240750/400/500
102440750/400/500
mS: milliseconds, CPU: central processing unit, GPU: graphical processing unit, and TPU: tensor processing units.
Table 15. Classification results of the SOTA systems.
Table 15. Classification results of the SOTA systems.
ModelEpochsSESPACCPRF1-ScoreTrainable Parameters
SqueezeNet-Light100949695.694.1295.23,182,412
SqueezeNet100878587.687877,037,504
Salama-ResNet-SVM [10]100808281798053,982,272
Ashraf-FRCNN [15]100828382808150,213,111
Naeem-VGG16 [26]100838583838448,222,341
Hosny-AlexNet [34]100848684848549,112,242
Fujisawa-DCNN [37]100798078798052,128,141
Harangi-GoogleNet-Inception [39]100858685848550,440,122
SE: sensitivity, SP: specificity, PR: precision, ACC: accuracy.
Table 16. Computational performance of different SOTA architectures.
Table 16. Computational performance of different SOTA architectures.
SOTA ArchitecturesComplexity (MFLOPs)Parameters
(M)
Model Size
(MB)
GPU Speed
(S)
SqueezeNet-Light68.320.189.30.7
SqueezeNet96.938.1114.52.6
Salama-ResNet-SVM [10]120.353.9822.32.7
Ashraf-FRCNN [15]140.950.2120.55.6
Naeem-VGG16 [26]150.348.2219.34.7
Hosny-AlexNet [34]122.349.1117.34.7
Fujisawa-DCNN [37]120.952.1316.53.6
Harangi-GoogleNet-Inception [39]150.350.4419.33.7
MFLOPS: Million floating-point operations per second, M: millions, MB: megabytes, S: seconds.
Table 17. Limitations and advantages of different SOTA architectures compared to proposed architecture.
Table 17. Limitations and advantages of different SOTA architectures compared to proposed architecture.
TL ArchitecturesLimitationsAdvantages
SqueezeNet-LightIt should be trained on more different datasets and this classifier can be tested on different modality of images to check the generalizability of the model.Tiny model, high speed, several datasets are evaluated, and identifies nine classes
SqueezeNetIt is small model but requires hyper-parameter tuningIt is better classifier compared to other pretrained TL models.
Salama-ResNet-SVM [10]Classes imbalance, binary classification only two classes, and computationally expensive.Integration of SVM and ResNet and good for binary decision.
Ashraf-FRCNN [15]Classes imbalance, evaluated on single dataset, classify only two classes, and computationally expensive.This approach used CNN and is better for features extraction.
Naeem-VGG16 [26]Four classes only, many hyper-parameters required, and tested on signal dataset so classifier is not generalized.This method used pre-trained TL VG-16 to recognize four classes of PSLs
Hosny-AlexNet [34]Three classes only, required huge parameters, evaluate on single dataset, and computationally expensive.This approach used CNN and is better for features extraction.
Fujisawa-DCNN [37]Two classes (benign vs. malignant), required huge parameters, evaluate on single dataset, and computationally expensive.This approach used CNN and is better for features extraction.
Harangi-GoogleNet-Inception [39]Seven classes, required huge parameters, evaluate on single dataset, and computationally expensive.Combining the GoogleNet and Inception pretrained TL to recognize nine classes.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Abbas, Q.; Daadaa, Y.; Rashid, U.; Ibrahim, M.E.A. Assist-Dermo: A Lightweight Separable Vision Transformer Model for Multiclass Skin Lesion Classification. Diagnostics 2023, 13, 2531. https://doi.org/10.3390/diagnostics13152531

AMA Style

Abbas Q, Daadaa Y, Rashid U, Ibrahim MEA. Assist-Dermo: A Lightweight Separable Vision Transformer Model for Multiclass Skin Lesion Classification. Diagnostics. 2023; 13(15):2531. https://doi.org/10.3390/diagnostics13152531

Chicago/Turabian Style

Abbas, Qaisar, Yassine Daadaa, Umer Rashid, and Mostafa E. A. Ibrahim. 2023. "Assist-Dermo: A Lightweight Separable Vision Transformer Model for Multiclass Skin Lesion Classification" Diagnostics 13, no. 15: 2531. https://doi.org/10.3390/diagnostics13152531

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop