Next Article in Journal
Position Estimator Design for a MEMS Top-Drive Electrostatic Rotary Actuator
Previous Article in Journal
Face Pose Alignment with Event Cameras
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Skin Lesion Classification Using Densely Connected Convolutional Networks with Attention Residual Learning

1
Hubei Province Key Laboratory of Intelligent Information Processing and Real-Time Industrial System, College of Computer Science and Technology, Wuhan University of Science and Technology, Wuhan 430081, China
2
School of Computer Science and Statistics, Trinity College Dublin, Dublin 2, Ireland
*
Authors to whom correspondence should be addressed.
Sensors 2020, 20(24), 7080; https://doi.org/10.3390/s20247080
Submission received: 18 October 2020 / Revised: 2 December 2020 / Accepted: 9 December 2020 / Published: 10 December 2020
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Skin lesion classification is an effective approach aided by computer vision for the diagnosis of skin cancer. Though deep learning models presented advantages over traditional methods and brought tremendous breakthroughs, a precise diagnosis is still challenging because of the intra-class variation and inter-class similarity caused by the diversity of imaging methods and clinicopathology. In this paper, we propose a densely connected convolutional network with an attention and residual learning (ARDT-DenseNet) method for skin lesion classification. Each ARDT block consists of dense blocks, transition blocks and attention and residual modules. Compared to a residual network with the same number of convolutional layers, the size of the parameters of the densely connected network proposed in this paper has been reduced by half, while the accuracy of skin lesion classification is preserved. Our improved densely connected network adds an attention mechanism and residual learning after each dense block and transition block without introducing additional parameters. We evaluate the ARDT-DenseNet model with the ISIC 2016 and ISIC 2017 datasets. Our method achieves an ACC of 85.7% and an AUC of 83.7% in skin lesion classification with ISIC 2016 and an average AUC of 91.8% in skin lesion classification with ISIC 2017. The experimental results show that the method proposed in this paper has achieved a significant improvement in skin lesion classification, which is superior to that of the state-of-the-art method.

1. Introduction

Melanoma is one of the deadliest skin cancers in the world [1]. It develops from the pigment-producing cell and typically occurs in the skin but is rarely found in other parts of the human body, such as the mouth, nose and eyes. Malignant melanoma can evolve congenitally, be acquired from benign melanocytic nevus, evolve malignantly from a dysplastic nevus or be new growth. In recent years, the incidence and mortality of malignant melanoma have increased annually, and the age of death is much younger compared with other types of cancers. Though melanoma is a deadly disease, it has a high probability of being cured if it is found and detected early [2].
Dermoscopy [3] is an effective inspection method to detect skin lesions. By eliminating the reflection on the skin surface and optically amplifying a specific area by many times, it is possible to observe the skin underneath the epidermis and the junction of the real epidermis and the upper dermis, which are structures usually invisible to the naked eye.
Dermoscopy for the observation of pigmented skin lesions, vascular diseases and inflammatory diseases, for the auxiliary diagnosis and differential diagnosis of benign and malignant tumors of the skin and for non-invasive skin monitoring and remote medical treatment avoids unnecessary skin biopsies to a certain extent. It provides patients with a new, non-invasive skin detection method and greatly reduces the pain from an invasive examination. It is also capable of assisting symptom diagnosis and helping doctors to evaluate the disease and analyze the cause.
In practice, the accurate classification of skin lesions in dermoscopy images remains as a challenging problem due to the large scale of intra-class variation and inter-class similarity caused by the differences in imaging modalities and clinical pathologies and the limited ability to focus on semantic regions [4,5].
Deep learning-based methods for skin lesion classification have been broadly accepted in recent years. Convolutional neural networks (CNN) have brought a series of breakthroughs to image classification. The depth of the network is one of the critical factors in the design of the CNN architecture. In general, we usually increase the depth of the network and then expect a better feature expression capability and an enhanced prediction accuracy.
A deep convolutional neural network (DCNN) trained for classification has significant localization ability and can highlight the image recognition area [6]. However, the deepening of the network will cause gradient vanishing or exploding, which hinders the convergence of DCNN and its training process. In recent research, this problem has been solved by normalization initialization [7,8,9], which enables networks with dozens of layers to begin to converge to random gradients using stochastic gradient descent (SGD) [10]. He et al. [11] introduced a deep residual learning framework to solve the degradation problem, making it possible to build very deep residual networks, such as AlexNet [12], VGGNet [13], GoogLeNet [14], Inception [15] and so on. Later on, DenseNet [16] proposed the idea of densely connected layers, wherein each layer can obtain the connection feature map of the previous layer. These two main ideas are applied to different application scenarios and network architectures [17,18,19,20].
The attention mechanism stems from the study of human vision. Humans selectively focus on certain visible information of interest and ignore what they feel is irrelevant. This allocates available processing resources to the most informative part of the input signal. Typically, it is implemented in combination with gated functions (such as softmax or sigmoid) technology. In recent years, attention mechanisms have been widely used in deep learning applications, such as image and video captioning [21], Visual Question Answering (VQA) [22], machine translation [23] and image analysis [24,25,26,27,28]. One study [28] proposed an VGGNet-based method combined with an attention module. The attention map is estimated by the attention module, which highlights the image area of interest related to the classification of the skin lesion. There are many other methods that adopt attention mechanisms to classify skin lesions [23,29]. In image classification, although the attention mechanism effectively improves the performance of the model, the attention weights in these models are learned through a learnable layer with a large number of additional parameters, which leads to increased computational cost.
To address these problems, we propose a skin lesion classification model based on an attention mechanism and the residual learning of densely connected convolutional neural networks. One study [30] proposed a densely connected residual block to ensure the deep supervision of the convolutional neural network, effective gradient flow and feature reuse capabilities. Another study [31] proposed an attention model for skin lesion classification without introducing additional parameters. Inspired by these studies [30,31], the improved densely connected network added the attention and residual learning after each dense block and transition block. The proposed method can reduce the number of parameters and calculation operations, while achieving feasible accuracy without using any special hardware-software equipment to reduce the cost of training and inference processes. Joining residual learning helps to reverse the gradient information flow of the deep network and to strengthen the deep supervision of the network. The attention module learned together with other network parameters can adaptively focus on the distinguishing parts of skin lesions, highlighting the image regions of interest related to skin lesion classification. It uses the inherent self-attention ability of the convolutional neural network and uses the feature map obtained by the high layer as the low attention mask, rather than learning attention masks with extra layers. We evaluated our method with the ISIC 2016 [32] and ISIC 2017 datasets [33]. There are two binary classification tasks: one is to classify the lesion type as melanoma or other and one is to classify it as seborrheic keratosis or other. The experimental results show that the method proposed in this paper yields state-of-the-art classification performance in the classification of skin lesions.
This paper makes the following contributions:
  • We propose a new densely connected convolutional network with an attention and residual learning (ARDT-DenseNet) model to precisely classify skin lesions in dermoscopy images. The model consists of multiple ARDT blocks, a global average pooling layer and a classification layer. Each ARDT block is combined with dense blocks, transition blocks and an attention residual learning module. Residual learning prevents the gradient from vanishing when the network becomes deeper. The attention map automatically highlights the image area related to the classification, with additional interpretable information over a class label.
  • Compared to the hundreds of layers of the residual network, the number of network model parameters with hundreds of layers of densely connected networks is less than half, reducing the computing costs while the accuracy of skin lesion recognition is comparable.
  • The proposed model is an end-to-end network architecture. Our method has achieved state-of-the art performance for skin lesion classification in the ISIC 2016 and ISIC 2017 datasets.
  • Next, the design and architecture of our ARDT-DenseNet model will be introduced in detail in Section 2. Section 3 is the experimental process and results, and Section 4 is the discussion and conclusions. Finally, Section 5 is about future work.

2. Materials and Methods

2.1. Network Architecture

The proposed ARDT-DenseNet model consists of multiple ARDT blocks, a global average pooling layer and a classification layer. Each ARDT block is a combination of dense blocks, transition blocks and attention and residual modules. The proposed architecture in this paper is shown in Figure 1, and the inner details of the dense blocks and transition blocks are shown in Figure 2.
The depth of DenseNet100 30 is 100 layers, which has nearly half of the model parameters of ResNet101. We adopt DenseNet100 with an improved first convolutional layer. In the original DenseNet, Conv1 has a convolutional kernel with dimensions of 7 × 7 and stride of 2 and dimensions of 1 × 1 and 2 × 2 in the transform block. The transition block was mainly composed of a convolutional layer and an average pooling layer, and we changed the output channel of the transition block to be consistent with the input channel of the current ARDT block. It is worth noting that each convolutional operation inspired by Reference [31] uses a pre-activation method, which first includes regularization processing, then activation and finally convolution.
In order to adapt to the skin lesion classification and expand the receptive field of the convolutional operation to obtain more information, the Conv1 was replaced with a 3 × 3 convolution, and the convolution layer had 4k (k is growth rate) convolution kernels of size 3 × 3 with padding and a stride of 1. Each dense block has 12 layers of 1 × 1 and 3 × 3 convolutions. The growth rate of the feature map following each layer is set to 12, and there are four dense blocks in total, as shown in Table 1.
As shown in Figure 2, the transition blocks are composed with a 1 × 1 convolutional layer and a 2 × 2 average pooling layer. The 1 × 1 convolution in each transition block comes with a 4 × 4 convolutional kernel, which, in order to adapt to the recognition of skin lesions, increases the receptive field of the convolution process and obtains more information. After four dense blocks, a global average pooling layer is followed. Since the purpose of the paper is to deal with the binary classification of skin lesions (melanoma vs. others and seborrheic keratosis vs. others.), this paper uses a fully connected layer softmax function that has 2 neurons. Moreover, our ARDT-DenseNet model reduces vanishing-gradient and strengthens the delivery of the feature, and the feature is used more effectively. At the same time, the number of parameters is reduced to a certain extent.
It is worth mentioning that the convolutional network generally adopts the order of convolutional operation, batch regularization and activation functions, while the convolutional operation in the dense blocks and transition blocks here uses the way of pre-activation, with batch regularization first, then the activation function and finally the convolutional operation. In the model proposed in this paper, each dense block uses 12 layers. There is a total number of five dense blocks and one hundred convolutional layers in depth.
The degradation is not caused by overfitting, and adding more layers to a suitable depth model will lead to higher training errors. We introduced residual learning, as shown in Figure 3, to avoid this problem. Rather than introducing a new convolutional layer, residual learning is added after each dense block and transition block. Adding residual learning increases the information transmission of the effective gradient stream, which is more conducive to the training of deep networks. Since the resolution of the image is doubled after the transition block, the corresponding residual learning uses a 2 × 2 average pooling layer, which reduces the resolution of the original image by half.
Not all of the regions in the image have equal significance to the classification. Only those areas related to the task need to be concerned and to gain attention from the network. This learning mechanism enhances the propagation of gradient information, is conducive to the training of deep network models and strengthens the networks attention to the critical semantic information of lesion images. The attention mechanism usually uses additional trainable layers to learn attention weights. Inspired by Reference [31], the attention mechanism uses the inherent self-attention ability of DCNNs. This not only reduces the number of parameters of the network but also makes the network pay more attention to the semantically important fields and speeds up the training of the network. Finally, the residual learning is combined with the self-attention learning mechanism to improve the ability of network identification, enhance the transmission of gradient information in the process of network training and realize deep supervision. We chose the spatial attention mechanism in the paper to generate an attention mask from a high-level attention map. The specific structure of the introduced attention and residual learning is shown in Figure 3. It is worth noting not only that our attention module does not introduce additional parameters but also that our overall structure reduces parameters compared with Reference [30].
Figure 3. Architecture of ARDT Blocks. The “⊗ “represents matrix multiplication.
Figure 3. Architecture of ARDT Blocks. The “⊗ “represents matrix multiplication.
Sensors 20 07080 g003
Suppose the input of the ARDT Block is recorded as x l 1 , the result of the residual learning output is recorded as G x l 1 and the outputs of the dense block and the transition block are recorded as F x l 1 .
The normalization function is shown as Equation (1):
N s p F x l - 1   = e F x l - 1 i , j c i j e F x l - 1 i j c
where i and j represent the spatial position information, c represents the output channel index of F x l - 1 and N s p · uses the softmax function to emphasize the important areas in each channel. The attention mask is N s p F x l - 1 . The calculation method of the attention feature map W is defined as:
W   =   N s p F x l - 1 · G x l - 1
The entire convolutional block learning is a combination of residual learning and attention mechanism learning. That is, the identity mapping is added at the element level.
Finally, the residual feature mapping of densely connected networks and the feature mapping of the attention mechanism are defined as follows:
x l   =   F x l - 1 + G x l - 1 + b W
where b is a hyperparameter to trade-off between the feature map of the attention mechanism and the other two feature maps and W is the output of attention learning.

2.2. Loss Function

In the model training process, two binary classifiers are trained to discriminate between melanoma and others and between seborrheic keratosis and others. Since both models solve the problem of binary classification in the training process, this paper does not use the cross-entropy loss function of multi-classification. The cross-entropy loss function specifically for binary classification is defined as follows:
L l o s s   =   w i [ y i l o g x n + ( 1 y i ) l o g 1 x i ]
where x i represents the probability that the i-th lesion image is predicted as a positive example by the model, y i represents the real label of the i-th image and w i indicates the weight.

3. Experiments and Results

3.1. Implementation Details

  • Data preparation
The dataset used in this paper is hosted by the International Skin Imaging Collaboration (ISIC) 2016 and 2017. The International Skin Imaging Collaboration (ISIC) is an international effort to improve melanoma diagnosis, sponsored by the International Society for Digital Imaging of the Skin (ISDIS). The organization provides the largest publicly available dermoscopy image dataset of skin lesions.
The proposed model was trained and evaluated using two different datasets: two types of ISIC 2016 and three types of International Standard Industrial Classification 2017, as shown in Figure 4 and Figure 5 correspondingly.
ISIC 2016 consists of 900 training images and 379 test images. The lesion images are divided into two categories, malignant and benign.
ISIC 2017 provides 2000 training images, 150 validation images and 600 test images. The number of various types of lesions in the training set, validation set and test set is shown in Table 2. The lesion images break down into three categories: melanoma, nevus and seborrheic keratosis. Melanoma is commonly known as malignant skin cancer; it is derived from melanocytes. Nevus is benign skin cancer, derived from melanocytes, and seborrheic keratosis is benign skin cancer, derived from keratinocytes (non-melanocytes). There are two binary classification subtasks on this dataset: the classification of melanoma (melanoma vs. others) and the classification of seborrheic keratosis (seborrheic keratosis vs. others).
As presented in Table 2, there is a severe class imbalance in the dataset. Therefore, we adopt an oversampling method to address this problem. By this method, we generate samples of the minority class. We repeat the samples of each class in the training set and randomly sample from the samples of the minority class to add new samples so that all classes have an equal number of samples for training. We randomly select patches of different scales (1/5, 2/5, 3/5 and 4/5 of the original image) from the center of each image in the training set and then adjust to 224 × 224 by using bilinear interpolation. Each size operation generated 15 pictures, and each training image generated a total of 60 pictures. For example, the ISIC 2017 training set has a total of 2000 images, so the training data set was expanded by 60 times, with a total of 12,000 pictures. Data augmentation includes random cropping, random rotation and flipping.
The training set contains skin lesion images in JPEG format and CSV files. The CSV file contains some clinical metadata for each image. The first column of each row contains a string in the form of ISIC_<image_id>, where <image_id> matches the corresponding training data image. The second column in each row belongs to the first task of classification: malignant (melanoma) vs. benign (seborrheic keratosis/nevus), containing 0 (the lesion is melanoma) or 1 (the lesion is nevus or seborrheic keratosis). In ISIC 2017, the third column in each row belongs to the second classification task: melanocytes (melanoma) vs. non-melanocytes (seborrheic keratosis/nevus), containing 0 (seborrheic keratosis) or 1 (melanoma/nevus).
  • Network training
During the training process, the batch size was set to 10, the initial learning rate was 0.0001 and the maximum epoch was set to 100. The learning rate was reduced by half every 30 epochs. The network was adopted by the SGD as the optimizer, the attenuation of the weights was set to 0.0001 and the momentum was set to 0.9. The initial weighting factor for attention learning was set to 0.001. Then the official validation set was used to verify the performance of the model. In the test phase, we also used the official test set and selected a patch from each test image in the same way as in the training process and input these into the trained network to obtain the prediction score of the image. The unknown skin lesion dermoscopy image was fed to the trained model to classify the type of lesions.
The aim of this paper is to implement our model based on densely connected network attention and residual learning for skin lesion classification in Python under Pytorch.
  • Metrics
To evaluate the performance, we use average precision (AP), classification accuracy (ACC), sensitivity (SE), specificity (SP) and the area under the ROC curve (AUC) as the metrics in this paper.
A C C   =   T P + T N T P + F N + T N + F P
S e n s i t i v i t y   =   T P T P + F N
S p e c i f i c i t y   =   T N T N + F P
where:
  • TP is true positive. It is judged as a positive sample; in fact it is also a positive sample.
  • TN is true negative. It is judged as a negative sample; in fact it is also a negative sample.
  • FN is false negative. It is judged as a negative sample, but in fact it is a positive sample.
  • FP is false positive. It is judged as a positive sample, but in fact it is a negative sample.

3.2. Results

The proposed method is compared with several baseline models for classification in ISIC 2016 and ISIC 2017, referring to Table 3. (1) VGGNet has 16 layers of convolution, and the size of the convolution kernel is 3 × 3. (2) GoogLeNet is a convolutional neural network with 21 layers, part of which is composed of inception modules. The network has only one layer of fully connected convolutional neural network. (3) ResNet network compares the 50-layer and 100-layer residual network models. (4) DenseNet is the backbone of our model.
This paper compares our model with these methods, illustrated in Table 4 and Table 5.

3.2.1. Compared with VGGNet, GooLeNet, ResNet and DenseNet in ISIC 2016

It can be seen from Table 3, VGGNet [13], GoogLeNet [14], ResNet [11] and DenseNet [16] that the depth of the convolutional neural network model is gradually increasing. The initial 16 layers gradually increased to hundreds of layers. Table 3 shows the number of convolution layers, the number of fully connected layers, the size of the convolution kernel and the size of the fully connected layers for each model.
Table 4 presents a comparison of model performance by applying VGGNet, GoogLeNet, ResNet50, ResNet101 and DenseNet100 models to the ISIC 2016 test set. The skin lesions on 379 test images are evaluated through SE, SP, ACC and AUC.
The classification performance can be seen so that when the network is deeper, the performance of the model becomes better across all metrics. It also shows that the proposed ARDT-DenseNet model achieves the best performance in skin lesion classification over other methods.

3.2.2. Compared with VGGNet, GooLeNet, ResNet and DenseNet in ISIC 2017

These models are in Task 1 (malignant (melanoma) vs. benign (seborrheic keratosis/nevus)) and Task 2 (melanoma/nevus (melanocytes) vs. non-melanocytes (seborrheic keratosis)) in ISIC 2017. It can be seen from Table 5 that the number of network layers increases, and the accuracy of each model also gradually increases. The accuracy of the original 16-layer VGGNet model in Tasks 1 and 2 is 75.7% and 74.6%, respectively. Once the number of layers increased to 22, the accuracy of GoogLeNet is 76.4% and 75.8%, respectively. When the number of network layers is increased to 50, the accuracy rate of ResNet50 is increased by 7.1% and 8.0%, respectively.
It can be seen that the performance of the network model grows better with the increase of the number of convolution layers. This paper further studies that, when the number of network layers reaches the scale of hundreds, the accuracy rate of ResNet101 in Tasks 1 and 2 is 83.9% and 84.3%, respectively, and the accuracy of DenseNet100 reaches 84.2% and 85.0%, respectively. Compared with the 50-layer residual network, the 200-layer network models have greatly improved accuracy.
In addition, we also compared the number of parameters of these models. The number of parameters of the proposed VGGNet model reaches at most 138 × 106, but the accuracy rate is only 75.7%to 74.6%. The optimized GoogLeNet not only improves its accuracy to 76.4% to 75.8% but also reduces the model parameters by about 20 times. Although the number of ResNet50 convolutional layers has increased, the number of parameters is only 25.5 × 106, and the accuracy rate has reached 83.5%. ResNet101 not only has hundreds of layers but also has relatively large improvement in AUC, ACC, SE and SP over ResNet50. However, the number of parameters is twice that of ResNet50.
However, the DenseNet100 network model, which also has hundreds of layers, not only compares with ResNet101 in each evaluation index but also has only 21.2 × 106 parameters, which is lower than the number of parameters of ResNet50 (25.5 × 106). However, the number of convolutional layers is twice that of ResNet50. It can be seen that DenseNet100 has achieved relatively competitive results when the model sizes are not that different. Therefore, we chose DenseNet100 as the backbone network model.
In this paper, after determining the reference network model as DenseNet100, the method for attention and residual learning is introduced into the reference network model, called ARDT-DenseNet. The results of the evaluation in the two classifications are shown in Table 5. It is worth noting that the changes in the transition block not only do not introduce additional layers but also reduce the number of parameters of our model. After introducing attention and residual learning, the number of model parameters is still 20.3 × 106. It can be clearly seen that, after the introduction of the new method, the ACC values of ARDT-DenseNet in Task 1 and Task 2 are 86.8% and 87.9%, respectively, which are 2.6% and 1.8% higher than for DenseNet100, respectively. The AUC values of ARDT-DenseNet in Task 1 and Task 2 are 87.9% and 95.7%, respectively, which are higher than those of DenseNet100.
In Task 1, ARDT-DenseNet was used to identify whether the skin of the lesion was melanoma (malignant) or seborrheic keratosis/nevus (benign). The value of AUC in Task 1 was 87.9%, which was 1.8% higher than that of DenseNet100. As shown in Figure 6, the red and black lines represent the result curves of ARDT-DenseNet and DenseNet100, respectively. In Task 2, the models are used to identify whether the skin lesion is non-melanocytes (seborrheic keratosis) or melanocytes (melanoma/nevus). The values of AUC in Task 2 are 93.9% and 95.7%, respectively. ARDT-DenseNet is improved over DenseNet100 by 1.8%. As shown in Figure 7, the pink and black lines represent the result curves of ARDT-DenseNet and DenseNet100, respectively.
The ARDT-DenseNet network model has made progress in both tasks of skin lesion classification, but the number of parameters has not increased relative to DenseNet100. This is because the attention learning introduced in ARDT-DenseNet does not introduce additional convolutional layers.

3.2.3. ARDT-DenseNet Compared with State-of-the-Art Models and Top-Ranking Challenge Models in ISIC 2017

Our model is also compared with advanced skin lesion classification methods. Table 6 presents the performance comparison of the models by applying the advanced model ARL-CNN [31] and the top models of the challenge rankings in the ISIC-17 competition leaderboard of the ISIC 2017 test set. The performance of the dermoscopy images is evaluated through the SE, SP, ACC and AUC of the classification performance for skin lesions.
It is worth noting that the proposed ARDT-DenseNet model without external data obtained the highest mean AUC and the highest SE, ACC and AUC values in Task 1, and, in Task 2, it obtained the highest value for SE. For Reference [34] and Reference [35], although they have a high SP value in Task 1, indicating a high probability of not being misdiagnosed, their SE is 10.3% and 43.6%, respectively, which means low sensitivity and that, when diagnosing disease, the probability of not being missed is low. For a single metric, if the sensitivity of the diagnosis is increased, the specificity of the diagnosis will inevitably decrease. In other words, reducing missed diagnoses will inevitably increase misdiagnoses. Our method can achieve a good balance between the two metrics.

4. Discussion and Conclusions

Our model is also compared with advanced ensemble models of skin lesion classification, as shown in Table 7. As a matter of fact, these models cannot be compared directly with our model because of the ensemble strategy. However, to compare the state-of-the art methods with advanced performance and enough comparison information, we provide the ensemble models as a reference in Table 7. It can be seen that the ensemble model has superior performance on the skin lesion classification task. However, we can see that the SE of Task 1 in Reference [36] is 37.6%, which means that, when diagnosing disease, the probability of not being missed is low. Moreover, the ensemble model has a complex structure, which is usually difficult to train and time-consuming and requires more resources.
This paper studies the classification of skin lesions. Due to the high degree of similarity between different levels of skin lesion types, it has brought certain challenges to the classification of skin lesions. The deep convolutional neural network significantly improved the classification ability of the model, but it also brings a dramatic increment in the number of model parameters, introduces the computation cost, makes the model more difficult to train and therefore affects the result. For the densely connected network with 100 layers proposed in this paper, the number of parameters is greatly reduced, whereases the accuracy of skin lesion classification is improved compared with a general network with the same number of layers. The introduction of attention and residual learning not only avoids additional parameters but also strengthens the deep supervision of the network and improves the networks ability to identify the interested subareas of the image. In this paper, two classification tasks are separately tested on the dataset. Compared with other methods, the ARDT-DenseNet proposed in this paper achieved state-of-the-art results for skin lesion classification in the ISIC 2016 and ISIC 2017 datasets.

5. Future Work

Although the method has made some progress in the task of automatic classification of skin lesions in dermoscopy images, there is still space for improvement in research work. The problems in the classification of skin lesion images need to be further improved through research. For example, in the new ISIC 2018 dataset, lesion images are divided into seven categories in more detail, which requires further research on multi-classification issues. In future research, we will pay more attention to the most advanced classification models in deep learning and develop higher-performance network structures. In addition, we will explore whether the model proposed in this paper can be applied to other types of skin lesion recognition and to more fields.

Author Contributions

J.W., W.T. and X.L.; methodology, W.H.; software, J.W.; validation, J.W.; formal analysis, J.W.; investigation, J.W.; resources, J.W.; data curation, J.W.; writing—original draft preparation, J.W. and Y.W.; writing—review and editing, J.W.; visualization, W.H.; supervision, W.H.; project administration. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by The National Natural Science Foundation of China, grant number U1803262.

Conflicts of Interest

The authors have no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Siegel, R.L.; Miller, K.D.; Jemal, A. Cancer statistics, 2017. CA Cancer J. Clin. 2017, 60, 277–300. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Apalla, Z.; Nashan, D.; Weller, R.B.; Castellsagué, X. Skin cancer: Epidemiology, disease burden, pathophysiology, diagnosis, and therapeutic approaches. Dermatol. Ther. 2017, 7, 5–19. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Kittler, H.; Pehamberger, H.; Wolff, K.; Binder, M. Diagnostic accuracy of dermoscopy. Lancet Oncol. 2002, 3, 159–165. [Google Scholar] [CrossRef]
  4. Song, Y.; Cai, W.; Huang, H.; Zhou, Y.; Feng, D.D.; Wang, Y.; Fulham, M.J.; Chen, M. Large Margin Local Estimate With Applications to Medical Image Classification. IEEE Trans. Med. Imaging 2015, 34, 1362–1377. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Yu, L.; Chen, H.; Dou, Q.; Qin, J.; Heng, P.A. Automated Melanoma Recognition in Dermoscopy Images via Very Deep Residual Networks. IEEE Trans. Med. Imaging 2017, 36, 994–1004. [Google Scholar] [CrossRef] [PubMed]
  6. Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Learning deep features for discriminative localization. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2921–2929. [Google Scholar]
  7. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics (AISTATS 2010), Sardinia, Italy, 13–15 May 2010. [Google Scholar]
  8. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on ImageNet classification. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  9. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on International Conference on Machine Learning (ICML 2015), Lille, France, 6–11 July 2015. [Google Scholar]
  10. LeCun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation Applied to Handwritten Zip Code Recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  11. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  12. Krizhevsky, A.; Sutskever, I.; Hinton, G. ImageNet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2012; pp. 1106–1114. [Google Scholar]
  13. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  14. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  15. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  16. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar]
  17. Lopez, A.R.; Giro-I-Nieto, X.; Burdick, J.; Marques, O. Skin lesion classification from dermoscopic Images using deep learning techniques. In Proceedings of the 13th IASTED International Conference on Biomedical Engineering (BioMed), Innsbruck, Australia, 20–21 February 2017. [Google Scholar]
  18. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Corrigendum: Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115. [Google Scholar] [CrossRef] [PubMed]
  19. Zhang, J.; Xie, Y.; Wu, Q.; Xia, Y. Skin lesion classification in dermoscopy images using synergic deep learning. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI 2018), Granada, Spain, 16–20 September 2018; pp. 12–20. [Google Scholar]
  20. Zhang, J.; Xie, Y.; Wu, Q.; Xia, Y. Medical image classification using synergic deep learning. Med. Image Anal. 2019, 54, 10–19. [Google Scholar] [CrossRef] [PubMed]
  21. Xu, K.; Ba, J.L.; Kiros, R.; Cho, K.; Courville, A.; Salakhutdinov, R.; Zemel, R.S.; Bengio, Y. Show, attend and tell: Neural image caption generation with visual attention. In Proceedings of the 32nd International Conference on Machine Learning (ICML 2015), Lille, France, 6–11 July 2015; pp. 2048–2057. [Google Scholar]
  22. Krishna, R.; Zhu, Y.; Groth, O. Visual Genome: Connecting Language and Vision Using Crowdsourced Dense Image Annotations. Int. J. Comput. Vis. 2017, 123, 32–73. [Google Scholar] [CrossRef] [Green Version]
  23. Bahdanau, D.; Cho, K.; Bengio, Y. Neural machine translation by jointly learning to align and translate. arXiv 2014, arXiv:1409.0473. [Google Scholar]
  24. Jetley, S.; Lord, N.A.; Lee, N.; Torr, P.H.S. Learn To Pay Attention. arXiv 2018, arXiv:1804.02391. [Google Scholar]
  25. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual attention network for image classification. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6450–6458. [Google Scholar]
  26. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  27. Chen, L.; Zhang, H.; Xiao, J.; Nie, L.; Shao, J.; Liu, W.; Chua, T.-S. SCACNN: Spatial and Channel-wise Attention in Convolutional Networks for image captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5659–5667. [Google Scholar]
  28. Yan, Y.; Kawahara, J.; Hamarneh, G. Melanoma recognition via visual attention. In International Conference on Information Processing in Medical Imaging; Springer: Cham, Switzerland, 2019; pp. 793–804. [Google Scholar]
  29. Gessert, N.; Sentker, T.; Madesta, F.; Schmitz, R.; Kniep, H.; Baltruschat, I.; Werner, R.; Schlaefer, A. Skin lesion classification using cnns with patch-based attention and diagnosis-guided loss weighting. IEEE Trans. Biomed. Eng. 2019, 67, 495–503. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Fooladgar, F.; Kasaei, S. Lightweight Residual Densely Connected Convolutional Neural Network. arXiv 2020, arXiv:2001.00526. [Google Scholar] [CrossRef]
  31. Zhang, J.; Xie, Y.; Xia, Y.; Shen, C. Attention Residual Learning for Skin Lesion Classification. IEEE Trans. Med. Imaging 2019, 38, 2092–2103. [Google Scholar] [CrossRef] [PubMed]
  32. Gutman, D.; Codella, C.F.; Celebi, E.; Helba, B.; Marchetti, M.; Mishra, N.; Halpern, A. Skin Lesion Analysis toward Melanoma Detection: A Challenge at the International Symposium on Biomedical Imaging (ISBI) 2016, hosted by the International Skin Imaging Collaboration (ISIC). arXiv 2016, arXiv:1605.01397. [Google Scholar]
  33. Codella, N.C.; Gutman, D.; Celebi, M.E.; Helba, B.; Marchetti, M.A.; Dusza, S.W.; Kalloo, A.; Liopyris, K.; Mishra, N.; Kittler, H.; et al. Skin lesion analysis toward melanoma detection: A challenge at the 2017 international symposium on biomedical imaging (ISBI), hosted by the international skin imaging collaboration (ISIC). arXiv 2017, arXiv:1710.05006. [Google Scholar]
  34. Díaz, I.G. Incorporating the knowledge of dermatologists to convolutional neural networks for the diagnosis of skin lesions. arXiv 2017, arXiv:1703.01976v1. [Google Scholar]
  35. DeVries, T.; Ramachandram, D. Skin lesion classification using deep multi- scale convolutional neural networks. arXiv 2017, arXiv:1703.01402v1. [Google Scholar]
  36. Tang, P.; Liang, Q.; Yan, X.; Xiang, S.; Zhang, D. GP-CNN-DTEL: Global-Part CNN Model with Data-Transformed Ensemble Learning for Skin Lesion Classification. IEEE J. Biomed. Health Inform. 2020, 24, 2870–2882. [Google Scholar] [CrossRef] [PubMed]
  37. Mahbod, A.; Schaefer, G.; Wang, C.; Ecker, R.; Dorffner, G.; Ellinger, L. Investigating and Exploiting Image Resolution for Transfer Learning-based Skin Lesion Classification. arXiv 2020, arXiv:2006.14715. [Google Scholar]
Figure 1. The architecture of the proposed network. The backbone network is DenseNet100. Dense blocks and transition blocks are described in Figure 2. ARDT blocks as shown in Figure 3.
Figure 1. The architecture of the proposed network. The backbone network is DenseNet100. Dense blocks and transition blocks are described in Figure 2. ARDT blocks as shown in Figure 3.
Sensors 20 07080 g001
Figure 2. Architecture of the dense blocks and transition blocks. The “⊕” represents element-wise addition.
Figure 2. Architecture of the dense blocks and transition blocks. The “⊕” represents element-wise addition.
Sensors 20 07080 g002
Figure 4. The International Skin Imaging Collaboration (ISIC) 2016 dataset contains two types of skin lesions: malignant (melanoma) (a) and benign (b). The lesions in these images are matched with the gold standard (definitive) malignant diagnosis, including 900 training images and 379 test dermoscopy images, which have been screened to ensure privacy and quality.
Figure 4. The International Skin Imaging Collaboration (ISIC) 2016 dataset contains two types of skin lesions: malignant (melanoma) (a) and benign (b). The lesions in these images are matched with the gold standard (definitive) malignant diagnosis, including 900 training images and 379 test dermoscopy images, which have been screened to ensure privacy and quality.
Sensors 20 07080 g004
Figure 5. The lesion images in the ISIC 2017 dataset of the International Skin Imaging Collaboration are divided into three categories: melanoma (a), nevus (b) and seborrheic keratosis (c). This includes 2000 training, 150 verification and 600 test dermoscopy lesion images, which are all paired with gold standard diagnoses and ensure privacy and quality.
Figure 5. The lesion images in the ISIC 2017 dataset of the International Skin Imaging Collaboration are divided into three categories: melanoma (a), nevus (b) and seborrheic keratosis (c). This includes 2000 training, 150 verification and 600 test dermoscopy lesion images, which are all paired with gold standard diagnoses and ensure privacy and quality.
Sensors 20 07080 g005
Figure 6. The red and black lines represent the result curves of ARDT-DenseNet100 and DenseNet100, respectively.
Figure 6. The red and black lines represent the result curves of ARDT-DenseNet100 and DenseNet100, respectively.
Sensors 20 07080 g006
Figure 7. The pink and blue lines represent the result curves of ARDT-DenseNet and DenseNet100, respectively.
Figure 7. The pink and blue lines represent the result curves of ARDT-DenseNet and DenseNet100, respectively.
Sensors 20 07080 g007
Table 1. The detailed structure of the DenseNet100 network.
Table 1. The detailed structure of the DenseNet100 network.
LayersInput SizeOutput SizeOperations
Conv1224 × 224112 × 1123 × 3 conv
Pooling112 × 11256 × 562 × 2 Avg pool
Dense Block 156 × 5656 × 561 × 1 conv, 3 × 3 conv, 12
Transition layer 156 × 5628 × 281 × 1 conv, 2 × 2 Avg pool
Dense Block 228 × 2828 × 281 × 1 conv, 3 × 3 conv, 12
Transition layer 228 × 2814 × 141 × 1 conv, 2 × 2 Avg pool
Dense Block 314 × 1414 × 141 × 1 conv, 3 × 3 conv, 12
Transition layer 314 × 147 × 71 × 1 conv, 2 × 2 Avg pool
Dense Block 47 × 77 × 71 × 1 conv, 3 × 3 conv, 12
GAP7 × 71 × 1GAP
FC layer1 × 11 × 1FC
softmax
Table 2. The number of various lesion types in the training set, validation set and test set.
Table 2. The number of various lesion types in the training set, validation set and test set.
MelanomaSeborrheic KeratosisNevusTotal
Training set37425413722000
Validation set304278150
Test set11790393600
Total52138618432750
Table 3. The architecture of the baseline model.
Table 3. The architecture of the baseline model.
VGGNetGoogLeNetResNet50ResNet101DenseNet100
Layer162250101100
Conv13214910099
FC31111
Convolutional kernel37,2,3,57,1,37,1,34,4,4
Table 4. Comparison of baseline models on ISIC 2016.
Table 4. Comparison of baseline models on ISIC 2016.
ModelsSESPACCAUC
VGGNet0.7680.6780.8120.802
GoogLeNet0.7700.6890.8170.816
ResNet500.7990.7140.8300.782
ResNet1010.8040.7390.8340.821
DenseNet1000.8120.7420.8420.817
ARDT-DenseNet0.8160.7560.8570.837
Table 5. Comparison of baseline models on ISIC 2017.
Table 5. Comparison of baseline models on ISIC 2017.
ModelsTask 1Task 2 Params   ( ×   10 6 )
SESPACCAUCSESPACCAUC
VGGNet0.5460.8010.7570.7420.8090.7230.7460.832138
GoogLeNet0.5910.7950.7640.7670.8240.7880.7580.84711.3
ResNet500.6280.8820.8350.8520.8650.8370.8380.94225.5
ResNet1010.6370.8900.8390.8610.8710.8490.8430.95052.3
DenseNet1000.6410.8870.8420.8610.8690.8520.8500.93921.2
ARDT-DenseNet0.6680.8960.8680.8790.8870.8730.8780.95720.3
Table 6. ARDT-DenseNet compared with state-of-the-art models and top-ranking models in ISIC 2017.
Table 6. ARDT-DenseNet compared with state-of-the-art models and top-ranking models in ISIC 2017.
ModelsExternal DataEnsemblesTask 1Task 2Mean AUC
SESPACCAUCSESPACCAUC
ARL-CNN [31]1320N0.6580.8960.8500.8750.8780.8670.868 0.958 0.917
#2 [34]900N0.103 0.998 0.8230.8560.178 0.998 0.875 0.965 0.911
#6 [35]0N0.4360.9250.8300.8300.7000.995 0.917 0.9420.886
ARL-CNN [31]0N0.5900.8960.8370.8590.7780.9310.9080.9510.905
ARDT-DenseNet0N 0.668 0.896 0.868 0.879 0.887 0.873 0.878 0.957 0.918
Table 7. ARDT-DenseNet compared with state-of-the-art ensemble learning models in ISIC 2017.
Table 7. ARDT-DenseNet compared with state-of-the-art ensemble learning models in ISIC 2017.
ModelsEnsemblesTask 1Task 2Mean AUC
SESPACCAUCSESPACCAUC
Reference [36]Y0.3760.9650.8500.8910.7220.9730.9350.9600.926
Reference [37]Y---0.892---0.9660.929
ARDT-DenseNetN0.6680.8960.8680.8790.8870.8730.8780.9570.918
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wu, J.; Hu, W.; Wen, Y.; Tu, W.; Liu, X. Skin Lesion Classification Using Densely Connected Convolutional Networks with Attention Residual Learning. Sensors 2020, 20, 7080. https://doi.org/10.3390/s20247080

AMA Style

Wu J, Hu W, Wen Y, Tu W, Liu X. Skin Lesion Classification Using Densely Connected Convolutional Networks with Attention Residual Learning. Sensors. 2020; 20(24):7080. https://doi.org/10.3390/s20247080

Chicago/Turabian Style

Wu, Jing, Wei Hu, Yuan Wen, Wenli Tu, and Xiaoming Liu. 2020. "Skin Lesion Classification Using Densely Connected Convolutional Networks with Attention Residual Learning" Sensors 20, no. 24: 7080. https://doi.org/10.3390/s20247080

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop