Next Article in Journal
Recurrent NOMO1 Gene Deletion Is a Potential Clinical Marker in Early-Onset Colorectal Cancer and Is Involved in the Regulation of Cell Migration
Next Article in Special Issue
Mitochondrial Protein Cox7b Is a Metabolic Sensor Driving Brain-Specific Metastasis of Human Breast Cancer Cells
Previous Article in Journal
Targeted Therapy for Hepatocellular Carcinoma: Old and New Opportunities
Previous Article in Special Issue
Prediction of Breast Cancer Histological Outcome by Radiomics and Artificial Intelligence Analysis in Contrast-Enhanced Mammography
 
 
Correction published on 11 April 2023, see Cancers 2023, 15(8), 2237.
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Connected-SegNets: A Deep Learning Model for Breast Tumor Segmentation from X-ray Images

1
Department of Electrical Engineering, National Taipei University of Technology, Taipei 10608, Taiwan
2
Division of General Surgery, Cheng Hsin General Hospital, Taipei 112, Taiwan
3
Department of Communications, Navigation and Control Engineering, National Taiwan Ocean University, Keelung 202301, Taiwan
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Cancers 2022, 14(16), 4030; https://doi.org/10.3390/cancers14164030
Submission received: 18 July 2022 / Revised: 17 August 2022 / Accepted: 18 August 2022 / Published: 20 August 2022 / Corrected: 11 April 2023
(This article belongs to the Special Issue Updates on Breast Cancer)

Abstract

:

Simple Summary

The segmentation of breast tumors is an important step in identifying and classifying benign and malignant tumors in X-ray images. Mammography screening has proven to be an effective tool for breast cancer diagnosis. However, the inspection of breast mammograms for early-stage cancer can be a challenging task due to the complicated structure of dense breasts. Several deep learning models have been proposed to overcome this particular issue; however, the false positive and false negative rates are still high. Hence, this study introduced a deep learning model, called Connected-SegNets, that combines two SegNet architectures with skip connections to provide a robust model to reduce false positive and false negative rates for breast tumor segmentation from mammograms.

Abstract

Inspired by Connected-UNets, this study proposes a deep learning model, called Connected-SegNets, for breast tumor segmentation from X-ray images. In the proposed model, two SegNet architectures are connected with skip connections between their layers. Moreover, the cross-entropy loss function of the original SegNet has been replaced by the intersection over union (IoU) loss function in order to make the proposed model more robust against noise during the training process. As part of data preprocessing, a histogram equalization technique, called contrast limit adapt histogram equalization (CLAHE), is applied to all datasets to enhance the compressed regions and smooth the distribution of the pixels. Additionally, two image augmentation methods, namely rotation and flipping, are used to increase the amount of training data and to prevent overfitting. The proposed model has been evaluated on two publicly available datasets, specifically INbreast and the curated breast imaging subset of digital database for screening mammography (CBIS-DDSM). The proposed model has also been evaluated using a private dataset obtained from Cheng Hsin General Hospital in Taiwan. The experimental results show that the proposed Connected-SegNets model outperforms the state-of-the-art methods in terms of Dice score and IoU score. The proposed Connected-SegNets produces a maximum Dice score of 96.34% on the INbreast dataset, 92.86% on the CBIS-DDSM dataset, and 92.25% on the private dataset. Furthermore, the experimental results show that the proposed model achieves the highest IoU score of 91.21%, 87.34%, and 83.71% on INbreast, CBIS-DDSM, and the private dataset, respectively.

1. Introduction

The United States of America reported a total of 43,250 female deaths and 530 male deaths due to breast cancer in 2022 [1]. Researchers are motivated by these statistics to develop accurate tools for early breast cancer diagnosis, which will offer physicians more options for treatment. Mammograms are still being widely used to detect the presence of any abnormalities in breasts [2,3,4]. Mammogram images show different types of breast tissues as pixel clusters with different intensities [5]. These tissues include fiber-glandular, fatty, and pectoral muscle tissues [6]. On mammography, abnormal tissues such as lesions, tumors, lumps, masses, or calcifications may be indicators of breast cancer [7,8]. However, there is always the possibility of human error when analyzing and diagnosing breast cancer due to dense breasts and the high variability between patients [9,10,11]. Additionally, mammography screening sensitivity is affected by image quality and radiologist experience [12,13].
Automated techniques are being developed to analyze and diagnose breast mammograms with the goal of counteracting this variability and standardizing diagnostic procedures [14,15]. The rapid emergence of artificial intelligence (AI) and deep learning (DL) has significant implications for breast cancer diagnosis [16,17,18]. The advancements in image segmentation using convolutional neural networks (CNNs) have been applied to segment breast cancer from X-ray images [19,20,21,22,23]. The earlier works on mass segmentation faced some challenges, such as low signal to noise ratio, indiscernible mass boundaries, high false positives, and high false negative rates. To address these challenges, one study proposed a deeply supervised UNet model (DS U-Net) coupled with dense conditional random fields (CRFs) for lesion segmentation from whole mammograms [19]. The DS U-Net model has produced a Dice score of 79% on the INbreast dataset and 83% on the CBIS-DDSM dataset, whereas its IoU score is 83% and 86% on the INbreast and CBIS-DDSM datasets, respectively. Another study [20] proposed an attention-guided dense up-sampling network (AU-Net) for accurate breast mass segmentation from mammograms. An asymmetrical encoder–decoder structure is employed in this AU-Net and it uses an effective up-sampling block and attention-guided dense up-sampling block (AU block). The AU block is designed to have three merits. First, dense upsampling compensates for the information loss experienced during bilinear up-sampling. Second, it integrates high- and low-level features more effectively. Third, it highlights channels with rich information via the channel attention function. Compared to the state-of-the-art FCNs, AU-Net achieved the best performance, with a Dice score of 90% on the INbreast dataset and 89% on the CBIS-DDSM dataset.
However, such models do not capture the features of different scales of masses effectively, and therefore they suffer from low segmentation accuracy. Hence, a new model, called UNet, was presented to mitigate the limitations of the previous models [21]. UNet integrates the high-level features of the encoder with the low-level features of the decoder. Through skip connections, the UNet architecture was able to maintain this form of fusion for a variety of medical applications. The UNet architecture achieves better performance on different biomedical segmentation applications. Asma Baccouche et al. [22] introduced Connected-UNets to segment breast masses. This method integrated atrous spatial pyramid pooling (ASPP) in the two standard UNets. The architecture of Connected-UNets was built on the attention network (AUNet) and residual network (ResUNet). To augment and enhance the images, cycle-consistent generative adversarial networks (CycleGANs) were used between two unpaired datasets. Additionally, a regional deep learning approach called you-only-look-once (YOLO) has been used to detect breast lesions from mammograms. Finally, a full-resolution convolutional network (FrCN) has been implemented to segment breast lesions. The Connected-UNets model has produced a Dice score of 94% and 92% on the INbreast and CBIS-DDSM datasets, respectively. Moreover, it has achieved an IoU score of 90% and 86% on INbreast and CBIS-DDSM, respectively. Badrinarayanan et al. [23] proposed a practical deep fully convolutional neural network architecture for semantic pixel-wise segmentation, termed SegNet. Its segmentation architecture consists of an encoder network and a decoder network followed by a pixel-wise classification layer. Topologically, the architecture of the encoder network matches that of the 13 convolutional layers in the VGG16 network. The role of the decoder network is to map the low-resolution encoder feature maps to full-input-resolution feature maps for pixel-wise classification. The SegNet model has achieved satisfactory segmentation performance. However, since the SegNet architecture does not consist of skip connections, incorporating fine multiscale information during the training process is challenging.
This study combines the characteristics of the Connected-UNets and SegNet models to form Connected-SegNets from two standard SegNets with skip connections for breast tumor segmentation from breast mammograms. The flow chart of the proposed system is illustrated in Figure 1. The major contributions of this study include the following.
  • This study proposes a deep learning model called Connected-SegNets for breast tumor segmentation from X-ray images.
  • The proposed model, Connected-SegNets, is designed using skip connections, which helps to recover the spatial information lost during the pooling operations.
  • The original SegNet cross-entropy loss function has been replaced by the IoU loss function to overcome any noisy features and enhance the detection of the false negative and false positive cases.
  • The histogram equalization method of the contrast limit adapt histogram equalization (CLAHE) is applied to all datasets to enhance the compressed areas and smooth the pixel distribution.
  • Image augmentation methods including rotation and flipping have been used to increase the number of training data and to reduce the impact of overfitting.
The rest of this paper is organized as follows. Section 2 describes the datasets and architectural details of the proposed method. Section 3 presents the experimental results. Section 4 discusses the merits of this study. Finally, the article is concluded with its primary findings in Section 5.

2. Materials and Methods

This research uses the two publicly available datasets of INbreast and CBIS-DDSM, and one private dataset obtained from Cheng Hsin General Hospital in Taiwan. Initially, a histogram equalization, CLAHE, is applied to all datasets to enhance the compressed areas and smooth the pixel distribution. Then, each X-ray dataset is randomly divided into 70%, 15%, and 15% for training, validation, and testing, respectively. Finally, the training and validation samples are augmented to increase the amount of data before feeding them to the proposed Connected-SegNets model.

2.1. Datasets

The proposed model, Connected-SegNets, has been evaluated on the following datasets.

2.1.1. INbreast Dataset

The INbreast dataset is a collection of mammograms from Centro de Mama Hospital de S. João, Breast Centres Network, Porto, Portugal. A total of 410 images with 115 cases were collected from August 2008 to July 2010 [24,25], and 95 of 115 cancer cases involved both breasts in women. Four different types of breast diseases are recorded in the database, including calcification, mass, distortions, and asymmetries. This database includes images from craniocaudal (CC) and mediolateral oblique (MLO) perspectives. Moreover, the breast density is divided into four categories according to the breast imaging reporting and data system (BI-RADS) assessment categories, which are: entirely fat (BI-RADS 1), scattered fibroglandular (BI-RADS 2), heterogeneously dense (BI-RADS 3), and extremely dense (BI-RADS 4). All the images were saved in two sizes: 3328 × 4084 or 2560 × 3328 pixels. Among the 410 mammograms, 107 images contain breast tumors. Hence, these 107 images were selected for this study. The 107 images were randomly split into 90 images for training and 17 images for testing, as shown in Table 1. The image augmentation methods, including rotation and flipping, were applied to the training data. The augmentation methods increased the number of breast tumor mammography images to 720 images. The 720 images were randomly split into 576 images for training data and 174 images for validation data, as shown in Table 2.

2.1.2. CBIS-DDSM Dataset

The DDSM is a public dataset provided by the University of South Florida Computer Science and Engineering Department, Sandia National Laboratories, and Massachusetts General Hospital [26]. The CBIS-DDSM is an updated and standardized version of the DDSM [27]. It contains a variety of pathologically verified cases, including malignant, benign, and normal cases. DDSM is an extremely useful database for the development and testing of computer-aided diagnosis (CAD) systems due to its scale and the ground truth validation it offers. The CBIS-DDSM collection includes a subset of the DDSM data organized by expert radiologists. It also comprises pathological diagnosis, bounding boxes, and region of interest (ROI) segmentation for training data. Among all mammography images with tumors in the CBIS-DDSM dataset, 838 images were selected for this study. The 838 images were randomly split into 728 images for training data and 110 images for testing data, as shown in Table 1. The image augmentation methods, including rotation and flipping, were applied to the training samples. Through image augmentation, the number of breast tumor mammography images was increased to 5824. The 5824 images were randomly split into 4659 images for training data and 1165 images for validation data, as shown in Table 2.

2.1.3. Private Dataset

The private dataset comprised mammography images from the Cheng Hsin General Hospital, Taipei City, Taiwan. Initially, VGG image annotator (VIA) software was used by an expert radiologist from the department of medical imaging to mark the tumor location based on the pathological data [28]. Then, all the labeled images were verified and confirmed by the department of hematology and oncology. Finally, the dataset was de-identified for patient privacy. A total of 196 mammography images were collected from January 2019 to December 2019. All the mammograms consist of tumors with a grade of breast imaging reporting and data system assessment category 4 (BIRADS 4) or higher. A total of 196 mammography images were randomly split into 148 images for training and 48 images for testing, as shown in Table 1. The image augmentation methods, including rotation and flipping, were applied to the training samples. Through image augmentation methods, the number of breast tumor mammography images was increased to 1184. The 1184 images were randomly split into 947 images for training and 237 images for validation, as shown in Table 2.

2.2. Data Preprocessing

This research study only focused on the segmentation step. Initially, the ROI of the tumor was cropped manually. The ROI of the tumor was resized into 256 × 256 . In order to eliminate additional noise and degradation caused by the scanning process of digital X-ray mammography, all images were preprocessed [29,30].

2.2.1. Histogram Equalization

Histogram equalization is a well-known technique widely used for contrast enhancement [31]. It is used in a variety of applications, including medical image processing and radar signal processing, due to its simple function and effectiveness [32,33,34,35]. Histogram equalization well distributes the pixels over the full dynamic intensity range. One drawback of histogram equalization is that the background noise can be increased when the image is too bright or too dark in the local area after the histogram equalization, which is mainly due to the flattening property of the histogram equalization. This study applied the local histogram equalization method called CLAHE to address the above challenges. CLAHE is an adaptive extension of histogram equalization. It helps in the dynamic preservation of the local contrast features of an image. CLAHE has been applied to all datasets of this study. The sample results on the datasets after applying the CLAHE are shown in Figure 2. From Figure 2, it is noted that the edges of the tumors became clearer after applying the CLAHE technique. A total of 107, 838, and 196 ROIs were obtained from the INbreast, CBIS-DDSM, and the private datasets, respectively. The complete details of the mammography datasets are listed in Table 1.

2.2.2. Image Augmentation

The most common problem that DL models might face is the overfitting problem due to the limited amount of training samples [36,37,38]. As a result of overfitting, a model might detect or classify features derived from the training samples, but the same model will not be able to detect or classify features derived from unseen samples. To address the issue of overfitting, this study has used two image augmentation methods, namely rotation and flipping. First, bi-linear interpolation has been used to rotate each image around its center point by a value of 90 ° degrees counter-clockwise up to 360 ° . By using the bi-linear interpolation method, the rotated image has the same aspect ratio as the original image, without losing any part of the image. Second, mirroring or flipping is the simplest augmentation approach. It results in a dataset with twice as many images. The flipping technique is basically the same as the rotation technique; however, it transforms rotation in the reverse direction. The sample results on the datasets after applying the augmentation methods are shown in Figure 3.
The raw ROIs of the training data were augmented by rotating at an angle of 90 ° and horizontal flipping. Hence, a total of 720, 5824, and 1184 ROIs were generated from the INbreast, CBIS-DDSM, and private datasets, respectively. Then, the data were randomly split into training and validation. Detailed information of the mammography datasets in terms of the training data is provided in Table 2.

2.3. Proposed Model

SegNet can record pooling indices when applying Max pooling. These pooling indices are used to up-sample the images to the original size. Hence, the required graphics processing unit (GPU) memory for training the model can be lower. Inspired by the success of SegNet and Connected-UNets, this research proposed a model, called Connected-SegNets, which connects two standard SegNets using additional adapted skip connections. The overall architecture of the proposed Connected-SegNets model is shown in Figure 4.
The proposed model consists of two encoder and two decoder networks. The first decoder network and the second encoder network are connected with additional skip connections after cascading a second SegNet. This helps to recover the fine-grained features that are lost in the encoding of the SegNet and apply them to encode the high-resolution features by connecting them to the previously decoded features. The proposed Connected-SegNets architecture is deepened by stacking two SegNets. The upper half of the proposed architecture is similar to SegNet, which uses the first 13 convolutional layers in the VGG16 network as the encoder network [39]. In the decoder network, the last convolutional layer is removed. Each encoder network comprises two convolutional kernels, which includes 3 × 3 convolutional layers followed by an activation rectified linear unit (ReLU) and a batch normalization (BN) layer. Then, a maximum pooling indices operation is applied to the output of each encoder network before passing the information to the next encoder. Each decoder network consists of a 2 × 2 transposed convolution unit that is concatenated with the previous encoder output, and then the result is fed into two convolution blocks, which consist of 3 × 3 convolutions followed by an activation ReLU and a BN layer. Additionally, a second SegNet is attached to the first SegNet through new skip connections that use information from the first up-sampling pathway. The result of the last decoder block is concatenated with the same result after being fed into a 3 × 3 convolution layer followed by an activation ReLU and a BN layer. This serves as the input of the first encoder network to the second SegNet. The output of the maximum pooling indices operations of each of the three encoder networks is fed into 3 × 3 convolution layers and then concatenated with the output of the last previous decoder network. The result is next down-sampled to the next encoder network. Finally, the last output is given to a dilation layer with a dilation rate of 3, followed by an advanced ReLU activation layer to generate the predicted mask. In order to obtain more features, a dilation layer with a dilation rate of 3 is used in the last layer. Moreover, an activation ReLU limits the maximum value to 1, which is called an advanced ReLU. The details of the Connected-SegNets layers are listed in Table 3.

2.4. Experimental Environment and Parameter Settings

All experiments were performed using a PC with an Intel i7-9700K CPU, 55 GB of DDR4 RAM, and an NVIDIA GeForce RTX 2080Ti GPU with 11 GB of memory. The software environment used a Windows 10 64-bit operating system, python 3.8.12, CUDA 10.1, cuDNN 7.6.5, and TensorFlow 2.8.0. The learning rate was set to 0.0001 using the Adam optimizer [40] and the batch size was 4. The loss function was the IoU loss function.

2.5. Evaluation Metrics

In this research, precision, recall, IoU score, and Dice score evaluation metrics have been used to evaluate the proposed model based on the confusion matrix. The confusion matrix is an evaluation metric often used to evaluate classification, detection, and segmentation algorithms. The confusion matrix shows information about the true classes and the predicted classes. The true class and the predicted class can be positive or negative. The true negative (TN) case is when both the true case and the predicted case are tumors. False negatives (FN) occur when the true case is not a tumor, but the predicted case is. The false positive (FP) case occurs when the true case is a tumor while the prediction is a non-tumor. True positives (TP) occur when the actual case is non-tumor and the predicted case is tumor. The Dice score is also known as the F1-score, which represents the harmonic mean of precision and recall, as expressed in Equation (3). Additionally, the IoU evaluation metric represents the percentage of overlap between the predicted classes and the true classes, as represented in Equation (4).
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
D i c e s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
I o U s c o r e = T P T P + F P + F N

3. Results

3.1. Results on INbreast Dataset

The confusion matrix results of Connected-SegNets on the INbreast dataset are listed in Table 4. From the Table 4, it is observed that the proportion of actual tumors that was correctly identified as tumors (TP) by Connected-SegNets is 96%. This is the highest TP rate compared to the other datasets. In addition, the proportion of non-tumors that was correctly identified as non-tumors (TN) by Connected-SegNets is 88%.

3.2. Results on CBIS-DDSM Dataset

The identification results of Connected-SegNets on the CBIS-DDSM dataset are listed in Table 5. From the Table 5, it can be seen that the proportion of true tumors that was correctly identified as tumors (TP) by Connected-SegNets is 93%. Moreover, the proportion of non-tumors that was correctly identified as non-tumors (TN) by Connected-SegNets is 87%.

3.3. Results on Private Dataset

The results of the Connected-SegNets model on the private dataset are listed in Table 6. It is observed that the proportion of actual tumors that was correctly identified as tumors (TP) by Connected-SegNets is 92%. On the other hand, the proportion of tumors that were not tumors and were correctly identified as non-tumors (TN) by Connected-SegNets is 89%. This TN rate is considered to be the highest compared to other datasets.
The accuracy and loss curves of the training and validation for Connected-SegNets are shown in Figure 5 and Figure 6, respectively. It can be noted from Figure 5 and Figure 6 that the training and validation curves behave similarly, which is an indication that the proposed Connected-SegNets can be generalized and does not suffer from overfitting.
A large number of epochs might cause a deep learning model to overfit the data, whereas a small number of epochs can lead to smooth convergence. Therefore, the early stop technique has been utilized during the model training to avoid overfitting. The validation dataset is used to track the model training performance. The early stop method can help to set a suitable training epoch by tracking the best performance on the validation dataset. Therefore, when the validation performance stops improving, an early stop mode of the training process will be activated. Moreover, using the early stop algorithm not only can avoid the overfitting problem, but it also can help with choosing the optimal hyperparameter configurations for training the model. The early stop algorithm steps are shown in Algorithm 1. In this research, the validation tracking, ActStepSetting, was set to 20 iterations. Hence, if the validation performance did not improve after 20 iterations, the training was stopped automatically.
Algorithm 1 Validation Loss Tracking for Early Stop
Input: LatestValLoss, ActStepSetting
Output: BestValLossScore
1:  E a r l y S t o p F a l s e ;
2: if B e s t V a l i d a t i o n R e p e a t N u m < = A c t S t e p S e t t i n g then
3:     if  L a t e s t V a l L o s s < B e s t V a l L o s s S c o r e  then
4:       B e s t V a l i d a t i o n R e p e a t N u m 0 ;
5:       B e s t V a l L o s s S c o r e L a t e s t V a l L o s s ;
6:    else
7:       B e s t V a l i d a t i o n R e p e a t N u m B e s t V a l i d a t i o n R e p e a t N u m + 1 ;
8:    end if
9: else
10:     E a r l y S t o p T r u e ;
11: end if
12: return  ( B e s t V a l L o s s S c o r e )

3.4. Comparison of Segmentation Results

As shown in Table 7, the segmentation results of each testing datum were evaluated by the two evaluation metrics, Dice score and IoU score, for the segmented maps per pixel, and compared with the original ground truth. It is noted that the proposed Connected-SegNets model produced the highest Dice score of 96.34%, 92.86%, and 92.25% on the INbreast, CBIS-DDSM, and private datasets, respectively. Moreover, the proposed model achieved the highest IoU Score of 91.21%, 87.34%, and 83.71% on the INbreast, CBIS-DDSM, and private datasets, respectively. Finally, the comparative results show that the proposed model, Connected-SegNets, outperformed the related models in terms of Dice score and IoU score on the three datasets.
Figure 7 shows some examples of the segmented ROI results generated by different models against their ground truth images. It is clearly observed that the quality of the segmentation maps of the Connected-SegNets model contain less error and produce more precise segmentation compared to other methods.

4. Discussion

In recent years, several DL models have been developed and applied for breast tumor segmentation. These DL models have achieved remarkable success in segmenting breast tumors in mammograms. Nevertheless, many of these DL models produce high false positive and false negative rates [41]. The SegNet model is considered to be one of the deep learning models that is easy to modify and further optimize to provide better segmentation performance in different fields. Therefore, this study proposed a DL model, called Connected-SegNets, based on SegNet, for better breast tumor segmentation. The main goal of the proposed Connected-SegNets model is to improve the overall performance of breast tumor segmentation. Hence, several techniques have been implemented and incorporated into the proposed method in order to achieve this goal. These techniques include deepening the architecture with two SegNets, replacing the cross-entropy loss function of the standard SegNet with the IoU loss function, applying histogram equalization (CLAHE), and performing image augmentation. Figure 7 illustrates the segmentation results of AUNet, Standard UNet, Connected-UNets, Standard SegNet, and the proposed Connected-SegNets on the testing data of the INbreast, CBIS-DDSM, and private datasets. The segmentation results of the proposed Connected-SegNets are the closest to the ground truth compared to those of the AUNet, UNet, Connected-UNets, and SegNet models. The proposed model fully connects two single SegNets using additional skip connections. These are helpful to recover the spatial information that is lost during the pooling operations. Moreover, the IoU loss function leads to a more robust model. Furthermore, the histogram equalization (CLAHE) has been applied to smoothen the distribution of the image pixels for better pixel segmentation. Additionally, image augmentation methods, including rotation and flipping, have been applied to increase the number of training samples and reduce the impact of overfitting. This has led to more accurate segmentation performance compared to the other models. The significant improvement is shown in Table 4, Table 5 and Table 6, where the Connected-SegNets model has the TP value of 96%, 93%, and 92%, on the INbreast, CBIS-DDSM, and private datasets, respectively. Similarly, the TN value is of 88%, 87%, and 89%, on INbreast, CBIS-DDSM, and the private dataset, respectively. The results of the proposed model, Connected-SegNets, showed a significant segmentation improvement compared to the other models, with a maximum Dice score of 96.34% on the INbreast dataset, 92.86% on the CBIS-DDSM dataset, and 92.25% on the private dataset. Similarly, the Connected-SegNets model has achieved the highest IoU score of 91.21% on the INbreast dataset, 87.34% on the CBIS-DDSM dataset, and 83.71% on the private dataset. Overall, the proposed Connected-SegNets model has outperformed DS U-Net, AUNet, UNet, Connected-UNets, and SegNet in terms of Dice score and IoU score. This shows the power of the proposed model to learn complex features through the connections added between the two SegNets in the proposed Connected-SegNets, which take advantage of the decoded features as another input in the encoder pathway.

5. Conclusions

This research proposed a deep learning model, namely Connected-SegNets, for breast tumor segmentation from X-ray images. Two SegNets were used in the proposed model, both of which were fully connected via additional skip connections. The cross-entropy loss function of the original SegNet was replaced by the IoU loss function to make the proposed model more robust against sparse data. Additionally, the contrast limit adapt histogram equalization (CLAHE) was applied to enhance the compressed areas and smooth the pixel distribution. Moreover, two augmentation methods including rotation and flipping were used to increase the number of training samples and prevent overfitting. The experimental results showed that Connected-SegNets outperformed the existing models, with the highest Dice scores of 96.34%, 92.86%, and 92.25%, and the highest IoU scores of 91.21%, 87.34%, and 83.71% on the INbreast, CBIS-DDSM, and private datasets, respectively. Future work will focus on implementing new deep learning algorithms for tumor detection and classification for automatic breast cancer diagnosis.

Author Contributions

Conceptualization, M.A., L.C. and Y.-L.C.; Data curation, C.-H.C. and T.-C.W.; Formal analysis, M.A., T.-H.T., C.-H.C., S.-C.M. and L.C.; Funding acquisition, C.-H.C., S.-C.M. and Y.-L.C.; Investigation, T.-H.T.; Methodology, M.A., T.-H.T. and T.-C.W.; Project administration, C.-H.C., S.-C.M. and Y.-L.C.; Resources, Y.-L.C.; Software, T.-C.W.; Supervision, C.-H.C.; Validation, T.-H.T., S.-C.M. and Y.-L.C.; Writing—original draft, T.-C.W. and L.C.; Writing–review and editing, M.A. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Ministry of Science and Technology, Taiwan, Grant Nos. MOST 110-2119-M-027-001, MOST 110-2221-E-027-101, MOST 110-2622-E-027-025, MOST 111-2622-8-038-004-TD2, and National Taipei University of Technology and Cheng Hsin General Hospital, Grant No. NTUT-CHGH-110-01.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Siegel, R.L.; Miller, K.D.; Fuchs, H.E.; Jemal, A. Cancer statistics, 2022. CA Cancer J. Clin. 2022, 72, 7–33. [Google Scholar] [CrossRef] [PubMed]
  2. Zou, R.; Loke, S.Y.; Tan, V.K.M.; Quek, S.T.; Jagmohan, P.; Tang, Y.C.; Madhukumar, P.; Tan, B.K.T.; Yong, W.S.; Sim, Y.; et al. Development of a microRNA panel for classification of abnormal mammograms for breast cancer. Cancers 2021, 13, 2130. [Google Scholar] [CrossRef] [PubMed]
  3. Li, J.; Guan, X.; Fan, Z.; Ching, L.M.; Li, Y.; Wang, X.; Cao, W.M.; Liu, D.X. Non-invasive biomarkers for early detection of breast cancer. Cancers 2020, 12, 2767. [Google Scholar] [CrossRef] [PubMed]
  4. Almalki, Y.E.; Soomro, T.A.; Irfan, M.; Alduraibi, S.K.; Ali, A. Computerized Analysis of Mammogram Images for Early Detection of Breast Cancer. Healthcare 2022, 10, 801. [Google Scholar] [CrossRef] [PubMed]
  5. Shi, P.; Zhong, J.; Rampun, A.; Wang, H. A hierarchical pipeline for breast boundary segmentation and calcification detection in mammograms. Comput. Biol. Med. 2018, 96, 178–188. [Google Scholar] [CrossRef]
  6. Waks, A.G.; Winer, E.P. Breast cancer treatment: A review. JAMA 2019, 321, 288–300. [Google Scholar] [CrossRef] [PubMed]
  7. Salgado, R.; Denkert, C.; Demaria, S.; Sirtaine, N.; Klauschen, F.; Pruneri, G.; Wienert, S.; Van den Eynden, G.; Baehner, F.L.; Pénault-Llorca, F.; et al. The evaluation of tumor-infiltrating lymphocytes (TILs) in breast cancer: Recommendations by an International TILs Working Group 2014. Ann. Oncol. 2015, 26, 259–271. [Google Scholar] [CrossRef]
  8. Tariq, M.; Iqbal, S.; Ayesha, H.; Abbas, I.; Ahmad, K.T.; Niazi, M.F.K. Medical image based breast cancer diagnosis: State of the art and future directions. Expert Syst. Appl. 2021, 167, 114095. [Google Scholar] [CrossRef]
  9. Petrillo, A.; Fusco, R.; Di Bernardo, E.; Petrosino, T.; Barretta, M.L.; Porto, A.; Granata, V.; Di Bonito, M.; Fanizzi, A.; Massafra, R.; et al. Prediction of Breast Cancer Histological Outcome by Radiomics and Artificial Intelligence Analysis in Contrast-Enhanced Mammography. Cancers 2022, 14, 2132. [Google Scholar] [CrossRef]
  10. Ahmed, L.; Iqbal, M.M.; Aldabbas, H.; Khalid, S.; Saleem, Y.; Saeed, S. Images data practices for semantic segmentation of breast cancer using deep neural network. J. Ambient. Intell. Humaniz. Comput. 2020, 1–17. [Google Scholar] [CrossRef]
  11. Le, E.; Wang, Y.; Huang, Y.; Hickman, S.; Gilbert, F. Artificial intelligence in breast imaging. Clin. Radiol. 2019, 74, 357–366. [Google Scholar] [CrossRef] [PubMed]
  12. Bi, W.L.; Hosny, A.; Schabath, M.B.; Giger, M.L.; Birkbak, N.J.; Mehrtash, A.; Allison, T.; Arnaout, O.; Abbosh, C.; Dunn, I.F.; et al. Artificial intelligence in cancer imaging: Clinical challenges and applications. CA Cancer J. Clin. 2019, 69, 127–157. [Google Scholar] [CrossRef] [PubMed]
  13. Shah, S.M.; Khan, R.A.; Arif, S.; Sajid, U. Artificial intelligence for breast cancer analysis: Trends & directions. Comput. Biol. Med. 2022, 142, 105221. [Google Scholar] [PubMed]
  14. Ketabi, H.; Ekhlasi, A.; Ahmadi, H. A computer-aided approach for automatic detection of breast masses in digital mammogram via spectral clustering and support vector machine. Phys. Eng. Sci. Med. 2021, 44, 277–290. [Google Scholar] [CrossRef]
  15. Hosny, A.; Parmar, C.; Quackenbush, J.; Schwartz, L.H.; Aerts, H.J. Artificial intelligence in radiology. Nat. Rev. Cancer 2018, 18, 500–510. [Google Scholar] [CrossRef]
  16. Vobugari, N.; Raja, V.; Sethi, U.; Gandhi, K.; Raja, K.; Surani, S.R. Advancements in Oncology with Artificial Intelligence—A Review Article. Cancers 2022, 14, 1349. [Google Scholar] [CrossRef]
  17. Alkhaleefah, M.; Wu, C.C. A hybrid CNN and RBF-based SVM approach for breast cancer classification in mammograms. In Proceedings of the 2018 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Miyazaki, Japan, 7–10 October 2018; pp. 894–899. [Google Scholar]
  18. Kallenberg, M.; Petersen, K.; Nielsen, M.; Ng, A.Y.; Diao, P.; Igel, C.; Vachon, C.M.; Holland, K.; Winkel, R.R.; Karssemeijer, N.; et al. Unsupervised deep learning applied to breast density segmentation and mammographic risk scoring. IEEE Trans. Med. Imaging 2016, 35, 1322–1331. [Google Scholar] [CrossRef]
  19. Ravitha Rajalakshmi, N.; Vidhyapriya, R.; Elango, N.; Ramesh, N. Deeply supervised u-net for mass segmentation in digital mammograms. Int. J. Imaging Syst. Technol. 2021, 31, 59–71. [Google Scholar]
  20. Sun, H.; Li, C.; Liu, B.; Liu, Z.; Wang, M.; Zheng, H.; Feng, D.D.; Wang, S. AUNet: Attention-guided dense-upsampling networks for breast mass segmentation in whole mammograms. Phys. Med. Biol. 2020, 65, 055005. [Google Scholar] [CrossRef]
  21. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  22. Baccouche, A.; Garcia-Zapirain, B.; Castillo Olea, C.; Elmaghraby, A.S. Connected-UNets: A deep learning architecture for breast mass segmentation. NPJ Breast Cancer 2021, 7, 1–12. [Google Scholar] [CrossRef]
  23. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  24. Moreira, I.C.; Amaral, I.; Domingues, I.; Cardoso, A.; Cardoso, M.J.; Cardoso, J.S. Inbreast: Toward a full-field digital mammographic database. Acad. Radiol. 2012, 19, 236–248. [Google Scholar] [CrossRef] [PubMed]
  25. Huang, M.L.; Lin, T.Y. Dataset of breast mammography images with masses. Data Brief 2020, 31, 105928. [Google Scholar] [CrossRef]
  26. Heath, M.; Bowyer, K.; Kopans, D.; Kegelmeyer, P.; Moore, R.; Chang, K.; Munishkumaran, S. Current status of the digital database for screening mammography. In Digital Mammography; Springer: Cham, Switzerland, 1998; pp. 457–460. [Google Scholar]
  27. Lee, R.S.; Gimenez, F.; Hoogi, A.; Miyake, K.K.; Gorovoy, M.; Rubin, D.L. A curated mammography data set for use in computer-aided detection and diagnosis research. Sci. Data 2017, 4, 1–9. [Google Scholar] [CrossRef] [PubMed]
  28. Dutta, A.; Zisserman, A. The VIA Annotation Software for Images, Audio and Video. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; ACM: New York, NY, USA, 2019. MM’19. [Google Scholar] [CrossRef]
  29. Al-Masni, M.A.; Al-Antari, M.A.; Park, J.M.; Gi, G.; Kim, T.Y.; Rivera, P.; Valarezo, E.; Choi, M.T.; Han, S.M.; Kim, T.S. Simultaneous detection and classification of breast masses in digital mammograms via a deep learning YOLO-based CAD system. Comput. Methods Programs Biomed. 2018, 157, 85–94. [Google Scholar] [CrossRef]
  30. Hai, J.; Qiao, K.; Chen, J.; Tan, H.; Xu, J.; Zeng, L.; Shi, D.; Yan, B. Fully convolutional densenet with multiscale context for automated breast tumor segmentation. J. Healthc. Eng. 2019, 2019, 8415485. [Google Scholar] [CrossRef] [PubMed]
  31. Dhal, K.G.; Das, A.; Ray, S.; Gálvez, J.; Das, S. Histogram equalization variants as optimization problems: A review. Arch. Comput. Methods Eng. 2021, 28, 1471–1496. [Google Scholar] [CrossRef]
  32. Huang, Z.; Zhang, Y.; Li, Q.; Zhang, T.; Sang, N. Spatially adaptive denoising for X-ray cardiovascular angiogram images. Biomed. Signal Process. Control. 2018, 40, 131–139. [Google Scholar] [CrossRef]
  33. Huang, Z.; Li, X.; Wang, N.; Ma, L.; Hong, H. Simultaneous denoising and enhancement for X-ray angiograms by employing spatial-frequency filter. Optik 2020, 208, 164287. [Google Scholar] [CrossRef]
  34. Huang, Z.; Zhang, Y.; Li, Q.; Li, X.; Zhang, T.; Sang, N.; Hong, H. Joint analysis and weighted synthesis sparsity priors for simultaneous denoising and destriping optical remote sensing images. IEEE Trans. Geosci. Remote Sens. 2020, 58, 6958–6982. [Google Scholar] [CrossRef]
  35. Rao, B.S. Dynamic histogram equalization for contrast enhancement for digital images. Appl. Soft Comput. 2020, 89, 106114. [Google Scholar] [CrossRef]
  36. Alkhaleefah, M.; Ma, S.C.; Chang, Y.L.; Huang, B.; Chittem, P.K.; Achhannagari, V.P. Double-shot transfer learning for breast cancer classification from X-ray images. Appl. Sci. 2020, 10, 3999. [Google Scholar] [CrossRef]
  37. Elasal, N.; Swart, D.M.; Miller, N. Frame augmentation for imbalanced object detection datasets. J. Comput. Vis. Imaging Syst. 2018, 4, 3. [Google Scholar]
  38. Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 1–48. [Google Scholar] [CrossRef]
  39. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  40. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  41. Al-Antari, M.A.; Al-Masni, M.A.; Choi, M.T.; Han, S.M.; Kim, T.S. A fully integrated computer-aided diagnosis system for digital X-ray mammograms via deep learning detection, segmentation, and classification. Int. J. Med. Infor. 2018, 117, 44–54. [Google Scholar] [CrossRef]
Figure 1. Flow chart of the proposed tumor segmentation system.
Figure 1. Flow chart of the proposed tumor segmentation system.
Cancers 14 04030 g001
Figure 2. Sample results after applying the histogram equalization (CLAHE) to random ROI images from the datasets.
Figure 2. Sample results after applying the histogram equalization (CLAHE) to random ROI images from the datasets.
Cancers 14 04030 g002
Figure 3. Random sample results after applying the rotation and flipping augmentation methods on the original ROIs. Arrows refer to the direction of the image.
Figure 3. Random sample results after applying the rotation and flipping augmentation methods on the original ROIs. Arrows refer to the direction of the image.
Cancers 14 04030 g003
Figure 4. Architecture of the proposed Connected-SegNets model.
Figure 4. Architecture of the proposed Connected-SegNets model.
Cancers 14 04030 g004
Figure 5. The training and validation accuracy curves of Connected-SegNets.
Figure 5. The training and validation accuracy curves of Connected-SegNets.
Cancers 14 04030 g005
Figure 6. The training and validation loss curves of Connected-SegNets.
Figure 6. The training and validation loss curves of Connected-SegNets.
Cancers 14 04030 g006
Figure 7. Example of the breast tumor segmentation results using AUNet, UNet, Connected-UNets, SegNet, and the proposed Connected-SegNets on the testing data of INbreast, CBIS-DDSM, and the private dataset.
Figure 7. Example of the breast tumor segmentation results using AUNet, UNet, Connected-UNets, SegNet, and the proposed Connected-SegNets on the testing data of INbreast, CBIS-DDSM, and the private dataset.
Cancers 14 04030 g007
Table 1. Distribution of the mammography datasets.
Table 1. Distribution of the mammography datasets.
DatasetRaw ROIsTraining SamplesTesting Samples
INbreast dataset1079017
CBIS-DDSM dataset838728110
Private dataset19614848
Total1141966175
Table 2. The number of training and validation samples before and after data augmentation.
Table 2. The number of training and validation samples before and after data augmentation.
DatasetRaw Images Augmented Images Training Validation
INbreast Dataset90720576144
CBIS-DDSM dataset728582446591165
Private dataset1481184947237
Total966772861821546
Table 3. The detailed architecture of the proposed Connected-SegNet.
Table 3. The detailed architecture of the proposed Connected-SegNet.
SegNet1
No.Layer NameOutputFilter SizeNo. of FiltersNo. of Layers
1Input256 × 256 × 1 1
2Conv1256 × 256 × 643 × 3642
3Maxpool 1128 × 128 × 64 1
4Conv2128 × 128 × 1283 × 31282
5Maxpool 164 × 64 × 128 1
6Conv364 × 64 × 2563 × 32563
7Maxpool 132 × 32 × 256 1
8Conv432 × 32 × 5123 × 35123
9Maxpool 116 × 16 × 512 1
10Conv516 × 16 × 5123 × 35123
11Maxpool 18× 8 × 512 1
12Upsampling 216 × 16 × 512 1
13Conv616 × 16 × 5123 × 35123
14Upsampling 232 × 32 × 512 1
15Conv732 × 32 × 5123 × 35122
16Conv832 × 32 × 2563 × 32561
17Upsampling 264 × 64 × 256 1
18Conv964 × 64 × 2563 × 32562
19Conv1064 × 64 × 1283 × 31281
20Upsampling 2128 × 128 × 128 1
21Conv11128 × 128 × 1283 × 31282
22Conv12128 × 128 × 643 × 3641
23Upsampling 2256 × 256 × 64 1
24Conv13256 × 256 × 643 × 3641
25Conv13 256 × 256 × 64
26Conv14 256 × 256 × 64 3 × 3 642
27Maxpool 1 128 × 128 × 64 1
28Concatenate 128 × 128 × 128 1
29Conv15 128 × 128 × 128 3 × 3 1282
30Maxpool 1 64 × 64 × 128 1
31Concatenate 64 × 64 × 256 1
32Conv16 64 × 64 × 256 3 × 3 2563
33Maxpool 1 32 × 32 × 256 1
34Concatenate 16 × 16 × 512 1
35Conv17 32 × 32 × 512 3 × 35123
36Maxpool 1 16 × 16 × 512 1
37Concatenate 16 × 16 × 1024 1
38Conv18 16 × 16 × 512 3 × 35123
39Maxpool 1 8 × 8 × 512 1
40Upsampling 2 16 × 16 × 512 1
41Conv19 16 × 16 × 512 3 × 3 5123
42Upsampling 2 32 × 32 × 512 1
43Conv20 32 × 32 × 512 3 × 3 5122
44Conv21 32 × 32 × 256 3 × 3 2561
45Upsampling 2 64 × 64 × 256 1
46Conv22 64 × 64 × 256 3 × 3 2562
47Conv23 64 × 64 × 128 3 × 3 1281
48Upsampling 2 128 × 128 × 128 1
49Conv24 128 × 128 × 128 3 × 3 1282
50Conv25 128 × 128 × 64 3 × 3 641
51Upsampling 2 256 × 256 × 64 1
52Conv26 256 × 256 × 64 3 × 3 641
53Conv27 256 × 256 × 64 3 × 3 (D 3 = 3)641
54Output 256 × 256 × 1 1 × 1 11
1 Maxpooling: Maxpooling and recording of the indices. 2 Upsampling: Upsampling with the recorded indices. 3 D: Dilation rate.
Table 4. Confusion matrix results of the proposed Connected-SegNets on INbreast dataset.
Table 4. Confusion matrix results of the proposed Connected-SegNets on INbreast dataset.
Connected-SegNets
Ground Truth
TumorNon-Tumor
PredictionTumor96% (TP)4% (FN)
Non-Tumor12% (FP)88% (TN)
Table 5. Confusion matrix results of the proposed Connected-SegNets on CBIS-DDSM dataset.
Table 5. Confusion matrix results of the proposed Connected-SegNets on CBIS-DDSM dataset.
Connected-SegNets
Ground Truth
TumorNon-Tumor
PredictionTumor93% (TP)7% (FN)
Non-Tumor13% (FP)87% (TN)
Table 6. Confusion matrix results of the proposed Connected-SegNets on the private dataset.
Table 6. Confusion matrix results of the proposed Connected-SegNets on the private dataset.
Connected-SegNets
Ground Truth
TumorNon-Tumor
PredictionTumor92% (TP)8% (FN)
Non-Tumor11% (FP)89% (TN)
Table 7. Comparison results between the proposed Connected-SegNets and the related segmentation models on the testing datasets of INbreast, CBIS-DDSM, and the private dataset, respectively.
Table 7. Comparison results between the proposed Connected-SegNets and the related segmentation models on the testing datasets of INbreast, CBIS-DDSM, and the private dataset, respectively.
ModelINbreast DatasetCBIS-DDSM DatasetPrivate Dataset
Dice Score (%)IoU Score (%)Dice Score (%)IoU Score (%)Dice Score (%)IoU Score (%)
DS U-Net [19]79.0083.4082.7085.70NANA
AUNet [20]90.1286.5189.0382.6589.4480.87
UNet [21]92.1488.2390.4784.7989.1180.21
Connected-UNets [22]94.4589.7290.6685.8190.4181.33
SegNet [23]92.0188.7790.5285.3088.4981.97
Connected-SegNets96.3491.2192.8687.3492.2583.71
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Alkhaleefah, M.; Tan, T.-H.; Chang, C.-H.; Wang, T.-C.; Ma, S.-C.; Chang, L.; Chang, Y.-L. Connected-SegNets: A Deep Learning Model for Breast Tumor Segmentation from X-ray Images. Cancers 2022, 14, 4030. https://doi.org/10.3390/cancers14164030

AMA Style

Alkhaleefah M, Tan T-H, Chang C-H, Wang T-C, Ma S-C, Chang L, Chang Y-L. Connected-SegNets: A Deep Learning Model for Breast Tumor Segmentation from X-ray Images. Cancers. 2022; 14(16):4030. https://doi.org/10.3390/cancers14164030

Chicago/Turabian Style

Alkhaleefah, Mohammad, Tan-Hsu Tan, Chuan-Hsun Chang, Tzu-Chuan Wang, Shang-Chih Ma, Lena Chang, and Yang-Lang Chang. 2022. "Connected-SegNets: A Deep Learning Model for Breast Tumor Segmentation from X-ray Images" Cancers 14, no. 16: 4030. https://doi.org/10.3390/cancers14164030

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop