Next Article in Journal
Usefulness of T2-Weighted Images with Deep-Learning-Based Reconstruction in Nasal Cartilage
Next Article in Special Issue
Efficient Skip Connections-Based Residual Network (ESRNet) for Brain Tumor Classification
Previous Article in Journal
The Value of Electroretinography in Identifying Candidate Genes for Inherited Retinal Dystrophies: A Diagnostic Guide
Previous Article in Special Issue
One-Stage Detection without Segmentation for Multi-Type Coronary Lesions in Angiography Images Using Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bone Metastases Lesion Segmentation on Breast Cancer Bone Scan Images with Negative Sample Training

1
Department of Biomedical Imaging and Radiological Science, China Medical University, Taichung 404, Taiwan
2
Department of Nuclear Medicine, Feng Yuan Hospital, Ministry of Health and Welfare, Taichung 420, Taiwan
3
Department of Nuclear Medicine and PET Center, China Medical University Hospital, Taichung 404, Taiwan
*
Authors to whom correspondence should be addressed.
Diagnostics 2023, 13(19), 3042; https://doi.org/10.3390/diagnostics13193042
Submission received: 31 August 2023 / Revised: 18 September 2023 / Accepted: 22 September 2023 / Published: 25 September 2023
(This article belongs to the Special Issue Artificial Intelligence in Clinical Medical Imaging)

Abstract

:
The use of deep learning methods for the automatic detection and quantification of bone metastases in bone scan images holds significant clinical value. A fast and accurate automated system for segmenting bone metastatic lesions can assist clinical physicians in diagnosis. In this study, a small internal dataset comprising 100 breast cancer patients (90 cases of bone metastasis and 10 cases of non-metastasis) and 100 prostate cancer patients (50 cases of bone metastasis and 50 cases of non-metastasis) was used for model training. Initially, all image labels were binary. We used the Otsu thresholding method or negative mining to generate a non-metastasis mask, thereby transforming the image labels into three classes. We adopted the Double U-Net as the baseline model and made modifications to its output activation function. We changed the activation function to SoftMax to accommodate multi-class segmentation. Several methods were used to enhance model performance, including background pre-processing to remove background information, adding negative samples to improve model precision, and using transfer learning to leverage shared features between two datasets, which enhances the model’s performance. The performance was investigated via 10-fold cross-validation and computed on a pixel-level scale. The best model we achieved had a precision of 69.96%, a sensitivity of 63.55%, and an F1-score of 66.60%. Compared to the baseline model, this represents an 8.40% improvement in precision, a 0.56% improvement in sensitivity, and a 4.33% improvement in the F1-score. The developed system has the potential to provide pre-diagnostic reports for physicians in final decisions and the calculation of the bone scan index (BSI) with the combination with bone skeleton segmentation.

1. Introduction

According to the gender statistics database published by the Gender Equality Committee of the Executive Yuan in Taiwan in 2023, breast cancer was ranked first among the top 10 cancer incidence rates in 2020 [1]. Breast cancer, prostate cancer, lung cancer, and other prevalent cancers account for more than 80% of cases of metastatic bone disease. For patients with breast cancer, late-stage bone metastasis is prone to occur, significantly reducing the prognosis of the patients. A study by Coleman and Rubens reported bone metastasis in 69% of breast cancer patients who died between 1979 and 1984, out of a total of 587 patients [2]. Bone metastasis in breast cancer most commonly occurs in the spine, followed by the ribs and sternum [3]. Radiologically, bone metastases in breast cancer are predominantly osteolytic, leading to severe complications such as bone pain, pathological fractures, spinal cord compression, hypercalcemia, and bone marrow suppression. Therefore, the early detection and treatment of bone metastasis in breast cancer patients are crucially important.
Current methods for detecting breast cancer metastasis include the clinical observation of distant organ involvement, organ biopsies, diagnostic imaging, and serum tumor markers. One of the primary imaging techniques used in clinics for bone metastasis diagnosis is the whole-body bone scan (WBBS) with vein injection using the Tc-99m MDP tracer [4,5]. WBBS offers the advantages of whole-body examination, cost-effectiveness, and high sensitivity, making it a preferred modality for bone metastasis screening [6]. Unlike X-radiography (XR) and computed tomography (CT) images, which can only detect changes in bone when there is approximately 40–50% mineralization [7], bone scans exhibit higher sensitivity in detecting bone changes, capable of detecting alterations as low as 5% in osteoblast activity. The reported sensitivity and specificity of skeletal scintigraphy for bone metastasis detection are 78% and 48%, respectively [8].
The bone scan index (BSI) is an image biomarker utilized in WBBS to evaluate the severity of bone metastasis in cancer patients. It enables a quantification on the degree of tumor involvement in the skeleton [9,10]. BSI is used for observing disease progression or treatment response. The commercial software EXINI bone (version 1 and version 2, including subversions), developed by EXINI Diagnostics AB, incorporates aBSI (automated bone scan index) technology for the comprehensive automated quantitative assessment of bone scan images [11]. In [11], there exists a strong correlation between manual and automated BSI assessment values (ρ = 0.80), which further strengthens (ρ = 0.93) when cases with BSI scores exceeding 10 (1.8%) are excluded. This indicates that automated BSI calculations can deliver clinical value comparable to manual calculations. Shimizu et al. has proposed an image interpretation system based on deep learning [12], using BtrflyNets for the hotspot detection of bone metastasis and bone segmentation, followed by automatic BSI calculation. The aBSI technology has now become a clinically valuable tool. Nevertheless, there are still challenges regarding recognition performance (sensitivity and precision) in this technique.
Cheng et al. applied a deep convolutional neural network (D-CNN) for the object detection of bone metastasis from prostate cancer in bone scan images [13]. Their investigation specifically focused on the chest and pelvic regions, and the sensitivity and precision for detecting and classifying chest bone metastasis were determined using bounding boxes to be 0.82 ± 0.08 and 0.70 ± 0.11, respectively. Regarding pelvic bone metastasis classification, the reported sensitivity and specificity were 0.87 ± 0.12 and 0.81 ± 0.11, respectively. Cheng et al. conducted a more detailed study on chest bone metastasis in prostate cancer patients [14]. The average sensitivity and precision for detecting and classifying chest bone metastasis based on lesion locations are reported as 0.72 ± 0.04 and 0.90 ± 0.04, respectively. For classifying chest bone metastasis based on patient-level outcomes, the average sensitivity and specificity are found to be 0.94 ± 0.09 and 0.92 ± 0.09, respectively. Patents filed by Cheng et al. are referenced as [15], which leverage deep learning for the identification of bone metastasis in prostate cancer bone scan images. Since they use bounding boxes, they are unable to calculate BSI.
In a related study [16], a neural network (NN) model based on U-Net++ is proposed for the automated segmentation of metastatic lesions in bone scan images. The anterior–posterior and posterior–anterior views are superimposed, and image segmentation is exclusively performed on the chest region of whole-body bone scan images. The achieved average F1-score is 65.56%.
In this study, we modified the Double U-Net [17] as the fundamental architecture to perform bone metastases segmentation on WBBS. We explored various methods to enhance network performance, including background pre-processing, adding negative samples, and transfer learning. We used Otsu thresholding [18] and negative mining [14] methods for background pre-processing and generating negative samples. Background pre-processing helped eliminate unnecessary background information, while adding negative samples reduced the model’s false positive rate. Both of these methods did not require manual labeling or modification, saving time and manpower. Previous studies in the same field [16,19,20,21,22,23] focused only on segmenting bone metastases in specific regions (chest or pelvis) and could only predict either the anterior or posterior view. The datasets we used only excluded the non-metastatic-prone areas below the knees and could simultaneously segment images in both the anterior and posterior views. In comparison, our model was able to provide a more comprehensive assessment of bone metastasis images.
The following points summarize the contributions of this paper:
  • We discuss the challenges of lesion segmentation in breast cancer bone scan images.
  • We compare and discuss the state-of-the-art methods in the same research field.
  • Our experiments have shown that background pre-processing significantly improves a model’s performance and adding negative samples enhances model precision. Both methods do not require manual labeling or label modification, saving time and manpower.
  • Our segmentation model offers greater comprehensiveness. It can perform lesion segmentation on WBBS images, predicting both anterior and posterior views simultaneously.

2. Related Work

Deep learning has found numerous applications in cancer detection tasks. The authors of Ref. [24] proposed an improved SIFT descriptor with Harris corner to form Bag-Of-Words features in image representation. This study made a significant contribution to medical image classification tasks. For skin lesions, the authors of Ref. [25] conducted a comprehensive comparative study of U-Net and attention-based methods for dermatological image segmentation, aiding in the diagnosis of skin lesions. The authors of Ref. [26] introduced an enhanced deep learning model, SBXception, based on the Xception network to improve skin cancer classification. In the realm of MRI, the authors of Ref. [27] presented a weighted ensemble deep learning model for brain tumor classification. The authors of Ref. [28] explored five machine learning techniques to deepen the understanding of brain tumor classification and enhance its scope and significance.
Some early work has been carried out on automatic segmentation of metastatic lesions using bone scan images [29,30,31,32,33]. The trend of using deep learning for bone scintigraphy image analysis is becoming increasingly evident. In classification tasks, the authors of Ref. [34] introduced an improved ResNet model that combines convolutional block attention module and contextual transformer attention mechanisms to achieve the accurate classification of SPECT images [35] based their work on widely used deep networks, including VGG, ResNet, and DenseNet, by fine-tuning their parameters and structures or by customizing new network architectures. The proposed classifiers performed well in identifying bone metastases through SPECT imaging. The authors of Ref. [36] presented an automated bone metastasis diagnostic model based on multi-view images. The authors of Ref. [37] introduced a new framework in this work, which included data preparation and image classification, for automatically classifying scintigraphy images collected from patients clinically diagnosed with lung cancer.
In object detection tasks, the authors of Ref. [38] employed scaled-YOLOv4 and Detectron2 object detection networks for bone metastasis localization in breast cancer patient nuclear imaging data and for detecting degenerative and pathological findings in whole-body scintigraphy scans. The authors of Ref. [39] proposed an automatic lesion detection model based on single shot multibox object detector for the automatic detection of lung cancer bone metastases in low-resolution SPECT bone scintigraphy images. The authors of Ref. [14] applied D-CNN for object detection of prostate cancer bone metastases in the chest and pelvic regions. As object detection uses bounding boxes, it cannot calculate the BSI as a subsequent quantitative measure.
Compared to classification and object detection tasks, segmentation tasks are more challenging. The authors of Ref. [19] introduced a model called MaligNet, which semantically segments abnormal hotspots in a semi-supervised manner and classifies bone cancer metastases in the chest region. The authors of Ref. [20] built a segmentation model based on U-Net and Mask R-CNN networks by fine-tuning their architectures for identifying and segmenting metastatic hotspots in bone SEPCT images. The authors of Ref. [21] added a methods attention mechanism on top of the original U-Net network’s skip connections to enhance feature selection, allowing for the automatic identification and segmentation of bone metastases. The authors of Ref. [16] proposed a neural network model based on U-Net++ for the automatic segmentation of metastatic lesions in bone scan images. The authors of Ref. [22] introduced an improved UNet3+ network that combines attention mechanisms for the automatic segmentation of bone metastatic lesions. The authors of Ref. [23] presented a bone imaging focus segmentation algorithm based on the Swin Transformer, which uses the swin transformer as the backbone network for extracting feature information from bone images. In current research in the same field, segmentation tasks are limited to predicting specific local regions, such as the chest or pelvis, and they cannot simultaneously predict both anterior and posterior views.

3. Materials and Methods

3.1. Materials

In this study, we collected 200 bone scan images from the Department of Nuclear Medicine of China Medical University Hospital. The details of the bone scan images are provided in Table 1. Specifically, D1 is defined as 90 images from breast cancer patients with bone metastasis. D2 is defined as 10 images from breast cancer patients without bone metastasis. D3 is defined as 50 images from prostate cancer patients with bone metastasis. D4 is defined as 50 images from prostate cancer patients without bone metastasis. Figure 1 shows bone scan images of breast cancer patients. This study has been approved by the Institutional Review Board (IRB) of China Medical University and Hospital Research Ethics Committee (CMUH106-REC2-130), approved on 27 September 2017.
The WBBS process can be described as follows. Patients undergo WBBS with a gamma camera (Millennium MG, Infinia Hawkeye 4, or Discovery NM/CT 670 system; GE Healthcare, Waukesha, WI, USA). Bone scans are acquired 2–4 h after the intravenous injection of 740–925 MBq (20–25 mCi) of technetium-99m methylene diphosphonate (Tc-99m MDP) with an acquisition time of 10–15 cm/min. The collected WBBS images are saved in DICOM format. The raw images include anterior–posterior (AP) and posterior–anterior (PA) views, with a matrix size of 1024 × 256 pixels.

3.2. Image Labeling

To facilitate labeling the bone scan images, the Labelme (version 4.5.9) software is used as the annotation tool. The manual annotation of bone metastasis images is carried out under the guidance and supervision of nuclear medicine physicians. This process is very time-consuming. The outputs generated by the Labelme software are saved in JSON format, and then converted to the PNG format. Figure 2 represents a schematic of the manually annotated results.

3.3. Image Pre-Processing

The raw images possess a large memory size and the DICOM format is not directly suitable for neural network training. Moreover, the raw images exhibit variations in brightness and contrast levels. Thus, the pre-processing of the raw images becomes imperative. The detection of the body range was accomplished using the projection profile, followed by the extraction of two views with dimensions of 950 × 256 pixels through cutting and centering. No scaling or other transformations were applied during this process. We utilized the brightness normalization method proposed in [14] for brightness pre-processing. This method uses a linear transformation to adjust the dynamic range of an image, with the objective of controlling the average intensity of each image within the range of (7, 14). The algorithm for the linear transformation is illustrated in Figure 3. The region below the knees, which is uncommon for bone metastasis, was excluded from the calculation of BSI. To obtain the region above the knees, pixels beyond row 640 were eliminated, resulting in two views with a spatial resolution of 640 × 256 pixels each. Finally, the pre-processed AP (anterior–posterior) and PA (posterior–anterior) view images were horizontally merged, generating images with a spatial resolution of 640 × 512 pixels.

3.4. Positive and Negative Samples

According to previous research [14], adding negative samples to the training dataset helps reduce false positives and improve model precision. In this study, we also used negative samples to enhance the model’s performance. Positive samples are defined as images with bone metastases (D1 and D3), while negative samples are defined as images without bone metastases (D2 and D4).
For the bone metastasis segmentation task in WBBS images, the background significantly interferes with network training. In this scenario, the background not only includes air but also the non-metastatic (NM) human body regions. Intuitively, NM regions of the human body contain information but do not contain air. Therefore, an alternative approach is to filter out the air to extract NM human body regions to generate NM masks. The generation of NM masks involves two methods, which we briefly explain below.
The Otsu thresholding method is used to generate NM masks for both positive and negative samples. It is important to note that the metastatic (M) regions must be manually excluded from positive samples beforehand. Otsu thresholding can automatically determine the threshold that separates air from the human body.
The negative mining method for generating NM masks involves two steps. First, the baseline network is trained using only positive samples. After training, this model is used to predict negative samples. Since negative samples do not have bone metastatic lesions, all segmentation results produced are false positives. These false positive segmentation results are then treated as NM regions to generate NM masks. The same model is also used to predict positive samples. It is worth noting that the metastatic (M) regions in the positive samples must be manually excluded beforehand.
The initial classes include background and metastasis. After generating NM masks using the two methods mentioned above, the number of classes increases from the original two to three, now including air-background (BG), non-metastatic (NM), and metastatic (M) classes.

3.5. Transfer Learning

Transfer learning is a widely used technique in neural networks to increase their performance. Before applying transfer learning, two crucial factors need to be considered: (1) the size of the target dataset and (2) the similarity between the target dataset and the pre-training dataset.
In this study, the Double U-Net network model was pre-trained using the D3 and D4 datasets. We chose a pre-training dataset that contains highly similar bone scan images of prostate cancer. Subsequently, the model was fine-tuned using the target dataset consisting of breast cancer bone scan images. By leveraging transfer learning and selecting a pre-training dataset closely related to the target dataset, our goal was to utilize shared features between the two datasets to enhance the model’s performance on the current specific task.

3.6. Neural Network Model

We adopted the Double U-Net architecture as our network framework. The original Double U-Net architecture was developed for binary segmentation tasks, which we refer to as the baseline network. To adapt the Double U-Net architecture for multi-class segmentation, we modified its network structure following the method described in our previous research [40]. Figure 4 illustrates the modified network architecture. We changed the output layer of network 1 to obtain a SoftMax activation function, enabling it to perform multi-class segmentation. With this modification, the Double U-Net architecture can handle three-class segmentation tasks involving the BG, NM, and M regions.

3.7. Loss Function

The selection of an appropriate loss function is a critical aspect in the design of deep learning architectures for image segmentation tasks, as it greatly impacts the learning dynamics of the algorithm. In our study, we consider two loss functions: the Dice loss (Equation (1)), as originally proposed in [17], and the Focal Tversky loss (Equation (2)). By comparing these loss functions, we aim to explore their respective influences on the model’s performance in the context of our specific task.
The dice coefficient is a widely adopted metric in computer vision for assessing the similarity between two images. In our study, we utilize a modified version of the dice coefficient known as the dice loss, which served as a loss function for our model.
DL ( y ,   p ) = 1 2 yp y + p
where y is true value and p is the predicted outcome.
The focal Tversky loss is particularly well-suited for solving highly imbalanced class scenarios. It incorporates a γ coefficient that allows for the down-weighting of easy samples. Additionally, by adjusting the α and β coefficients, different weights can be assigned to false positives (FP) and false negatives (FN).
FTL ( y ,   p ) = ( 1 yp yp + α ( 1 y ) p + β y ( 1 p ) ) γ
where γ = 0.75, α = 0.3, and β = 0.7.
When performing the three-class segmentation task for the BG, NM, and M regions, calculating loss and back-propagating for the BG class is unnecessary and would make model training difficult. Therefore, during the execution of the three-class segmentation task, we do not calculate loss or perform backpropagation for the BG class.

3.8. Experimental Configuration and Evaluation Metrics

All experiments were conducted on four Intel Xeon Gold 6154 CPUs and a 32 GB Nvidia Tesla V100 GPU. The memory capacity configured was 90 GB. Our segmentation system was implemented in Python using Keras with TensorFlow 2.4.1.
The evaluation metrics employed in this study include precision (Equation (3)), sensitivity (Equation (4)), and the overall model assessment based on the F1-score (Equation (5)). The terms true positive (TP), false positive (FP), true negative (TN), and false negative (FN) were defined at the pixel level.
Precision = TP TP + FP
Sensitivity = TP TP + FN
F 1 -score = 2 × Precision × Sensitivity Precision + Sensitivity

4. Results

All experimental results in tables are obtained through 10-fold cross-validation, with a ratio of 8:1:1 for the training, validation, and testing sets, respectively. The learning rate used for training was 0.0001, batch size was set to 4, and the number of iterations was 500.

4.1. Negative Samples

The qualitative results of two negative samples are illustrated in Figure 5. The Otsu thresholding can extract NM masks easily and produce three classes: BG, NM, and M. Its results are shown in Figure 5a. Nevertheless, negative mining requires two steps, as described in the method. Its results are shown in Figure 5b.

4.2. Results of the Baseline Network

The original Double U-Net network was trained using the D1 dataset and utilized the dice loss function. The objective of this experiment was to establish the baseline performance of the baseline network.
For comparison, it is essential to evaluate the performance of deep learning models in each task using quantitative metrics. Here, the precision, sensitivity, and F1-score are utilized for performance evaluation. Figure 6 shows the qualitative results, and the quantitative results are shown in Table 2.

4.3. The Baseline Network Using Otsu Thresholding

The modified Double U-Net network was trained using the D1 dataset. Before training, we used the Otsu thresholding method on the D1 dataset for background pre-processing, generating the NM mask. Figure 7 illustrates a training sample from the D1 dataset, which includes three classes.
The model’s performance is shown in Table 3. In Table 3, we included the focal Tversky loss for comparison. Compared to Table 2, we observe that using the Otsu thresholding method for background pre-processing on the D1 dataset significantly improves the model’s performance. In both the dice loss and focal Tversky loss models, the F1-score improved by 3.12% and 4.16%, respectively.
We wanted to investigate the impact of adding negative samples to the training dataset on the model. In this experiment, we first used the Otsu thresholding method for background pre-processing on the D2 dataset, generating the NM mask. Then, we added the D2 dataset to the D1 training dataset in each fold. Figure 8 shows an example training sample from the D2 dataset, which contains three classes.
Table 4 presents the model performance when adding the D2 dataset using the Otsu thresholding method. In both the dice loss and focal Tversky loss models, precision improved by 2.61% and 2.09%, respectively. From the results, we did not observe any significant improvement in the F1 score. However, adding negative samples did indeed increase precision, which aligns with our expectations.

4.4. The Baseline Network Using Negative Mining

The modified Double U-Net network was trained using the D1 dataset. Prior to training, we applied the negative mining method to pre-process the D1 dataset and generate the NM mask. Figure 9 illustrates a training sample from the D1 dataset, including three classes.
The model performance is shown in Table 5. Compared to the baseline (Table 2), negative mining indeed shows significant improvement. In the dice loss and focal Tversky loss models, F1-score improved by 2.14% and 2.27%, respectively.
Next, in this experiment, we first used the negative mining method to pre-process the D2 dataset and generate the NM mask. Then, we added the D2 dataset to the training data of D1 in each fold. Figure 10 shows a training sample from the D2 dataset (containing three classes).
The quantitative results of this experiment are shown in Table 6. In both the dice loss and focal Tversky loss models, precision improved by 1.85% and 1.64%, respectively. Similar to Table 4, adding negative samples to the training set led to a slight improvement in precision but a slight decrease in sensitivity. The F1-score remained unchanged, as expected.

4.5. Model Performance after Transfer Learning

Based on the previous experiments, we found that using the Otsu threshold method to generate the NM mask leads to better performance improvement. To understand the impact of transfer learning, we pre-trained the modified Double U-Net network using the D3 and D4 datasets. Before pretraining, we used the Otsu threshold method to pre-process the D3 and D4 datasets and generate the NM masks. Then, we added the D4 dataset to the training data of D3 in each fold.
The pre-trained model was fine-tuned by learning from breast cancer patient images. Before fine-tuning, we used the Otsu threshold method to pre-process the D1 and D2 datasets and generate the NM masks. Then, we added the D2 dataset to the training data of D1 in each fold.
The qualitative results of segmentation are shown in Figure 11. Specifically, we compare two loss functions: dice and focal Tversky.
The quantitative results are shown in Table 7. Compared to the results without transfer learning (Table 4), a slight improvement can be seen in the F1-score.

5. Discussion

In this study, the raw Double U-Net architecture served as the baseline model for performance comparison. Subsequently, two schemes on negative sample extraction are explored to see the impact on the model performance.
Otsu thresholding can easily separate air and body, thus removing the air background. The air background contains no information and wastes computation time. Although we define three classes in training, the BG class does not count into the loss. Another profit is to extract negative samples. In our previous study [14] we found that training with only metastasis (positive) class is not a good idea. It is better to train models with positive and negative samples simultaneously. Our results shown in Table 3 and Table 4 have confirmed this again; they are better than the baseline shown in Table 2.
Table 5 and Table 6 show the results obtained using negative mining. The model performances seem slightly worse than Otsu thresholding. This might be due to the fact that negative sample areas are significantly smaller than those negative samples produced using the thresholding technique. Thus, they contained less information for training. Moreover, negative mining requires a pre-trained model, and the thresholding technique does not. Our study indicates that while adding negative samples is necessary, there is not only one way to do so. There could be many other ways to create negative samples for training.
We compared our research with other relevant studies in terms of network architecture and results, as shown in Table 8. This table summarizes studies that used deep learning methods for the segmentation of bone metastases in bone scan images. The authors of [20] proposed an improved ResU-Net model for the segmentation of metastatic hotspots in thorax SPECT images, achieving precision, sensitivity, and IoU scores of 77.21%, 67.88%, and 61.03%, respectively. The authors of [21] added a methods attention mechanism to the original U-Net network’s skip connections to enhance model performance, resulting in an F1-score of 57.10% and an IoU of 63.30%. The authors of [16] introduced a neural network model based on U-Net++, achieving segmentation performance with a precision of 68.85%, a sensitivity of 62.57%, and an F1-score of 65.56%. The authors of [22] combined the UNet3+ network with an attention mechanism, proposing an improved UNet3+ network that achieved segmentation performance with a precision of 61.20%, a sensitivity of 68.33%, and an F1-score of 64.33%. The authors of [23] utilized a swin transformer as the backbone network and proposed a bone imaging focus segmentation algorithm, achieving an F1-score of 77.81% and an IoU of 35.59%.
In Table 8, the improved ResU-Net model used in [20] achieved the best model performance in the segmentation of metastatic lesions in the thorax region. The model based on the swin transformer in [23] achieved the highest F1-score. Our segmented region is the widest in Table 8, and our model’s F1-score of 66.60 is second only to [23].
The selection of loss function might play a crucial role in model performance. For complex tasks like segmentation, there is no universally applicable loss function. It largely depends on the properties of the training dataset, such as distribution, skewness, boundaries, etc. For segmentation tasks with extreme class imbalance, focal-related loss functions are more appropriate [41]. Additionally, since the vanilla Double U-Net model has a higher precision than sensitivity, we are keen to use Tversky-related loss functions to balance the false positives (FP) and false negatives (FN) rates. Therefore, we adopt focal Tversky loss as the compared loss function. In the future, further exploration and research should be conducted on the selection of optimizers.
Not all hotspots in bone scan images represent bone metastases; normal bone absorption, renal metabolism, inflammation, and injuries can also cause hotspots in images, leading to false positives in segmentation. In addition to the inherent imaging principles of bone scan images that make training the model challenging, the presence of artifacts in the images is also a crucial factor leading to misclassification. Examples of such artifacts include high-activity areas like the kidneys and bladder, the injection site of the radioactive isotope, and motion artifacts, as shown in Figure 12. Apart from artifacts in breast cancer images, prostate cancer bone scan images also exhibit high-activity artifacts from catheters, urine bags, and diapers, as shown in Figure 13. In the future, appropriate pre-processing can be applied to minimize the impact from artifacts, or additional classes such as benign lesions and artifacts can be introduced to train the model more accurately.
The image pre-processing is usually important before using neural networks. Our previous study proposed a pre-processing method where the original images were combined into a 3D image to alleviate the issue of spatial connectivity loss [14]. View aggregation, an operation applied to bone scan images, has been used to enhance areas of high absorption [16]. This method enhances lesions that appear in both anterior and posterior view images and maps lesions that only appear in either anterior or posterior view images. However, that method cannot be applied in this study, since we calculate every pixel here and all errors (sensitivity, precision) are calculated in pixel-wise scale.

6. Conclusions

In this study, we confirm the validity of using negative samples in the task of bone metastatic lesions detection in breast cancer whole body bone scan images. The model is trained using positive and negative samples. We used background pre-processing to remove excess air background information. Adding negative samples improved the model’s precision. The images we used only excluded the less common regions below the knees for bone metastatic lesions and could simultaneously perform image segmentation for both anterior and posterior views. Our model is able to provide a more comprehensive evaluation of bone metastasis images. The precision, sensitivity, and F1-score for the segmentation of bone metastatic lesions are calculated on a pixel-level scale and the best results reach 70.24%, 61.80%, and 65.75% for dice loss and 69.96%, 63.55%, and 66.60% for focal Tversky loss, respectively.
The limitation of this study is the use of a small, single-center dataset, comprising only 100 breast cancer patients. This may result in limited model performance and generalizability. In the dataset, only 10 negative samples were collected from breast cancer patients, and the class imbalance between positive and negative samples could also pose a challenge to model performance.
There is still significant room for improvement in the model’s performance in this study. In the future, we plan to collect more WBBS images from different centers to further validate the proposed model’s performance. We will focus on fine-tuning the hyperparameters of the neural network and optimizing the choice of optimizers to enhance segmentation performance and reduce computational costs. Noise and artifacts in WBBS images are inevitable issues, and we plan to explore more image pre-processing methods to remove false artifacts and image noise to improve image quality, thus enhancing segmentation capabilities. Finally, we will use the interpretations of nuclear medicine physicians as the gold standard to compare the final model with the decisions made by nuclear medicine physicians, aiming to assess any discrepancies in decisions and evaluate the clinical utility of the model.

Author Contributions

Conceptualization, D.-C.C.; methodology, D.-C.C.; software, P.-N.Y. and Y.-Y.C.; validation, P.-N.Y. and Y.-Y.C.; formal analysis D.-C.C.; investigation, Y.-C.L. and T.-C.H.; resources, T.-C.H. and D.-C.C.; data curation, P.-N.Y. and Y.-Y.C.; writing—original draft preparation, Y.-Y.C.; writing—review and editing, D.-C.C.; visualization, D.-C.C.; supervision, D.-C.C.; project administration, D.-C.C.; funding acquisition, D.-C.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Science and Technology Council (NSTC), Taiwan, grant number MOST 111-2314-B-039-040.

Institutional Review Board Statement

The study was approved by the Institutional Review Board (IRB) and the Hospital Research Ethics Committee (CMUH106-REC2-130) of China Medical University.

Informed Consent Statement

Patient consent was waived by IRB due to this being a retrospective study and images were only used without patients’ identification.

Data Availability Statement

Not applicable.

Acknowledgments

We thank the National Center for High-performance Computing (NCHC) for providing computational and storage resources.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gender Equality Committee of the Executive Yuan. Available online: https://www.gender.ey.gov.tw/gecdb/Stat_Statistics_DetailData.aspx?sn=nLF9GdMD%2B%2Bv41SsobdVgKw%3D%3D (accessed on 24 February 2023).
  2. Coleman, R.E.; Rubens, R.D. The clinical course of bone metastases from breast cancer. Br. J. Cancer 1987, 55, 61–66. [Google Scholar] [CrossRef] [PubMed]
  3. Kakhki, V.R.D.; Anvari, K.; Sadeghi, R.; Mahmoudian, A.S.; Torabian-Kakhki, M. Pattern and distribution of bone metastases in common malignant tumors. Nucl. Med. Rev. 2013, 16, 66–69. [Google Scholar] [CrossRef] [PubMed]
  4. Hamaoka, T.; Madewell, J.E.; Podoloff, D.A.; Hortobagyi, G.N.; Ueno, N.T. Bone imaging in metastatic breast cancer. J. Clin. Oncol. 2004, 22, 2942–2953. [Google Scholar] [CrossRef]
  5. Even-Sapir, E.; Metser, U.; Mishani, E.; Lievshitz, G.; Lerman, H.; Leibovitch, I. The detection of bone metastases in patients with high-risk prostate cancer: 99mTc-MDP Planar bone scintigraphy, single- and multifield-of-viewSPECT, 18F-fluoride PET, and 18F-fluoride PET/CT. J. Nucl. Med. 2006, 47, 287–297. [Google Scholar]
  6. Costelloe, C.M.; Rohren, E.M.; Madewell, J.E.; Hamaoka, T.; Theriault, R.L.; Yu, T.K.; Ueno, N.T. Imaging bone metastases in breast cancer: Techniques and recommendations for diagnosis. Lancet Oncol. 2009, 10, 606–614. [Google Scholar] [CrossRef] [PubMed]
  7. Vijayanathan, S.; Butt, S.; Gnanasegaran, G.; Groves, A.M. Advantages and limitations of imaging the musculoskeletal system by conventional radiological, radionuclide, and hybrid modalities. Semin. Nucl. Med. 2009, 39, 357–368. [Google Scholar] [CrossRef]
  8. O’Sullivan, G.J.; Carty, F.L.; Cronin, C.G. Imaging of bone metastasis: An update. World J. Radiol. 2015, 7, 202. [Google Scholar] [CrossRef]
  9. Imbriaco, M.; Larson, S.M.; Yeung, H.W.; Mawlawi, O.R.; Erdi, Y.; Venkatraman, E.S.; Scher, H.I. A new parameter for measuring metastatic bone involvement by prostate cancer: The Bone Scan Index. Clin. Cancer Res. 1998, 4, 1765–1772. [Google Scholar]
  10. Erdi, Y.E.; Humm, J.L.; Imbriaco, M.; Yeung, H.; Larson, S.M. Quantitative bone metastases analysis based on image segmentation. J. Nucl. Med. 1997, 38, 1401–1406. [Google Scholar]
  11. Ulmert, D.; Kaboteh, R.; Fox, J.J.; Savage, C.; Evans, M.J.; Lilja, H.; Larson, S.M. A novel automated platform for quantifying the extent of skeletal tumour involvement in prostate cancer patients using the Bone Scan Index. Eur. Urol. 2012, 62, 78–84. [Google Scholar] [CrossRef]
  12. Shimizu, A.; Wakabayashi, H.; Kanamori, T.; Saito, A.; Nishikawa, K.; Daisaki, H.; Kawabe, J. Automated measurement of bone scan index from a whole-body bone scintigram. Int. J. Comput. Assist. Radiol. Surg. 2020, 15, 389–400. [Google Scholar] [CrossRef] [PubMed]
  13. Cheng, D.C.; Liu, C.C.; Hsieh, T.C.; Yen, K.Y.; Kao, C.H. Bone metastasis detection in the chest and pelvis from a whole-body bone scan using deep learning and a small dataset. Electronics 2021, 10, 1201. [Google Scholar] [CrossRef]
  14. Cheng, D.C.; Hsieh, T.C.; Yen, K.Y.; Kao, C.H. Lesion-based bone metastasis detection in chest bone scintigraphy images of prostate cancer patients using pre-train, negative mining, and deep learning. Diagnostics 2021, 11, 518. [Google Scholar] [CrossRef]
  15. Cheng, D.C.; Liu, C.C.; Kao, C.H.; Hsieh, T.C. System of Deep Learning Neural Network in Prostate Cancer Bone Metastasis Identification Based on Whole Body Bone Scan Images. U.S. Patent US11488303B2, 1 November 2022. [Google Scholar]
  16. Cao, Y.; Liu, L.; Chen, X.; Man, Z.; Lin, Q.; Zeng, X.; Huang, X. Segmentation of lung cancer-caused metastatic lesions in bone scan images using self-defined model with deep supervision. Biomed. Signal Process. Control 2023, 79, 104068. [Google Scholar] [CrossRef]
  17. Jha, D.; Riegler, M.A.; Johansen, D.; Halvorsen, P.; Johansen, H.D. Doubleu-net: A deep convolutional neural network for medical image segmentation. In Proceedings of the IEEE 33rd International Symposium on Computer-Based Medical Systems (CBMS), Rochester, MN, USA, 28–30 July 2020; pp. 558–564. [Google Scholar]
  18. Otsu, N. A threshold selection method from gray-level histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef]
  19. Apiparakoon, T.; Rakratchatakul, N.; Chantadisai, M.; Vutrapongwatana, U.; Kingpetch, K.; Sirisalipoch, S.; Chuangsuwanich, E. MaligNet: Semisupervised learning for bone lesion instance segmentation using bone scintigraphy. IEEE Access 2020, 8, 27047–27066. [Google Scholar] [CrossRef]
  20. Lin, Q.; Luo, M.; Gao, R.; Li, T.; Man, Z.; Cao, Y.; Wang, H. Deep learning based automatic segmentation of metastasis hotspots in thorax bone SPECT images. PLoS ONE 2020, 15, e0243253. [Google Scholar] [CrossRef]
  21. Zhang, J.; Huang, M.; Deng, T.; Cao, Y.; Lin, Q. Bone metastasis segmentation based on Improved U-NET algorithm. In Proceedings of the 4th International Conference on Advanced Algorithms and Control Engineering (ICAACE), Sanya, China, 29–31 January 2021. [Google Scholar]
  22. Liu, C.; Cao, Y.; Lin, Q.; Man, Z.; He, Y.; Peng, L. Segmentation of metastatic lesions on bone scan images based on improved UNet3+ network. In Proceedings of the 4th International Conference on Computer Engineering and Application (ICCEA), Hangzhou, China, 7–9 April 2023; pp. 916–920. [Google Scholar]
  23. Wu, T.; Luo, R.; Lin, H.; Yu, H.; Wang, Q.; Liu, H. Research on focal segmentation of bone scan based on Swin Transformer. In Proceedings of the 4th International Conference on Computer Vision, Image and Deep Learning (CVIDL), Zhuhai, China, 12–14 May 2023; pp. 426–430. [Google Scholar]
  24. Khan, S.A.; Gulzar, Y.; Turaev, S.; Peng, Y.S. A modified HSIFT Descriptor for medical image classification of anatomy objects. Symmetry 2021, 13, 1987. [Google Scholar] [CrossRef]
  25. Gulzar, Y.; Khan, S.A. Skin lesion segmentation based on vision transformers and convolutional neural networks—A comparative study. Appl. Sci. 2022, 12, 5990. [Google Scholar] [CrossRef]
  26. Mehmood, A.; Gulzar, Y.; Ilyas, Q.M.; Jabbari, A.; Ahmad, M.; Iqbal, S. SBXception: A Shallower and Broader Xception Architecture for Efficient Classification of Skin Lesions. Cancers 2023, 15, 3604. [Google Scholar] [CrossRef]
  27. Anand, V.; Gupta, S.; Gupta, D.; Gulzar, Y.; Xin, Q.; Juneja, S.; Shaikh, A. Weighted Average Ensemble Deep Learning Model for Stratification of Brain Tumor in MRI Images. Diagnostics 2023, 13, 1320. [Google Scholar] [CrossRef] [PubMed]
  28. Khan, F.; Ayoub, S.; Gulzar, Y.; Majid, M.; Reegu, F.A.; Mir, M.S.; Elwasila, O. MRI-Based Effective Ensemble Frameworks for Predicting Human Brain Tumor. J. Imaging 2023, 9, 163. [Google Scholar] [CrossRef] [PubMed]
  29. Sadik, M.; Jakobsson, D.; Olofsson, F.; Ohlsson, M.; Suurkula, M.; Edenbrandt, L. A new computer-based decision-support system for the interpretation of bone scans. Nucl. Med. Commun. 2006, 27, 417–423. [Google Scholar] [CrossRef] [PubMed]
  30. Sadik, M.; Hamadeh, I.; Nordblom, P.; Suurkula, M.; Höglund, P.; Ohlsson, M.; Edenbrandt, L. Computer-assisted interpretation of planar whole-body bone scans. J. Nucl. Med. 2008, 49, 1958–1965. [Google Scholar] [CrossRef]
  31. Aslantas, A.; Dandil, E.; Saǧlam, S.; Çakiroǧlu, M. CADBOSS: A computer-aided diagnosis system for whole-body bone scintigraphy scans. J. Cancer Res. Ther. 2016, 12, 787–792. [Google Scholar] [CrossRef]
  32. Elfarra, F.G.; Calin, M.A.; Parasca, S.V. Computer-aided detection of bone metastasis in bone scintigraphy images using parallelepiped classification method. Ann. Nucl. Med. 2019, 33, 866–874. [Google Scholar] [CrossRef]
  33. Calin, M.A.; Elfarra, F.G.; Parasca, S.V. Object-oriented classification approach for bone metastasis mapping from whole-body bone scintigraphy. Phys. Medica 2021, 84, 141–148. [Google Scholar] [CrossRef]
  34. Feng, Q.; Cao, Y.; Lin, Q.; Man, Z.; He, Y.; Liu, C. SPECT bone scan image classification by fusing multi-attention mechanism with deep residual networks. In Proceedings of the 4th International Conference on Computer Vision, Image and Deep Learning (CVIDL), Zhuhai, China, 12–14 May 2023; pp. 47–51. [Google Scholar]
  35. Lin, Q.; Li, T.; Cao, C.; Cao, Y.; Man, Z.; Wang, H. Deep learning based automated diagnosis of bone metastases with SPECT thoracic bone images. Sci. Rep. 2021, 11, 4223. [Google Scholar] [CrossRef]
  36. Pi, Y.; Zhao, Z.; Xiang, Y.; Li, Y.; Cai, H.; Yi, Z. Automated diagnosis of bone metastasis based on multi-view bone scans using attention-augmented deep neural networks. Med. Image Anal. 2020, 65, 101784. [Google Scholar] [CrossRef]
  37. Li, T.; Lin, Q.; Guo, Y.; Zhao, S.; Zeng, X.; Man, Z.; Hu, Y. Automated detection of skeletal metastasis of lung cancer with bone scans using convolutional nuclear network. Phys. Med. Biol. 2022, 67, 015004. [Google Scholar] [CrossRef]
  38. Moustakidis, S.; Siouras, A.; Papandrianos, N.; Ntakolia, C.; Papageorgiou, E. Deep learning for bone metastasis localisation in nuclear imaging data of breast cancer patients. In Proceedings of the 12th International Conference on Information, Intelligence, Systems & Applications (IISA), Chania Crete, Greece, 12–14 July 2021; pp. 1–8. [Google Scholar]
  39. Lin, Q.; Chen, X.; Liu, L.; Cao, Y.; Man, Z.; Zeng, X.; Huang, X. Detecting multiple lesions of lung cancer-caused metastasis with bone scans using a self-defined object detection model based on SSD framework. Phys. Med. Biol. 2022, 67, 225009. [Google Scholar] [CrossRef] [PubMed]
  40. Yu, P.N.; Lai, Y.C.; Chen, Y.Y.; Cheng, D.C. Skeleton segmentation on bone scintigraphy for BSI computation. Diagnostics 2023, 13, 2302. [Google Scholar] [CrossRef] [PubMed]
  41. Jadon, S. A survey of loss functions for semantic segmentation. In Proceedings of the IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB), Via del Mar, Chile, 27–29 October 2020; pp. 1–7. [Google Scholar]
Figure 1. Bone scan images of breast cancer patients. (a) With metastasis; (b) without metastasis.
Figure 1. Bone scan images of breast cancer patients. (a) With metastasis; (b) without metastasis.
Diagnostics 13 03042 g001
Figure 2. The schematic of the manually annotated results. (a) Bone scan image; (b) overlay of bone scan image with ground truth; (c) ground truth.
Figure 2. The schematic of the manually annotated results. (a) Bone scan image; (b) overlay of bone scan image with ground truth; (c) ground truth.
Diagnostics 13 03042 g002
Figure 3. Flowchart of brightness normalization.
Figure 3. Flowchart of brightness normalization.
Diagnostics 13 03042 g003
Figure 4. The modified architecture diagram of Double U-Net, the baseline network.
Figure 4. The modified architecture diagram of Double U-Net, the baseline network.
Diagnostics 13 03042 g004
Figure 5. Illustration of negative sample productions. Notably, the metastasis hotspots are eliminated (the black holes), if the image has metastasis. (a) Otsu thresholding; (b) negative mining.
Figure 5. Illustration of negative sample productions. Notably, the metastasis hotspots are eliminated (the black holes), if the image has metastasis. (a) Otsu thresholding; (b) negative mining.
Diagnostics 13 03042 g005
Figure 6. The qualitative result of the baseline network. (a) Ground truth; (b) segmentation results (precision: 79.14; sensitivity: 78.22; F1-score: 78.68).
Figure 6. The qualitative result of the baseline network. (a) Ground truth; (b) segmentation results (precision: 79.14; sensitivity: 78.22; F1-score: 78.68).
Diagnostics 13 03042 g006
Figure 7. Illustration of applying Otsu thresholding to positive samples to generate NM masks. Three classes are included: BG, NM, and M.
Figure 7. Illustration of applying Otsu thresholding to positive samples to generate NM masks. Three classes are included: BG, NM, and M.
Diagnostics 13 03042 g007
Figure 8. Illustration of applying Otsu thresholding to negative samples to generate NM masks. Three classes are included: BG, NM, and M.
Figure 8. Illustration of applying Otsu thresholding to negative samples to generate NM masks. Three classes are included: BG, NM, and M.
Diagnostics 13 03042 g008
Figure 9. Illustration of applying negative mining to positive samples to generate NM masks. Three classes are included: BG, NM, and M.
Figure 9. Illustration of applying negative mining to positive samples to generate NM masks. Three classes are included: BG, NM, and M.
Diagnostics 13 03042 g009
Figure 10. Illustration of applying negative mining to negative samples to generate NM masks. Three classes are included: BG, NM, and M.
Figure 10. Illustration of applying negative mining to negative samples to generate NM masks. Three classes are included: BG, NM, and M.
Diagnostics 13 03042 g010
Figure 11. The qualitative results after transfer learning. (a) Ground truth; (b) segmentation results with dice loss (precision: 79.14, sensitivity: 73.41, F1-score: 76.17); (c) segmentation results with focal Tversky loss (precision: 74.02, sensitivity: 86.24, F1-score: 79.67).
Figure 11. The qualitative results after transfer learning. (a) Ground truth; (b) segmentation results with dice loss (precision: 79.14, sensitivity: 73.41, F1-score: 76.17); (c) segmentation results with focal Tversky loss (precision: 74.02, sensitivity: 86.24, F1-score: 79.67).
Diagnostics 13 03042 g011
Figure 12. Mis-segmentation of non-metastatic lesions. (a) Bone fracture (head region) (precision: 88.46; sensitivity: 60.97; F1-score: 72.19); (b) motion artifact (head region) (precision: 69.32; sensitivity: 47.84; F1-score: 56.61); (c) injection site (wrist) (precision: 43.55; sensitivity: 70.65; F1-score: 53.88); (d) injection site (elbow) (precision: 82.81; sensitivity: 55.52; F1-score: 66.47); (e) kidney (precision: 51.85; sensitivity: 47.89; F1-score: 49.79); (f) bladder (precision: 47.47; sensitivity: 78.28; F1-score: 59.10).
Figure 12. Mis-segmentation of non-metastatic lesions. (a) Bone fracture (head region) (precision: 88.46; sensitivity: 60.97; F1-score: 72.19); (b) motion artifact (head region) (precision: 69.32; sensitivity: 47.84; F1-score: 56.61); (c) injection site (wrist) (precision: 43.55; sensitivity: 70.65; F1-score: 53.88); (d) injection site (elbow) (precision: 82.81; sensitivity: 55.52; F1-score: 66.47); (e) kidney (precision: 51.85; sensitivity: 47.89; F1-score: 49.79); (f) bladder (precision: 47.47; sensitivity: 78.28; F1-score: 59.10).
Diagnostics 13 03042 g012
Figure 13. Artifacts in bone scan images of prostate cancer. (a) Catheter; (b) urinary bag; (c) diaper.
Figure 13. Artifacts in bone scan images of prostate cancer. (a) Catheter; (b) urinary bag; (c) diaper.
Diagnostics 13 03042 g013
Table 1. The details of the bone scan images.
Table 1. The details of the bone scan images.
Breast CancerProstate Cancer
w/metastasisD1:90D3:50
w/o metastasisD2:10D4:50
Total100100
Table 2. The quantitative results of the baseline network (dice loss).
Table 2. The quantitative results of the baseline network (dice loss).
Fold NumberPrecisionSensitivityF1-Score
149.2179.1960.70
258.7464.9861.70
370.5660.0164.86
481.6952.2063.70
560.5754.0657.13
672.5545.9256.24
743.6383.3257.27
849.0368.2157.05
961.7360.8461.28
1067.8961.1364.34
Mean61.5662.9962.27
Table 3. The quantitative results for this experiment. Using the Otsu thresholding method for background pre-processing on the D1 dataset.
Table 3. The quantitative results for this experiment. Using the Otsu thresholding method for background pre-processing on the D1 dataset.
Fold
Number
Dice LossFocal Tversky Loss
PrecisionSensitivityF1-ScorePrecisionSensitivityF1-Score
168.7467.7268.2366.9568.8167.86
271.2956.7563.2069.9160.7064.98
370.1164.5767.2369.6968.4969.09
483.9258.3968.8685.1956.3467.82
564.0658.3261.0665.5260.1762.73
677.0153.1062.8679.2451.8862.70
760.5273.9966.5860.6476.0567.47
851.3065.1957.4152.3668.9059.50
963.2464.2463.7362.5566.4064.42
1066.1470.5868.2966.6672.6769.54
Mean67.6363.2965.3967.8765.0466.43
Table 4. The quantitative results for this experiment. Using the Otsu thresholding method for background pre-processing on the D2 dataset and adding it to the training set.
Table 4. The quantitative results for this experiment. Using the Otsu thresholding method for background pre-processing on the D2 dataset and adding it to the training set.
Fold
Number
Dice LossFocal Tversky Loss
PrecisionSensitivityF1-ScorePrecisionSensitivityF1-Score
170.4165.4267.8269.6667.9468.79
270.8560.7265.3970.0963.1666.44
373.8864.5568.9072.2967.4369.78
485.2154.9166.7885.2956.5568.01
571.9555.8962.9172.8653.8261.91
682.0949.9362.1082.5451.3863.34
762.1467.7664.8362.2272.5867.00
854.3165.7959.5054.4264.2558.93
965.3763.0964.2163.2364.9464.07
1066.1469.9067.9766.9873.4970.08
Mean70.2461.8065.7569.9663.5566.60
Table 5. The quantitative results for this experiment. Using the negative mining method for background pre-processing on the D1 dataset.
Table 5. The quantitative results for this experiment. Using the negative mining method for background pre-processing on the D1 dataset.
Fold
Number
Dice LossFocal Tversky Loss
PrecisionSensitivityF1-ScorePrecisionSensitivityF1-Score
165.9472.6369.1365.0170.9567.85
271.4759.2664.8065.7062.1863.89
370.6361.1365.5467.4968.3367.91
482.8548.7761.4080.0460.9269.19
557.7257.8157.7640.4378.8853.46
678.3245.9957.9564.7660.6562.64
749.5082.3061.8251.5080.9962.97
850.2573.0659.5548.9770.1157.66
959.4268.0663.4552.0474.0061.11
1070.0063.5766.6357.8179.9567.10
Mean65.6163.2664.4159.3870.7064.54
Table 6. The quantitative results for this experiment. Using the negative mining method for background pre-processing on the D2 dataset and adding it to the training set.
Table 6. The quantitative results for this experiment. Using the negative mining method for background pre-processing on the D2 dataset and adding it to the training set.
Fold
Number
Dice LossFocal Tversky Loss
PrecisionSensitivityF1-ScorePrecisionSensitivityF1-Score
164.7969.7967.1959.5976.9267.16
270.6759.4064.5559.2771.7564.92
377.2748.0559.2667.9662.6165.18
485.5653.5265.8583.3654.1065.62
557.4663.4760.3246.5269.6155.77
680.6946.1058.6874.9259.4566.29
757.0673.5264.2548.7981.9361.16
854.7262.9858.5650.7971.0059.22
965.6062.0763.7960.2665.4362.74
1060.7574.8967.0858.7173.3565.22
Mean67.4661.3864.2861.0268.6264.59
Table 7. The model performance with transfer learning.
Table 7. The model performance with transfer learning.
Fold
Number
Dice LossFocal Tversky Loss
PrecisionSensitivityF1-ScorePrecisionSensitivityF1-Score
167.8170.2869.0266.5172.0469.17
271.4463.7567.3870.8060.6465.33
376.7560.7367.8069.8769.2569.56
486.5651.5464.6181.8060.2869.41
568.8862.8465.7251.2866.7858.01
684.1443.4657.3180.0052.2063.18
762.0766.3464.1450.6986.0463.80
851.9072.1760.3843.8285.6257.97
962.3064.4463.3552.0777.7862.38
1064.9874.4769.4063.9277.5670.09
Mean69.6863.0066.1763.0870.8266.72
Table 8. Comparison with network architecture and analysis results from related studies.
Table 8. Comparison with network architecture and analysis results from related studies.
MethodRegionPrecisionSensitivityF1-ScoreIoU
ResU-Net [20]Thorax77.2167.88-61.03
U-Net [21]Thorax--57.1063.30
U-Net++ [16]Thorax68.8562.5765.56-
UNet3+ [22]Thorax61.2068.3364.33-
Swin
Transformer [23]
Thorax + Pelvis--77.8135.59
OursWhole body
excluded below knees
69.9663.5566.60-
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, Y.-Y.; Yu, P.-N.; Lai, Y.-C.; Hsieh, T.-C.; Cheng, D.-C. Bone Metastases Lesion Segmentation on Breast Cancer Bone Scan Images with Negative Sample Training. Diagnostics 2023, 13, 3042. https://doi.org/10.3390/diagnostics13193042

AMA Style

Chen Y-Y, Yu P-N, Lai Y-C, Hsieh T-C, Cheng D-C. Bone Metastases Lesion Segmentation on Breast Cancer Bone Scan Images with Negative Sample Training. Diagnostics. 2023; 13(19):3042. https://doi.org/10.3390/diagnostics13193042

Chicago/Turabian Style

Chen, Yi-You, Po-Nien Yu, Yung-Chi Lai, Te-Chun Hsieh, and Da-Chuan Cheng. 2023. "Bone Metastases Lesion Segmentation on Breast Cancer Bone Scan Images with Negative Sample Training" Diagnostics 13, no. 19: 3042. https://doi.org/10.3390/diagnostics13193042

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop