Next Article in Journal
Diagnosis and Management of Gastrointestinal Manifestations in Children with Cystic Fibrosis
Previous Article in Journal
Exceptional Liver Transplant Indications: Unveiling the Uncommon Landscape
Previous Article in Special Issue
Application of Machine Learning Algorithm in Predicting Axillary Lymph Node Metastasis from Breast Cancer on Preoperative Chest CT
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Enhanced LightGBM-Based Breast Cancer Detection Technique Using Mammography Images

by
Abdul Rahaman Wahab Sait
1,* and
Ramprasad Nagaraj
2
1
Department of Documents and Archive, Center of Documents and Administrative Communication, King Faisal University, P.O. Box 400, Hofuf 31982, Al-Ahsa, Saudi Arabia
2
Department of Biochemistry, S S Hospital, S S Institute of Medical Sciences & Research Centre, Rajiv Gandhi University of Health Sciences, Davangere 577005, Karnataka, India
*
Author to whom correspondence should be addressed.
Diagnostics 2024, 14(2), 227; https://doi.org/10.3390/diagnostics14020227
Submission received: 4 December 2023 / Revised: 18 January 2024 / Accepted: 19 January 2024 / Published: 22 January 2024
(This article belongs to the Special Issue Artificial Intelligence in Cancers)

Abstract

:
Breast cancer (BC) is the leading cause of mortality among women across the world. Earlier screening of BC can significantly reduce the mortality rate and assist the diagnostic process to increase the survival rate. Researchers employ deep learning (DL) techniques to detect BC using mammogram images. However, these techniques are resource-intensive, leading to implementation complexities in real-life environments. The performance of convolutional neural network (CNN) models depends on the quality of mammogram images. Thus, this study aimed to build a model to detect BC using a DL technique. Image preprocessing techniques were used to enhance image quality. The authors developed a CNN model using the EfficientNet B7 model’s weights to extract the image features. Multi-class classification of BC images was performed using the LightGBM model. The Optuna algorithm was used to fine-tune LightGBM for image classification. In addition, a quantization-aware training (QAT) strategy was followed to implement the proposed model in a resource-constrained environment. The authors generalized the proposed model using the CBIS-DDSM and CMMD datasets. Additionally, they combined these two datasets to ensure the model’s generalizability to diverse images. The experimental findings revealed that the suggested BC detection model produced a promising result. The proposed BC detection model obtained an accuracy of 99.4%, 99.9%, and 97.0%, and Kappa (K) values of 96.9%, 96.9%, and 94.1% in the CBIS-DDSM, CMMD, and combined datasets. The recommended model streamlined the BC detection process in order to achieve an exceptional outcome. It can be deployed in a real-life environment to support physicians in making effective decisions. Graph convolutional networks can be used to improve the performance of the proposed model.

1. Introduction

BC is a neoplastic condition characterized by the formation of a malignant tumor originating from the breast cells [1]. It is the most prevalent form of cancer observed in women, with a lesser incidence in men across the globe [1,2]. In 2020, the International Agency for Research on Cancer reported more than 2.26 million new cases and about 685,000 mortalities globally [2]. The incidence of BC has increased over time [3]. After leukemia, BC is the leading cause of mortality in Saudi Arabia [3]. Globally, more than 100 million mammograms are performed annually to detect breast cancer [4]. Each mammography necessitates a minimum of two evaluations by expert radiologists to detect abnormalities and provide a detailed analysis of the mammographic image [4,5]. These factors render BC screening increasingly expensive and resource-intensive.
Chemotherapy and radiation therapy can significantly affect an individual’s physical and mental health [6]. Radiation or surgery can affect the lymphatic system, resulting in lymphedema [7]. Advanced treatment may cause a financial burden to individuals. As a result, individuals may face challenges in obtaining critical medical care. Earlier detection of BC can reduce its financial burden and BC’s mortality rate [8]. In addition, healthcare centers can provide personalized diagnoses based on BC stages.
Healthcare centers employ various techniques to screen and diagnose BC. Sound waves are used in ultrasound scanning techniques to represent breast tissue visually [8]. However, the expertise of the operators plays a crucial role in effective BC detection. Magnetic resonance imaging employs magnetic fields and radio waves to identify tumors [8]. It is widely used to detect abnormalities in advanced stages of BC. In addition, biopsy procedures are used to extract breast tissue specimens for BC staging [8]. Nonetheless, the invasive nature of this procedure may affect patients’ physical and mental health.
Mammography is a specialized form of medical imaging that uses low-dose X-ray to visualize the internal organs of the breasts [8,9,10]. It is common practice to utilize mammography as a screening method for women who do not exhibit any symptoms of BC, particularly in age groups where the likelihood of acquiring BC is higher [11]. Regular screenings have the potential to detect anomalies prior to the manifestation of symptoms. There have been an extensive number of studies suggesting that routine mammographic screening programs lead to a reduction in BC mortality rates [12,13,14]. Mammography screening is affordable and readily accessible [14]. Advancements in imaging technologies have facilitated the rapid development of artificial intelligence (AI) for breast imaging. Computer-aided detection identifies potentially abnormal lesions, such as masses and microcalcifications [15]. On the other hand, computer-aided diagnosis analyzes a lesion’s suspicious region and estimates the likelihood of its occurrence.
The sensitivity, specificity, and capability to manage the complexity of mammogram images present potential obstacles to conventional BC screening approaches [16]. Early-stage BC patterns may be hidden in complex mammograms. The level of complexity exceeds the capabilities of the current image processing methods, which could cause diagnostic delays or false negatives (FNs) [17]. Traditional BC detection techniques are not generalizable to the broader population. They require substantial computational resources and expert opinions are needed to make a decision.
Researchers have recently employed DL models for detecting BC with optimal outcomes [17]. The detection and diagnosis of BC can differ among radiologists due to variances in human mammography interpretation. DL models can learn complicated hierarchical features, potentially enhancing breast cancer detection sensitivity and specificity [18]. This facilitates an earlier identification of BC conditions and improves the precision of diagnostic procedures. The number of highly specialized professionals can be minimized using DL models [18]. A DL model can be integrated with edge and mobile devices to remotely screen individuals [18]. In addition, it can identify subtle signs of abnormalities in mammography images. In the context of BC diagnosis, it is possible to employ a pretrained CNN model on a substantial dataset, such as ImageNet, and subsequently fine-tune it using a smaller dataset consisting of mammography images [19]. Transfer learning (TL) can improve generalization using the pretrained model’s weights to identify BC’s low- and mid-level features [19]. Physicians can employ DL-based models to detect BC using mammography images. The models can handle a massive volume of images, which supports healthcare centers in diagnosing large numbers of individuals [19]. The pretrained models can extract critical features with limited resources. In addition, fine-tuning and image preprocessing techniques can improve the feature extraction process [20].
By automating the identification of BC with DL models, healthcare providers may process multiple mammograms in a shorter period, substantially enhancing screening efficiency. A lightweight BC detection model can be implemented in edge and mobile devices. Healthcare centers can benefit from this model and render their services in remote areas. The proposed study aims to offer reliable and unbiased assessments in BC detection. Thus, the authors build a DL-based BC detection model with limited computational resources. This study’s contributions are as follows:
i.
A feature engineering technique using QAT strategy to generate critical features of BC.
The authors built a CNN model with the EfficientNet B7 model’s weights. QAT strategy was used to improve the performance of the EfficientNet B7 model in a resource-constrained environment.
ii.
Development of a fine-tuned LightGBM-based BC detection model.
The proposed feature engineering presents valuable features and passes them to the LightGBM model in order to detect BC from mammography images. In addition, the authors fine-tuned the LightGBM model using the Optuna algorithm.
The remaining parts of this study are organized as follows: Section 2 offers the research methodology for BC detection using mammography images. The experimental outcomes are outlined in Section 3. Section 4 discusses the significance of the proposed study in detecting BC. Lastly, Section 5 presents the features and limitations of the proposed research.

2. Materials and Methods

The authors built a DL-based BC detection model using effective image preprocessing, augmentation, and feature extraction techniques. Image preprocessing techniques were used to improve the quality of mammography images. In addition, image augmentation techniques were employed to generate additional training samples and improve prediction accuracy. DL approaches offer the integration of numerous models in order to provide a final prediction. Improving the efficiency of CNN models can address complex problems, improve predictive accuracy, and provide deeper insights into the relationships between structured and unstructured data in various domains, including healthcare, finance, manufacturing, and more. Additionally, overall performance of the BC detection model was improved by utilizing a number of distinct models. Using critical features of BC images, the DL approach can learn minor details to predict BC.
The EfficientNet B7 model can learn complex patterns from its training data [21]. Compared to traditional CNN models, EfficientNet B7 offers an impressive performance with few parameters. It captures intricate image patterns and representations with top-tier performance on benchmark datasets. It offers remarkable efficacy across multiple image classification tasks, rendering it a comprehensive model suitable for medical imaging classification. In addition, in order to achieve a high level of precision, a compound scaling technique can be used to scale the depth and width of the model [22]. However, training the EfficientNet B7 model can be significantly resource-intensive and time-consuming. EfficientNet B7 may not be able to generalize effectively to completely new or unknown data. In a real-world setting, this constraint may affect model performance. Improving the EfficienNet B7 model’s performance can support the proposed model to generalize on real-time data. The LightGBM model encompasses the ensemble tree approach to handle the massive datasets [23].
The authors were motivated to construct a feature engineering model using EfficientNet B7 and employ the LightGBM model to predict BC using its features. Figure 1 presents the proposed BC detection model.

2.1. Image Acquisition

In order to train the proposed BC detection model, the authors utilized the CBIS-DDSM and CMMD datasets, which are available in repositories [24,25]. The CBIS-DDSM dataset contains enhanced images of the DDSM dataset [26,27]. The dataset providers addressed the limitations of the DDSM dataset. They removed low-quality images and reannotated the images. They converted the quality of the primary images to 16-bit grayscale TIFF images. Spurious ground-labeled images were removed from the dataset. The interoperability and reusability of the dataset were improved by converting analog images into digitized images.
The CMMD dataset includes two datasets. CMMD1 encompasses mammography and clinical data, whereas CMMD2 covers malignant BC images. The dataset owner obtained the images from Sun-Yat-Sen University Cancer Center in Guangzhou and Nanhai Affiliated Hospital of Southern Medical University, Guangdong, China. The data were collected between July 2012 and July 2016. To ensure the efficiency of the proposed BC detection model and to use a unique dataset, the authors combined the CBIS-DDSM and CMMD datasets. They included benign and malignant images from the MIAS [28] and BCDR-D [29] datasets. Table 1 highlights the details of the datasets. Figure 2a,b show sample images from the CBIS-DDSM and CMMD datasets.

2.2. Image Preprocessing

Noises and artifacts can reduce the performance of BC detection models in identifying the abnormalities. Artifacts include grid lines, scratches, and distortions that can interfere with BC identification. These may mimic features related to lesions and tumors that may influence the BC detection model to generate false positives (FPs) and FNs. Image preprocessing techniques can improve the performance of the CNN models [20]. For instance, the EfficientNet model requires an image size of 224 × 224. In addition, image normalization helps pretrained models to extract key features of BC images. Thus, image preprocessing is required to improve EfficientNet B7-based feature extraction. The authors employed multiple image preprocessing techniques to overcome the shortcomings of BC images. Contrast-limited adaptive histogram equalization (CLAHE) was employed to enhance the local contrast of an image by preventing noise amplification. It applies histogram equalization to the smaller regions of BC images. Equation (1) presents the mathematical form of the image resizing process with the CLAHE method.
I i = R e s i z e _ C L A H E I i , c , t ,   i = 1 ,   2 ,     n
where I i is the BC image, c is the clip limit, and t is the size of the tile grid.

2.3. Image Augmentation

In order to improve the training environment, the authors employed generative adversarial networks (GANs) [21,30]. They followed previous studies [21,30] to generate mammogram images. A Unet-based generator was used to produce synthetic images for multiple classes. A CNN model was employed to build a discriminator model in order to authenticate the synthetic images. Thus, extended training was not required for image generation.

2.4. Feature Engineering

The authors constructed a CNN model with four convolutional layers, batch normalization, and Leaky ReLu. Figure 3 highlights the proposed feature engineering process for extracting crucial features. The EfficientNet B7 model’s weights were used for the feature extraction.
EfficientNet B7’s architecture relies heavily on inverted residuals. It uses a lightweight depth-wise separable convolution, succeeded by a linear bottleneck and a skip connection. This architecture facilitates the efficient capturing of intricate patterns. Compound scaling scales the model in depth, width, and resolution. It ensures that the model’s power increases with its size while retaining its efficiency. Mobile inverted convolution is a composite operation that integrates the principles of depth-wise separable convolution and linear bottleneck. This procedure possesses high computational efficiency and contributes to reducing the number of parameters while preserving meaningful patterns.
The EfficientNet models were trained using the larger datasets. The authors froze the lower-level layers of the EfficientNet B7 model to prevent overfitting and promote better generalization. The QAT strategy was used to improve the efficiency of the proposed BC detection model [31]. This prepares the proposed model for deployment in resource-constrained environments. In addition, it reduces the computational cost and memory requirements of the proposed model. The authors applied dropout layers to address overfitting by using assigned probabilities. The extracted features were flattened and forwarded to fully connected (FCN) layers. The authors integrated the FCN layers to collect the outcomes for each class.

2.5. BC Classification

LightGBM is an ensemble learning-based gradient-boosting or tree-based learning model. It improves prediction efficiency, manages extensive datasets, and minimizes training time. It is commonly suggested for the analysis of tabular datasets. LightGBM uses leaf-wise splits to generate more complex trees, reducing loss while strengthening accuracy. The partitioning process uses a unique sampling technique called gradient-based one-side sampling (GOSS). This method involves excluding data points with tiny gradients and utilizing the rest of the data for estimating information gain and facilitating tree development. Equation (2) shows the mathematical form of computing the gradients of the loss function.
g b c = g b c + λ × m a x g b c , γ  
where g b c is the gradient of the loss function in classifying BC, λ is the variable to control the number of one-side sampling, and γ is the gradient threshold.
After the training phase, the FCN layer was replaced with LightGBM to achieve optimal accuracy. The authors utilized the GOSS technique to classify features. The proposed technique excludes smaller gradients and considers the remaining features to generate an outcome. It can produce significant information gain in resource-constrained environments. In addition, it requires less training time and low memory and is compatible with smaller and larger datasets. Figure 4 highlights multi-class classification using the LightGBM model. Equation (3) shows the LightGBM-based BC detection model.
C l a s s i f i c a t i o n = L i g h t G B M   X i ,   L ,   N ,   M ,   i = 1 ,   2 ,     n
where X is the BC image feature, L is the learning rate, n is the number of estimators, and M is the maximum depth.
To fine-tune the performance of the proposed BC model, the authors applied Optuna hyperparameter optimization. Optuna is an open-source framework that can be implemented in the TensorFlow and PyTorch frameworks. It employs tree-structured parzen estimators to determine the hyperparameters of the LightGBM model. Equation (4) outlines Optuna hyperparameter optimization.
Best _ Parameters = Optuna LightGBM Parameters
where Best_Parameters is the optimized hyperparameters of the LightGBM model, Optuna is the hyperparameter optimization function, and LightGBM is the BC classification model.

2.6. Performance Evaluation

The authors employed multiple evaluation metrics to measure the performance of the BC detection model. The accuracy (Accy) metric identifies the model’s capability to detect BC. Table 2 describes the notations of true positives (TPs), true negatives (TNs), FPs, and FNs. Equation (5) presents the mathematical form of the accuracy metric.
A c c u r a c y = B T P + B T N B T P + B T N + B F P + B F N
The precision (Pren) metric determines the number of FPs in the outcome, whereas the recall (Recl) metric measures the model’s ability to find TPs. F1-Score (F1) is used to ensure that there is no uneven class distribution in the outcome. Equations (6)–(8) outline the computational forms of Pren, Recl, and F1.
P r e n = B T P B T P + B F P
R e c n = B T P B T P + B F N
F 1 = 2 × P r e n × R e c n P r e n + R e c n
Furthermore, the authors used Cohen’s Kappa (K) and Matthew correlation coefficient (MCC) to evaluate the multi-class classification performance of the proposed model. The confidence interval (CI), SD, and computational loss were calculated to measure the uncertainties of the proposed model. Equations (9) and (10) present the mathematical forms of K and MCC.
K = 2 × B T P × B T N B F N × B F P B T P + B F P + B F P + B T N + B T P + B F N + B F N + B T N
M C C = B T N × B F P B F N × B F P   B T P + B F P + B T P + B F N + B T N + B F P + B T N + B F N

3. Results

The authors employed Windows 10 Professional, Intel i7 with 16 GB RAM, and NVIDIA GeForce RTX 3050 to implement the proposed BC detection model. The source codes of the EfficientNet B7 and LightGBM models were extracted from the Github repositories [32,33]. The Optuna algorithm was used to fine-tune the hyperparameters of the LightGBM model. The PyTorch 2.0 and TensorFlow v2.15.0 libraries were used to implement the proposed model. The computational settings for the model implementation are presented in Table 3.
Table 4 outlines the findings of the proposed model’s performance in the analysis of the CBIS-DDSM dataset. The suggested QAT strategy supported the proposed feature extraction to extract meaningful features. The model produced an optimal outcome for each class.
Table 5 reveals a significant improvement in the multi-class classification ability of the proposed model. The recommended image preprocessing and augmentation techniques assisted the EfficientNet B7 model in extracting key features. In addition, the suggested hyperparameter optimization technique improved the proposed model’s classification accuracy.
The outcomes of the model’s performance using the combined dataset are presented in Table 6. The QAT strategy assisted the recommended model in achieving a better outcome in the combined dataset. In addition, the proposed feature extraction improved the BC detection model’s efficiency. Figure 5 highlights the performance outcomes of the proposed model.
Table 7 highlights the batch-wise performance of the suggested BC detection model on the CBIS-DDSM dataset. The recommended EL approach improved the performance of the proposed model in addressing the existing limitations of BC detection.
Table 8 reveals the batch-wise performance of the proposed BC detection model. The model achieved reasonable results in a batch size of 12. However, the authors extended the analysis to a batch size of 16. The QAT strategy helped the proposed model to produce better results by addressing data imbalance.
The suggested model’s batch-wise performance is listed in Table 9. It achieved a significant improvement at batch 16. The fine-tuned LightGBM model classified the BC images with an optimal accuracy.
The generalization of the BC detection model to the CBIS-DDSM dataset is presented in Table 10. The proposed model outperformed the existing models, obtaining a promising result. The fine-tuned LightGBM model classified the images with optimal accuracy.
Table 11 presents the outcome of generalizing the proposed BC model to the CMMD dataset. The number of images in the CMDD dataset is larger compared to the CBIS-DDSM dataset. However, the proposed model obtained a superior outcome compared to existing models.
The outcomes of comparative analysis using the combined dataset are revealed in Table 12. The proposed BC detection model addresses the challenges in detecting BC using BC images using the fine-tuned EfficientNet B7 and LightGBM models.
Table 13 highlights the computational complexities of BC detection models. The proposed BC model requires fewer parameters and floating-point operations (FLOPs) for image classification. Inception V3 and EfficientNet B7 needed additional parameters to achieve a reasonable outcome.
Lastly, Table 14 shows the statistical significance of BC detection models. The findings highlight the reliability and trustworthiness of the proposed model. Based on the outcomes, the proposed model can handle out-of-distribution samples and produce reliable results in a real-life environment. In addition, healthcare practitioners can make effective decisions using the proposed BC detection model.

4. Discussions

The authors built a BC detection model to identify BC in resource-constrained environments. The proposed model includes image preprocessing and augmentation techniques, feature engineering, and classifier models. Using the image preprocessing technique, the authors enhanced image quality. An image augmentation technique was used to improve the prediction accuracy of the proposed model. CNN-based feature engineering was used to extract the key features of BC images. The authors used the weights of the EfficientNet B7 model for feature extraction. The Optuna algorithm was employed to fine-tune the hyperparameters of the LightGBM model.
Batch-wise performance analysis outcomes suggest that the model effectively addresses overfitting and class imbalance. Table 10, Table 11 and Table 12 highlight the findings of generalizing the BC model to the CBIS-DDSM, CMMD, and Combined datasets. It is evident that the proposed model obtained a superior outcome. The suggested image preprocessing and feature extraction techniques significantly improved the proposed model’s classification performance. Table 13 reveals the computational complexities of implementing BC detection models. In all three datasets, the proposed model demanded few parameters and FLOPs to achieve an accuracy of 99.4%, 99.9%, and 97.0% and a K value of 96.9%, 96.9%, and 94.1%. In contrast, existing models required a higher number of parameters for feature extraction. The authors employed the Optuna algorithm to fine-tune the LightGBM model’s hyperparameters to classify BC images in a resource-constrained environment. Table 14 discusses the statistical significance of BC detection models. The proposed model generated outcomes with minimal losses of 0.4, 0.3, and 0.3, SD of 0.0003, 0.0002, and 0.0003, and CI range of [98.1–98.5], [97.6–98.2], and [98.1–98.4] using the three datasets. The minimal computational loss and optimal accuracy revealed that there is no overfitting in the proposed model.
The findings reveal that the model is reliable and can handle variations in BC images. The proposed model can seamlessly integrate into the clinical workflow and support healthcare professionals in making effective decisions. The suggested approach can reduce the risk of errors and improve the overall outcome of the proposed BC detection model. The fine-tuned LightGBM model reduced the number of FPs and FNs and accurately identified normal, malignant, and benign images. The BC detection model enables healthcare providers to prioritize patient care, consultation, and complex decision-making while effectively managing routine assessments. It achieves a high performance level and contributes to BC’s reliable and standardized identification. Its few requirements for retesting, greater efficiency, and the possibility of earlier intervention can lead to overall cost-effectiveness. To improve BC screening services, the suggested BC detection model can be a valuable tool to supplement existing expertise. By facilitating prompt treatment with the potential for improved outcomes, early and precise detection can empower patients.
The authors generalized the model using three datasets. The generalization findings highlighted the significance of the suggested BC detection model in detecting benign, malignant, and normal images. Feature extraction assisted the LightGBM model in producing a superior outcome. Integrating diverse images into the combined dataset reinforced the model’s ability to identify BC’s intricate patterns. The inherent features of QAT strategies contributed to a significant performance improvement. The suggested feature extraction technique enabled the LightGBM model to overcome class imbalances. The regularization features of the LightGBM model supported the proposed model in generalizing to unseen images.
El Houby and Yassin [9] classified malignant and nonmalignant breast lesions using a CNN model. The model required a larger number of parameters and FLOPs to classify the images. Umer et al. [12] built a BC detection model using convoluted features. Falconi et al. [22] fine-tuned pretrained models for classifying BC images. Pretrained models are designed for a specific set of classes. They may face challenges in handling multiple classes of BC. Agarwal et al. [34] used the Inception V3 and ResNet 50 models for BC detection. The complex architecture of these CNN models required additional training time and a larger number of parameters for BC detection. In addition, the ResNet 50 model demands huge computational resources to store and process the parameters. Zahoor et al. [35] applied a deep neural network and entropy-controlled whale optimization algorithms to classify BC images. They used Whale optimization for feature extraction. The MobileNet V2 and NasNet models were used for classification. The integration of CNN models and the implementation of Whale optimization required additional memory resources. Boudouh et al. [36] employed a TL technique to detect abnormalities in BC images. Bai et al. [37] developed a BC detection model using a feature fusion Siamese network. Their model faced challenges in capturing minor details of BC. The limitations of CNN models reduced their performance in BC detection.
Furthermore, Bobowicz et al. [38] employed an attention-based DL system for BC classification. Levy and Jain [39] used the CBIS-DDSM dataset to train the VGG-16, ResNet 50, and Inception V3 models. They achieved an average accuracy of 84.16%. Ahmed et al. [40] proposed semantic and instance segmentation techniques for classifying BC images. Hameed et al. [41] proposed a BC detection model using histopathology images. In addition, the authors employed recent models, including Yolo-V8, CatBoost, and ShuffleNet V2, for comparison analysis using combined datasets. The current models required additional computational resources, restricting them from producing an exceptional outcome. Compared to our proposed BC detection model, the architecture of Inception v3 is complex, which may affect the interpretation of the outcomes. Implementing a BC detection model based on Inception V3 may demand substantial computational resources. Extended training is required for the Inception V3 model to produce a reasonable outcome. The EfficientNet B7 model demands considerable computation time for image classification. Training and deploying the model may require significant computer power and memory, which can be resource-intensive. In order to maintain the configurations of EfficientNet B7, a considerable amount of memory is required. This can make its implementation less straightforward on devices with limited memory capacity. In contrast, the recommended BC detection model requires limited computational resources for BC detection. It achieved an extraordinary outcome by addressing the existing limitations using effective image preprocessing and augmentation techniques.
To deploy the proposed BC detection model, healthcare centers require DL professionals. The proposed BC detection model may demand additional training to streamline the detection process in a real-life environment. In real-time data, mammography images may include unique sizes, noises, and artifacts. A substantial image preprocessing technique is needed to improve EfficientNet B7-based feature extraction. Fine-tuning processes are required to implement the recommended model in resource-constrained environments. In addition, healthcare centers should integrate an application into the proposed model to maintain patient data privacy and security.
The proposed model produced promising results in detecting BC. However, it has certain limitations. Massive volumes of labeled data may be essential in order to train the suggested model to generalize effectively to novel and unseen cases. An imbalance in class distribution can impact the model’s accuracy. The model’s efficacy may vary depending on the population, ethnicity, or imaging modalities. Continuous monitoring is required to generalize the proposed BC detection model to a broader population. It is necessary to update the proposed model to handle new information as healthcare data evolve over time. In order to promote its smooth adoption, the proposed model should align with healthcare practitioners’ specific requirements. In the future, the performance of the proposed BC model can be improved using a graph convolutional network.

5. Conclusions

The proposed study addressed the existing limitations in detecting BC using mammography images. Early detection of BC can reduce the mortality rate and support physicians to provide effective treatments. The authors followed an image preprocessing technique to enhance image quality. Key features were generated using the EfficientNet B7 model. The fine-tuned LightGBM detected BC with limited computational resources. The authors generalized the proposed model using the CBIS-DDSM and CMMD datasets. The experimental outcome revealed a significant improvement in the performance of the proposed model. The proposed model obtained an exceptional result compared to other recent models. The findings highlighted that the model can be implemented in healthcare centers to identify BC in earlier stages. The authors faced a few challenges in deploying the proposed model. Substantial training is required to overcome the shortcomings of mammography images. The complexity of EfficientNet B7 may reduce the performance of the feature extraction process. Generalizing the proposed model using diverse datasets can minimize the limitations in identifying crucial patterns of BC. However, the authors improved the training process using the QAT approach. The proposed model can be extended using graph convolutional networks.

Author Contributions

Conceptualization, A.R.W.S. and R.N.; methodology, A.R.W.S.; software, A.R.W.S.; validation, A.R.W.S. and R.N.; formal analysis, A.R.W.S.; investigation, A.R.W.S. and R.N.; resources, A.R.W.S.; data curation, A.R.W.S. and R.N.; writing—original draft preparation, A.R.W.S. and R.N.; writing—review and editing, A.R.W.S. and R.N.; visualization, A.R.W.S.; supervision, A.R.W.S.; project administration, A.R.W.S.; funding acquisition, A.R.W.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Deanship of Scientific Research, Vice Presidency for Graduate Studies and Scientific Research, King Faisal University, Saudi Arabia (Grant No. 5499).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

1. CBIS-DDSM Dataset. Available online: https://www.kaggle.com/datasets/awsaf49/cbis-ddsm-breast-cancer-image-dataset (accessed on 3 March 2023). 2. CMMD Dataset. Available online: https://wiki.cancerimagingarchive.net/pages/viewpage.action?pageId=70230508 (accessed on 5 March 2023).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Basudan, A.M. Breast Cancer Incidence Patterns in the Saudi Female Population: A 17-Year Retrospective Analysis. Medicina 2022, 58, 1617. [Google Scholar] [CrossRef]
  2. Ministry of Health, Saudi Arabia. Available online: https://www.moh.gov.sa/en/HealthAwareness/healthDay/2023/Pages/HealthDay-2023-10-01-31.aspx (accessed on 2 January 2023).
  3. Ramadan, M.; Alsiary, R.; Alsaadoun, N.; Alhusseini, N.; Raihan Sajid, M.; Mohamed Hamed, N.; Ziad Arabi, T.; Nedal Sabbah, B. Risk of Breast Cancer Progression after Treatment in the Western Region of Saudi Arabia. Breast Cancer Basic Clin. Res. 2023, 17, 11782234231158270. [Google Scholar] [CrossRef] [PubMed]
  4. Baek, S.J.; Ko, K.H.; Jung, H.K.; Park, A.Y.; Koh, J. Comparison of abbreviated MRI with mammography and ultrasound in women with a personal history of breast cancer. Acad. Radiol. 2022, 29, S19–S25. [Google Scholar] [CrossRef] [PubMed]
  5. Wong, D.J.; Gandomkar, Z.; Wu, W.J.; Zhang, G.; Gao, W.; He, X.; Wang, Y.; Reed, W. Artificial intelligence and convolution neural networks assessing mammographic images: A narrative literature review. J. Med. Radiat. Sci. 2020, 67, 134–142. [Google Scholar] [CrossRef] [PubMed]
  6. Altameem, A.; Mahanty, C.; Poonia, R.C.; Saudagar, A.K.J.; Kumar, R. Breast cancer detection in mammography images using deep convolutional neural networks and fuzzy ensemble modeling techniques. Diagnostics 2022, 12, 1812. [Google Scholar] [CrossRef] [PubMed]
  7. Shen, L.; Margolies, L.R.; Rothstein, J.H.; Fluder, E.; McBride, R.; Sieh, W. Deep learning to improve breast cancer detection on screening mammography. Sci. Rep. 2019, 9, 12495. [Google Scholar] [CrossRef]
  8. Pereira, R.D.O.; Luz, L.A.D.; Chagas, D.C.; Amorim, J.R.; Nery-Júnior, E.D.J.; Alves, A.C.B.R.; Abreu-Neto, F.T.D.; Oliveira, M.D.C.B.; Silva, D.R.C.; Soares-Júnior, J.M.; et al. Evaluation of the accuracy of mammography, ultrasound and magnetic resonance imaging in suspect breast lesions. Clinics 2020, 75, e1805. [Google Scholar] [CrossRef]
  9. El Houby, E.M.; Yassin, N.I. Malignant and nonmalignant classification of breast lesions in mammograms using convolutional neural networks. Biomed. Signal Process. Control 2021, 70, 102954. [Google Scholar] [CrossRef]
  10. Sannasi Chakravarthy, S.; Bharanidharan, N.; Rajaguru, H. Multi-deep CNN based experimentations for early diagnosis of breast cancer. IETE J. Res. 2022, 69, 7326–7341. [Google Scholar] [CrossRef]
  11. Logan, J.; Kennedy, P.J.; Catchpoole, D. A review of the machine learning datasets in mammography, their adherence to the FaIR principles and the outlook for the future. Sci. Data 2023, 10, 595. [Google Scholar] [CrossRef]
  12. Umer, M.; Naveed, M.; Alrowais, F.; Ishaq, A.; Hejaili, A.A.; Alsubai, S.; Eshmawi, A.A.; Mohamed, A.; Ashraf, I. Breast Cancer Detection Using Convoluted Features and Ensemble Machine Learning Algorithm. Cancers 2022, 14, 6015. [Google Scholar] [CrossRef] [PubMed]
  13. Singh, H.; Sharma, V.; Singh, D. Comparative analysis of proficiencies of various textures and geometric features in breast mass classification using k-nearest neighbor. Vis. Comput. Ind. Biomed. Art 2022, 5, 3. [Google Scholar] [CrossRef] [PubMed]
  14. Lou, M.; Wang, R.; Qi, Y.; Zhao, W.; Xu, C.; Meng, J.; Deng, X.; Ma, Y. MGBN: Convolutional neural networks for automated benign and malignant breast masses classification. Multimed. Tools Appl. 2021, 80, 26731–26750. [Google Scholar] [CrossRef]
  15. Ragab, M.; Albukhari, A.; Alyami, J.; Mansour, R.F. Ensemble deep-learning-enabled clinical decision support system for breast cancer diagnosis and classification on ultrasound images. Biology 2022, 11, 439. [Google Scholar] [CrossRef] [PubMed]
  16. Barsha, N.A.; Rahman, A.; Mahdy, M.R.C. Automated detection and grading of Invasive Ductal Carcinoma breast cancer using ensemble of deep learning models. Comput. Biol. Med. 2021, 139, 104931. [Google Scholar] [CrossRef] [PubMed]
  17. Haq, I.U.; Ali, H.; Wang, H.Y.; Lei, C.; Ali, H. Feature fusion and Ensemble learning-based CNN model for mammographic image classification. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 3310–3318. [Google Scholar] [CrossRef]
  18. Humayun, M.; Khalil, M.I.; Almuayqil, S.N.; Jhanjhi, N.Z. Framework for detecting breast cancer risk presence using deep learning. Electronics 2023, 12, 403. [Google Scholar] [CrossRef]
  19. Yengec Tasdemir, S.B.; Tasdemir, K.; Aydin, Z. A review of mammographic region of interest classification. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2020, 10, e1357. [Google Scholar] [CrossRef]
  20. Ebenezer, A.S.; Kanmani, S.D.; Sivakumar, M.; Priya, S.J. Effect of image transformation on EfficientNet model for COVID-19 CT image classification. Mater. Today Proc. 2022, 51, 2512–2519. [Google Scholar] [CrossRef]
  21. Joseph, A.J.; Dwivedi, P.; Joseph, J.; Francis, S.; Pournami, P.N.; Jayaraj, P.B.; Shamsu, A.V.; Sankaran, P. Prior-guided generative adversarial network for mammogram synthesis. Biomed. Signal Process. Control 2024, 87, 105456. [Google Scholar] [CrossRef]
  22. Falconi, L.; Pérez, M.; Aguilar, W.G.; Conci, A. Transfer learning and fine tuning in breast mammogram abnormalities classifcation on cbis-ddsm database. Adv. Sci. Technol. Eng. Syst. 2020, 5, 154–165. [Google Scholar] [CrossRef]
  23. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.Y. Lightgbm: A highly efficient gradient boosting decision tree. In Advances in Neural Information Processing Systems 30; Neural Information Processing Systems Foundation, Inc.: South Lake Tahoe, NV, USA, 2017. [Google Scholar]
  24. Lee, R.S.; Gimenez, F.; Hoogi, A.; Miyake, K.K.; Gorovoy, M.; Rubin, D.L. A curated mammography data set for use in computer-aided detection and diagnosis research. Sci. Data 2017, 4, 170177. [Google Scholar] [CrossRef] [PubMed]
  25. Cai, H.; Wang, J.; Dan, T.; Li, J.; Fan, Z.; Yi, W.; Cui, C.; Jiang, X.; Li, L. An Online Mammography Database with Biopsy Confirmed Types. Sci. Data 2023, 10, 123. [Google Scholar] [CrossRef]
  26. Heath, M.; Bowyer, K.; Kopans, D.; Kegelmeyer, W.P.; Moore, R.; Chang, K.; MunishKumaran, S. Current status of the Digital Database for Screening Mammography. In Digital Mammography, Proceedings of the Fourth International Workshop on Digital Mammography, Nijmegen, The Netherlands, 7–10 June 1998; Kluwer Academic Publishers: Dordrecht, The Netherlands, 1998; pp. 457–460. [Google Scholar]
  27. Bowyer, K.; Kopans, D.; Moore, R.; Kegelmeyer, W.P. The Digital Database for Screening Mammography, Michael Heath. In Proceedings of the Fifth International Workshop on Digital Mammography, Toronto, ON, Canada, 11–14 June 2000; Yaffe, M.J., Ed.; Medical Physics Publishing: Madison, WI, USA, 2001; pp. 212–218. ISBN 1-930524-00-5. [Google Scholar]
  28. MIAS Dataset. Available online: https://www.repository.cam.ac.uk/handle/1810/250394 (accessed on 5 January 2024).
  29. BCDR-D Dataset. Available online: https://bcdr.eu/ (accessed on 5 January 2024).
  30. Zhang, F.; Luo, L.; Sun, X.; Zhou, Z.; Li, X.; Yu, Y.; Wang, Y. Cascaded generative and discriminative learning for microcalcification detection in breast mammograms. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 12578–12586. [Google Scholar]
  31. Huang, X.; Liu, Z.; Liu, S.-Y.; Cheng, K.-T. Efficient Quantization-aware Training with Adaptive Coreset Selection. arXiv 2023, arXiv:2306.07215. [Google Scholar]
  32. EfficientNet B7. Available online: https://github.com/qubvel/efficientnet (accessed on 3 March 2023).
  33. LightGBM. Available online: https://github.com/microsoft/LightGBM (accessed on 3 March 2023).
  34. Agarwal, R.; Diaz, O.; Lladó, X.; Yap, M.H.; Mart, R. Automatic mass detection in mammograms using deep convolutional neural networks. J. Med. Imaging 2019, 6, 031409. [Google Scholar] [CrossRef] [PubMed]
  35. Zahoor, S.; Shoaib, U.; Lali, I.U. Breast cancer mammograms classification using deep neural network and entropy-controlled whale optimization algorithm. Diagnostics 2022, 12, 557. [Google Scholar] [CrossRef]
  36. Boudouh, S.S.; Bouakkaz, M. Breast cancer: Toward an accurate breast tumor detection model in mammography using transfer learning techniques. Multimed. Tools Appl. 2023, 82, 34913–34936. [Google Scholar] [CrossRef]
  37. Bai, J.; Jin, A.; Wang, T.; Yang, C.; Nabavi, S. Feature fusion Siamese network for breast cancer detection comparing current and prior mammograms. Med. Phys. 2022, 49, 3654–3669. [Google Scholar] [CrossRef]
  38. Bobowicz, M.; Rygusik, M.; Buler, J.; Buler, R.; Ferlin, M.; Kwasigroch, A.; Szurowska, E.; Grochowski, M. Attention-Based Deep Learning System for Classification of Breast Lesions—Multimodal, Weakly Supervised Approach. Cancers 2023, 15, 2704. [Google Scholar] [CrossRef]
  39. Lévy, D.; Jain, A. Breast mass classification from mammograms using deep convolutional neural networks. arXiv 2016, arXiv:1612.00542. [Google Scholar]
  40. Ahmed, L.; Iqbal, M.M.; Aldabbas, H.; Khalid, S.; Saleem, Y.; Saeed, S. Images data practices for semantic segmentation of breast cancer using deep neural network. J. Ambient. Intell. Humaniz. Comput. 2020, 14, 15227–15243. [Google Scholar] [CrossRef]
  41. Hameed, Z.; Zahia, S.; Garcia-Zapirain, B.; Javier Aguirre, J.; Maria Vanegas, A. Breast cancer histopathology image classification using an ensemble of deep learning models. Sensors 2020, 20, 4373. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The proposed BC detection model.
Figure 1. The proposed BC detection model.
Diagnostics 14 00227 g001
Figure 2. (a) CBIS-DDSM dataset. (b) CMMD dataset.
Figure 2. (a) CBIS-DDSM dataset. (b) CMMD dataset.
Diagnostics 14 00227 g002
Figure 3. Feature engineering process.
Figure 3. Feature engineering process.
Diagnostics 14 00227 g003
Figure 4. Fine-tuned LightGBM-based multi-class classification.
Figure 4. Fine-tuned LightGBM-based multi-class classification.
Diagnostics 14 00227 g004
Figure 5. Findings of performance analysis.
Figure 5. Findings of performance analysis.
Diagnostics 14 00227 g005
Table 1. Dataset characteristics.
Table 1. Dataset characteristics.
DatasetsYearNumber of ImagesNumber of PatientsClassification
CBIS-DDSM201716442890Benign, malignant, and normal
CMMD202222141026Benign and malignant
Combined Dataset-4900-Benign, malignant, and normal
Table 2. Notations and definitions.
Table 2. Notations and definitions.
NotationsDefinition
B T P Correctly identified BC images
B T N Correctly identified normal or benign images
B F P Incorrectly identified BC images
B F N Incorrectly identified normal or benign images
Table 3. Computational settings.
Table 3. Computational settings.
ParametersValues
BC image size224 × 224
Strides3
Number of convolutional layers6
Learning rate1 × 10−4
Decay rate0.97 per 2 Epochs
Batch size128
Fully connected layer1
Activation functionSoftmax
Table 4. Performance analysis outcomes—CBIS-DDSM.
Table 4. Performance analysis outcomes—CBIS-DDSM.
ClassesAccyPrenReclF1KMCC
Benign99.299.199.399.296.896.7
Malignant99.599.399.399.397.196.5
Normal99.598.798.898.796.996.7
Average99.499.099.199.196.996.6
Table 5. Performance analysis outcomes—CMMD.
Table 5. Performance analysis outcomes—CMMD.
ClassesAccyPrenReclF1KMCC
Benign99.998.899.699.696.797.4
Malignant99.999.499.199.197.297.3
Average99.999.199.399.296.997.3
Table 6. Performance analysis outcomes—combined dataset.
Table 6. Performance analysis outcomes—combined dataset.
ClassesAccyPrenReclF1KMCC
Benign96.895.895.995.894.293.8
Malignant97.596.196.396.293.894.1
Normal96.997.497.197.294.393.4
Average97.096.496.496.494.193.7
Table 7. Batch-wise performance analysis—CBIS-DDSM.
Table 7. Batch-wise performance analysis—CBIS-DDSM.
BatchesAccyPrenReclF1KMCC
497.197.897.697.792.591.8
897.898.198.598.393.492.8
1298.198.698.398.494.693.7
1698.498.998.898.894.895.1
1899.499.099.199.196.996.6
Table 8. Batch-wise performance analysis—CMMD.
Table 8. Batch-wise performance analysis—CMMD.
BatchesAccyPrenReclF1KMCC
497.597.597.297.393.494.1
897.996.998.297.594.795.1
1298.898.698.698.695.396.4
1699.999.199.399.296.997.3
Table 9. Batch-wise performance analysis—combined dataset.
Table 9. Batch-wise performance analysis—combined dataset.
BatchesAccyPrenReclF1KMCC
495.294.294.594.393.193.4
895.894.994.794.892.892.1
1296.495.695.395.493.592.6
1697.096.496.496.494.193.7
Table 10. Findings of comparative analysis—CBIS-DDSM.
Table 10. Findings of comparative analysis—CBIS-DDSM.
Models/MetricsAccyPrenReclF1KMCC
Proposed BC detection model99.499.099.199.196.996.6
Falconi et al., model [22]95.895.595.795.689.188.5
Agarwal et al., model [34]96.796.996.996.991.289.4
Zahoor et al., model [35]96.795.794.995.390.790.4
EfficientNet B797.196.897.397.092.391.7
Table 11. Findings of comparative analysis—CMMD.
Table 11. Findings of comparative analysis—CMMD.
Models/MetricsAccyPrenReclF1KMCC
Proposed BC detection model99.999.199.399.296.997.3
Boudouh et al., model [36]99.899.499.699.590.589.4
Bai et al., model [37]98.498.298.298.290.888.7
Inception V396.797.596.897.191.190.3
Bobowicz et al., model [38]97.197.297.197.191.390.6
EfficientNet B797.597.497.397.392.190.8
Table 12. Findings of comparative analysis—combined dataset.
Table 12. Findings of comparative analysis—combined dataset.
Models/MetricsAccyPrenReclF1KMCC
Proposed BC detection model97.096.496.496.494.193.7
Yolo—V8 model92.490.590.790.689.290.4
El Houby and Yassin, model [9]96.295.295.295.293.791.3
Singh et al., [13]84.381.682.281.979.875.8
CatBoost model95.194.994.694.792.889.9
ShuffleNet V2 model91.390.891.291.090.288.5
Table 13. Computational strategies.
Table 13. Computational strategies.
ModelsCBIS-DDSMCMMDCombined Dataset
Learning RateParameters
(in Millions ( m ))
FLOPs
(in Giga ( G ))
Learning RateParameters
(in Millions ( m ))
FLOPs
(in Giga ( G ))
Learning RateParameters
(in Millions ( m ))
FLOPs
(in Giga ( G ))
Proposed BC detection model1 × 10−4211.81 × 10−4181.41 × 10−4272.3
Falconi et al., model [22]1 × 10−2484.1------
Agarwal et al., model [34]1 × 10−2343.4------
Zahoor et al., model [35]1 × 10−4713.2------
EfficientNet B71 × 10−3312.21 × 10−4271.9---
Boudouh et al., model [36]---1 × 10−3565.1---
Bai et al., model [37]---1 × 10−3653.8---
Inception V31 × 10−3753.71 × 10−3584.1---
Bobowicz et al., model [38]---1 × 10−3695.2---
Yolo—V8 model------1 × 10−4384.2
El Houby and Yassin, model [9]------1 × 10−3433.7
Singh et al., [13]------1 × 10−3484.1
CatBoost Model------1 × 10−4312.7
ShuffleNet V2 model------1 × 10−4523.9
Table 14. Findings of uncertainty analysis.
Table 14. Findings of uncertainty analysis.
ModelsCBIS-DDSMCMMDCombined Dataset
LossCISDLossCISDLossCISD
Proposed BC detection model0.4[98.1–98.5]0.00030.3[97.6–98.2]0.00020.3[98.1–98.4]0.0003
Falconi et al., model [22]0.8[96.1–96.7]0.0004------
Agarwal et al., model [34]1.2[95.8–96.2]0.0005------
Zahoor et al., model [35]0.9[95.9–96.4]0.0005------
EfficientNet B70.7[97.1–97.6]0.00030.8[97.5–98.3]0.0003---
Boudouh et al., model [36]---1.4[96.1–96.7]0.0005---
Bai et al., model [37]---1.1[96.8–97.7]0.0002---
Inception V30.7[97.3–98.2]0.00030.7[95.7–96.6]0.0002---
Bobowicz et al., model [38]---0.9[97.1–97.8]0.0004---
Yolo—V8 model------0.9[95.3–96.4]0.0003
El Houby and Yassin, model [9] ------1.2[96.1–96.5]0.0007
Singh et al., [13]------1.4[95.8–96.3]0.0005
CatBoost Model------0.5[95.3–96.1]0.0003
ShuffleNet V2 model------0.7[95.4–96.2]0.0004
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sait, A.R.W.; Nagaraj, R. An Enhanced LightGBM-Based Breast Cancer Detection Technique Using Mammography Images. Diagnostics 2024, 14, 227. https://doi.org/10.3390/diagnostics14020227

AMA Style

Sait ARW, Nagaraj R. An Enhanced LightGBM-Based Breast Cancer Detection Technique Using Mammography Images. Diagnostics. 2024; 14(2):227. https://doi.org/10.3390/diagnostics14020227

Chicago/Turabian Style

Sait, Abdul Rahaman Wahab, and Ramprasad Nagaraj. 2024. "An Enhanced LightGBM-Based Breast Cancer Detection Technique Using Mammography Images" Diagnostics 14, no. 2: 227. https://doi.org/10.3390/diagnostics14020227

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop