Next Article in Journal
Erdheim–Chester Disease of the Breast: First Review and First Case of Isolated Severe Gynecomastia
Previous Article in Journal
Quantitative MR Neurography in Multifocal Motor Neuropathy and Amyotrophic Lateral Sclerosis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

BC2NetRF: Breast Cancer Classification from Mammogram Images Using Enhanced Deep Learning Features and Equilibrium-Jaya Controlled Regula Falsi-Based Features Selection

1
Department of Computer Science, HITEC University, Taxila 47080, Pakistan
2
College of Computer Science, King Khalid University, Abha 61413, Saudi Arabia
3
Higher Institute of Applied Science and Technology of Sousse (ISSATS), Cité Taffala (Ibn Khaldoun) 4003 Sousse, University of Souse, Sousse 4000, Tunisia
4
College of Computer Science and Engineering, University of Ha’il, Ha’il 81451, Saudi Arabia
5
Department of Information Systems, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
6
Faculty of Information Technology and Computing, Arab Open University, Ardiya 92400, Kuwait
7
Department of Management, CoBA, Prince Sattam bin Abdulaziz University, Al-Kharj 11942, Saudi Arabia
8
Department of Computer Science, Hanyang University, Seoul 04763, Republic of Korea
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(7), 1238; https://doi.org/10.3390/diagnostics13071238
Submission received: 1 March 2023 / Revised: 13 March 2023 / Accepted: 23 March 2023 / Published: 25 March 2023
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
One of the most frequent cancers in women is breast cancer, and in the year 2022, approximately 287,850 new cases have been diagnosed. From them, 43,250 women died from this cancer. An early diagnosis of this cancer can help to overcome the mortality rate. However, the manual diagnosis of this cancer using mammogram images is not an easy process and always requires an expert person. Several AI-based techniques have been suggested in the literature. However, still, they are facing several challenges, such as similarities between cancer and non-cancer regions, irrelevant feature extraction, and weak training models. In this work, we proposed a new automated computerized framework for breast cancer classification. The proposed framework improves the contrast using a novel enhancement technique called haze-reduced local-global. The enhanced images are later employed for the dataset augmentation. This step aimed at increasing the diversity of the dataset and improving the training capability of the selected deep learning model. After that, a pre-trained model named EfficientNet-b0 was employed and fine-tuned to add a few new layers. The fine-tuned model was trained separately on original and enhanced images using deep transfer learning concepts with static hyperparameters’ initialization. Deep features were extracted from the average pooling layer in the next step and fused using a new serial-based approach. The fused features were later optimized using a feature selection algorithm known as Equilibrium-Jaya controlled Regula Falsi. The Regula Falsi was employed as a termination function in this algorithm. The selected features were finally classified using several machine learning classifiers. The experimental process was conducted on two publicly available datasets—CBIS-DDSM and INbreast. For these datasets, the achieved average accuracy is 95.4% and 99.7%. A comparison with state-of-the-art (SOTA) technology shows that the obtained proposed framework improved the accuracy. Moreover, the confidence interval-based analysis shows consistent results of the proposed framework.

1. Introduction

Cancer affects people all over the world. Breast cancer is the most prevalent type among women [1]. However, according to a Breast Cancer Care (BCC) survey, 42% of NHS trusts claim they do not have enough staff to assign people, citing minimal specialist nursing experience in breast cancer. It is the main factor that contributes to the low global survival rate of breast cancer [2]. A shortage of breast cancer specialists among medical professionals will delay disease detection, encourage noncompliance with the best screening and treatment methods, and lead to unequal access to the best care [3]. Breast cancer detection was established to effectively detect abnormalities and categorize breast cancer. This is done to help in breast cancer diagnosis [4]. Early detection is crucial to lowering the death rate; however, due to the modest sizes of possible nodules concerning the overall breast, early diagnosis of breast cancer by screening mammography is difficult [5]. Compared to other cancer types, breast cancer has the highest probability of being treated (around 90%). Cancer does not produce early pain, so it is not noticed until serious health issues arise [6]. Age is a factor in both breast cancer mortality and incidence rates. The average age of breast cancer diagnosis between 2010 and 2014 is 62.
Pakistan has Asia’s highest breast cancer incidence rate—annual reports of cases totaling about 90,000 results include a fatality rate of 40,000 [7]. The patient survival rate is the percentage of patients expected to live after a diagnosis for a pre-determined amount of time with the expectation of leading a normal life. The stage at which cancer is discovered affects the survival rate [8]. Due to its ability and low cost to satisfy medical requirements, mammography has appeared as the most reliable tool for detecting breast cancer.
Mammography analysis is the doctor’s primary method to determine a diagnosis, but this method is vulnerable to bias and doctor tiredness [1]. Unfortunately, mammography has a low detection rate. It can produce false-negative results in the range of 5% to 30%, based on the type of lesion, the density of the breasts and the patient’s age [9]. Therefore, low-dose radiography is used in mammography, which allows us to see the inside structure of the breast. Various signal processing methods are used to detect breast cancer, including ultrasound imaging [10], microwave imaging [11], and curvelet transform [8,12].
Computer-aided detection (CAD) technologies for breast cancer are crucial therapeutically in order to lessen radiologists’ labor and increase their detection precision [5]. Pattern recognition is the foundation of the conventional method for classifying medical infections, such as breast masses, skin lesions, and brain tumors. For breast cancer, the mammogram features are manually extracted, and the extracted features are then entered into a machine learning classifier for categorization [1]. However, gaining an accurate classification is still challenging due to several image issues and variations in the tumor regions. Therefore, AI has been significantly involved in detecting and classifying medical infections in the last decade, especially for breast cancer [13].
Thanks to the CAD system, this is based on several intermediate stages, such as the preprocessing of original images, feature learning and extraction, feature selection and reduction, and classification [14]. In the preprocessing step, the researcher attempts to produce good-quality images and overcome the noise if it exists. The objective of preprocessing is to make the tumor region more visible, which can later help in accurately detecting a region of interest (ROI) [15]. Several conventional approaches have been described in the literature for ROI detection, such as fuzzy, clustering, and saliency-based techniques, to name but a few. The next important step is the feature extraction, in which each image’s key properties are computed. Several conventional feature extraction techniques, such as the shape, texture, and point features, are introduced in the literature. Some researchers also focused on feature reduction and selection for improved accuracy and less computational time [16]. The last important step in AI is the machine learning techniques based on classifying cancer regions into relevant categories, such as cancerous or non-cancerous [17].
Recently, a convolutional neural network (CNN) has shown a remarkable performance in medical imaging for detecting and classifying cancers. The effectiveness of the deep learning models usually depends on the size of the training datasets. The conventional techniques did not perform well for the complex nature datasets; however, on the other side, the deep learning-based techniques showed impressive performance. Truly, deep learning uses the idea of CNN to classify breast cancer. A CNN model contains layers (hidden layers), such as convolutional, pooling, activation, and fully connected layers. The last layer of a CNN model is Softmax, which works as a classifier. Deep learning empowers automated AI techniques in medical imaging. Several deep learning-based architectures have been introduced in the literature for diagnosing and classifying medical infections [13]. The researchers introduced many deep learning approaches for breast cancer classification and diagnosis; however, they still face challenges, such as imbalanced datasets, noisy imaging data, and the downsampling of important features [18]. They focused on the learning task of deep models based on transfer learning. Transfer learning is a method of reusing a pre-trained model for another purpose or task [19]. Several hyperparameters have been employed for the training process, such as the learning rate and mini-batch size, etc.; however, for breast cancer, it is difficult to initialize the manual values of each parameter. Obayya et al. [20] presented an optimized hyperparameter-based deep-learning framework for classifying breast cancer. After the training, the researcher extracted the deep features from the fully connected layer; however, based on the analysis, it was noticed that several features were redundant and affected the classification accuracy of breast cancer [21]. Recently, Atban et al. [22] presented an optimized deep learning approach for improved breast cancer classification. Peirera et al. [23] presented a dialectical feature selection technique for enhanced breast cancer classification; however, these techniques face the problem of termination where the optimal values have been retrieved.

1.1. Major Challenges and Contributions

Several challenges exist in the medical image processing domain, especially for breast cancer classification when using deep learning. The first challenge is fewer amounts of available mammography image datasets because a deep learning model requires a larger dataset for the training and a better understanding of the classification task [24]. The second challenge is feature engineering. This step also extracts several redundant features that cause a false-negative rate and high computational time [25]. In this paper, we proposed a new framework based on original and enhanced mammogram images’ deep-learning optimal features aggregation. Our main contributions to this work are listed as follows:
  • We proposed a contrast enhancement approach based on the haze reduction concept called haze-reduced local-global (HRLG).
  • We performed data augmentation and trained the fine-tuned EfficientNet-b0 deep learning model. For the training, hit- and trial-based values were selected for the hyperparameters of the network. This model was trained using original and enhanced images and extracted deep features from the average pool layer instead of a fully connected layer. The extracted deep features were later fused using a serial nature approach.
  • A feature selection technique is proposed and called Equilibrium-Jaya controlled Regula Falsi. The best features are selected and fused using a new, short serial-based technique.

1.2. Justification of the Contributions

The major objective of the proposed contrast enhancement strategy is to increase the local information of the image, which is later helpful in the better learning of a deep model and improved classification (the results-based justification is provided under Section 4). The reason behind the selection of EfficientNet-b0 is based on the lesser number of parameters and a better top-five accuracy. In addition, the hit- and trial-based hyperparameters selection shows better model learning that increases the training accuracy. In addition, the proposed feature selection technique reduces the problem of overfitting and lessens the proposed framework’s computational cost (the results are given in Section 4).
The remainder of the article is arranged in the following order. Section 2 defines the related work of this manuscript. The proposed methodology, discussed in Section 3, includes the proposed contrast enhancement, augmentation, feature extraction, optimization, fusion, and classification. The outcomes, in the form of numerical values, are presented in Section 4. Finally, Section 5 concludes the manuscript.

2. Related Work

The most common type of cancer worldwide is breast cancer, with about 1.7 million women affected by cancer being reported in 2012. Age, family history, and medical history are just a few risk factors for breast cancer [4]. Most patients who die from cancer are women, and 2.1 million of them are identified with breast cancer each year. According to a recent survey, 627,000 women are estimated to have died from cancer in 2018, making up 15% of all cancer fatalities among women [5]. Breast cancer detection and its classification via computer visualization are commonly performed through a deep learning-based model. However, due to the complexity of early breast cancer and the dimming of mammography images, it is challenging for clinicians to diagnose cancer from these images. As a result, it is crucial to improve a doctor’s detection effectiveness using the CAD system of deep learning approaches [26].
Tan et al. [4] suggested a CNN-based framework for classifying mammography images into normal, benign, and malignant to classify breast cancer. First, to visualize the mammogram images, preprocessing was performed. Then, the preprocessed images were trained on the deep learning model that extracted the features. The extracted features of the last layer were finally classified using a CNN classifier called Softmax. The selected model increased the classification accuracy of mammography images using the introduced framework. The results clearly show that the suggested framework was more accurate than other existing methods, having an accuracy of 0.8585 and 0.8271, respectively. Falconi et al. [27] presented the preliminary findings for the classification of breast abnormalities as malignancies using transfer learning. They used several deep learning models and found two top models, ResNet50 and MobileNet. Both models produced the best outcomes, with 78.4% and 74.3% accuracy, respectively. Additionally, they also applied several preprocessing techniques to improve the classification’s precision. Finally, Samee et al. [28] presented a novel hybrid processing method, which is based on both principal component analysis (PCA) and logistic regression (LR).
The CAD system was evaluated using the INbreast and mini-MIAS benchmark datasets. The presented CAD system achieved the best accuracies of 98.60% using the INbreast dataset and 98.80% using the mini-MIAS dataset. Next, Hekal et al. [5] presented a new computer-aided detection technique for classifying two mammography cancers. The automated optimal Otsu thresholding approach was used in a CAD system to identify tumor-like regions. Then, deep CNNs were used to investigate the AlexNet and ResNet50 architectures, which processed the retrieved TLRs to extract the relevant mammography features. The experiment was acted upon on two datasets, yielding 91% and 84% accuracy rates, respectively. Finally, Siddeeq et al. [29] presented a framework with a ResNet-based customized neural network that was applied to an unbalanced dataset using the data augmentation and pyramid of scales approaches. The results obtained from the mammograms in the INbreast dataset show improved performance when the training dataset is increased.
Hikmah et al. [30] improved the diagnostic outcomes using an image-processing architecture for multi-view screening. The tumor regions were segmented using first-order local entropy, a texture-based method. The feature extraction results were used to calculate the radius and area of likely malignancy. According to the outcomes of this suggested method, the detection accuracy of the CC and MLO views for breast cancer was 88.0% and 80.5%, respectively. Alruwaili et al. [31] presented a framework that emphasized transferable learning. In order to avoid overfitting and achieve reliable results, a variety of augmentation techniques are used to enhance the number of mammograms. Almalki et al. [32] presented a technique on the large mammography images dataset. In the first stage, classification was performed, followed by extracting the pectoral muscle in the next step. In the third stage, abnormal spots in a well-enhanced image were detected using a new segmentation module to recognize breast cancer. The Breast Imaging and Reporting and Data System’s dataset had five categories and obtained an accuracy of 92%.
Moreover, on the Mammographic Image Analysis Society database, the suggested technique yields a result of about 97%. Karthiga et al. [33] presented a methodology combining two main components: transfer learning and CNNs. The hyperparameters of the CNN model were modified to enhance the classification performance. The outcome showed that the presented strategies significantly improved accuracy for the combined datasets (92.27%), for MIAS (95.95%), for DDSM (99.39%), and for INbreast (96.53%). Several other methods were also introduced to classify breast cancer [34]. Few other studies also existed in the literature that proposed deep learning-based frameworks for breast cancer classification, such as the optimized stacking learning approach [35], fuzzy c-mean and median support value-based CNN approach [36], and named a few more [37].
The above studies focused on the model’s information fusion, selection of manual hyperparameter values, data augmentation, and tumor identification using thresholding and CNN techniques. However, it was observed that they missed several important steps that could help improve accuracy. Those steps are contrast enhancement and the optimization of the extracted features. In the deep model, the SGD and ADAM optimizers are normally employed for the weight optimizations. However, we also included a feature optimization technique after the feature extraction process to overcome the computational time, problem of overfitting, and improvement in accuracy.

3. Proposed Novel Framework

This section presents the proposed framework for classifying breast cancer using mammogram images. Figure 1 highlights the architecture of this cancer classification. According to this figure, for the experimentation, datasets from INbreast and CBIS-DDSM are utilized. In the first phase, a contrast enhancement technique is employed on both datasets and then data augmentation is performed. A modified EfficientNet-b0 model is employed for training on both the original and improved datasets. For training purposes, a deep transfer learning technique is used. After that, feature extraction is performed from the average pool layer. Then, a hybrid optimization algorithm is employed to select the best features. By using a serial-based approach, the best-selected features are fused and classified using machine learning classifiers.

3.1. Datasets

In this section, the details of selected datasets have been presented. Two datasets have been employed for the experimentation of this work, such as the CBIS-DDSM (the Curated Breast Imaging Subset of DDSM), where the digital database is a revised and standardized version of scanning for breast cancer (DDSM) and INbreast. The DDSM dataset comprises two classes, benign and malignant, as shown in Figure 2.
The INbreast dataset comprises two classes that are benign and malignant. Figure 3 illustrates the sample images of this dataset, and it contains 410 images of 115 patients. The size of the images is 2560 × 3328 and 3328 × 4084 pixels. For the experiment, 108 mass mammogram images were employed.

3.2. Novelty 1: Contrast Enhancement and Augmentation

Traditional haze-removal techniques aim to produce a high-quality rebuilt image by adjusting the contrast and saturation. The visibility of the scene in the image can be made substantially better by using the haze reduction procedure. In this work, we proposed a new hybrid contrast enhancement technique based on haze removal and local-global transformation. Mathematically, this technique is described as follows:
Consider Δ as an entire image database having N number of images. Let I x , y be an original image of dimensional N × M × 3 and C F ˜ x , y is the final enhanced image. In the first step, we employed a haze reduction technique based on the dark channel and applied it to the original image. Mathematically, the haze reduction process is defined as follows:
H x = Y x t x + L 1 t x
where H denotes the observed intensity value, Y denotes the scene radiance, t x denotes the transmission map, and L denotes the atmospheric light. The employed dehazing algorithm recovers the scene radiance Y from the estimation of the transmission map and atmospheric light as follows:
Y x = H x α max ( t x ,   t 0 + α
The resultant Y x is later utilized to obtain the global contrast of an image based on the following formulation:
g 0 = 1 + C k × g i k m e a n + σ
where g 0 is a resultant global contrast image, C k denotes the global contrast gain factor, g i denotes the input pixel value of Y x , k m e a n is a global mean value of Y x , and σ denotes the standard deviation of Y x . In the latter step, we computed the local contrast of the haze reduction image by employing the following mathematical function:
L x , y = ϕ x , y + L C σ i , j + α × ϕ x , y μ x , y
where ϕ x , y denotes the grayscale pixel of the dehazed image Y x , L C denotes local contrast, α denotes a small parameter value, and μ x , y denotes the mean value of the dehazed image, respectively. Finally, we fused the local and global contrast resultant images into a single image by employing the following mathematical equation and obtaining the final enhanced image.
C F ˜ x , y = g x , y + L x , y I x , y
The final enhanced image is C F ˜ x , y , which is further utilized for the augmentation process. Visually, these enhanced images are shown in Figure 4.

3.3. Data Augmentation

The small number of image datasets is useful for training traditional machine learning techniques, such as shape features (HOG), point features, color features, and more. For deep learning models, it is always essential to generate or collect some larger datasets. However, the publicly available datasets for breast cancer are not large enough; therefore, we performed data augmentation in this work. In addition to growing the dataset, data augmentation reduces the overfitting issues and strengthens the deep learning model’s robustness. Eight more photos for each identified patch were created by rotating each image four times at the angles of 0°, 90°, 180°, and 270°, and then by flipping the resulting four images from left to right. After the process of augmentation, the summary of images is presented in Table 1. Moreover, Figure 5 displays the sample images.

3.4. EfficientNet-B0 Pre-Trained Model

EfficientNet uses a compound coefficient to scale both the design and sizing of its convolutional neural network. This compound coefficient uniformly scales the depth, width, and resolution parameters to maintain consistency throughout the network. Unlike the traditional approach of unconstrained scaling of these parameters, the EfficientNet scaling approach employs a fixed set of scaling coefficients to uniformly adjust the resolution, width, and depth of the network. The concept underlying the compound scaling method is that, when the input images become larger, the network requires additional layers to expand its receptive field and more channels to detect finer details within the larger image. The fundamental EfficientNet-b0 network is composed of squeeze-and-excitation blocks, as well as MobileNetV2 inverted bottleneck residual blocks. The architecture of the EfficientNet-b0 model is shown in Figure 6.
Fine-Tuned Model: Originally, the EfficientNet-b0 network was trained using more than a million images of the ImageNet dataset that contains 1000 different object categories. For fine-tuning, we initially excluded the last three layers and added three new layers, such as the fully connected, softmax, and output layers. After that, we initialized some hyperparameters, such as a learning rate of 0.005, momentum of 0.703, epochs of 100, and the stochastic gradient descent as an optimizer. Finally, we trained this fine-tuned model using deep transfer learning.
Knowledge transfer from one field or domain to another is called transfer learning. Deep learning involves a great deal of training data, making it difficult and time-consuming to learn specific patterns, especially in medical imaging. A labeled dataset serves as the “source,” which is indicated as:
F $ = G $ , H $
where H $ I r 1 is the label, and G $ I X r 1 is a d-dimensional feature space. The r 1 is the total number of source samples. The “target” dataset is identified as F τ = G τ , where G τ ϵ I X r 2 and r 2 are the total number of target samples. There are two important definitions as follows:
The two primary components of a “domain” F = G , ρ i are the feature data G and the associated marginal probability distribution ρ i , which has a range of [0, 1]. The (Task) can be defined as labels H and their matching function J i , which anticipates them, making up a “task” τ = H ,   J i . Overall, this process is shown in Figure 7.
After the training of the deep learning models using deep transfer learning, deep features were extracted. Deep features were extracted for both datasets, such as CBIS-DDSM and INbreast. As shown in Figure 1, the training of the fine-tuned model is conducted separately on original and enhanced images. The features were extracted from each trained model’s global average pool layers. After the feature extraction, we obtained feature vectors of dimensions N × 1280 for the CBIS-DDSM dataset (two vectors of the same dataset) and N × 1280 for the INbreast dataset (two vectors of the same dataset).

3.5. Novelty 2: Binomial Probability Serially Fusion

Consider that we have two feature vectors 1 and 2 of CBIS-DDSM, which have the dimensions N × 1280 and N × 1280 . Similarly, we have two feature vectors ˇ 1 and ˇ 2 of the INbreast dataset, with the dimensions N × 1280 and N × 1280 , respectively. Suppose that ˇ x and ˇ x 1 denote the fused feature vectors using a simple serial-based approach and are mathematically defined as follows:
ˇ x = 1 2 N × 1280 + N × 1280
ˇ x 1 = ˇ 1 ˇ 2 N × 1280 + N × 1280
Both fused vectors ˇ x and ˇ x 1 are analyzed and further refined using the proposed serially extended mean deviation approach. In this approach, the binomial probability distribution is initially computed for the entire feature vector and then defined as a final activation function based on the resultant value used for the fusion process. Mathematically, this process is defined as follows:
P ˇ x = n C ˇ x p ˇ x 1 p n ˇ x
P ˇ x 1 = n C ˇ x 1 p ˇ x 1 1 p n ˇ x 1
Using P ˇ x and P ˇ x 1 , the final activation functions have been defined for the final fusion.
A c t 1 = F U 1                 f o r ˇ x P ˇ x I g n o r e E l s e w h e r e
A c t 2 = F U 2                 f o r ˇ x 1 P ˇ x 1 I g n o r e E l s e w h e r e
where F U 1 and F U 2 denote the fused feature vectors of the CBIS-DDSM and INbreast datasets, respectively. In our work, the dimensions of these feature vectors were N × 1726 and N × 1702 , respectively. Furthermore, the resultant feature vectors were optimized further using the proposed Equilibrium-Jaya controlled Regula Falsi method.

3.6. Novelty 3: Proposed Features Selection Method

This work selects the best features by employing a hybrid optimization algorithm named Equilibrium-Jaya controlled Regula Falsi. The main purpose of the hybridization of both algorithms is to obtain the most optimal features that can help in improved accuracy and a reduction in computational time.

Equilibrium Optimizer Algorithm

Initialization: In this stage, the equilibrium optimizer employs a collection of particles, each representing a concentration vector that holds the answer to the optimization issue. By using the following formula, a random initial concentrations vector is created in the search space [41]:
x j = u m i n + u m a x u m i n v                             j = 1 , 2 , 3 , n
where x j represents the particle n ’s concentration vector and u m i n ,   u m a x set the upper and lower bounds, respectively. For each dimension in the challenge, v is an arbitrary number in the range of [0, 1], and n indicates the number of particles present.
Equilibrium Pool and Candidates: Each meta-heuristic algorithm attempts to attain a goal based on its nature. For instance, WOA looks for predators. An artificial bee colony (ABC) looks for a food supply, and in relation to EO, it looks for the system’s equilibrium state. The near-optimal solution to the optimization issue may be reached by EO when it reaches the equilibrium state. During optimization, EO is unaware of the degree of levels of concentration that reach equilibrium. The four best particles identified in the community at equilibrium have members assigned, along with another candidate with the average of the four best particles. These five equilibrium members assist EO in its role as an exploitation and exploration operator, with the first four members assisting EO in improving its capacity for diversification and exploitation on average. These five candidates are kept in an equilibrium pool vector as follows [41]:
ω E q , p o o l = ω E q 1 , ω E q 2 , ω E q 3 , ω E q 4 , ω E q A v g
Updating the Concentration: Using the following phrase aids EO in maintaining a tenable balance between intensification and diversity. Since the turnover rate in a real control volume can change over time, k is meant to be a random vector between 0 and 1.
H = e δ I T I T 0
I T = 1 I t e r I T m a x α 2 I t e r I T m a x
where the current and maximum iterations are denoted by I t e r and I T m a x , respectively. Additionally, the exploitation capability is managed by the constant variable α 2 . Another factor is α 1 , which is employed to enhance the diversification and intensity of EO:
I T 0 = 1 δ ln α 1   S i g n   v 0.5   [ 1 e δ I T + I T
where α 1 is a constant used to control the exploration capabilities; as α 1 increases, diversification improves while the intensity decreases. Unlike α 1 , α 2 is a fixed value utilized to regulate the exploitation capacity. The intensification capability is good, and the diversification capability is worse when α 2 is higher. Another term that is used to increase the intensification operator is the generation rate (GR), which is defined as follows [41]:
G R = G R 0 e δ I T I T 0
where, G R 0 is the beginning value, and δ is an arbitrary vector with a range from 0 to 1:
G R 0 = G R C u E q δ U
G R C = 0.65 v 1 v 2 > R C 0 O t h e r w i s e
where the random values v 1 and v 2 range from 0 to 1. The generation rate control parameter in this equation is denoted by G R C , that, based on a probability, R C determines whether the generation rate will be implemented in the updating process. Last but not least, the EO updating equation is as follows:
U = u E q + U u E q + H + G R δ x 1 H
The output of this algorithm U is again analyzed and modified with the Jaya algorithm. The working principle of the algorithm is that the seeking agents or particles continuously try to approach the goal by ignoring the worst solution in each iteration. By retaining the best position and completely discarding all other positions, the seeking agent updates its position based on this method. As a result, all solutions produced by iteration are better than the prior worst solution. Every suggested solution or searching agent in JA is known as a particle V σ , , ρ . Each particle searches for the best and optimal solution and avoids the worst solution of the objective or cost function in the search region. With a candidates i . e . ,   σ = 1 ,   2 , ,   a and b design variables i . e . ,   = 1 ,   2 , ,   b , this is accomplished in order to maximize the objective function (J). The particle locations are mathematically updated as follows:
V σ , , ρ = V σ , , ρ + z 1 , σ , ρ V σ , b e s t , ρ V σ , , ρ z 2 , σ , ρ V σ , w o r s t , ρ V σ , , ρ
where V q , r , s represents the t h candidate’s σ t h variable during the ρ t h iteration. Random numbers with values between [0, 1] are z 1 , σ , ρ and z 2 , σ , ρ . The V σ , b e s t , ρ and V σ , w o r s t , ρ , respectively, are the best and worst candidate values. The updated value of V σ , , ρ is V σ , , ρ . If the value returned by V σ , , ρ is better than V σ , , ρ then the V σ , , ρ value is retained. The cost function is employed in each iteration, and the resultant features are used to compute the best value using the Regula Falsi method. Mathematically, the Regula Falsi method is defined as follows:
R F = x 0 f x 0 × x 1 x 0 f x 1 f x 0
R F n = x n 2 f x n 2 × x n 1 x n 2 f x n 1 f x n 2
The output of this function is passed to the fitness function to check the accuracy. The fine KNN is employed in this work as a function. The selected features, after this process, are finally passed to the classifiers for classification. The dimensions of the proposed selected features are N × 826 and N × 812 , which was previously N × 1726 (the fusion of dataset 1, and 1702 for the fusion of dataset 2). This shows that the proposed technique significantly reduced the size of the feature vector.
The reasons for feature selection (FS): The aim of feature selection is to reduce the problem of overfitting and the overall computational cost. In this work, the purpose of the proposed FS technique is to choose the most optimal features that maintain the classification accuracy or improve the accuracy but significantly reduce the computational time. We compared the proposed FS technique with other techniques, such as the principal component analysis (PCA), genetic algorithm (GA), and PSO. In PCA, the features are reduced as per the experience, but in the GA and PSO, 20–30% of the features are reduced, and 70% (maximum) of the features are selected (in our experiments).

4. Results

The proposed architecture results for breast cancer classification have been presented in this section in terms of numeric values, confusion matrix, and plots. The augmented CBIS-DDSM and INbreast publicly available datasets have been employed for the experimental process (details of the datasets have been given under Section 3.1). The training and testing ratio is defined as 50: 50, with a cross-validation value of 10. Several classifiers have been utilized for classification, such as ensemble subspace KNN, fine KNN, cubic SVM, medium Gaussian SVM, medium neural networks, wide neural networks, and weighted KNN, to name but a few. The outcomes are calculated using several experiments: (i) a classification using deep features on the original dataset; (ii) a classification using deep features on an enhanced dataset; (iii) a fusion of the original and improved dataset deep features using the proposed fusion technique; and (iv) a feature selection using the proposed Equilibrium-Jaya controlled Regula Falsi method. The entire experimental process was conducted on MATLAB 2022a using a desktop computer with 16 GB of RAM and an 8 GB graphics card.

4.1. CBIS-DDSM Results

Table 2 describes the outcomes of the classification of the CBIS-DDSM dataset using the original dataset’s deep features. The ensemble subspace KNN classifier obtained the best-obtained accuracy of 92%. The sensitivity rate of this classifier is 92.05, the precision rate is 92.05, the F1 score is 92.05, and the FNR is 7.95%, respectively. The second best-obtained accuracy is 91.8%, achieved by fine KNN. The newly added classifiers, such as the narrow neural network (N3), medium NN, wide NN, and bi-layered NN, obtained 87.5, 88.7, 88.4, and 87.2% accuracy, respectively. The computational time of each classifier is listed in this table. Based on the mentioned time, the fine KNN classifier execution is less than the other classifiers, such as 50.021 (s); however, the maximum noted time for the ensemble subspace KNN classifier is 584.69 (s).
Table 3 describes the outcomes of the classification of the CBIS-DDSM dataset after the proposed contrast enhancement technique. The best-obtained accuracy of 95% was achieved by the ensemble subspace KNN classifier. The sensitivity rate of this classifier is 95.05, the precision rate is 95, the F1 score is 95.02, and the FNR is 4.95%. The second best-obtained accuracy is 94.7%, achieved by fine KNN. The noted accuracy for N3 is 90.1, the medium NN is 90.6, the wide NN is 90.5, and the bi-layered NN is 89.9%, respectively. Comparing these obtained accuracies after the proposed contrast enhancement step shows that the performance is more improved than the original dataset, as shown in Table 2. Computationally, the fine KNN classifier execution is less than the other classifiers, such as 54.658 (s); however, the maximum noted time is 633.3 (s) for the ensemble subspace KNN. Compared to Table 2, it is observed that the accuracy of this experiment is improved, which shows the strength of this classifier.
Table 4 describes the outcomes of the classification of the CBIS-DDSM dataset using the proposed feature fusion approach. The best-obtained accuracy after the feature fusion is 94.1% by ensemble subspace KNN. The sensitivity rate of this classifier is 94.15, the precision rate is 94.15, the F1 score is 94.15, and the FNR is 5.85%. The second best-obtained accuracy is 93.8%, achieved by fine KNN. The accuracy is also computed for several neural networks, such as N3, MNN, WNN, and Bi-NN, which are 90.3, 92.2, 92.2, and 90.5%, respectively. A comparison of these values with Table 2 shows that the accuracy is enhanced after the fusion process. Figure 8 illustrates the confusion matrix of the ESKNN classifier. This figure shows that the malignant class has a correct prediction rate of 93.4%. Moreover, Figure 9 displays that the computational time after the fusion process is increased, which is a drawback of this step; thus, this is why we proposed a feature selection technique.
Table 5 describes the outcomes of the classification of the CBIS-DDSM dataset using the proposed feature selection algorithm. The best-obtained accuracy of 95.4%, achieved by ensemble subspace KNN, is more improved than the originally extracted features, enhanced image features, and proposed fusion. The sensitivity rate of this classifier is 95.4, the precision rate is 95.35, the F1 score is 95.37, and the FNR is 4.6%. The classification accuracy for the neural networks is also improved after the proposed selection method, such as N3 is 94.1, MNN is 94.7, WNN is 90.7, and Bi-NN is 94.9%, respectively. Figure 10 shows the confusion matrix of the ESKNN classifier that shows the correct predicted values in a diagonal. The computational time is also reduced after the proposed selection method and is plotted in Figure 11. Overall, the accuracy is enhanced after the proposed feature selection method. Moreover, time is also reduced when compared to the previous experiments.

4.2. INbreast Dataset Results

The outcomes of the classification of the INbreast cancer dataset have been discussed in this subsection. Table 6 describes the outcomes of the classification of the INbreast dataset using the original images dataset. Deep features are extracted from the deep model and obtained the maximum accuracy of 98.3% on medium Gaussian SVM (MGSVM). The sensitivity rate of this classifier is 98.25, the precision rate is 98.25, the F1 score is 98.25, and the FNR is 1.75%. The computational time is noted for each classifier, listed in this table. Based on the mentioned time, the quadratic SVM (QSVM) classifier execution is less than the other classifiers.
Table 7 describes the outcomes of the classification of the INbreast dataset using the proposed enhanced images. After the training on enhanced images, we obtained a testing accuracy of 98.1% on MGSVM. The sensitivity rate of MGSVM for this experiment is 98.05, the precision rate is 98.15, the F1 score is 98.09, and the FNR is 1.95%, respectively. The computational time is noted for each classifier, listed in this table. However, it was discovered that the accuracy remains consistent after the enhancement, but the computational time is reduced.
Table 8 displays the outcomes of the proposed feature fusion for the INbreast dataset. After the proposed fusion, we obtained an accuracy of 99.6% on MGSVM. The sensitivity rate of this classifier is 99.55, the precision rate is 99.6, the F1 score is 99.57, and the FNR is 0.45%. These values can be further calculated through a confusion matrix, given in Figure 12. In this figure, the diagonal values display the correct or accurate predicted values. The computational time is noted for each classifier, and it is observed that the time is increased after the proposed fusion process. However, the accuracy is significantly increased, which is the strength of this step. We proposed a feature selection approach to resolve the high computational time problem.
The outcomes of the proposed features are presented in Table 9 and achieved the maximum accuracy of 99.4% on cubic SVM. The sensitivity rate of this classifier is 99.4, the precision rate is 99.4, the F1 score is 99.4, and the FNR is 0.6%. A confusion matrix is illustrated in Figure 13, in which the diagonal values show the correct prediction. The computational time is noted for each classifier, listed in this table, and the noted minimum time is 2.1975 for the cubic SVM classifier. Compared to the original image features, contrast-enhanced image features, and fused features, the proposed selection method shows improved performance for both accuracy and time. The visual comparison, in terms of the computational time, is shown in Figure 14, which shows the clear strength of the proposed feature selection.
Finally, we compare the proposed framework with some state-of-the-art techniques which used similar datasets. Table 10 compares the proposed framework with the recent techniques, and it is noted that Surendiren et al. [42] obtained an accuracy of 93.3% on the CBIS-DDSM dataset. Houbey et al. [43] obtained an improved accuracy of 96.52% on the INbreast dataset. In [44], the authors obtained 85.38% and 99% accuracy for the CBIS-DDSM and INbreast datasets. This work achieves an accuracy of 95.4% for CBIS-DDSM and 99.7% for the INbreast cancer dataset, which shows an improvement.

5. Conclusions

This work proposes a novel framework for breast cancer classification from mammography images. The proposed framework comprises important steps, starting with image acquisition and classification. In the first phase, a contrast enhancement technique is proposed. The resultant enhanced images have been used to train the deep learning model (EfficientNet-b0) and compare the results with the original image’s deep feature accuracy. The results show that the accuracy of the proposed enhancement technique is better, but the recently obtained accuracy was not met; therefore, a new fusion technique is proposed. The original image and enhanced image features have been fused using the proposed fusion technique and thus show a significant improvement in accuracy. The drawback of this step was the increase in computational time; therefore, a new feature selection technique is proposed, called Equilibrium-Jaya controlled Regula Falsi. After employing the proposed selection technique, time is significantly reduced for both datasets.

Limitations and Future Directions

The limitation of this work is the fusion process that consumes much time compared to the feature selection technique. In addition, the manual initialization of the hyperparameters is not an efficient method. These limitations will be resolved in the future by employing the following steps:
-
The tumor segmentation step will be added using image fusion techniques [49], and later-stage segmented regions will be considered for the feature extraction.
-
An optimized feature fusion technique will be considered for resolving the computational time problem.
-
The Bayesian optimization technique will have opted for the hyperparameters’ initialization.

Author Contributions

All authors contributed equally. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the “Human Resources Program in Energy Technology” of the Korea Institute of Energy Technology Evaluation and Planning (KETEP) and was granted financial resources from the Ministry of Trade, Industry & Energy, Republic of Korea. (No. 20204010600090). The funding for this work was provided by the Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R410), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used in this work are publicly available.

Acknowledgments

The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work through the Large Group Project under grant number (R.G.P.2/112/43). The authors are thankful to the Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R410), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, H.; Niu, J.; Li, D.; Zhang, C. Classification of breast mass in two-view mammograms via deep learning. IET Image Process. 2021, 15, 454–467. [Google Scholar] [CrossRef]
  2. Azamjah, N.; Soltan-Zadeh, Y.; Zayeri, F. Global trend of breast cancer mortality rate: A 25-year study. Asian Pac. J. Cancer Prev. 2019, 20, 2015–2020. [Google Scholar] [CrossRef]
  3. Medeiros, G.; Thuler, L.; Bergmann, A. Delay in breast cancer diagnosis: A Brazilian cohort study. Public Health 2019, 167, 88–95. [Google Scholar] [CrossRef]
  4. Tan, Y.J.; Sim, K.S.; Ting, F.F. Breast cancer detection using convolutional neural networks for mammogram imaging system. In Proceedings of the 2017 International Conference on Robotics, Automation and Sciences (ICORAS), Melaka, Malaysia, 27–29 November 2017; pp. 1–5. [Google Scholar]
  5. Hekal, A.A.; Elnakib, A.; Moustafa, H.E.-D. Automated early breast cancer detection and classification system. Signal Image Video Process. 2021, 15, 1497–1505. [Google Scholar] [CrossRef]
  6. Kurumety, S.K.; Howshar, J.T.; Loving, V.A. Breast Cancer Screening and Outcomes Disparities Persist for Native American Women. J. Breast Imaging 2022, 5, 3–10. [Google Scholar] [CrossRef]
  7. Menhas, R.; Umer, S. Breast cancer among Pakistani women. Iran. J. Public Health 2015, 44, 586–587. [Google Scholar]
  8. Charan, S.; Khan, M.J.; Khurshid, K. Breast cancer detection in mammograms using convolutional neural network. In Proceedings of the 2018 International Conference on Computing, Mathematics and Engineering Technologies (iCoMET), Sukkur, Pakistan, 3–4 March 2018; pp. 1–5. [Google Scholar]
  9. Ramadan, S.Z. Methods used in computer-aided diagnosis for breast cancer detection using mammograms: A review. J. Healthc. Eng. 2020, 2020, 9162464. [Google Scholar] [CrossRef] [Green Version]
  10. Guo, R.; Lu, G.; Qin, B.; Fei, B. Ultrasound imaging technologies for breast cancer detection and management: A review. Ultrasound Med. Biol. 2018, 44, 37–70. [Google Scholar] [CrossRef]
  11. AlSawaftah, N.; El-Abed, S.; Dhou, S.; Zakaria, A. Microwave imaging for early breast cancer detection: Current state, challenges, and future directions. J. Imaging 2022, 8, 123. [Google Scholar] [CrossRef]
  12. Ayana, G.; Dese, K.; Choe, S.-w. Transfer learning in breast cancer diagnoses via ultrasound imaging. Cancers 2021, 13, 738. [Google Scholar] [CrossRef]
  13. Aljuaid, H.; Alturki, N.; Alsubaie, N.; Cavallaro, L.; Liotta, A. Computer-aided diagnosis for breast cancer classification using deep neural networks and transfer learning. Comput. Methods Programs Biomed. 2022, 223, 106951. [Google Scholar] [CrossRef]
  14. Michael, E.; Ma, H.; Li, H.; Qi, S. An optimized framework for breast cancer classification using machine learning. BioMed Res. Int. 2022, 2022, 8482022. [Google Scholar] [CrossRef]
  15. Wang, Y.; Acs, B.; Robertson, S.; Liu, B.; Solorzano, L.; Wählby, C.; Hartman, J.; Rantalainen, M. Improved breast cancer histological grading using deep learning. Ann. Oncol. 2022, 33, 89–98. [Google Scholar] [CrossRef]
  16. Zahoor, S.; Shoaib, U.; Lali, I.U. Breast cancer mammograms classification using deep neural network and entropy-controlled whale optimization algorithm. Diagnostics 2022, 12, 557. [Google Scholar] [CrossRef]
  17. Jabeen, K.; Khan, M.A.; Alhaisoni, M.; Tariq, U.; Zhang, Y.-D.; Hamza, A.; Mickus, A.; Damaševičius, R. Breast cancer classification from ultrasound images using probability-based optimal deep learning feature fusion. Sensors 2022, 22, 807. [Google Scholar] [CrossRef]
  18. Dhar, T.; Dey, N.; Borra, S.; Sherratt, R.S. Challenges of Deep Learning in Medical Image Analysis-Improving Explainability and Trust. IEEE Trans. Technol. Soc. 2023, 1. [Google Scholar] [CrossRef]
  19. Ayana, G.; Dese, K.; Dereje, Y.; Kebede, Y.; Barki, H.; Amdissa, D.; Husen, N.; Mulugeta, F.; Habtamu, B.; Choe, S.-w. Vision-Transformer-Based Transfer Learning for Mammogram Classification. Diagnostics 2023, 13, 178. [Google Scholar] [CrossRef]
  20. Obayya, M.; Maashi, M.S.; Nemri, N.; Mohsen, H.; Motwakel, A.; Osman, A.E.; Alneil, A.A.; Alsaid, M.I. Hyperparameter Optimizer with Deep Learning-Based Decision-Support Systems for Histopathological Breast Cancer Diagnosis. Cancers 2023, 15, 885. [Google Scholar] [CrossRef]
  21. Awotunde, J.B.; Panigrahi, R.; Khandelwal, B.; Garg, A.; Bhoi, A.K. Breast cancer diagnosis based on hybrid rule-based feature selection with deep learning algorithm. Res. Biomed. Eng. 2023, 39, 115–127. [Google Scholar] [CrossRef]
  22. Atban, F.; Ekinci, E.; Garip, Z. Traditional machine learning algorithms for breast cancer image classification with optimized deep features. Biomed. Signal Process. Control. 2023, 81, 104534. [Google Scholar] [CrossRef]
  23. Pereira, J.M.S.; de Santana, M.A.; de Lima, C.L.; de Lima, R.d.C.F.; de Lima, S.M.L.; dos Santos, W.P. Feature Selection Based on Dialectical Optimization Algorithm for Breast Lesion Classification in Thermographic Images. In Research Anthology on Medical Informatics in Breast and Cervical Cancer; IGI Global: Hershey, PA, USA, 2023; pp. 96–119. [Google Scholar]
  24. Wetstein, S.C.; de Jong, V.M.T.; Stathonikos, N.; Opdam, M.; Dackus, G.M.H.E.; Pluim, J.P.W.; van Diest, P.J.; Veta, M. Deep learning-based breast cancer grading and survival analysis on whole-slide histopathology images. Sci. Rep. 2022, 12, 15102. [Google Scholar] [CrossRef]
  25. Din, N.M.U.; Dar, R.A.; Rasool, M.; Assad, A. Breast cancer detection using deep learning: Datasets, methods, and challenges ahead. Comput. Biol. Med. 2022, 149, 106073. [Google Scholar] [CrossRef]
  26. Umer, M.J.; Sharif, M.; Kadry, S.; Alharbi, A. Multi-Class Classification of Breast Cancer Using 6B-Net with Deep Feature Fusion and Selection Method. J. Pers. Med. 2022, 12, 683. [Google Scholar] [CrossRef]
  27. Falconí, L.G.; Pérez, M.; Aguilar, W.G. Transfer learning in breast mammogram abnormalities classification with mobilenet and nasnet. In Proceedings of the 2019 International Conference on Systems, Signals and Image Processing (IWSSIP), Osijek, Croatia, 5–7 June 2019; pp. 109–114. [Google Scholar]
  28. Samee, N.A.; Alhussan, A.A.; Ghoneim, V.F.; Atteia, G.; Alkanhel, R.; Al-Antari, M.A.; Kadah, Y.M. A Hybrid Deep Transfer Learning of CNN-Based LR-PCA for Breast Lesion Diagnosis via Medical Breast Mammograms. Sensors 2022, 22, 4938. [Google Scholar] [CrossRef]
  29. Siddeeq, S.; Li, J.; Bhatti, H.M.A.; Manzoor, A.; Subhan Malhi, U. Deep Learning RN-BCNN Model for Breast Cancer BI-RADS Classification. In Proceedings of the 2021 The 4th International Conference on Image and Graphics Processing, Sanya, China, 1–3 January 2021; pp. 219–225. [Google Scholar]
  30. Hikmah, N.F.; Sardjono, T.A.; Mertiana, W.D.; Firdi, N.P.; Purwitasari, D. An Image Processing Framework for Breast Cancer Detection Using Multi-View Mammographic Images. Emit. Int. J. Eng. Technol. 2022, 10, 136–152. [Google Scholar] [CrossRef]
  31. Alruwaili, M.; Gouda, W. Automated Breast Cancer Detection Models Based on Transfer Learning. Sensors 2022, 22, 876. [Google Scholar] [CrossRef]
  32. Almalki, Y.E.; Soomro, T.A.; Irfan, M.; Alduraibi, S.K.; Ali, A. Computerized Analysis of Mammogram Images for Early Detection of Breast Cancer. Healthcare 2022, 10, 801. [Google Scholar] [CrossRef]
  33. Karthiga, R.; Narasimhan, K.; Amirtharajan, R. Diagnosis of breast cancer for modern mammography using artificial intelligence. Math. Comput. Simul. 2022, 202, 316–330. [Google Scholar] [CrossRef]
  34. Fahrozi, F.; Hadiyoso, S.; Hariyani, Y.S. Breast Cancer Detection in Mammography Image using Convolutional Neural Network. Jurnal Rekayasa Elektrika 2022, 18, 35–42. [Google Scholar] [CrossRef]
  35. Kumar, M.; Singhal, S.; Shekhar, S.; Sharma, B.; Srivastava, G. Optimized Stacking Ensemble Learning Model for Breast Cancer Detection and Classification Using Machine Learning. Sustainability 2022, 14, 13998. [Google Scholar] [CrossRef]
  36. Girija, O.K.; Elayidom, S. Mammogram Pectoral Muscle Removal Using Fuzzy C-Means ROI Clustering and MS-CNN Based Multi Classification. Optik 2022, 170465, in press. [Google Scholar] [CrossRef]
  37. Mechria, H.; Hassine, K.; Gouider, M.S. Effect of Denoising on Performance of Deep Convolutional Neural Network For Mammogram Images Classification. Procedia Comput. Sci. 2022, 207, 2345–2352. [Google Scholar] [CrossRef]
  38. Falconi, L.G.; Pérez, M.; Aguilar, W.G.; Conci, A. Transfer learning and fine tuning in breast mammogram abnormalities classification on CBIS-DDSM database. Adv. Sci. Technol. Eng. Syst. 2020, 5, 154–165. [Google Scholar] [CrossRef] [Green Version]
  39. Huang, M.-L.; Lin, T.-Y. Dataset of breast mammography images with masses. Data Brief 2020, 31, 105928. [Google Scholar] [CrossRef]
  40. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; pp. 6105–6114. [Google Scholar]
  41. Faramarzi, A.; Heidarinejad, M.; Stephens, B.; Mirjalili, S. Equilibrium optimizer: A novel optimization algorithm. Knowl.-Based Syst. 2020, 191, 105190. [Google Scholar] [CrossRef]
  42. Surendiran, B.; Ramanathan, P.; Vadivel, A. Effect of BIRADS shape descriptors on breast cancer analysis. Int. J. Med. Eng. Inform. 2015, 7, 65–79. [Google Scholar] [CrossRef]
  43. El Houby, E.M.; Yassin, N.I. Malignant and nonmalignant classification of breast lesions in mammograms using convolutional neural networks. Biomed. Signal Process. Control. 2021, 70, 102954. [Google Scholar] [CrossRef]
  44. Baccouche, A.; Garcia-Zapirain, B.; Elmaghraby, A.S. An integrated framework for breast mass classification and diagnosis using stacked ensemble of residual neural networks. Sci. Rep. 2022, 12, 12259. [Google Scholar] [CrossRef]
  45. Ragab, D.A.; Sharkas, M.; Marshall, S.; Ren, J. Breast cancer detection using deep convolutional neural networks and support vector machines. PeerJ 2019, 7, e6201. [Google Scholar] [CrossRef]
  46. Falconí, L.; Pérez, M.; Aguilar, W.; Conci, A. Transfer learning and fine tuning in mammogram bi-rads classification. In Proceedings of the 2020 IEEE 33rd International Symposium on Computer-Based Medical Systems (CBMS), Rochester, MN, USA, 28–30 July 2020; pp. 475–480. [Google Scholar]
  47. Mohiyuddin, A.; Basharat, A.; Ghani, U.; Peter, V.; Abbas, S.; Naeem, O.B.; Rizwan, M. Breast tumor detection and classification in mammogram images using modified YOLOv5 network. Comput. Math. Methods Med. 2022, 2022, 1359019. [Google Scholar] [CrossRef]
  48. Muduli, D.; Dash, R.; Majhi, B. Automated diagnosis of breast cancer using multi-modal datasets: A deep convolution neural network based approach. Biomed. Signal Process. Control. 2022, 71, 102825. [Google Scholar] [CrossRef]
  49. Bhalla, K.; Koundal, D.; Sharma, B.; Hu, Y.-C.; Zaguia, A. A fuzzy convolutional neural network for enhancing multi-focus image fusion. J. Vis. Commun. Image Represent. 2022, 84, 103485. [Google Scholar] [CrossRef]
Figure 1. The proposed framework of an optimal deep learning feature fusion of the classification of breast cancer.
Figure 1. The proposed framework of an optimal deep learning feature fusion of the classification of breast cancer.
Diagnostics 13 01238 g001
Figure 2. Sample images of the CBIS-DDSM dataset [38].
Figure 2. Sample images of the CBIS-DDSM dataset [38].
Diagnostics 13 01238 g002
Figure 3. Sample images of the INbreast dataset [39].
Figure 3. Sample images of the INbreast dataset [39].
Diagnostics 13 01238 g003
Figure 4. Visual illustration of proposed enhancement technique.
Figure 4. Visual illustration of proposed enhancement technique.
Diagnostics 13 01238 g004
Figure 5. Visual illustration of data augmentation step.
Figure 5. Visual illustration of data augmentation step.
Diagnostics 13 01238 g005
Figure 6. The architecture of EfficientNet-b0 [40].
Figure 6. The architecture of EfficientNet-b0 [40].
Diagnostics 13 01238 g006
Figure 7. Visual process of deep transfer learning for breast cancer classification.
Figure 7. Visual process of deep transfer learning for breast cancer classification.
Diagnostics 13 01238 g007
Figure 8. Confusion matrix of ESKNN after the proposed feature fusion.
Figure 8. Confusion matrix of ESKNN after the proposed feature fusion.
Diagnostics 13 01238 g008
Figure 9. Comparison of proposed fusion step in terms of computational time with Table 2 and Table 3.
Figure 9. Comparison of proposed fusion step in terms of computational time with Table 2 and Table 3.
Diagnostics 13 01238 g009
Figure 10. Confusion matrix of ESKNN classifier using proposed feature selection method.
Figure 10. Confusion matrix of ESKNN classifier using proposed feature selection method.
Diagnostics 13 01238 g010
Figure 11. Comparison of proposed feature selection technique in terms of computational time with Table 2, Table 3 and Table 4.
Figure 11. Comparison of proposed feature selection technique in terms of computational time with Table 2, Table 3 and Table 4.
Diagnostics 13 01238 g011
Figure 12. Confusion matrix MGSVM after employing proposed feature fusion.
Figure 12. Confusion matrix MGSVM after employing proposed feature fusion.
Diagnostics 13 01238 g012
Figure 13. Confusion matrix of CSVM after employing the proposed feature selection.
Figure 13. Confusion matrix of CSVM after employing the proposed feature selection.
Diagnostics 13 01238 g013
Figure 14. Computational time-based visual comparison among all four experiments.
Figure 14. Computational time-based visual comparison among all four experiments.
Diagnostics 13 01238 g014
Table 1. Summary of augmented and original datasets.
Table 1. Summary of augmented and original datasets.
CBIS-DDSM
Original DatasetAugmented Dataset
Class# ImagesClass# Images
Benign557Benign4939
Malignant637Malignant5096
INbreast Dataset
Original DatasetAugmented Dataset
Class# ImagesClass# Images
Benign76Benign1216
Malignant70Malignant1120
Table 2. Classification results of original dataset after deep feature extraction using fine-tuned EfficientNet-b0. * ESKNN denotes the ensemble subspace KNN, F-KNN denotes fine KNN, C-SVM denotes cubic SVM, MGSVM denotes medium Gaussian SVM, MNN denotes the medium neural network, WNN denotes wide neural network, N3 denotes the narrow neural network, and BN2 denotes the bi-layered neural network.
Table 2. Classification results of original dataset after deep feature extraction using fine-tuned EfficientNet-b0. * ESKNN denotes the ensemble subspace KNN, F-KNN denotes fine KNN, C-SVM denotes cubic SVM, MGSVM denotes medium Gaussian SVM, MNN denotes the medium neural network, WNN denotes wide neural network, N3 denotes the narrow neural network, and BN2 denotes the bi-layered neural network.
* ClassifierSensitivity Rate %Precision Rate %F1 ScoreAUCAccuracy %FNRTime (s)
ESKNN92.0592.0592.050.96927.95584.69
F-KNN91.991.8591.870.9291.88.150.021
C-SVM90.3590.3590.350.9690.49.6595.967
MGSVM88.788.788.70.9588.711.3120.35
MNN88.788.6588.670.9388.711.355.385
Wide NN88.488.4588.420.9488.411.6103.12
WKNN88.188.188.10.9588.111.951.004
QSVM87.687.687.60.9587.612.497.295
N387.5587.5587.550.9287.512.45100.66
BN287.2587.2587.250.9287.212.75247.92
Table 3. Classification results after employing the proposed contrast enhancement step for deep features extraction.
Table 3. Classification results after employing the proposed contrast enhancement step for deep features extraction.
ClassifierSensitivity Rate (%)Precision Rate (%)F1 Score (%)AUCAccuracy (%)FNR (%)Time Complexity (s)
ESKNN95.059595.020.9795.04.95633.3
F-KNN94.794.794.70.9594.75.354.658
C-SVM92.1592.1592.150.9792.17.8590.261
MGSVM91.591.591.50.9791.58.5114.64
MNN90.5590.5590.550.9490.69.4586.577
Wide NN90.590.590.50.9690.59.599.719
WKNN9090900.9689.91060.308
QSVM90.490.4590.420.9690.49.692.179
N390.190.1590.120.9490.19.9105.76
BN289.8589.8589.850.9489.910.15308.88
Table 4. Classification results of CBIS-DDSM using proposed feature fusion approach. The bold represent the significant value.
Table 4. Classification results of CBIS-DDSM using proposed feature fusion approach. The bold represent the significant value.
ClassifierSensitivity Rate %Precision Rate %F1 ScoreAUCAccuracy %FNRTime Complexity (sec)
ESKNN94.1594.1594.150.9794.15.85961.1
F-KNN93.7593.7593.750.9493.86.25127.76
C-SVM93.693.693.60.9993.66.4167.34
MGSVM93.193.193.10.9893.16.9170.19
MNN92.292.2592.220.9892.27.894.742
Wide NN92.192.192.10.9892.17.9145.807
WKNN91.7591.7591.750.9791.78.25140.53
QSVM93.1593.293.170.9893.26.85175.34
N390.3590.390.320.9490.39.65149.832
BN290.5590.590.520.9490.59.45405.4
Table 5. Classification of CBIS-DDSM using the proposed feature selection algorithm.
Table 5. Classification of CBIS-DDSM using the proposed feature selection algorithm.
ClassifierSensitivity Rate (%)Precision Rate (%)F1 Score (%)AUCAccuracy (%)FNR (%)Time Complexity (sec)
Classifier95.495.3595.370.9895.44.6164.06
ESKNN95.2595.2595.250.9595.34.7525.859
F-KNN91.7591.7591.750.9791.88.2537.899
C-SVM90.7590.7590.750.9690.79.2552.97
MGSVM92.0592.0592.050.97927.9536.853
MNN94.794.794.70.9594.75.347.295
Wide NN90.790.6590.670.9690.79.325.815
WKNN89.7589.7589.750.9589.110.2539.806
QSVM91.591.591.50.9791.58.550.785
N394.994.8594.870.9894.95.1117.13
Table 6. Classification results of INbreast dataset using original images based on fine-tuned deep model features.
Table 6. Classification results of INbreast dataset using original images based on fine-tuned deep model features.
ClassifierSensitivity Rate %Precision Rate %F1 ScoreAUCAccuracy %FNRTime Complexity (sec)
MGSVM98.2598.2598.250.9998.31.757.1611
ESKNN98.1598.1598.150.9898.21.8538.082
QSVM98.0898.0598.06198.11.956.6397
F-KNN98.0598.0598.050.9898.11.959.1516
LSVM97.9597.9597.951982.057.5417
Bi-NN989898198220.558
EBT97.8597.8597.850.9997.92.1511.611
Tri-NN97.997.997.9197.92.130.736
C-SVM97.797.797.70.9997.72.36.837
N397.797.797.70.9997.72.317.319
Table 7. Classification results of INbreast dataset using proposed enhanced images based on fine-tuned deep model features.
Table 7. Classification results of INbreast dataset using proposed enhanced images based on fine-tuned deep model features.
ClassifierSensitivity Rate %Precision Rate %F1 ScoreAUCAccuracy %FNRTime Complexity (sec)
MGSVM98.0598.1598.090.9998.11.956.8656
ESKNN9898.198.040.9898.0238.111
QSVM97.99897.940.9997.92.16.3209
F-KNN97.99897.940.9897.92.19.3578
LSVM97.2595.596.36197.32.756.844
Bi-NN97.3597.397.32197.32.6518.814
EBT96.4596.596.470.9996.53.5511.588
Tri-NN97.3597.3597.350.9997.32.6527.701
C-SVM97.8597.8597.850.9997.72.156.3782
N397.797.797.70.9997.72.314.795
Table 8. Classification results of proposed feature fusion on INbreast dataset.
Table 8. Classification results of proposed feature fusion on INbreast dataset.
ClassifierSensitivity Rate (%)Precision Rate (%)F1 ScoreAUCAccuracy (%)FNRTime Complexity (sec)
MGSVM99.5599.699.57199.60.4520.066
ESKNN97.99897.940.9997.92.154.093
QSVM99.4599.599.47199.50.5515.09
F-KNN97.897.997.840.9897.92.220.399
LSVM99.4599.599.47199.50.5515.352
Bi-NN99.1599.299.17199.10.8513.076
EBT98.8598.998.87198.91.1518.83
Tri-NN99.1599.299.17199.10.8513.374
CSVM99.4599.599.47199.50.5516.174
N399.399.3599.32199.30.713.36
Table 9. Classification results of proposed feature selection technique on INbreast dataset.
Table 9. Classification results of proposed feature selection technique on INbreast dataset.
ClassifierSensitivity Rate %Precision Rate %F1 ScoreAUCAccuracy %FNRTime Complexity (sec)
MGSVM99.0599.199.07199.10.952.2698
ESKNN9898.198.04198.0215.915
QSVM99.299.2599.22199.20.82.2299
F-KNN97.5597.5597.550.9897.62.453.5232
LSVM99.299.299.2199.20.82.51
Bi-NN98.298.298.2198.21.82.4121
EBT98.298.298.2198.31.87.1051
Tri-NN98.698.698.60.9998.51.42.896
CSVM99.499.499.4199.40.62.1975
N399.0599.0599.05199.10.952.6468
Table 10. Comparison of the proposed framework with recent state-of-the-art techniques.
Table 10. Comparison of the proposed framework with recent state-of-the-art techniques.
ReferenceYearDatasetsAccuracy
[42]2015CBIS-DDSM93.3%
[45]2019CBIS-DDSM87.2%
[46]2020INbreast90.9
[47]2020CBIS-DDSM93.47
[1]2021CBIS-DDSM94.7
[43]2021INbreast96.52%
[48]2022CBIS-DDSM, INbreast90.68%, 91.28%
[44]2022CBIS-DDSM, INbreast85.38%, 99%
Proposed CBIS-DDSM, INbreast95.4%, 99.7%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jabeen, K.; Khan, M.A.; Balili, J.; Alhaisoni, M.; Almujally, N.A.; Alrashidi, H.; Tariq, U.; Cha, J.-H. BC2NetRF: Breast Cancer Classification from Mammogram Images Using Enhanced Deep Learning Features and Equilibrium-Jaya Controlled Regula Falsi-Based Features Selection. Diagnostics 2023, 13, 1238. https://doi.org/10.3390/diagnostics13071238

AMA Style

Jabeen K, Khan MA, Balili J, Alhaisoni M, Almujally NA, Alrashidi H, Tariq U, Cha J-H. BC2NetRF: Breast Cancer Classification from Mammogram Images Using Enhanced Deep Learning Features and Equilibrium-Jaya Controlled Regula Falsi-Based Features Selection. Diagnostics. 2023; 13(7):1238. https://doi.org/10.3390/diagnostics13071238

Chicago/Turabian Style

Jabeen, Kiran, Muhammad Attique Khan, Jamel Balili, Majed Alhaisoni, Nouf Abdullah Almujally, Huda Alrashidi, Usman Tariq, and Jae-Hyuk Cha. 2023. "BC2NetRF: Breast Cancer Classification from Mammogram Images Using Enhanced Deep Learning Features and Equilibrium-Jaya Controlled Regula Falsi-Based Features Selection" Diagnostics 13, no. 7: 1238. https://doi.org/10.3390/diagnostics13071238

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop