Next Article in Journal
Comprehensive Genomic Profiling and Therapeutic Implications for Patients with Advanced Cancers: The Experience of an Academic Hospital
Previous Article in Journal
Impact of [18F]FDG PET/CT in the Assessment of Immunotherapy-Induced Arterial Wall Inflammation in Melanoma Patients Receiving Immune Checkpoint Inhibitors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

BRMI-Net: Deep Learning Features and Flower Pollination-Controlled Regula Falsi-Based Feature Selection Framework for Breast Cancer Recognition in Mammography Images

1
Department of Computer Science, HITEC University, Taxila 47080, Pakistan
2
Department of Circulation and Medical Imaging, Faculty of Medicine and Health Sciences, Norwegian University of Science and Technology (NTNU), 7491 Trondheim, Norway
3
Department of Information Systems, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
4
College of Computer Science, King Khalid University, Abha 61413, Saudi Arabia
5
College of Computer Science and Engineering, University of Ha’il, Ha’il 81451, Saudi Arabia
6
Management Information System Department, College of Business Administration, Prince Sattam Bin Abdulaziz University, Al-Kharj 16278, Saudi Arabia
7
Department of Informatics, University of Leicester, Leicester LE1 7RH, UK
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(9), 1618; https://doi.org/10.3390/diagnostics13091618
Submission received: 27 February 2023 / Revised: 16 April 2023 / Accepted: 26 April 2023 / Published: 3 May 2023
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
The early detection of breast cancer using mammogram images is critical for lowering women’s mortality rates and allowing for proper treatment. Deep learning techniques are commonly used for feature extraction and have demonstrated significant performance in the literature. However, these features do not perform well in several cases due to redundant and irrelevant information. We created a new framework for diagnosing breast cancer using entropy-controlled deep learning and flower pollination optimization from the mammogram images. In the proposed framework, a filter fusion-based method for contrast enhancement is developed. The pre-trained ResNet-50 model is then improved and trained using transfer learning on both the original and enhanced datasets. Deep features are extracted and combined into a single vector in the following phase using a serial technique known as serial mid-value features. The top features are then classified using neural networks and machine learning classifiers in the following stage. To accomplish this, a technique for flower pollination optimization with entropy control has been developed. The exercise used three publicly available datasets: CBIS-DDSM, INbreast, and MIAS. On these selected datasets, the proposed framework achieved 93.8, 99.5, and 99.8% accuracy, respectively. Compared to the current methods, the increase in accuracy and decrease in computational time are explained.

1. Introduction

Breast cancer is the most common disease among women, and it affects the breast region and spreads to other body parts. Breast cancer affects the breast tissue and is known to be the second most widely spread disease in the world [1]. In an investigation by the World Health Organization (WHO), 8.4% of breast cancer patients received a diagnosis, and 6.6% passed away [2]. Breast cancer affects more than 8% of women at some point in time [3]. Breast cancer claimed the lives of 43,250 women in 2022. Breast cancer-related tumors are classified as benign or malignant [4]. A malignant tumor spreads to other organs, whereas a benign tumor does not spread to the rest of the body [5]. There are numerous imaging tools available for early breast cancer treatment. A pathological biopsy is one of the most accurate methods; according to evidence-based medicine, patients with cancerous tumors should avoid having too many biopsies performed in order to stop cancer cell growth and metastasis. As a result, pathological examinations are frequently performed prior to chemotherapy [6]. Furthermore, computer-aided radiologists (CAD) discovered that diagnosis aids in the early detection of breast cancer in ultrasound images (US) while decreasing operator-dependent US imaging behavior [7].
Mammography monitoring is one of the most important methods for preventing breast cancer. An X-ray is used to examine the breast in this technique [8]. Asymptomatic women are subjected to mammography screenings and procedures in order to detect early, clinically undiagnosed breast cancer [9]. The best imaging method for detecting microcalcifications and ductal carcinoma in situ is mammography (DCIS) [10]. The widespread adoption of screening mammography is likely to blame for the rise in the overall incidence of DCIS during the past 20 years [11]. Prior to classification, previous research used a variety of techniques to develop computer-aided diagnosis (CAD) systems, including images. To improve the images’ ability to identify benign or malignant tissue, edge detection, wavelet transform, microcalcification detection, and region of interest (ROI) segmentation were used [12].
Causes such as noise, poor image quality, and unrelated parts will have an impact on the most common characteristics in the field of medical images [13]. Preprocessing techniques are used to address this issue [14]. The preprocessing technique is the technique that improves the image quality [15]. The next process is to extract features from the targeted region [16]. After passing these features to a classifier, it determines whether these mammogram images are normal or abnormal [17]. Feature selection is a procedure that follows feature extraction and is frequently used in machine learning [18]. The primary goal of feature selection is to remove redundant features from the original data [19]. The many characteristics of machine learning algorithms include effective performance on healthcare datasets involving images such as X-rays, and various blood samples [20].
Deep learning has recently significantly improved in several areas, including cell segmentation, skin melanoma, and hemorrhage detection [21]. Deep learning has been shown to be effective in medical imaging, particularly in COVID-19, breast cancer detection, the diagnosis of brain tumors, Alzheimer’s disease, and a variety of other conditions. A convolutional neural network (CNN) is a deep learning-based architecture with many layers. CNN converts pixels in an image into features [22]. Later, the characteristics are used to identify and classify infections. The features were extracted by CNN from the original images. The collection of irrelevant features extracted from the raw images also influences how well the classification performs. Choosing only the most relevant features is critical for improving classification precision. An active research area is the selection of the best features from the extracted initial features. Particle swarm optimization (PSO), genetic algorithm (GA) [23], and a few other selection algorithms are described and used in the literature for medical imaging [24]. These methods focus on the most beneficial subset of features rather than the entire feature field. The ability of feature selection methods to improve system accuracy while reducing computation time is their primary advantage. However, when selecting the best features, some critical features are sometimes overlooked, reducing the system’s precision. As a result, experts in computer vision developed feature fusion techniques. The system is more precise and has more predictors as a result of the fusion procedure. Two popular fusion techniques are serial-based fusion and parallel fusion.
Several computer vision-based methods for detecting and classifying breast cancer using mammography images have been presented in recent years [25]. Some worked on segmentation, while others focused on feature extraction and feature extraction from raw images [26]. In addition, researchers used and worked on the preprocessing step to enhance the affected area for better feature extraction and to improve the contrast and noisy images [27]. For example, Dilovan et al. [28] presented a method for deep learning in breast cancer. In this suggested approach, the region is selected using machine learning techniques of interest. The region of interest is separated and divided into different blocks. The researchers extracted deep features and optimized them by utilizing a genetic algorithm. As a result, they achieved 98.7% accuracy. The limitation of this proposed method was that there were multiple features from each diagnosed block. Shahan et al. [29] presented a method that utilizes deep learning for breast cancer. In the proposed methodology, they collected information from various sources, fused the datasets, and simulated them using breast image data. They obtained 99% accuracy as a result. The proposed method’s drawback was that computational intelligence techniques could be used to increase the system’s accuracy.
Huynh et al. [30] described a procedure for treating breast cancer using transfer learning from a deep convolution network. They used and extracted features from CNN to train the support vector machine according to this proposed approach. A 90% accuracy rate was achieved. The deep learning strategy used in the suggested method, which uses large datasets and potent computational resources, has a drawback. Ghada et al. [31] reported a deep learning-based method for classifying breast cancer cases. Retina net, a deep learning-based model, was used in this suggested classification technique. The results were 97% accurate as a result. The shortcomings of the proposed technique were data duplication, imbalance, and inconsistent object detectors. Jing et al. [32] reported a technique for identifying breast cancer using the deep learning Ada Boost algorithm. They used CNN LSTM and advanced computational methods in the suggested methodology. They had a 98.3% success rate. The suggested approach’s disadvantage was the required sizeable quantity of training data. Neslihan et al. [33] presented a deep-learning image classification method for treating breast cancer. In the suggested approach, they used a CNN-based model to forecast both the likelihood of malignant tumors and the degree of image malignancy. They were accurate to 82.13%. The task-wise early stopping in the multitasking design was the method’s flaw. Naresh et al. [34] presented the use of a deep learning system for detecting breast cancer. In the proposed approach, they compared deep learning algorithms to machine learning algorithms using a deep learning neural network model. Their success percentage was 99.67%. The lack of data for expansion was a problem with the suggested approach. A breast cancer treatment method reported by Yeman et al. [35] used deep learning. This technique used a patch-based multi-input CNN to detect breast tumors. As a result, they had a 92% accuracy rate. The proposed method’s disadvantage was that reliability and detection accuracy could be improved by training with a larger data set. Luqman et al. [36] presented a method for the semantic segmentation of image data for breast cancer using deep learning. They used RCNN and a transfer learning preprocessing algorithm in this approach. They were, therefore, 98% accurate. The suggested methodology’s limitations were fewer training instances, sampling disease, and class imbalance. Aruna et al. [37] described a technique for the early detection of breast cancer using statistical tools. Support vector machines and Naive Bayes were contrasted with other data mining methods. They discovered the best predictor using WEKA. They were entirely accurate. The SVMs, both linear and non-linear, could not be analyzed using the suggested approach. Kundan et al. [38] presented a convolutional neural network-based treatment for breast cancer. They used a CNN-based model and classified the features in the suggested methodology. An accuracy of 90% was attained. The proposed method’s limitation was the small size of the data collection. Bindu et al. [39] described a method for categorizing breast cancer using an artificial neural network with the ideal amount of inputs. They classified the features using the feed-forward algorithm in the suggested approach. They obtained 90% accuracy as a result. The method’s flaw was that determining the tumor variety took longer and cost more money. Umar et al. [40] presented a technique for detecting breast cancer that classified mammography images using a convolution neural network. They used the CNN-based model in this methodology for training, and applied the winner filter’s preprocessing procedure. They then divided the crucial data into segments (masses). They were, therefore, 98% accurate.

1.1. Major Challenges

The approaches discussed above emphasized preprocessing and deep learning-based techniques to increase classification accuracy. However, they had to deal with limited data, parallels between benign and malignant masses, and redundant feature data. The most crucial stage in any computer-aided diagnostic system is feature engineering, according to literature reviews. Additionally, it is noted that the experts did not consider the optimization process, which can improve the accuracy and shorten the computation time. The key challenges which we consider in this work are as follows: (i) It is difficult to manually separate malignant and benign extracted slices from original compact images. (ii) Because there are fewer malignant images than benign lesions, this affects the selected deep learning model during the training process. (iii) Regarding the small growth of lesions, the visual has a high degree of similarity between benign and malignant lesions, which may distort the correct classification. (iv) To improve accuracy, selecting important features ignores some essential features found by several optimization algorithms [41], due to their stopping condition and search criteria [42]. We can address these issues using a mammography image classification for breast cancer using a deep-learning approach.

1.2. Major Contributions

Here is a list of this study paper’s main contributions:
  • A fusion-based contrast enhancement technique was proposed for lesion contrast enhancement of the original images.
  • We used a deep learning model called ResNet50 that had already been trained to perform the fine-tuning. After that, deep transfer learning with set hyper-parameters was used to train both the original and enhanced images.
  • After that, a fusing method known as serial mid-value feature fusion was proposed.
  • Flower pollination was proposed, using a controlled regula falsi-based best features selection method before using machine learning classifiers to classify the results.

2. Materials and Methods

This section presents a proposed breast cancer classification method based on mammogram images. Figure 1 depicts the proposed framework. The initial mammography images of the datasets are first augmented, and the augmented datasets are then contrasted. The refined Resnet50 deep network was then fed the original and improved images, which were then taught using deep transfer learning. The trained models are then used to determine the characteristics of the global average pool layer. Following feature extraction, the proposed technique is used for fusion. In the following step, we created a feature selection method to select the best features from the fused vector. Finally, machine learning classifiers are used to categorize the best features. Each stage is explained in detail below.

2.1. Dataset Collection

In this work, we consider three mammography image datasets: CBIS-DDSM [43], INbreast [44], and MIAS [45]. There are two classes in the CBIS-DDSM dataset, malignant and benign. The second INbreast dataset uses the classifications benign and malignant. The MIAS collection uses the classifications benign, malignant, and normal. Table 1 provides a short description of the images in these datasets. We used data augmentation to carry out all three procedures shown in Figure 2 because these datasets were unbalanced. First, we turned them 90 degrees, and then flipped them to the left and right.

2.2. Dataset Augmentation

Recently, several researchers performed data augmentation to improve the learning capability of deep learning methods [46]. For the deep learning model, the data sets currently available in the medical industry are from low-resource areas, but a significant amount of training data are needed [47]. Therefore, a data augmentation process must be used to increase the diversity of the initial datasets. Three datasets are used for validation in this study, as mentioned above. There are 1134 average-sized images in the collection of 500 × 500 pixels in the first CBIS-DDSM dataset. There are two categories in this dataset: Figure 3 shows the number of benign (557) and malignant (637) lesions. The second dataset of breast images contains 146 images with an average resolution of 500 × 500 pixels. This dataset consists of two categories: benign (76) and malignant (70), as shown in Figure 3.
In the third dataset of MIAS, there are 300 images with average sizes from the collection of 500 × 500 Pixels. There are three categories collectively in this dataset: benign, malignant, and normal (209), as seen in Figure 3. We split the entire dataset into training and testing groups (50:50). After training the images for each class, the dataset is inadequate to train the deep learning model. Therefore, three operations—horizontal flip, vertical flip, and rotation 90°—are performed on the original mammography images to increase the dataset diversity. Multiple iterations of these procedures are performed until each class of images contains at least 6000 for CBIS-DDSM and 4000 for INbreast and MIAS. After the augmentation method, the number of images in CBIS-DDSM is 12,000, 8000 in the INbreast dataset, and 12,000 in the MIAS dataset.
Figure 3. Illustration of mammography images datasets: (a) benign class data sample; (b) malignant class data sample; (c) normal class data sample.
Figure 3. Illustration of mammography images datasets: (a) benign class data sample; (b) malignant class data sample; (c) normal class data sample.
Diagnostics 13 01618 g003

2.3. Contrast Enhancement

Enhancing an input image is crucial for improving the original image’s quality. Recently, a lot of technology devices have been introduced for improving image acquisition, but they still they pose several challenges. The challenges include those specific to medical imaging, especially mammography, and are contrast imperfections, chromatic anomalies, and noise. In this proposed technique, our main goal is to improve the infected region and enhance the overall pixels of the image that are more pleasant to the human eye [48]. The objective of this study is to visualize datasets with breast cancer images rather than healthy ones. We created a hybrid approach that relies on combining various filtering results because the breast cancer mammography images’ poor quality and low contrast required us to do so. Considering we have an input image denoted by ϕ x , y of size M × N × 3 , let us have a grayscale image of size M × N denoted by ϕ ˜ x , y . In the first step, the probability density function is applied to the grayscale image as follows:
p ( ϕ ˜ k ) = h k h , k = 0 , 1 , 2 , 3 , ,   L 1
where h k denotes the number of pixels having an intensity ϕ ˜ k and h represents the total number of images Δ . In the next step, CDF is computed as follows:
c ( ϕ ˜ k ) i = 0 k p ( ϕ ˜ k )
Using the CDF, the transformation function is defined as follows:
F ( ϕ ˜ k ) = ϕ ˜ 0 + ( ϕ ˜ L 1 ϕ ˜ 0 ) c ϕ ˜ k
ψ = F ( ϕ ˜ x , y /   ϕ ˜ x , y ) ϕ ˜
After that, the spatial domain transformation is applied on F ( ϕ ˜ k ) .
G ( x , y ) = τ F ( x , y )
G ( x , y ) = Q ( x , y ) F ( x , y ) c × μ ( x , y ) + μ ( x , y ) a
Here, G ( x , y ) is a transformed image and Q ( x , y ) is a contrast stretching function that is mathematically defined as follows:
Q ( x , y ) = K × g m σ ( x , y ) + β
g m = 1 M × N x = 0 M 1 y = 0 N 1 F x , y
where g m denotes the global mean value, σ denotes the standard deviation, k and β denote the constant parameters and values manually assigned. Hence, the final transformation function is defined as follows:
G x , y = K × g m σ x , y + β F x , y c × μ x , y + μ x , y a
The visual output of the resultant enhanced image G x , y is illustrated in Figure 4.

2.4. ResNet50 Deep Learning Features

Deep learning networks acknowledge images as inputs and extract features from them that are later fed to the classifier for classification. This process of feature extraction and classification is conducted automatically [49]. Deep learning methods lower classification error rates to 5%, comparable to human mistakes [50].
ResNet50 [51] is the most frequently utilized subclass of convolution neural networks for picture classification. Figure 5 depicts the ResNet50 design. In this applicable ResNet50 deep learning model, the training was carried out with various weights. A broader network can benefit from extensive training, improving accuracy and outcomes. The ResNet50 network consists of several deep layers with a max-pool layer kernel size of 3 ×   3 and a small, related field of 7 × 7 in the input layer.
In this research, we improved this deep algorithm for breast cancer classification. We eliminated the first three layers and introduced three new ones to accomplish this. Later, we clarified the chosen dataset for breast cancer. The training process was conducted using deep transfer learning. Additionally, we did not freeze any single layer during the fine-tuning process.
Deep Transfer Learning: A pre-trained model is reused using the machine learning method known as transfer learning for another task [47]. The sampling efficiency could be greatly increased by transferring or reusing data from previously taught tasks for the newly learned activities [52]. TL is used in this case to extract deep features. To perform this, a pre-trained model is initially tweaked before being trained via TL [53]. The process of TL is defined as follows:
The two parameters that comprise the domain D = 2 ,   p z are the feature space z and the distributions of the marginal probabilities f z , where z = z 1 , z 2 , . z n z . If two domains are distinct, they either have different marginal probabilities p z p p z q or feature space z p z q . Given a specific domain, there are two parts to this task t   s , g . : the label space s , and a prediction function g (.); this is not immediately apparent but can be obtained using training data   ( k j , u j j 1 , 2 , 3 , 4 . N , w h e r e   k j   a n d u j   s ; thus, a probabilistic approach c k j can be written as p u j | k j , and so we rewrite this function v as V = s , p s | z . When two functions are different from each other, their label space is also different s p s q or produces different distributions with conditional probability p s p | z p p s q | z q .
Figure 6 gives a detailed overview of the visual transfer learning method. The modified deep model receives the original model’s (source domain) information (target domain) to train this updated model. The following hyperparameters are used in the model: the learning rate is 0.001, the mini-batch size is 16, and there are 200 epochs when using the stochastic gradient descent learning method. The features are extracted from the updated deep model’s Global Average Pooling (GAP) layer. Two feature vectors have been returned in an output that is finally fused using a serial-based mid-value approach.

2.5. Novelty 2: Serial Based Mid Value Fusion

Improving an object’s information through the fusion of data from two or more sources is known as information fusion. A serial-based feature is used when putting information from different sources together into a single matrix; the feature fusion method is simple to use while maintaining all features. However, this process added all relevant and irrelevant features. Adding all features increases the computational time and results in maximum chances of a high error rate for the classification purpose. Therefore, in this work, a new mid-value-based function is developed and finds a middle value used for the fusion of final features.
Consider that we have two feature vectors of original and enhanced datasets denoted by f 1 and f 2 . The dimensions of the extracted feature vectors are N × 2048 and N × 2048 , respectively. The serially fused vector dimension will be N × 4098 based on the following equation:
k f q v = f 1 f 2 N × j 1 + w × j 2
This process combined all of the extracted features. Using this fused vector, we computed the mid-value using the following formula:
F n c = M i d V k f q v
M i d V = l w + h w 2
The returned middle value is passed to the threshold function for finding the final fused vector.
T h e r s = F u s f   f o r   k f q v j F n c I g n o r e   E l s e w h e r e
where F u s f denotes the fused feature vector of dimension N × 2772 . Further enhancement of this outcome vector was carried out with the flower pollination-controlled regula falsi method.

2.6. Feature Optimization

Feature selection involves selecting the best features to improve accuracy and reduce computation time in computer vision. The size of the solution space grows exponentially as the number of features in the dataset increases. According to the theory, using fewer features improves classifier performance. It speeds up classification, resulting in accuracy rates equal to or higher than those obtained when all features are used. Based on the flower pollination-controlled regula falsi, this study developed a feature selection technique (FPcRF). Laws and floral reproduction inspire the flower optimization method. The FPA [54] is used for optimization, feature selection, and other optimization techniques in order to minimize its search space. The final selection requires a fitness function, and in this work, we used Fine-KNN. The initial representation of the flower consistency is
  y i j + 1 = y i j + ϑ L λ y i j B  
where y i j is a pollen i or solution vector y i and iteration j and B is the best solution. The symbol   ϑ is a given factor of scaling. The Levy Flight Distribution is defined as follows:
L ~ λ Γ λ sin Π 2   / Π 1 s 1 + λ ,   s 0 > > s 0 > 0
where Γ λ is also due to the accuracy of this distribution and the gamma function for significant steps s > 0 ; the local fertilization is shown as follows:
y i j + 1 = y i j + µ y z j y k j  
where y i j and y k j are pollens of different flowers of the same plant.
False Regulation Formula: assume that the number of y i j + 1 is denoted by r 1 , the greatest number is represented by r 2 , and then the root value’s approximation is calculated as follows:
V 1 = r 1 f r 2 r 2 f r 1 f r 2 f r 1
V n = r n 1 f r n r n f r n 1 f r n f r n 1
where V n is the final approximate root value of selected vector y i j + 1 . Based on V n , the final selection is performed and passed to the fitness function again to check the performance. The selected features are finally classified using machine learning classifiers for the final classification accuracy. Moreover, a detailed algorithm is presented in Algorithm 1.
Algorithm 1 Proposed Breast Cancer Classification Algorithm
Input: Original Image ϕ x , y
Output :   Labeled   Image   ϕ ^ x , y
Step 1: Dataset Augmentation
Step 2: Contrast Enhanced using Equations (1)–(9)
    -  The resultant Image is denoted by G x , y
Step 3: Trained Deep Learning Model
Step 4: Deep Features Extraction from Original and Enhanced Datasets
Step 5: Features Fusion using Equations (10)–(13)
    -   T h e r s = F u s f         f o r       k f q v j F n c I g n o r e                                             E l s e w h e r e
Step 6: Best Feature Selection
    -  Objective function: minimum and maximum f x , x = x 1 , x 2 ̤ x d
    -  Initialization: A population of n flower with a random solution
    -  Find the best solution: g   initial population
    -  Probability: p 0 , 1
While  j < max _ iteration
for i = 1: n
if rand < p ,
Draw step levy distribution is given as
Global pollination via y i j + 1 = y i j + ϑ L λ y i j B
Else
Draw Q from a uniform distribution 0 , 1
Local pollination via L ~   λ   Γ λ sin Π 2   / Π   1 s 1 + λ , s 0 > > s 0 > 0
end if
Evaluate new solution
end
for best solution
Find best root using Equations (17)–(18)
Final Best Solution
end while

3. Experimental Results and Discussion

3.1. Datasets and Experiments

Extensive experiments have been performed for the analysis of the proposed framework. Graphical representations and tabular data are used to show the findings. For the experimental procedure, three datasets have been used, and details are provided in Section 3.1. Furthermore, several experiments have been performed, such as proposed fusion and feature selection. In addition, the intermediate results are also provided to show the efficiency of the proposed framework.

3.2. Experimental Setup

The suggested framework’s training-to-testing ratio is set at 50/50. Several hyperparameters have been used during the design and experimental process, including a learning rate of 0.0002, epochs of 100, mini-batch size of 32, momentum value of 0.7223, and stochastic gradient descent (SGD) as an optimizer. During the testing procedure, a 10-fold cross-validation was used for each experiment. Various metrics were used to evaluate the success of the various classifiers, including F1-Score, FPR, Kappa, MCC, Accuracy, and Time as examples of statistical measures. A desktop computer was used, by one individual, with 16 GB of RAM and an 8 GB graphics card running MATLAB2022a, which simulated the full experimental procedure.

3.3. CBIS-DDSM Breast Cancer Dataset Results

The breast cancer information from the CBIS-DDSM dataset that was recommended for merging is displayed in Table 2. The cubic SVM (CSVM) had a maximum accuracy of 93.8%. The other calculated CSVM performance metrics had a precision rate of 93.09%, F1-score of 93.89%, FPR of 0.07, Kappa and MCC of 86.87%, and Sensitivity rates of 94.70% and 87.68%, respectively. These results for the remaining classifications demonstrate that CSVM surpasses them all. Figure 7 also displays the CSVM confusion matrix, which can be used to confirm the claimed performance metrics. The computation times for each classifier are also given; the MNN classifier recorded the lowest computation time of 249.39 s (seconds). Figure 8 shows the computational time mentioned previously.
Based on the suggested FPcRF, Table 3 presents the optimal feature selection results from the CBIS-DDSM. The CSVM classifier has a maximum accuracy of 93.3%. Additional metrics include a precision rate of 92.26%, sensitivity rate of 94.57%, F1-score of 93.40%, FPR of 0.07, Kappa of 86.63%, and MCC of 86.66%, supporting the proposed findings. Figure 9 depicts the CSVM’s confusion matrix, which can be used to verify computed performance measures. The assessment procedure also tracks how long each classification’s computations take. From lowest to highest, the spread is 82.532 (s) to 503.23 (s). Figure 8 visually compares the stated computational times for each classifier for both trials. The computational time is drastically decreased when the suggested FPcRF selection method is used, as is abundantly clear from this figure. Additionally, overall accuracy is improved after employing the recommended feature selection approach.

3.4. INbreast Breast Cancer Dataset Results

Table 4 displays the outcomes of the INbreast breast cancer dataset’s suggested fusion. The narrow neural network (NNN) classification achieved the highest accuracy of 99.5%, with values of 99.35%, 99.65%, 99.50%, and 99.00% for the precision rate, sensitivity, F1-score, Kappa, and MCC, respectively. These numbers are also calculated for the remaining classifiers, and the NNN’s best performance is discovered. Figure 9 shows a confusion matrix for NNN. The malignant class accurate prediction rate is shown in this figure to be above 99%. Each classifier’s computational time is mentioned, and the min-max range is 100.94–289.57 (s). Overall, the NNN produced better outcomes but took longer, which is a disadvantage.
We suggested a feature selection technique to address this flaw, and the results are shown in Table 5. This table shows that the best accuracy, 99.6%, which was attained using a bi-layered NN. A 99.45% accuracy rate, a 99.65% sensitivity rate, an F1-score of 99.55%, an FPR of 0.05, a Kappa of 99.10%, and an MCC of 99.10% are all the values reported. These metrics are calculated using various algorithms, including SVM, neural networks, and fine-KNN. Figure 10 and Figure 11 depicts the confusion matrices that displays the correct rate for each class, such as 99.45% for the malignant class.
It should be observed that the accuracy for the maximum classifier is consistent with Table 4, and for some classifiers it is improved. Each classifier’s computational duration, which can range from 127 s to 39.88 s, is also noted. Figure 12 compares proposed fusion and selection techniques and demonstrates how much less time is needed when using FPcRF selection.
Table 4. Classification results of proposed feature fusion for INbreast dataset.
Table 4. Classification results of proposed feature fusion for INbreast dataset.
ClassifierPrecisionSensitivityF1-ScoreFPRKappaMCCAccuracyTime (s)
CSVM99.1599.7099.430.0898.8598.8599.4123.89
LSVM98.3799.6098.980.0197.9597.9699.0121.42
QSVM99.2099.5599.380.0898.7598.7599.4116.32
MGSVM98.7199.4599.080.0198.1598.1599.1179.52
MNN99.2599.7099.480.0798.9598.9599.5108.77
WNN99.2599.6099.430.0798.8598.8599.4150.39
FKNN98.7099.0098.850.0197.7097.7098.9289.57
NNN99.3599.6599.500.0699.0099.0099.5101.52
BNN98.4599.4099.420.0598.8598.8599.4103.49
TNN99.3599.4099.380.0698.7598.7599.4100.94
Table 5. Classification results for INbreast dataset using proposed feature selection technique.
Table 5. Classification results for INbreast dataset using proposed feature selection technique.
ClassifierPrecisionSensitivityF1-ScoreFPRKappaMCCAccuracyTime (s)
CSVM99.3599.6599.500.0699.099.099.549.65
LSVM98.3299.6098.960.0197.9097.9199.054.65
QSVM99.3099.6599.480.0798.9598.9599.546.45
MGSVM98.6699.5599.100.0198.2098.2099.178.90
MNN99.2599.5099.380.0798.7598.7599.442.14
WNN99.3599.4599.400.0698.8098.8099.463.26
FKNN98.9099.2099.050.0198.1098.1099.1127.00
NNN99.5599.5099.520.0499.0599.0599.540.01
BNN99.4599.6599.550.0599.1099.1099.639.88
TNN99.3599.4599.400.0698.8098.8099.440.75
Figure 10. Confusion matrix of NNN classifier for INbreast dataset using proposed fusion technique.
Figure 10. Confusion matrix of NNN classifier for INbreast dataset using proposed fusion technique.
Diagnostics 13 01618 g010
Figure 11. Confusion matrix of BNN classifier using proposed feature selection technique on CBIS-DDSM dataset.
Figure 11. Confusion matrix of BNN classifier using proposed feature selection technique on CBIS-DDSM dataset.
Diagnostics 13 01618 g011
Figure 12. Comparison of proposed fusion and selection steps in terms of computational time for INbreast dataset.
Figure 12. Comparison of proposed fusion and selection steps in terms of computational time for INbreast dataset.
Diagnostics 13 01618 g012

3.5. Mias Dataset Results

Table 6 shows the outcomes of the suggested fusion technique on the MIAS dataset. The NNN classifier achieved maximum accuracy at 99.8%. Several additional performance metrics have also been calculated, including the precision rate, sensitivity rate, F1-score, Kappa value, and MCC value, which are all 99.82, 99.60, 99.82, and 99.73%, respectively. Figure 13, which displays the computed values as a confusion matrix, can be used to validate them. The benign and malignant classes in this figure had 100% and 99.7% accurate prediction rates. Each classifier’s computational duration is also listed, taking a maximum of 556.93 s and a minimum of 106.68 s.
Overall, the CSVM produced better outcomes, but this experiment’s disadvantage was that it took longer. To address this issue, a feature selection method is suggested. The outcomes are shown in Table 7. The CSVM yields the highest accuracy of 99.8% for this experiment. In particular, the Kappa is 99.50%, the MCC is 99.67%, the sensitivity is 99.78%, the F1-score is 99.78%, and the precision rate is 99.78%. SVM, neural networks, and fine KNN are just a few algorithms used to calculate all these metrics. Figure 14 shows a confusion matrix with the accurate prediction rates for each class, respectively, 99.7% and 100% for benign and malignant classes. It should be noted that the accuracy is consistent with Table 6, while the computational time is considerably shorter. Finally, the proposed fusion and selection techniques are compared in Figure 15, which illustrates a substantial reduction in computational time.

3.6. Comparison with Other State-of-the-Art Techniques

The proposed strategy is contrasted with a few contemporary techniques in Table 8. In [23], the authors proposed a multi-fractal and fusion approach and obtained an accuracy of 98% using three mammography breast cancer datasets, including CBIS-DDSM, INbreast, and MIAS. The precision for CBIS-DDSM, INbreast, and MIAS was 98.4% using the CNN-based model [46]. In this work, the authors used capsule network based architecture. The CNN-based architecture and fusion method were used in [47]. They obtained 96% accuracy, later improved by [48] using modified deep learning architecture, and achieved an average of 99.0% accuracy. The proposed framework obtained an accuracy of 93% on the CBIS-DDSM dataset, 99.5% on the INbreast dataset, and 99.8% on the MIAS dataset. These values show the improvement in accuracy compared to the existing techniques.

3.7. Confidence Interval-Based Analysis

A detailed confidence interval-based analysis is conducted in this section for all three datasets. The confidence interval (CI) is computed based on the Kappa measure. For the CBIS-DDSM dataset, the minimum and maximum noted values are 77.13% and 86.63%. Using these values, the obtained margin of error (MoE) is illustrated in Figure 16. In this figure, it is noted that at a confidence level of 68.3%, σx̄, the margin of error is 82.88 ± 4.066 (±4.91%). For a confidence level of 95%, 1.960σx̄, the margin of error is 82.88 ± 7.969 (±9.62%). This shows that the error dropped to 5% when a confidence level changed. Similarly, the MoE is computed for the rest of both datasets (Inbreast and MIAS), and the values are illustrated in Figure 17. For the Inbreast dataset, the MoE is 98.5 ± 0.832 (±0.84%) at a confidence level of 95%, 1.960σx̄. For the MIAS dataset, the MoE is 97.975 ± 2.114 (±2.16%) at a confidence level of 95%, 1.960σx̄. This shows that the performance of the proposed method is sufficient.

3.8. Visual Facts

In the end, the visualization is performed for the proposed deep learning framework. The Grad-CAM visualization is employed for the gradient visualization of the proposed deep framework, as shown in Figure 18. In this figure, a heatmap is applied to the important region based on the grad-CAM. The brown color shows the most important region that reveals this framework’s performance. In addition, the labeled prediction results are illustrated in Figure 19. Based on this figure, we can see that the proposed framework correctly predicts the benign and malignant classes. Finally, we also added a brief analysis for the selection of training and testing ratios. As shown in Figure 20, accuracy is consistent and only a minor change occurred for our selected ratio of 50:50. However, it is noted that the accuracy is going to decrease when the ratio is 40:60 or 30:70.

4. Conclusions

Breast cancer is the leading cause of cancer death in women worldwide. The early detection of malignant and benign cancers can help patients receive timely treatment. This study proposed a new framework based on fusion-based contrast enhancement using deep learning features from the average pool layer. It proposed feature selection for breast cancer classification based on mammography images. While some of the original points were lost during the fusion process as a result of the contrast enhancement phase, the performance of the fine-tuned deep learning model outperformed that of the original images. The fusion of deep features reduced the proposed framework’s effectiveness while improving accuracy and overall computational time. As a result, we proposed a feature selection method that improved precision while reducing computational time. Furthermore, the proposed framework improves accuracy for all three selected datasets by 93, 99.5%, and 99.8%, respectively. This work’s main limitations are as follows: i) the addition of a contrast enhancement technique improves image quality but increases computational time; ii) two consecutive constant values are not always obtained in the selection step, and the entire algorithm is executed for all initialized iterations. This takes longer and can result in the loss of important features.

Future Scope

  • Propose a new method for improving the contrast of the infected region pixels, which will improve the visibility of the lesion region and help with the correct segmentation.
  • Propose a segmentation technique using deep learning and a saliency map for tumor detection. The residual block will be added in the deep learning model, which can aid in better learning for the detection process.
  • Develop a fusion-based deep learning architecture with Bayesian optimization-based hyperparameters initialization.
  • Propose a new feature selection technique that will stop the iteration in a maximum of two constant cost values.

Author Contributions

Each contributor made an equal contribution. All authors have read and agreed to the published version of the manuscript.

Funding

The funding for this work was provided by the Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R410), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia. The King Khalid University Deanship of Scientific Research also provided funding for this research through a large group project with grant number (R.G.P.2/112/43), which is acknowledged by the authors.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used in this work are publicly available.

Acknowledgments

The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work through large group Research Project under grant number RGP.2/139/44. The authors are thankful to the Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R410), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jabeen, K.; Khan, M.A.; Alhaisoni, M.; Tariq, U.; Zhang, Y.-D.; Hamza, A.; Mickus, A.; Damaševičius, R. Breast Cancer Classification from Ultrasound Images Using Probability-Based Optimal Deep Learning Feature Fusion. Sensors 2022, 22, 807. [Google Scholar] [CrossRef] [PubMed]
  2. Naseem, U.; Rashid, J.; Ali, L.; Kim, J.; Haq, Q.E.U.; Awan, M.J.; Imran, M. An Automatic Detection of Breast Cancer Diagnosis and Prognosis Based on Machine Learning Using Ensemble of Classifiers. IEEE Access 2022, 10, 78242–78252. [Google Scholar] [CrossRef]
  3. Arooj, S.; Rahman, A.U.; Zubair, M.; Khan, M.F.; Alissa, K.; Mosavi, A. Breast Cancer Detection and Classification Empowered With Transfer Learning. Front. Public Health 2022, 10, 924432. [Google Scholar] [CrossRef] [PubMed]
  4. Aljuaid, H.; Alturki, N.; Alsubaie, N.; Cavallaro, L.; Liotta, A. Computer-aided diagnosis for breast cancer classification using deep neural networks and transfer learning. Comput. Methods Programs Biomed. 2022, 223, 106951. [Google Scholar] [CrossRef]
  5. Rashid, A.; Farhad, S.S.B.; Bhuyian, A.; Yeasmin, N.; Azim, M.A.; Alom, Z. A Comparative Analysis of Machine Learning techniques on Breast Cancer diagnosis using WEKA. In Proceedings of the 2022 25th International Conference on Computer and Information Technology (ICCIT), Cox's Bazar, Bangladesh, 17–19 December 2022; pp. 663–668. [Google Scholar] [CrossRef]
  6. Ye, G.; He, S.; Pan, R.; Zhu, L.; Zhou, D.; Lu, R. Research on DCE-MRI Images Based on Deep Transfer Learning in Breast Cancer Adjuvant Curative Effect Prediction. J. Health Eng. 2022, 2022, 4477099. [Google Scholar] [CrossRef]
  7. Sultana, Z.; Khan, A.R.; Jahan, N. Early Breast Cancer Detection Utilizing Artificial Neural Network. WSEAS Trans. Biol. Biomed. 2021, 18, 32–42. [Google Scholar] [CrossRef]
  8. Ganesan, K.; Acharya, U.R.; Chua, C.K.; Min, L.C.; Abraham, K.T.; Ng, K.-H. Computer-Aided Breast Cancer Detection Using Mammograms: A Review. IEEE Rev. Biomed. Eng. 2012, 6, 77–98. [Google Scholar] [CrossRef]
  9. Bhidé, A.; Datar, S.; Stebbins, K. Mammography: Case histories of significant medical advances. Harv. Bus. Sch. Account. Manag. Unit Work. Pap. 2021, 20-002. [Google Scholar] [CrossRef]
  10. Hooley, R.J.; Andrejeva, L.; Scoutt, L.M. Breast Cancer Screening and Problem Solving Using Mammography, Ultrasound, and Magnetic Resonance Imaging. Ultrasound Q. 2011, 27, 23–47. [Google Scholar] [CrossRef]
  11. Kelly, K.M.; Dean, J.; Lee, S.-J.; Comulada, W.S. Breast cancer detection: Radiologists’ performance using mammography with and without automated whole-breast ultrasound. Eur. Radiol. 2010, 20, 2557–2564. [Google Scholar] [CrossRef]
  12. Rehman, K.U.; Li, J.; Pei, Y.; Yasin, A.; Ali, S.; Mahmood, T. Computer Vision-Based Microcalcification Detection in Digital Mammograms Using Fully Connected Depthwise Separable Convolutional Neural Network. Sensors 2021, 21, 4854. [Google Scholar] [CrossRef]
  13. Ramani, R.; Vanitha, N.; Valarmathy, S. The Pre-Processing Techniques for Breast Cancer Detection in Mammography Images. Int. J. Image, Graph. Signal Process. 2013, 5, 47. [Google Scholar] [CrossRef]
  14. Tripathy, S.; Swarnkar, T. Unified Preprocessing and Enhancement Technique for Mammogram Images. Procedia Comput. Sci. 2020, 167, 285–292. [Google Scholar] [CrossRef]
  15. Zahoor, S.; Lali, I.U.; Khan, M.A.; Javed, K.; Mehmood, W. Breast Cancer Detection and Classification using Traditional Computer Vision Techniques: A Comprehensive Review. Curr. Med. Imaging 2020, 16, 1187–1200. [Google Scholar] [CrossRef]
  16. Mousa, R.; Munib, Q.; Moussa, A. Breast cancer diagnosis system based on wavelet analysis and fuzzy-neural. Expert Syst. Appl. 2005, 28, 713–723. [Google Scholar] [CrossRef]
  17. Udayakumar, E.; Santhi, S.; Vetrivelan, P. An investigation of Bayes algorithm and neural networks for identifying the breast cancer. Indian J. Med. Paediatr. Oncol. 2017, 38, 340–344. [Google Scholar] [CrossRef]
  18. Luo, S.-T.; Cheng, B.-W. Diagnosing Breast Masses in Digital Mammography Using Feature Selection and Ensemble Methods. J. Med. Syst. 2012, 36, 569–577. [Google Scholar] [CrossRef]
  19. Aalaei, S.; Shahraki, H.; Rowhanimanesh, A.; Eslami, S. Feature selection using genetic algorithm for breast cancer diagnosis: Experiment on three different datasets. Iranian J. Basic Med. Sci. 2016, 19, 476–482. [Google Scholar] [CrossRef]
  20. Abreu, P.H.; Santos, M.S.; Abreu, M.H.; Andrade, B.; Silva, D.C. Predicting breast cancer recurrence using machine learning techniques: A systematic review. ACM Comput. Surv. (CSUR) 2016, 49, 1–40. [Google Scholar] [CrossRef]
  21. Sharma, N.; Sharma, K.P.; Mangla, M.; Rani, R. Breast cancer classification using snapshot ensemble deep learning model and t-distributed stochastic neighbor embedding. Multimedia Tools Appl. 2023, 82, 4011–4029. [Google Scholar] [CrossRef]
  22. Houssein, E.H.; Emam, M.M.; Ali, A.A.; Suganthan, P.N. Deep and machine learning techniques for medical imaging-based breast cancer: A comprehensive review. Expert Syst. Appl. 2021, 167, 114161. [Google Scholar] [CrossRef]
  23. Ali, W.; Saeed, F. Hybrid Filter and Genetic Algorithm-Based Feature Selection for Improving Cancer Classification in High-Dimensional Microarray Data. Processes 2023, 11, 562. [Google Scholar] [CrossRef]
  24. Nadira, T.; Rustam, Z. Classification of cancer data using support vector machines with features selection method based on global artificial bee colony. In AIP Conference Proceedings; AIP Publishing LLC: New York, NY, USA, 2018; p. 020205. [Google Scholar] [CrossRef]
  25. Jiménez-Sánchez, A.; Tardy, M.; Ballester, M.A.G.; Mateus, D.; Piella, G. Memory-aware curriculum federated learning for breast cancer classification. Comput. Methods Programs Biomed. 2023, 229, 107318. [Google Scholar] [CrossRef] [PubMed]
  26. Khashei, M.; Bakhtiarvand, N. A novel discrete learning-based intelligent methodology for breast cancer classification purposes. Artif. Intell. Med. 2023, 7, 102492. [Google Scholar] [CrossRef] [PubMed]
  27. Loizidou, K.; Elia, R.; Pitris, C. Computer-aided breast cancer detection and classification in mammography: A comprehensive review. Comput. Biol. Med. 2023, 153, 106554. [Google Scholar] [CrossRef]
  28. Zebari, D.A.; Ibrahim, D.A.; Zeebaree, D.Q.; Mohammed, M.A.; Haron, H.; Zebari, N.A.; Damaševičius, R.; Maskeliūnas, R. Breast cancer detection using mammogram images with improved multi-fractal dimension approach and feature fusion. Appl. Sci. 2021, 11, 12122. [Google Scholar] [CrossRef]
  29. Siddiqui, S.Y.; Haider, A.; Ghazal, T.M.; Khan, M.A.; Naseer, I.; Abbas, S.; Rahman, M.; Khan, J.A.; Ahmad, M.; Hasan, M.K.; et al. IoMT Cloud-Based Intelligent Prediction of Breast Cancer Stages Empowered With Deep Learning. IEEE Access 2021, 9, 146478–146491. [Google Scholar] [CrossRef]
  30. Huynh, B.; Drukker, K.; Giger, M. MO-DE-207B-06: Computer-Aided Diagnosis of Breast Ultrasound Images Using Transfer Learning From Deep Convolutional Neural Networks. Med. Phys. 2016, 43, 3705. [Google Scholar] [CrossRef]
  31. Hamed, G.; Marey, M.A.E.-R.; Amin, S.E.-S.; Tolba, M.F. Deep learning in breast cancer detection and classification. In Proceedings of the International Conference on Artificial Intelligence and Computer Vision (AICV2020), Proceedings of the Artificial Intelligence and Computer Vision (AICV 2020), Cairo, Egypt, 8–10 April 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 322–333. [Google Scholar]
  32. Zheng, J.; Lin, D.; Gao, Z.; Wang, S.; He, M.; Fan, J. Deep Learning Assisted Efficient AdaBoost Algorithm for Breast Cancer Detection and Early Diagnosis. IEEE Access 2020, 8, 96946–96954. [Google Scholar] [CrossRef]
  33. Bayramoglu, N.; Kannala, J.; Heikkilä, J. Deep learning in magnification independent breast cancer histopathology image classification. In Proceedings of the 2016 23rd International conference on pattern recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 2440–2445. [Google Scholar]
  34. Khuriwal, N.; Mishra, N. Breast cancer diagnosis using deep learning algorithm. In Proceedings of the 2018 International Conference on Advances in Computing, Communication Control and Networking (ICACCCN), Greater Noida, India, 12–13 October 2018; pp. 98–103. [Google Scholar] [CrossRef]
  35. Mambou, S.J.; Maresova, P.; Krejcar, O.; Selamat, A.; Kuca, K. Breast Cancer Detection Using Infrared Thermal Imaging and a Deep Learning Model. Sensors 2018, 18, 2799. [Google Scholar] [CrossRef]
  36. Ahmed, L.; Iqbal, M.M.; Aldabbas, H.; Khalid, S.; Saleem, Y.; Saeed, S. Images data practices for Semantic Segmentation of Breast Cancer using Deep Neural Network. J. Ambient. Intell. Humaniz. Comput. 2020, 1–17. [Google Scholar] [CrossRef]
  37. Aruna, S.; Rajagopalan, S.; Nandakishore, L. Knowledge based analysis of various statistical tools in detecting breast cancer. Comput. Sci. Inf. Technol. 2011, 2, 37–45. [Google Scholar]
  38. Kumar, K.; Rao, A.C.S. Breast cancer classification of image using convolutional neural network. In Proceedings of the 2018 4th International Conference on Recent Advances in Information Technology (RAIT), Dhanbad, India, 15–17 March 2018; pp. 1–6. [Google Scholar]
  39. Khalid, A.; Noureldien, N.A. Determining the efficient structure of feed-forward neural network to classify breast cancer dataset. Int. J. Adv. Comput. Sci. Appl. 2014, 5. [Google Scholar] [CrossRef]
  40. Albalawi, U.; Manimurugan, S.; Varatharajan, R. Classification of breast cancer mammogram images using convolution neural network. Concurr. Comput. Pract. Exp. 2022, 34, e5803. [Google Scholar] [CrossRef]
  41. Ganesh, N.; Shankar, R.; Čep, R.; Chakraborty, S.; Kalita, K. Efficient Feature Selection Using Weighted Superposition Attraction Optimization Algorithm. Appl. Sci. 2023, 13, 3223. [Google Scholar] [CrossRef]
  42. Kurman, S.; Kisan, S. An in-depth and contrasting survey of meta-heuristic approaches with classical feature selection techniques specific to cervical cancer. Knowl. Inf. Syst. 2023, 65, 1881–1934. [Google Scholar] [CrossRef]
  43. Falconi, L.G.; Perez, M.; Aguila, W.G.; Conci, A. Transfer Learning and Fine Tuning in Breast Mammogram Abnormalities Classification on CBIS-DDSM Database. Adv. Sci. Technol. Eng. Syst. J. 2020, 5, 154–165. [Google Scholar] [CrossRef]
  44. Huang, M.-L.; Lin, T.-Y. Dataset of breast mammography images with masses. Data Brief 2020, 31, 105928. [Google Scholar] [CrossRef]
  45. Charan, S.; Khan, M.J.; Khurshid, K. Breast Cancer Detection in Mammograms Using Convolutional Neural Network. In Proceedings of the 2018 International Conference on Computing, Mathematics and Engineering Technologies (iCoMET), Wuhan, China, 7–8 February 2018; pp. 1–5. [Google Scholar]
  46. Fatima, N.; Liu, L.; Hong, S.; Ahmed, H. Prediction of Breast Cancer, Comparative Review of Machine Learning Techniques, and Their Analysis. IEEE Access 2020, 8, 150360–150376. [Google Scholar] [CrossRef]
  47. Munien, C.; Viriri, S. Classification of Hematoxylin and Eosin-Stained Breast Cancer Histology Microscopy Images Using Transfer Learning with EfficientNets. Comput. Intell. Neurosci. 2021, 2021, 5580914. [Google Scholar] [CrossRef]
  48. Lobbes, M.; Smidt, M.; Houwers, J.; Tjan-Heijnen, V.; Wildberger, J. Contrast enhanced mammography: Techniques, current results, and potential indications. Clin. Radiol. 2013, 68, 935–944. [Google Scholar] [CrossRef] [PubMed]
  49. Behar, N.; Shrivastava, M. ResNet50-Based Effective Model for Breast Cancer Classification Using Histopathology Images. Comput. Model. Eng. Sci. 2022, 130, 823–839. [Google Scholar] [CrossRef]
  50. Ashtaiwi, A. Optimal Histopathological Magnification Factors for Deep Learning-Based Breast Cancer Prediction. Appl. Syst. Innov. 2022, 5, 87. [Google Scholar] [CrossRef]
  51. Ayana, G.; Park, J.; Jeong, J.-W.; Choe, S.-W. A Novel Multistage Transfer Learning for Ultrasound Breast Cancer Image Classification. Diagnostics 2022, 12, 135. [Google Scholar] [CrossRef] [PubMed]
  52. Zhang, Y.; Chen, J.-H.; Lin, Y.; Chan, S.; Zhou, J.; Chow, D.; Chang, P.; Kwong, T.; Yeh, D.-C.; Wang, X.; et al. Prediction of breast cancer molecular subtypes on DCE-MRI using convolutional neural network with transfer learning between two centers. Eur. Radiol. 2021, 31, 2559–2567. [Google Scholar] [CrossRef]
  53. Rezaeijo, S.M.; Ghorvei, M.; Mofid, B. Predicting breast cancer response to neoadjuvant chemotherapy using ensemble deep transfer learning based on CT images. J. X-ray Sci. Technol. 2021, 29, 835–850. [Google Scholar] [CrossRef]
  54. Yang, X.-S.; Karamanoglu, M.; He, X. Flower pollination algorithm: A novel approach for multiobjective optimization. Eng. Optim. 2014, 46, 1222–1237. [Google Scholar] [CrossRef]
  55. Kavitha, T.; Mathai, P.P.; Karthikeyan, C.; Ashok, M.; Kohar, R.; Avanija, J.; Neelakandan, S. Deep Learning Based Capsule Neural Network Model for Breast Cancer Diagnosis Using Mammogram Images. Interdiscip. Sci. Comput. Life Sci. 2021, 123–129. [Google Scholar] [CrossRef]
  56. Shen, L. End-to-end training for whole image breast cancer diagnosis using an all convolutional design. arXiv 2017, arXiv:1711.05775. [Google Scholar]
  57. Mobark, N.; Hamad, S.; Rida, S.Z. CoroNet: Deep Neural Network-Based End-to-End Training for Breast Cancer Diagnosis. Appl. Sci. 2022, 12, 7080. [Google Scholar] [CrossRef]
Figure 1. The proposed framework for classifying breast cancer from mammography images.
Figure 1. The proposed framework for classifying breast cancer from mammography images.
Diagnostics 13 01618 g001
Figure 2. Visual samples of data augmentation on the selected dataset. * Started from the original image and obtained three new images after employing operations such as flip left to right, flip vertical, and rotate 90.
Figure 2. Visual samples of data augmentation on the selected dataset. * Started from the original image and obtained three new images after employing operations such as flip left to right, flip vertical, and rotate 90.
Diagnostics 13 01618 g002
Figure 4. For improved visualization of the images, contrast-enhancing methods were used. Left side: Original images before any image enhancement technique; Right Side: improved images after applying the proposed fused contrast enhancement method.
Figure 4. For improved visualization of the images, contrast-enhancing methods were used. Left side: Original images before any image enhancement technique; Right Side: improved images after applying the proposed fused contrast enhancement method.
Diagnostics 13 01618 g004
Figure 5. The architecture of ResNet50 model.
Figure 5. The architecture of ResNet50 model.
Diagnostics 13 01618 g005
Figure 6. Transfer learning-based training of modified model and feature extraction.
Figure 6. Transfer learning-based training of modified model and feature extraction.
Diagnostics 13 01618 g006
Figure 7. Confusion matrix of CSVM on CBIS-DDSM dataset using proposed fusion technique.
Figure 7. Confusion matrix of CSVM on CBIS-DDSM dataset using proposed fusion technique.
Diagnostics 13 01618 g007
Figure 8. Comparison of proposed fusion and selection steps in terms of computational time.
Figure 8. Comparison of proposed fusion and selection steps in terms of computational time.
Diagnostics 13 01618 g008
Figure 9. Confusion matrix of CSVM on CBIS-DDSM dataset using proposed feature selection technique.
Figure 9. Confusion matrix of CSVM on CBIS-DDSM dataset using proposed feature selection technique.
Diagnostics 13 01618 g009
Figure 13. Confusion matrix of MIAS dataset of CSVM Classifier after fusion.
Figure 13. Confusion matrix of MIAS dataset of CSVM Classifier after fusion.
Diagnostics 13 01618 g013
Figure 14. Confusion matrix of MIAS dataset for CSVM Classifier after proposed feature selection technique.
Figure 14. Confusion matrix of MIAS dataset for CSVM Classifier after proposed feature selection technique.
Diagnostics 13 01618 g014
Figure 15. Time-based comparison among proposed fusion and selection techniques for MIAS dataset.
Figure 15. Time-based comparison among proposed fusion and selection techniques for MIAS dataset.
Diagnostics 13 01618 g015
Figure 16. CI-based analysis of proposed framework on CBIS-DDSM dataset using Kappa measure.
Figure 16. CI-based analysis of proposed framework on CBIS-DDSM dataset using Kappa measure.
Diagnostics 13 01618 g016
Figure 17. CI-based analysis of proposed framework on INbreast and MIAS datasets.
Figure 17. CI-based analysis of proposed framework on INbreast and MIAS datasets.
Diagnostics 13 01618 g017
Figure 18. Grad-CAM-based visualization of proposed deep learning framework.
Figure 18. Grad-CAM-based visualization of proposed deep learning framework.
Diagnostics 13 01618 g018
Figure 19. Label-based visual prediction results of the proposed framework.
Figure 19. Label-based visual prediction results of the proposed framework.
Diagnostics 13 01618 g019
Figure 20. Accuracy analysis for CBIS-DDSM dataset after employing several training and testing ratios.
Figure 20. Accuracy analysis for CBIS-DDSM dataset after employing several training and testing ratios.
Diagnostics 13 01618 g020
Table 1. Brief description of selected datasets.
Table 1. Brief description of selected datasets.
Dataset NameClassesImagesAugmentedTraining/Testing
CBIS-DDSMBenign55760003000/3000
Malignant63760003000/3000
INbreastBenign7640002000/2000
Malignant7040002000/2000
MIASBenign5240002000/2000
Malignant3940002000/2000
Normal20940002000/2000
Table 2. Classification results of proposed feature fusion CBIS-DDSM dataset.
Table 2. Classification results of proposed feature fusion CBIS-DDSM dataset.
ClassifierPrecisionSensitivityF1-ScoreFPRKappaMCCAccuracyTime (s)
CSVM93.0994.7093.890.0787.6787.6893.8743.76
LSVM89.8390.6790.250.1080.4080.4090.2598.94
QSVM92.0393.2092.610.0885.1385.1492.6598.17
MGSVM91.5592.4792.010.0883.9383.9492.0824.48
MNN90.7992.0391.410.0982.7082.7191.3249.39
WNN91.4592.0091.720.0883.4083.4091.7255.09
FKNN91.9693.7392.840.0885.5385.5592.8531.48
NNN90.0390.5790.300.1080.5380.5390.3344.00
BNN89.3789.6389.500.0178.9778.9789.5419.86
TNN88.8689.9089.380.1178.6378.6489.3764.00
Table 3. Classification results of proposed FPcRF-based feature selection for CBIS-DDSM dataset.
Table 3. Classification results of proposed FPcRF-based feature selection for CBIS-DDSM dataset.
ClassifierPrecisionSensitivityF1-ScoreFPRKappaMCCAccuracyTime (s)
CSVM92.2694.5793.400.0786.6386.6693.3330.67
LSVM89.0489.8789.450.1178.8078.8089.4295.41
QSVM90.9592.7791.850.0985.5383.5591.8287.11
MGSVM91.0392.3091.660.09183.2083.2191.6411.00
MNN89.9990.5090.240.1080.4380.4390.282.53
WNN90.4790.8090.630.0981.2381.2390.6128.72
FKNN92.0693.5092.770.0885.4385.4492.7264.14
NNN88.4889.3788.920.1177.7377.7488.9171.24
BNN88.8189.4089.100.1178.1378.1489.1400.99
TNN87.9189.4388.660.1277.1377.1488.6503.23
Table 6. Classification results of proposed features fusion for MIAS dataset.
Table 6. Classification results of proposed features fusion for MIAS dataset.
ClassifierPrecisionSensitivityF1-ScoreFPRKappaMCCAccuracyTime (s)
CSVM99.8299.8299.820.0899.6099.7399.8364.76
LSVM99.4999.4999.490.0298.8599.2399.5224.13
QSVM99.7699.7699.760.0199.4599.6399.8325.27
MGSVM99.8299.8299.820.0899.6099.7399.8556.93
MNN99.5699.5699.560.0299.0099.3399.6107.21
WNN99.6799.6799.90.0199.2599.5099.7183.39
FKNN98.5098.4798.460.0796.5597.7298.5340.18
NNN99.5199.5199.510.0298.9099.2799.5104.89
BNN99.4799.4799.470.0298.8099.2099.5106.68
TNN98.9898.9899.980.0597.7098.4799.0131.41
Table 7. Classification results of proposed feature selection technique for MIAS dataset.
Table 7. Classification results of proposed feature selection technique for MIAS dataset.
ClassifierPrecisionSensitivityF1-ScoreFPRKappaMCCAccuracyTime (s)
CSVM99.7899.7899.780.0199.5099.6799.8173.23
LSVM99.3499.3399.330.0398.5099.0099.3102.25
QSVM99.6999.6999.690.0199.3099.5399.7146.78
MGSVM99.8299.8299.820.0899.6099.7399.8313.13
MNN99.4599.4499.440.0298.7599.1799.441.72
WNN99.5199.5199.510.0298.9099.2799.565.96
FKNN98.4798.4298.420.0796.4597.6598.3170.29
NNN99.3699.3699.360.0398.5599.0499.441.62
BNN98.9698.9698.960.0597.6598.4499.043.81
TNN98.6998.6999.690.0697.0598.0498.750.72
Table 8. Comparison with other state-of-the-art techniques.
Table 8. Comparison with other state-of-the-art techniques.
ModelYearMethodF1-ScoreAccuracy
[55]2022Capsule Neural Network ModelCBIS-DDSM/
INbreast/MIAS
98.4
[56]2017All Convolutional DesignCBIS-DDSM/
INbreast/MIAS
96
[57]2022CoroNetCBIS-DDSM/
INbreast/MIAS
99.7
Proposed Method-Fine-Tuned ResNet50 Model, Fusion & Flower Pollination
Optimization Algorithm
CBIS-DDSM/
INbreast/MIAS
93.3%
99.5%
99.8%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rehman, S.u.; Khan, M.A.; Masood, A.; Almujally, N.A.; Baili, J.; Alhaisoni, M.; Tariq, U.; Zhang, Y.-D. BRMI-Net: Deep Learning Features and Flower Pollination-Controlled Regula Falsi-Based Feature Selection Framework for Breast Cancer Recognition in Mammography Images. Diagnostics 2023, 13, 1618. https://doi.org/10.3390/diagnostics13091618

AMA Style

Rehman Su, Khan MA, Masood A, Almujally NA, Baili J, Alhaisoni M, Tariq U, Zhang Y-D. BRMI-Net: Deep Learning Features and Flower Pollination-Controlled Regula Falsi-Based Feature Selection Framework for Breast Cancer Recognition in Mammography Images. Diagnostics. 2023; 13(9):1618. https://doi.org/10.3390/diagnostics13091618

Chicago/Turabian Style

Rehman, Shams ur, Muhamamd Attique Khan, Anum Masood, Nouf Abdullah Almujally, Jamel Baili, Majed Alhaisoni, Usman Tariq, and Yu-Dong Zhang. 2023. "BRMI-Net: Deep Learning Features and Flower Pollination-Controlled Regula Falsi-Based Feature Selection Framework for Breast Cancer Recognition in Mammography Images" Diagnostics 13, no. 9: 1618. https://doi.org/10.3390/diagnostics13091618

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop