Next Article in Journal
The Effect of Repeated Restraint Stress on Neuroglobin-Oligodendrocytes Functions in the CA3 Hippocampal Area and Their Involvements in the Signaling Pathways of the Stress-Induced Anxiety
Next Article in Special Issue
A Hybrid Linear Iterative Clustering and Bayes Classification-Based GrabCut Segmentation Scheme for Dynamic Detection of Cervical Cancer
Previous Article in Journal
Planet Optimization with Deep Convolutional Neural Network for Lightweight Intrusion Detection in Resource-Constrained IoT Networks
Previous Article in Special Issue
A Deep Learning Method for Early Detection of Diabetic Foot Using Decision Fusion and Thermal Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Aquila Optimizer with Bayesian Neural Network for Breast Cancer Detection on Ultrasound Images

1
Department of Biomedical Engineering, College of Engineering, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
2
Department of Information Systems, College of Science & Art at Mahayil, King Khalid University, Abha 62529, Saudi Arabia
3
Department of Information Technology, College of Computers and Information Technology, Taif University, P.O. Box 11099, Taif 21944, Saudi Arabia
4
Department of Computer Sciences, College of Computing and Information System, Umm Al-Qura University, Makkah 24382, Saudi Arabia
5
Research Centre, Future University in Egypt, New Cairo 11845, Egypt
6
Department of Computer and Self Development, Preparatory Year Deanship, Prince Sattam Bin Abdulaziz University, AlKharj 16278, Saudi Arabia
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(17), 8679; https://doi.org/10.3390/app12178679
Submission received: 30 July 2022 / Revised: 25 August 2022 / Accepted: 27 August 2022 / Published: 30 August 2022
(This article belongs to the Special Issue Applications of Artificial Intelligence in Medical Imaging)

Abstract

:
Breast cancer is the second most dominant kind of cancer among women. Breast Ultrasound images (BUI) are commonly employed for the detection and classification of abnormalities that exist in the breast. The ultrasound images are necessary to develop artificial intelligence (AI) enabled diagnostic support technologies. For improving the detection performance, Computer Aided Diagnosis (CAD) models are useful for breast cancer detection and classification. The current advancement of the deep learning (DL) model enables the detection and classification of breast cancer with the use of biomedical images. With this motivation, this article presents an Aquila Optimizer with Bayesian Neural Network for Breast Cancer Detection (AOBNN-BDNN) model on BUI. The presented AOBNN-BDNN model follows a series of processes to detect and classify breast cancer on BUI. To accomplish this, the AOBNN-BDNN model initially employs Wiener filtering (WF) related noise removal and U-Net segmentation as a pre-processing step. Besides, the SqueezeNet model derives a collection of feature vectors from the pre-processed image. Next, the BNN algorithm will be utilized to allocate appropriate class labels to the input images. Finally, the AO technique was exploited to fine-tune the parameters related to the BNN method so that the classification performance is improved. To validate the enhanced performance of the AOBNN-BDNN method, a wide experimental study is executed on benchmark datasets. A wide-ranging experimental analysis specified the enhancements of the AOBNN-BDNN method in recent techniques.

1. Introduction

Despite substantial technological advancements, cancer detection at an initial stage remains a challenging one, and present cancer detection methods are time-consuming, expensive, complex, and uncomfortable [1]. Growths in organic electronic materials with optical imaging modalities, improvised models of different optical properties, optical biosensors, and biocompatibility are auspicious approaches for initial cancer detection and that of other diseases. Latest imaging techniques are compiled with the progression of ultrasound, positron emission tomography (PET), computed tomography (CT), and magnetic resonance imaging (MRI) for enhanced diagnosis of cancer and treatment which will help screen patients more accurately [2]. Developments in optical biosensors or organic electronics help in the segregation between healthy and cancer cells and the remodeling point-of-care gadgets can be used in cancer diagnosis [3]. Presently, utilizing ultrasound approaches for cancer detection depends on the experience of the clinician, particularly for the measurements and marks of cancers. To be specific, a clinician generally makes use of ultrasound instruments for cancer recognition by identifying a good angle for a clear vision of cancer displayed on the screen, and after continuing investigation fixing this for a longer period by utilizing one hand, with the other hand marking and measuring the cancers on the screen [4,5]. It becomes a tough task, because slight shaking of the hand which holds the probe may have a significant influence on the quality of breast ultrasound imagery; depending on this, computer-aided automated detection technologies are in high demand to locate regions of interest (ROIs), i.e., cancers, in breast ultrasound images [6].
Many researchers have involved computer-aided diagnosis (CAD) methods for breast cancer detection; such methods invoke the usage of Bayesian networks, artificial neural networks (ANN), k-means clustering, decision trees (DT), and fuzzy logic (FL) [7,8]. However, some of the authors have applied CAD techniques with DOT for diagnosing breast cancer. In recent times, convolutional neural networks (CNNs) have proved to be effective in distinguishing between malignant and benign breast lesions [9]. In comparison with classical techniques, CNNs eliminate the stages involved in feature extracting of an image; on the other hand, they provide images straight to the network which could automatically study discriminatory features [10]. CNN architecture was specifically adapted to have the benefit of 2D structures for input images.
This article introduces an Aquila Optimizer with Bayesian Neural Network for Breast Cancer Detection (AOBNN-BDNN) model on BUI. The presented AOBNN-BDNN model follows a series of processes to detect and classify breast cancer on BUI. To accomplish this, the AOBNN-BDNN model initially employs Wiener filtering (WF) related noise removal and U-Net segmentation as a pre-processing step. Besides, the SqueezeNet model derives a collection of feature vectors from the pre-processed image. Next, the BNN method was utilized for allocating suitable class labels to the input images. Finally, the AO technique can be employed to fine-tune the parameters relevant to the BNN method so that the classification performance will be enhanced, showing the novelty of our work. To validate the enriched performance of the AOBNN-BDNN algorithm, a wide experimental study is executed on a benchmark dataset.
The rest of the paper is organized as follows. Section 2 offers a brief survey of breast cancer classification using ultrasound images. Section 3 elaborates on the proposed model and Section 4 validates the performance of the proposed model. Lastly, Section 5 concludes the study.

2. Related Works

Hijab et al. [11] present a DL technique for managing this issue. The trained data, which has numerous images of malignant and benign cases, has been employed for training a deep CNN. Three training techniques were devised: a baseline technique in which the CNN structure can be well-trained from scratch, a TL method in which the pretrained VGG16 CNN structure can be additionally well-trained with the ultrasound imagery, and a finely tuned learning method in which the DL variables were optimally tuned to overcome overfitting. Kalafi et al. [12] introduce a novel structure for classifying breast cancer (BC) lesions with an attention module in an adapted VGG16 structure. The implemented attention system will enhance the feature discrimination among the target and background lesions in ultrasound. The author devises a novel ensemble loss function, which can be a grouping of the logarithm and binary cross-entropy of hyperbolic cosine loss, to enhance the method discrepancy among labels and classified lesions.
Lee et al. [13] devised a channel attention module including multiscale grid average pooling (MSGRAP) for segmenting BC sections accurately in ultrasound imagery. The author establishes the efficiency of the channel attention elements by adding MSGRAP for semantic segmentation and advances a new semantic segmentation network including the presented attention modules for segmenting BC regions accurately in ultrasound images. Whereas a standard convolutional function does not use global spatial data on input and just uses small local data in a kernel belonging to a convolution filter, the presented attention component permits utilization of global as well as local spatial data. Xie et al. [14] develop a technique called Domain Guided-CNN (DG-CNN) for incorporating margin data, a feature defined in the consensus for radiotherapists for diagnosing tumors in breast ultrasound (BUS) imagery. In DG-CNN, attention maps which emphasize marginal regions of cancers were initially produced and combined through various techniques into the networks.
Zhu et al. [15] attempted to advance an automated technique for distinguishing thyroid and breast lesions in ultrasound images with the help of deep CNNs (DCNN). To be specific, the author modeled a generic DCNN structure including TL and similar architectural variable settings for training methods for thyroid and BC (BNet and TNet) correspondingly, and to test the feasibility of these GA with an ultrasound image accumulated from medical practice. In [16], an RDAU-NET (Residual-Dilated-Attention-Gate-UNet) method was introduced and used for segmenting the cancers in BUS imagery. The method depends on the classical U-Net, but plain neural components were substituted by residual elements to enhance edge data and overcome the network performance deprivation issue linked with the deep network system. To raise the receptive domain and obtain more characteristic data, dilated convolutions can be utilized for processing the feature maps acquired from the encoder phases.

3. The Proposed Model

In this study, a new AOBNN-BDNN algorithm was modelled for the recognition and classification of breast cancer on BUI. The suggested AOBNN-BDNN model follows a series of processes to detect and classify breast cancer on BUI. At the initial stage, the WF-based noise removal and U-Net segmentation is a pre-processing step. In addition, the SqueezeNet model derives a collection of feature vectors from the pre-processed image. Finally, the AO with the BNN method can be utilized to allocate suitable class labels to the input images. Figure 1 depicts the overall process of the AOBNN-BDNN approach.

3.1. Image Pre-Processing

At the initial stage, the AOBNN-BDNN algorithm applies WF-based noise removal and U-Net segmentation as a pre-processing step. The WF technique carries out 2D adaptive noise removal filtering using a window size of 3 × 3 that eliminates the blur generated in Gaussian smoothing [17]. As a result, images are restored without noise. The local mean ( μ ) and variance ( σ 2 ) of every pixel can be defined as follows.
μ = 1 N M n 1 , n 2 ϵ θ α ( n 1 , n 2 )  
σ 2 = 1 N M n 1 , n 2 ϵ θ α 2 ( n 1 , n 2 ) μ 2
Next, the WF can be defined as follows:
b ( n 1 , n 2 ) = μ + ( σ 2 β 2 ) σ 2 ( α ( n 1 , n 2 ) μ )  
where b ( n 1 , n 2 ) refers to a recovered image, θ refers to the N × M local neighborhood of every pixel in a scaled input image and β 2 stands for the noise variance.
Next, the preprocessed images are passed onto the U-Net architecture to perform the segmentation process. This is a fundamental structure in the medical imaging automation society and has massive applications in the domain. The network structure consists of contractive and expansive paths [18]. The contracting path involves numerous patches of convolution with a filter size of 3 × 33 × 3 and unity strides in both directions, after the ReLU layer. Next, it draws data from the feature vector through up-convolution, and generates, through a consecutive operation, an output segmenting map, and from the contractive path, through cropping and copying. The building block of this architecture is the operation connecting both paths. This connection enables the network to accomplish a precise dataset from the contractive path, therefore producing the segmentation mask for the output.

3.2. Feature Extraction Using SqueezeNet

After pre-processing, the SqueezeNet model derives a collection of feature vectors from the pre-processed image. SqueezeNet is a lightweight and efficient CNN mechanism [19]. Using this model, we accomplish a 50× reduction in module size in comparison with AlexNet, while exceeding or meeting the top-1 and -5 accurateness of AlexNet. The SqueezetNet module accomplishes remarkable outcomes with a small number of variables. There are two significant portions of CNN: classification and feature extraction. The extracted feature was employed for the precise image classification. Especially, these two portions of CNN accomplish the primary operation of CNN. The feature extraction of CNN encompasses convolution and sampling layers. The filtering of convolution layers is applied to diminish the noise in an image; later, the features of the image were enhanced. The convolution process was completed between the presentation layer convolutional kernel layer and the upper layer feature vector. Lastly, the activation function of CNN decides the computation of the convolution process. The effectiveness of training a NN was noticeable while applying the cost function ( Z ), which indicates the proportion between the reached output and trained instance [20].
Z = 1 m [ x   l n   a ( 1 x ) ln ( 1 β ) ]
In Equation (4), m characterizes the trainable dataset amount, x denotes the foreseeable value, and β indicates the original values in the resulting layer.
The activation function role is a major portion of the classification technique with weighting the result of the CNN method and transmission kernel size. The ReLU activation function is in the middle of the commonly applied activation function. It is exploited from almost every CNN method for setting each negative value corresponding to zero. The zero setting inhibits several nodes contributed by the learning algorithm. An additional function, called ELU and LReLU, offers a slightly negative value and was hardly employed in the classification technique. ReLU activation function shows optimal results when compared to the LReLU activation function from the classification that is applied in this technique [21]. The ReLU activation function can be arithmetically formulated in the subsequent equation.
R e L U ( x ) = m a x ( 0 ;   x )
This is more commonly applied in embedded settings; it involves distinct models of compression techniques. For example, 3 × 3 convolution kernels in the proposed technique are replaced by 1 × 1 convolution kernels. By using these techniques, the variable number for the individual convolution function is minimalized through a factor of 9. Furthermore, the 3 × 3 convolution kernel is reduced and down-sampling is hindered in the network. As a result, the suggested technique reduces the computation efforts and the number of trained variables. For that reason, it is practicable to establish SqueezeNet in memory constraint hardware devices. In contrast with the contemporary AI model, we perceived that SqueezeNet has a minimal variable amount; consequently, it is the better choice for robot vacuum applications. Inappropriately, the model size (viz., 6.1 MB) was higher in comparison with the memory space existing in the robot vacuum.

3.3. Image Classification Using BNN

During the classification process, the BNN model is utilized to allocate proper class labels to the input images. The BNN is a kind of probability distribution through network weight, immediate advantages being the fully probabilistic treatment, therefore the approximation of uncertainty in prediction [22]. Here, we consider previous knowledge or p 0 ( W ) distribution of the weight and evaluate the posterior weight distribution p ( W | D ) afterward regarding the D dataset and it can be assessed by
p ( W | D ) = p ( D | W ) p 0 ( W ) p ( D )
In Equation (6), p 0 ( W ) refers to prior weight distribution, generally an isotropic Gaussian, and p ( D ) represents the normalized distribution constant. In these settings, prediction is performed by the posterior from Equation (6):
p ( Y * | X * , D ) = p ( Y * | , X * , W ) p ( W | D ) d W
In this study, the uncertainty arises because of sampling the batch normalization (BN) and dropout weight, and prediction can be performed through averaging T forwarded passes through the network:
p ( Y * | X * , D ) = 1 T t = 1 T p ( Y * | X * , D , W t )
The network layer (max pool, BN) is presented as follows [23].
  • Batch Normalization is a process to accelerate network training via decreasing the internal covariate shift (which defines the variations in the distribution of activation unit because of parameter changes) performed by normalizing the hidden layer activation through an evaluated μ β and σ β from every mini-batch.
  • Dropout is a regularization method, also regarded as a Bayesian methodology: the process eliminates part of the network in a random fashion which makes the weight stochastic quantity: W ^ = W α , where α B e r n o u l l i ( p ) ,   W refers to the first weight of the network, and denotes the direct product using binary vector randomly.
The network can be trained by the W dropout and it has a stochastic gradient descent, resulting in robustness and uncertainty.

3.4. Parameter Optimization

Finally, the AO technique is utilized to fine-tune the parameters related to the BNN method so that the classification performance will be improved [24]. The AO technique simulates AO’s social activity for catching their prey. AO is a population-based optimized approach, related to other meta-heuristic systems, which start with establishing a primary population X with N agents. The subsequent formula is utilized for executing this process. Figure 2 demonstrates the flowchart of the AO technique.
X i j = r 1 × ( U B j L B j ) + L B j ,   i = 1 , 2 ,   N j = 1 , 2 ,   ,   D i m
In Equation (9), U B j and L B j demonstrate the restrictions of searching space. r 1 [ 0 , 1 ] implies the arbitrary value and D i m denotes the dimensional of agents. In the AO approach, the following step is to perform exploration and exploitation until the optimal solution is established [25]. Two stages exist in exploitation and exploration. The X b optimal agent and (X) the average of agents are applied in the exploration, and it is mathematically expressed in the following:
X i ( t + 1 ) = X b ( t ) × ( 1 t T ) + ( X M ( t ) X b ( t ) r a n d ) ,
X M ( t ) = 1 N i = 1 N X ( t ) ,   j = 1 , 2 ,   ,   D i m
The exploration stage can be controlled using ( 1 t T ) in Equation (10). The maximal amount of generations is represented as T . The exploration stage applies the Levy flight (Levy(D)) distribution as well as X b to upgrade the solution as follows:
X i ( t + 1 ) = X b ( t ) × L e v y ( D ) + X R ( t ) + ( y x ) r a n d ,
L e v y   ( D ) = s × u × σ | v | 1 β , σ = ( Γ ( 1 + β ) × s i n e ( π β 2 ) Γ ( 1 + β 2 ) × β × 2 ( β 1 2 ) )  
In Equation (13), s = 0.01 and β = 1.5 u and v indicate the arbitrary values. X R refers to arbitrarily selected agents. Additionally, y and x refer to two variables utilized for stimulating the spiral shape:
y = r × c o s ( θ ) ,   x = r × s i n ( θ )  
r = r 1 + U × D 1 ,   θ = ω × D 1 + θ 1 ,   θ 1 = 3 × π 2
In Equation (15), ω = 0.005 and U = 0.00565 r 1 [ 0 ,   20 ] stands for an arbitrary value. The initial method utilized to improve the agent in the exploitation stage depends on X b and X M , comparable to exploration as follows:
X i ( t + 1 ) = ( X b ( t ) X M ( t ) ) × α r n d + ( U B × r n d + L B ) × δ
In Equation (16), U B = ( U B L B ) ,   α , and δ refer to the exploitation adjustment parameter. r n d [ 0 ,   1 ] denotes a random number. The agent is upgraded by X b , Levy, and the quality function Q F in the next exploitation phase:
X i ( t + 1 ) = Q P × X b ( t ) G X G 2 × L e v y ( D ) + r n d × G 1  
G X = ( G 1 × X ( t ) × r n d )
Q P ( t ) = t 2 × r n d (   ) 1 ( 1 T ) 2
Moreover, G 1 denotes the motion applied to track the optimum individual solution, as follows:
G 1 = 2 × r n d (   ) 1 ,   G 2 = 2 × ( 1 t T )
In Equation (19), r n d refers to a random number. Furthermore, G 2 indicates a variable decreased from two to zero as follow:
G 2 = 2 × ( 1 t T )
The AO method derives a fitness function (FF) to exhibit the better performance of the classification. It defines a positive integer to characterize the improved performance of the solution candidate. In this study, the reduction of classification error rate was regarded as the FF, as follows.
f i t n e s s ( x i ) = C l a s s i f i e r E r r o r R a t e ( x i ) = n u m b e r   o f   m i s c l a s s i f i e d   s a m p l e s T o t a l   n u m b e r   o f   s a m p l e s 100  

4. Results and Discussion

The experimental validation of the AOBNN-BDNN method is tested using the breast ultrasound image dataset [26]. It contains a total of 780 images with three class labels, as shown in Table 1.
Figure 3 exhibits the confusion matrices created by the AOBNN-BDNN algorithm on the applied data. On the entire dataset, the AOBNN-BDNN method has recognized 429 samples as benign, 210 samples as malignant, and 132 samples as normal. Simultaneously, on 70% of training (TR) data, the AOBNN-BDNN technique has recognized 311 samples as benign, 135 samples as malignant, and 92 samples as normal. Concurrently, on 30% of testing (TS) data, the AOBNN-BDNN method has recognized 118 samples as benign, 75 samples as malignant, and 40 samples as normal.
Table 2 and Figure 4 portray a brief classification outcome of the AOBNN-BDNN method on the entire dataset. The results implied that the AOBNN-BDNN method has effectually recognized all three classes. For instance, the AOBNN-BDNN model has identified benign class samples with a c c u y of 98.85%, p r e c n of 99.77%, r e c a l of 98.17%, F s c o r e of 98.96%, and MCC of 97.68%. Moreover, the AOBNN-BDNN algorithm has identified Malignant class samples with a c c u y of 98.97%, p r e c n of 96.33%, r e c a l of 100%, F s c o r e of 98.13%, and MCC of 97.46%. Furthermore, the AOBNN-BDNN technique has identified normal class samples with a c c u y of 99.87%, p r e c n of 100%, r e c a l of 99.25%, F s c o r e of 99.62%, and MCC of 99.55%.
Table 3 and Figure 5 display a detailed classification outcome of the AOBNN-BDNN algorithm on 70% of TR data. The results denoted the AOBNN-BDNN methodology has effectually recognized all three classes. For example, the AOBNN-BDNN technique has identified benign class samples with a c c u y of 98.53%, p r e c n of 99.68%, r e c a l of 97.80%, F s c o r e of 98.73%, and MCC of 97.02%. Additionally, the AOBNN-BDNN approach has identified Malignant class samples with a c c u y of 98.72%, p r e c n of 95.07%, r e c a l of 100%, F s c o r e of 97.47%, and MCC of 96.67%. Besides, the AOBNN-BDNN technique has identified normal class samples with a c c u y of 99.82%, p r e c n of 100%, r e c a l of 98.92%, F s c o r e of 99.46%, and MCC of 99.35%.
Table 4 and Figure 6 exhibit brief classification results of the AOBNN-BDNN technique on 30% of the TS dataset. The outcomes implied that the AOBNN-BDNN approach has effectually recognized all three classes. For example, the AOBNN-BDNN methodology has identified benign class samples with a c c u y of 99.57%, p r e c n of 100%, r e c a l of 99.16%, F s c o r e of 99.58%, and MCC of 99.15%. along with that, the AOBNN-BDNN algorithm has identified Malignant class samples with a c c u y of 99.57%, p r e c n of 98.68%, r e c a l of 100%, F s c o r e of 99.34%, and MCC of 99.03%. In addition, the AOBNN-BDNN methodology has identified normal class samples with a c c u y of 100%, p r e c n of 100%, r e c a l of 100%, F s c o r e of 100%, and MCC of 100%.
The training accuracy (TA) and validation accuracy (VA) obtained by the AOBNN-BDNN methodology on the test dataset is established in Figure 7. The experimental result denoted the AOBNN-BDNN algorithm has reached higher values of TA and VA. In Particular, the VA is greater than TA.
The training loss (TL) and validation loss (VL) obtained by the AOBNN-BDNN approach on the test dataset are displayed in Figure 8. The experimental outcome represented that the AOBNN-BDNN method has exhibited minimal values of TL and VL. To be specific, the VL is lesser than the TL.
A clear precision-recall analysis of the AOBNN-BDNN technique on the test dataset is represented in Figure 9. The figure inferred the AOBNN-BDNN technique has resulted in enhanced values of precision-recall values in every class label.
A brief ROC study of the AOBNN-BDNN method on the test dataset is shown in Figure 10. The outcomes signified that the AOBNN-BDNN methodology has displayed its capability in classifying distinct classes on the test dataset.
Table 5 provides an overall comparative inspection of the AOBNN-BDNN method with recent approaches [11]. Figure 11 renders a brief study of the AOBNN-BDNN method with existing models in terms of p r e c n and r e c a l . The experimental outcomes reported the betterment of the AOBNN-BDNN model. With respect to p r e c n , the AOBNN-BDNN model has attained increased p r e c n of 99.56% whereas the SEODTL-BDC, ESD, LSVM, ESKNN, FKNN, and LD methods have obtained reduced p r e c n of 99.18%, 98.89%, 98.41%, 97.55%, 96.93%, and 97.41% respectively. In addition, with regard to r e c a l , the AOBNN-BDNN method has achieved increased r e c a l of 99.72% whereas the SEODTL-BDC, ESD, LSVM, ESKNN, FKNN, and LD algorithms have gained reduced r e c a l of 98.18%, 98.11%, 97.87%, 98.27%, 97%, and 98.05% correspondingly.
Figure 12 offers a comparative examination of the AOBNN-BDNN method with existing models with a c c u y and F s c o r e . The experimental outcomes reported the betterment of the AOBNN-BDNN method. With a c c u y , the AOBNN-BDNN algorithm has gained increased a c c u y of 99.72% whereas the SEODTL-BDC, ESD, LSVM, ESKNN, FKNN, and LD methodologies have reached reduced a c c u y of 99.12%, 99.07%, 98.73%, 97.70%, 97.42%, and 97.92%, correspondingly. Additionally, concerning F s c o r e , the AOBNN-BDNN method has received increased F s c o r e of 99.64% whereas the SEODTL-BDC, ESD, LSVM, ESKNN, FKNN, and LD methodologies have gained reduced F s c o r e of 98.14%, 98.88%, 98.29%, 98.61%, 96.93%, and 96.99%, correspondingly.
Finally, a classification time (CT) inspection of the AOBNN-BDNN model with recent models is carried out in Table 6 and Figure 13. The attained values implied the ESKNN method has shown poor results with a higher CT of 2.28 min. In the meantime, the ESD and FKNN methods have obtained slightly reduced outcomes with closer CT of 2.28 min and 2.24 min, respectively.
In addition, the LSVM and LD models have accomplished moderately reduced CT of 1.95 min and 1.75 min, respectively. However, the AOBNN-BDNN model has gained maximum outcome with minimal CT of 1.01 min. Therefore, the experimental values guaranteed that the AOBNN-BDNN method was found to be an effective tool compared to other approaches.

5. Conclusions

In this study, a new AOBNN-BDNN method was developed for the recognition and classification of breast cancer on BUI. The suggested AOBNN-BDNN method follows a series of processes to detect and classify breast cancer on BUI, primarily the WF-based noise removal and U-Net segmentation as a pre-processing steps. Besides, the SqueezeNet model derives a collection of feature vectors from the pre-processed image. Next, the BNN method is utilized to allocate suitable class labels to the input images. Finally, the AO technique is utilized to fine-tune the parameters related to the BNN algorithm so that the classification performance will be improved. To validate the enhanced performance of the AOBNN-BDNN method, a wide-ranging experimental analysis has been conducted on the benchmark dataset. An extensive experimental analysis stated the enhancements of the AOBNN-BDNN method over recent approaches with maximum accuracy of 99.72%. In the future, advanced DL models can be designed to enhance breast cancer classification performance. Besides, the proposed model can be tested on a real-time large-scale dataset in the future.

Author Contributions

Conceptualization, S.B.H.H. and S.A.; methodology, M.O.; software, A.A.A.; validation, G.P.M., M.O. and S.A.; formal analysis, I.Y.; investigation, A.S.Z.; resources, I.Y.; data curation, A.S.Z.; writing—original draft preparation, M.O., S.A. and M.K.N.; writing—review and editing, A.M. (Abdullah Mohamed) and A.M. (Abdelwahed Motwakel); visualization, I.Y.; supervision, M.O.; project administration, M.K.N.; funding acquisition, M.O. All authors have read and agreed to the published version of the manuscript.

Funding

The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work through the Large Groups Project under grant number (25/43). Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2022R203), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia. The authors would like to thank the Deanship of Scientific Research at Umm Al-Qura University for supporting this work by Grant Code: 22UQU4310373DSR31.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing does not apply to this article as no datasets were generated during the current study.

Conflicts of Interest

The authors declare that they have no conflict of interest. The manuscript was written with the contributions of all authors. All authors have approved the final version of the manuscript.

References

  1. Muhammad, M.; Zeebaree, D.; Brifcani, A.M.A.; Saeed, J.; Zebari, D.A. Region of interest segmentation based on clustering techniques for breast cancer ultrasound images: A review. J. Appl. Sci. Technol. Trends 2020, 1, 78–91. [Google Scholar]
  2. Sun, Q.; Lin, X.; Zhao, Y.; Li, L.; Yan, K.; Liang, D.; Sun, D.; Li, Z.C. Deep learning vs. radiomics for predicting axillary lymph node metastasis of breast cancer using ultrasound images: Don’t forget the peritumoral region. Front. Oncol. 2020, 10, 53. [Google Scholar] [CrossRef] [PubMed]
  3. Ayana, G.; Dese, K.; Choe, S.W. Transfer learning in breast cancer diagnoses via ultrasound imaging. Cancers 2021, 13, 738. [Google Scholar] [CrossRef] [PubMed]
  4. Qian, X.; Pei, J.; Zheng, H.; Xie, X.; Yan, L.; Zhang, H.; Han, C.; Gao, X.; Zhang, H.; Zheng, W.; et al. Prospective assessment of breast cancer risk from multimodal multiview ultrasound images via clinically applicable deep learning. Nat. Biomed. Eng. 2021, 5, 522–532. [Google Scholar] [CrossRef]
  5. Zhang, X.; Li, H.; Wang, C.; Cheng, W.; Zhu, Y.; Li, D.; Jing, H.; Li, S.; Hou, J.; Li, J.; et al. Evaluating the accuracy of breast cancer and molecular subtype diagnosis by ultrasound image deep learning model. Front. Oncol. 2021, 11, 623506. [Google Scholar] [CrossRef]
  6. Khairalseed, M.; Javed, K.; Jashkaran, G.; Kim, J.W.; Parker, K.J.; Hoyt, K. Monitoring Early Breast Cancer Response to Neoadjuvant Therapy Using H-Scan Ultrasound Imaging: Preliminary Preclinical Results. J. Ultrasound Med. 2019, 38, 1259–1268. [Google Scholar] [CrossRef]
  7. Zhang, T.; Jiang, Z.; Xve, T.; Sun, S.; Li, J.; Ren, W.; Wu, A.; Huang, P. One-pot synthesis of hollow PDA@ DOX nanoparticles for ultrasound imaging and chemo-thermal therapy in breast cancer. Nanoscale 2019, 11, 21759–21766. [Google Scholar] [CrossRef]
  8. Wang, Y.; Choi, E.J.; Choi, Y.; Zhang, H.; Jin, G.Y.; Ko, S.B. Breast cancer classification in automated breast ultrasound using multiview convolutional neural network with transfer learning. Ultrasound Med. Biol. 2020, 46, 1119–1132. [Google Scholar] [CrossRef]
  9. Yan, Y.; Liu, Y.; Wu, Y.; Zhang, H.; Zhang, Y.; Meng, L. Accurate segmentation of breast tumors using AE U-net with HDC model in ultrasound images. Biomed. Signal Process. Control 2022, 72, 103299. [Google Scholar] [CrossRef]
  10. Zhang, G.; Zhao, K.; Hong, Y.; Qiu, X.; Zhang, K.; Wei, B. SHA-MTL: Soft and hard attention multi-task learning for automated breast cancer ultrasound image segmentation and classification. Int. J. Comput. Assist. Radiol. Surg. 2021, 16, 1719–1725. [Google Scholar] [CrossRef]
  11. Hijab, A.; Rushdi, M.A.; Gomaa, M.M.; Eldeib, A. Breast cancer classification in ultrasound images using transfer learning. In Proceedings of the 2019 Fifth International Conference on Advances in Biomedical Engineering (ICABME), Tripoli, Lebanon, 17–19 October 2019; pp. 1–4. [Google Scholar]
  12. Kalafi, E.Y.; Jodeiri, A.; Setarehdan, S.K.; Lin, N.W.; Rahmat, K.; Taib, N.A.; Ganggayah, M.D.; Dhillon, S.K. Classification of breast cancer lesions in ultrasound images by using attention layer and loss ensemble in deep convolutional neural networks. Diagnostics 2021, 11, 1859. [Google Scholar] [CrossRef] [PubMed]
  13. Lee, H.; Park, J.; Hwang, J.Y. Channel attention module with multiscale grid average pooling for breast cancer segmentation in an ultrasound image. IEEE Trans. Ultrason. Ferroelectr. Freq. Control. 2020, 67, 1344–1353. [Google Scholar] [PubMed]
  14. Xie, X.Z.; Niu, J.W.; Liu, X.F.; Li, Q.F.; Wang, Y.; Han, J.; Tang, S. DG-CNN: Introducing Margin Information into Convolutional Neural Networks for Breast Cancer Diagnosis in Ultrasound Images. J. Comput. Sci. Technol. 2022, 37, 277–294. [Google Scholar] [CrossRef]
  15. Zhu, Y.C.; AlZoubi, A.; Jassim, S.; Jiang, Q.; Zhang, Y.; Wang, Y.B.; Ye, X.D.; Hongbo, D.U. A generic deep learning framework to classify thyroid and breast lesions in ultrasound images. Ultrasonics 2021, 110, 106300. [Google Scholar] [CrossRef] [PubMed]
  16. Zhuang, Z.; Li, N.; Joseph Raj, A.N.; Mahesh, V.G.; Qiu, S. An RDAU-NET model for lesion segmentation in breast ultrasound images. PLoS ONE 2019, 14, e0221535. [Google Scholar] [CrossRef] [PubMed]
  17. Lahmiri, S. An iterative denoising system based on Wiener filtering with application to biomedical images. Opt. Laser Technol. 2017, 90, 128–132. [Google Scholar] [CrossRef]
  18. Saood, A.; Hatem, I. COVID-19 lung CT image segmentation using deep learning methods: U-Net versus SegNet. BMC Med. Imaging 2021, 21, 19. [Google Scholar] [CrossRef]
  19. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5 MB model size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
  20. Koonce, B. SqueezeNet. In Convolutional Neural Networks with Swift for Tensorflow; Apress: Berkeley, CA, USA, 2021; pp. 73–85. [Google Scholar]
  21. Ullah, A.; Elahi, H.; Sun, Z.; Khatoon, A.; Ahmad, I. Comparative analysis of AlexNet, ResNet18 and SqueezeNet with diverse modification and arduous implementation. Arab. J. Sci. Eng. 2022, 47, 2397–2417. [Google Scholar] [CrossRef]
  22. Wu, A.; Nowozin, S.; Meeds, E.; Turner, R.E.; Hernández-Lobato, J.M.; Gaunt, A.L. Deterministic variational inference for robust bayesian neural networks. arXiv 2018, arXiv:1810.03958. [Google Scholar]
  23. Meng, X.; Babaee, H.; Karniadakis, G.E. Multi-fidelity Bayesian neural networks: Algorithms and applications. J. Comput. Phys. 2021, 438, 110361. [Google Scholar] [CrossRef]
  24. Abualigah, L.; Yousri, D.; Abd Elaziz, M.; Ewees, A.A.; Al-Qaness, M.A.; Gandomi, A.H. Aquila optimizer: A novel meta-heuristic optimization algorithm. Comput. Ind. Eng. 2021, 157, 107250. [Google Scholar] [CrossRef]
  25. AlRassas, A.M.; Al-qaness, M.A.; Ewees, A.A.; Ren, S.; Abd Elaziz, M.; Damaševičius, R.; Krilavičius, T. Optimized ANFIS model using Aquila Optimizer for oil production forecasting. Processes 2021, 9, 1194. [Google Scholar] [CrossRef]
  26. Al-Dhabyani, W.; Gomaa, M.; Khaled, H.; Fahmy, A. Dataset of breast ultrasound images. Data Brief 2020, 28, 104863. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Overall process of AOBNN-BDNN algorithm.
Figure 1. Overall process of AOBNN-BDNN algorithm.
Applsci 12 08679 g001
Figure 2. Flowchart of AO technique.
Figure 2. Flowchart of AO technique.
Applsci 12 08679 g002
Figure 3. Confusion matrices of AOBNN-BDNN approach: (a) entire dataset, (b) 70% of TR data, and (c) 30% of TS data.
Figure 3. Confusion matrices of AOBNN-BDNN approach: (a) entire dataset, (b) 70% of TR data, and (c) 30% of TS data.
Applsci 12 08679 g003
Figure 4. Result analysis of AOBNN-BDNN approach under entire dataset.
Figure 4. Result analysis of AOBNN-BDNN approach under entire dataset.
Applsci 12 08679 g004
Figure 5. Result analysis of AOBNN-BDNN approach under 70% of TR data.
Figure 5. Result analysis of AOBNN-BDNN approach under 70% of TR data.
Applsci 12 08679 g005
Figure 6. Result analysis of AOBNN-BDNN approach under 30% of TS data.
Figure 6. Result analysis of AOBNN-BDNN approach under 30% of TS data.
Applsci 12 08679 g006
Figure 7. TA and VA analysis of AOBNN-BDNN methodology.
Figure 7. TA and VA analysis of AOBNN-BDNN methodology.
Applsci 12 08679 g007
Figure 8. TL and VL analysis of AOBNN-BDNN methodology.
Figure 8. TL and VL analysis of AOBNN-BDNN methodology.
Applsci 12 08679 g008
Figure 9. Precision-recall curve analysis of AOBNN-BDNN methodology.
Figure 9. Precision-recall curve analysis of AOBNN-BDNN methodology.
Applsci 12 08679 g009
Figure 10. ROC curve analysis of AOBNN-BDNN methodology.
Figure 10. ROC curve analysis of AOBNN-BDNN methodology.
Applsci 12 08679 g010
Figure 11. P r e c n and R e c a l analysis of the AOBNN-BDNN approach with existing methodologies.
Figure 11. P r e c n and R e c a l analysis of the AOBNN-BDNN approach with existing methodologies.
Applsci 12 08679 g011
Figure 12. A c c u y and F s c o r e analysis of the AOBNN-BDNN approach with existing methodologies.
Figure 12. A c c u y and F s c o r e analysis of the AOBNN-BDNN approach with existing methodologies.
Applsci 12 08679 g012
Figure 13. CT analysis of AOBNN-BDNN approach with recent algorithms.
Figure 13. CT analysis of AOBNN-BDNN approach with recent algorithms.
Applsci 12 08679 g013
Table 1. Dataset details.
Table 1. Dataset details.
ClassNo. of Images
Benign437
Malignant210
Normal133
Total Number of Images780
Table 2. Result analysis of the AOBNN-BDNN approach with different measures under the entire dataset.
Table 2. Result analysis of the AOBNN-BDNN approach with different measures under the entire dataset.
Entire Dataset
Class LabelsAccuracyPrecisionRecallF-ScoreMCC
Benign98.8599.7798.1798.9697.68
Malignant98.9796.33100.0098.1397.46
Normal99.87100.0099.2599.6299.55
Average99.2398.7099.1498.9198.23
Table 3. Result analysis of the AOBNN-BDNN approach with various measures under 70% of TR data.
Table 3. Result analysis of the AOBNN-BDNN approach with various measures under 70% of TR data.
Training Phase (70%)
Class LabelsAccuracyPrecisionRecallF-ScoreMCC
Benign98.5399.6897.8098.7397.02
Malignant98.7295.07100.0097.4796.67
Normal99.82100.0098.9299.4699.35
Average99.0298.2598.9198.5597.68
Table 4. Result analysis of the AOBNN-BDNN approach with various measures under 30% of TS data.
Table 4. Result analysis of the AOBNN-BDNN approach with various measures under 30% of TS data.
Testing Phase (30%)
Class LabelsAccuracyPrecisionRecallF-ScoreMCC
Benign99.57100.0099.1699.5899.15
Malignant99.5798.68100.0099.3499.03
Normal100.00100.00100.00100.00100.00
Average99.7299.5699.7299.6499.39
Table 5. Comparative analysis of AOBNN-BDNN approach with existing methodologies.
Table 5. Comparative analysis of AOBNN-BDNN approach with existing methodologies.
MethodsPrecisionRecallAccuracyF-Score
AOBNN-BDNN99.5699.7299.7299.64
SEODTL-BDC99.1898.1899.1298.14
ESD model98.8998.1199.0798.88
LSVM model98.4197.8798.7398.29
ESKNN model97.5598.2797.7098.61
FKNN algorithm96.9397.0097.4296.93
LD algorithm97.4198.0597.9296.99
Table 6. Classification time analysis of AOBNN-BDNN approach with recent algorithms.
Table 6. Classification time analysis of AOBNN-BDNN approach with recent algorithms.
MethodsClassifification Time (min)
AOBNN-BDNN1.01
SEODTL-BDC1.18
ESD model2.28
LSVM model1.95
ESKNN model2.80
FKNN algorithm2.24
LD algorithm1.75
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Obayya, M.; Haj Hassine, S.B.; Alazwari, S.; K. Nour, M.; Mohamed, A.; Motwakel, A.; Yaseen, I.; Sarwar Zamani, A.; Abdelmageed, A.A.; Mohammed, G.P. Aquila Optimizer with Bayesian Neural Network for Breast Cancer Detection on Ultrasound Images. Appl. Sci. 2022, 12, 8679. https://doi.org/10.3390/app12178679

AMA Style

Obayya M, Haj Hassine SB, Alazwari S, K. Nour M, Mohamed A, Motwakel A, Yaseen I, Sarwar Zamani A, Abdelmageed AA, Mohammed GP. Aquila Optimizer with Bayesian Neural Network for Breast Cancer Detection on Ultrasound Images. Applied Sciences. 2022; 12(17):8679. https://doi.org/10.3390/app12178679

Chicago/Turabian Style

Obayya, Marwa, Siwar Ben Haj Hassine, Sana Alazwari, Mohamed K. Nour, Abdullah Mohamed, Abdelwahed Motwakel, Ishfaq Yaseen, Abu Sarwar Zamani, Amgad Atta Abdelmageed, and Gouse Pasha Mohammed. 2022. "Aquila Optimizer with Bayesian Neural Network for Breast Cancer Detection on Ultrasound Images" Applied Sciences 12, no. 17: 8679. https://doi.org/10.3390/app12178679

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop