Next Article in Journal
Acknowledgment to the Reviewers of Diagnostics in 2022
Next Article in Special Issue
Accuracy Analysis of Deep Learning Methods in Breast Cancer Classification: A Structured Review
Previous Article in Journal
Ultra-Low Dose CT Chest in Acute COVID-19 Pneumonia: A Pilot Study from India
Previous Article in Special Issue
Feature Transformation for Efficient Blood Glucose Prediction in Type 1 Diabetes Mellitus Patients
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

White Blood Cells Classification Using Entropy-Controlled Deep Features Optimization

1
Department of Computer Science, Iqra University, Islamabad 44800, Pakistan
2
Department of Computer Science, COMSATS University Islamabad, Wah Campus, Wah Cantt 47010, Pakistan
3
Department of Electrical & Computer Engineering, COMSATS University Islamabad, Wah Campus, Wah Cantt 47010, Pakistan
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(3), 352; https://doi.org/10.3390/diagnostics13030352
Submission received: 23 November 2022 / Revised: 13 January 2023 / Accepted: 13 January 2023 / Published: 18 January 2023
(This article belongs to the Special Issue Medical Data Processing and Analysis)

Abstract

:
White blood cells (WBCs) constitute an essential part of the human immune system. The correct identification of WBC subtypes is critical in the diagnosis of leukemia, a kind of blood cancer defined by the aberrant proliferation of malignant leukocytes in the bone marrow. The traditional approach of classifying WBCs, which involves the visual analysis of blood smear images, is labor-intensive and error-prone. Modern approaches based on deep convolutional neural networks provide significant results for this type of image categorization, but have high processing and implementation costs owing to very large feature sets. This paper presents an improved hybrid approach for efficient WBC subtype classification. First, optimum deep features are extracted from enhanced and segmented WBC images using transfer learning on pre-trained deep neural networks, i.e., DenseNet201 and Darknet53. The serially fused feature vector is then filtered using an entropy-controlled marine predator algorithm (ECMPA). This nature-inspired meta-heuristic optimization algorithm selects the most dominant features while discarding the weak ones. The reduced feature vector is classified with multiple baseline classifiers with various kernel settings. The proposed methodology is validated on a public dataset of 5000 synthetic images that correspond to five different subtypes of WBCs. The system achieves an overall average accuracy of 99.9 % with more than 95 % reduction in the size of the feature vector. The feature selection algorithm also demonstrates better convergence performance as compared to classical meta-heuristic algorithms. The proposed method also demonstrates a comparable performance with several existing works on WBC classification.

1. Introduction

Blood is a crucial fluid in the human body that is essential for life. Human blood is made up of plasma and blood cells. Plasma is the yellowish liquid component of blood that is largely water and accounts for 55 % of blood volume [1]. The blood also includes proteins, carbohydrates, minerals, hormones, carbon dioxide, and blood cells. Red blood cells (RBCs), white blood cells (WBCs), and platelets (thrombocytes) are the three different types of cellular components found in the blood, each distinguished by their color, texture, and appearance. RBCs, also known as erythrocytes, carry hemoglobin, an iron-containing protein that aids in the delivery of oxygen from the lungs to the tissues. WBCs, also known as leukocytes, are an essential component of the human immune system, assisting the body in fighting infectious diseases and foreign substances [2,3]. Figure 1 demonstrates a classification of WBCs on the basis of their structure. WBCs are primarily of two types, i.e., granulocytes and agranulocytes. The granulocytes have their origin in the bone marrow and are present within the cytoplasm in the form of granules of protein. There are three types of granulocyte cells, namely basophils, eosinophils, and neutrophils. Agranulocytes, which are defined as cells without granules in their cytoplasm, are further divided into two types, i.e., lymphocytes and monocytes [4]. Each type of cell has a unique role in the immune system of the body. For example, neutrophils act as scavengers that surround and destroy bacteria and fungi present in the body. Eosinophils play a role in the general immune and inflammatory responses of the body. An increased level of basophils results in a blood disorder after an allergic reaction. Monocytes fight against infections, remove dead or damaged tissues, and kill cancerous cells; lymphocytes combat bacteria, viruses, and other cells that pose a threat to the body’s ability to function [5]. A detailed analysis of WBCs is very important to assess the overall condition of the human immune system. In particular, WBC analysis is crucial in the diagnosis of leukemia, a type of blood cancer that occurs due to the excessive production of malignant WBCs in the bone marrow. Leukemia diagnosis is performed by one of three main clinical tests, i.e., physical test, complete blood count (CBC) test, and bone marrow test. The first step of CBC is to determine different types of WBCs from the blood samples. This task is mainly performed by hematologists through the visual examination of microscopic images of blood smears. This manual method is labor-intensive, time-consuming, and prone to inaccuracy due to judgment errors influenced by several external factors.
With the recent advancement in digital image processing technology, the automated classification of WBCs using computer vision techniques has attracted significant research interest. However, due to morphological overlap between different subclasses and their structural irregularities, the machine learning-based classification and localization of WBCs is challenging. Deep learning with convolutional neural networks (CNNs) is the most promising method for classification and detection tasks in the field of contemporary medical imaging [6,7]. Despite the fact that CNNs perform best on large datasets, training them takes a lot of data and computational power. The dataset is frequently small and may not be sufficient to train a CNN from scratch. In such a case, transfer learning is frequently used to maximize the effectiveness of CNNs while also decreasing the computational costs [8]. In this approach, the CNN is initially pre-trained on a large dataset consisting of a diverse range of classes and then applied to a specific task [9]. There are various pre-trained neural networks that have won international contests, including VGGNet [10], Resnet [11], Darknet [12], Densenet [13], Mobilenet [14], Inception [15], Xception, [16] etc. Through their capacity for self-learning, these models are able to extract a rich set of features from images that contain substantial semantic information. This helps to achieve a significant level of accuracy for a variety of image classification scenarios. In modern deep learning applications, feature selection is a crucial step which reduces the difficulty of model learning by removing irrelevant or redundant features. The present research is focused on achieving a high level of accuracy with a smaller feature set to reduce the computation costs and memory requirements of expert systems.
The existing works on WBC classification are broadly classified into two categories, i.e., (a) classical methods and (b) deep learning methods. The classical methods consist of approaches which propose efficient preprocessing techniques to extract strong features from WBC images and classify them using baseline classifiers. Some remarkable works in this domain are discussed as follows. In [17], the authors proposed a method which selects the eigenvectors from color images of blood cells based on the minimization of similarities. The Bayesian classifier is then used to classify the eigen cells on the basis of density and color information. In [18], Fuzzy C-means clustering is applied to separate the nucleus and cytoplasm of leukocytes. Then, various geometric, color, and statistical properties are extracted and classified by support vector machines (SVMs). In [19], an image segmentation method is proposed based on mean-shift clustering and boundary removal rules with a gradient vector flow. An ensemble of features is extracted from the segmented nucleus and cytoplasm, which is then classified using a random forest algorithm. In [20], the authors tested the performance of six different machine learning algorithms on 35 different geometric and statistical features. The multinomial logistic regression algorithm outperformed other methods. A stepwise linear discriminant analysis method is proposed in [21], which extracts specific features from blood structure images and classifies them using reversion values such as partial F values. In [22], the authors presented a WBC cancer detection method which combines various morphological, clustering, and image pre-processing steps with random forest classifier. The suggested method uses a decision tree learning method, which uses predictors at each node to make better decisions, in order to categorize various types of cancer.
The second category of works is based on deep learning approaches for WBC classification. The works in this category primarily employ transfer learning of a pretrained deep neural network for feature extraction or classification. Some important works are discussed as follows. In [23], the authors proposed a deep learning method that uses the DenseNet121 [13] model to classify WBC subtypes. The model is optimized with the preprocessing techniques of normalization and data augmentation applied to a Kaggle dataset. The work in [24] first applies a thresholding-based segmentation on the WBC images. Feature extraction from segmented images is performed using VGG16 CNN [10] model learning. The extracted feature vectors are classified using the K-nearest neighbor (KNN) algorithm. In [25], the authors investigated generative adversarial networks (GANs) for data augmentation and employed the DenseNet169 [13] network for WBC classification. In [26], the authors applied Gaussian and median filtering before training the images using multiple deep neural networks. The authors in [27] applied a you-only-look-once (YOLO) algorithm for the detection of blood cells from a smear images. In [28], two techniques are proposed for blood cell identification, namely single-shot multibox detector and an incrementally improved version of YOLO.
Although modern approaches based on transfer learning on deep CNN models achieve a decent level of accuracy for a variety of classification tasks, they all share the use of a large number of features extracted from deep neural networks. This suffers from high computational cost and memory requirements for practical deployment. In most biomedical scenarios, many of these deep features are redundant or contain zeros. Effective dimensionality reduction, or choosing only powerful, discriminant features, increases classifier accuracy while decreasing computational time and expense. WBC classification using deep feature selection is an emerging research area. Few works have reported population-based meta-heuristics for deep feature selection. The authors of [29] have proposed a leukemia detection system in which various features, such as color, texture, shape, and hybrid features, are first extracted from WBC images and then an optimization algorithm inspired by social spiders is used to select the most useful features. In [30], a leukemia detection approach is proposed which combines deep feature extraction using VGGNet and a statistically enhanced salp swarm algorithm for feature selection. Furthermore, the classification of reduced feature vectors was performed using a baseline classifier. The work in [31] proposes a self-designed neural network named W-Net to classify five subtypes of WBCs. The authors also generated a synthetic WBC dataset using a generative adversarial network (GAN).
In this study, we have proposed a hybrid approach for WBCs classification. The proposed approach first creates an ensemble of deep features extracted by applying transfer learning of multiple deep CNNs on WBC images and then performs feature selection using an entropy-controlled nature-inspired algorithm. The main contributions of this work can be summarized in the following steps.
  • Using a synthetic, real-world, large-scale dataset of five WBC types, transfer learning is performed using two deep CNNs, namely Darknet53 and Densenet201, followed by their feature fusion;
  • For feature selection, a nature-inspired meta-heuristic named entropy-controlled marine predators algorithm (ECMPA) is proposed. The proposed algorithm effectively selects only the most dominant features;
  • The reduced feature set is classified using various baseline classifiers with multiple kernel settings;
  • The proposed feature selection algorithm demonstrates a high accuracy with significant reduction in feature size. The algorithm also achieves a better convergence rate as compared to classical population-based selection methods.
The main focus of our manuscript is to present a novel method of deep-feature selection using an entropy-controlled population-based algorithm and show its effectiveness in the domain of WBCs classification for leukemia detection. Since the definition of appropriate image features is a very difficult task due to the morphological similarity of images and subject variability, WBC classification is a pertinent design case for such an approach. The rest of this paper is organized as follows. Section 2 discusses all steps of the proposed WBC classification pipeline, Section 3 presents the results and analysis, and Section 4 concludes the paper.

2. Materials and Methods

This section provides a description of all steps of the proposed WBC classification system which are discussed in the following subsections.

2.1. Dataset Description

This work uses the public dataset in [31], which was generated synthetically from a real-world dataset [32] of five sub-types of WBC images, namely neutrophil, eosinophil, basophil, lymphocyte, and monocyte. The synthetic dataset was generated with the help of a deep convolutional generative adversarial network (DCGAN) trained on the original real-world dataset [32] of blood smear images, captured by a Sysmex DI-69 machine and provided by the Catholic University of Korea. The synthetic dataset is composed of 5000 images each of size ( 128 × 128 × 3 ), with 1000 images belonging for each class. Figure 2 shows samples belonging to all classes of the dataset of [32] used in this work.

2.2. WBCs Classification Pipeline

Figure 3 shows the proposed WBCs classification pipeline, whose main computation steps are discussed as follows.

2.2.1. Preprocessing

Image contrast enhancement is a fundamental pre-processing step of many digital image-processing applications. In this work, the input image enhancement is performed with the help of color histogram equalization. The classical method of histogram equalization is applied to grayscale images and performs as redistribution of their intensity. In case of color images, performing histogram equalization on R, G, and B components independently will not necessarily enhance the image. Color histogram equalization can be achieved by converting a color image into a HSV/HSI image and enhancing the intensity while preserving hue and saturation components. The main steps of color histogram equalization are as follows.
  • Convert the RGB image into HSI image;
  • Obtain the intensity matrix from the HSI image matrix;
  • Perform histogram equalization on the intensity matrix;
  • Replace the intensity matrix of the HSI image with the histogram-equalized intensity matrix;
  • Convert HSI image back to RGB image.

2.2.2. Feature Extraction Using Transfer Learning

The pre-processed image dataset is now subjected to the feature extraction using the transfer learning step. In this work, we performed transfer learning on two well-known deep CNNs, namely DarkNet53 and DenseNet201, which are discussed as follows.
DarkNet53 is a convolutional neural network proposed as a feature extractor in YOLO3 image detection workflow [12]. It is pretrained on more than a million images from ImageNet database [33]. The pretrained network is able to classify up to 1000 categories of image objects. Details about the various layers in the DarkNet CNN architecture are shown in Table 1. The network has an input layer with a size of 256 × 256 × 3 and is primarily made up of convolution layers with sizes of 1 × 1 and 3 × 3 , totaling 53 layers, including the final fully connected layer but excluding the residual layer. Each convolutional layer is composed of a Conv2d layer followed by a batch normalization (BN) [34] and LeakyReLU [11] layer. The residual layer is added to solve the gradient disappearance or gradient explosion problems in the network [12]. In Darknet53, a significant reduction in parameters is achieved as compared to its previous version, i.e., Darknet19.
In order to perform transfer learning of Darknet53, the last learnable layer of Darknet53, i.e., “Conv5”, is replaced with a new fully connected layer with the number of outputs equal to the number of classes in our WBCs dataset (5 classes). A new softmax layer is created and replaced with the original softmax layer of the network. Similarly, the classification layer of the network is replaced with a new classification layer without class labels. To perform the network training, first the dataset images are resized to 256 × 256 × 3 using the nearest neighbor interpolation method, followed by image augmentation which performs random rotation of images in the range of [ 0 , 360 ] degrees and scaling with a factor in the range of [ 0.5 , 1 ] . The deep features are extracted from the “GlobalAvgPool” layer. The DarkNet53 returns a deep feature vector of size 1024 per image.
DenseNet201. This deep convolutional neural network is 201 layers deep [13]. It is also pre-trained on Imagenet [33] dataset. DenseNet is designed to overcome the vanishing gradient problem in high-level neural networks. In DenseNet, each layer receives new inputs from all preceding levels and passes on its own feature maps to all following layers. Concatenation is utilized. Each layer receives “collective knowledge” from all preceding levels. This results in a thinner and compact network that achieves a high computational efficiency and memory saving. Table 2 shows the layer details of DenseNet201.
In order to perform transfer learning using DenseNet201, the last learnable layer of the network, i.e., “fc1000” is replaced with a new fully connected layer with 5 classes of the WBCs dataset used in this work. A new softmax layer is created and replaced with the original softmax layer of the network. Similarly, the classification layer of the network is replaced with a new classification layer without class labels. The dataset images are first resized to 224 × 224 and augmented in a way similar to DarkNet53. From the trained network, the deep features are extracted from the global average pooling layer. DenseNet201 returns a deep feature vector of size 1920 per image. There was no layer freezing carried out during the training process. As a result, the highest possible numbers of trainable parameters, i.e., 18.1 million for DenseNet201 and 41.6 million for DarkNet53, were considered.

2.2.3. Feature Fusion

The deep features extracted from both the Darknet53 and Densenet201 networks discussed above are concatenated together to form a fused feature vector. Let X and Y be the feature vectors of Darknet53 and Densenet201, respectively, and the fused feature vector Z is given as
Z = [ X Y ]
The total size of fused feature vector Z is 2944 features per image.

2.2.4. Feature Selection Using Marine Predators Algorithm

Feature selection is an important step which significantly alleviates the curse of dimensionality by reducing the size of the feature vector, selecting only relevant features. The classical feature selection methods based on search algorithms such as complete and sequential search suffer from high computational cost. Population-based meta-heuristic algorithms have been demonstrated as an effective way to solve complex optimization problems [35,36].
The marine predators algorithm (MPA) is a meta-heuristic algorithm that draws inspiration from nature and models the foraging behavior of marine predators (MPs) to find their prey [37]. In the aquatic environment, both the prey and the predator are looking for food at the same time. Position updates for the predator and prey are based on Brownian or Lévy movement, depending on the relative velocities of the two. The goal of MPA, like other swarm optimization techniques, is to choose the best solution (elite) from the population of predators. The MP with the strongest foraging capacity is called elite predator. The MPA is based on the observation that MPs move in Lévy patterns when there are few prey items present and in Brownian patterns when there are many prey items present. Additionally, predators alter their behavior and move to areas with different prey concentrations in the presence of environmental effects. As a result, there are three phases to the position updates in MPA optimization based on the relative velocities of predator and prey: low velocity ratio, unit velocity ratio, and high velocity ratio. The velocity ratio vs. is defined as the ratio between the velocity of prey and predator.
  • In low velocity ratio ( v < = 0.1 ) , the most suitable movement strategy for the MP is Lévy, whereas the prey moves in Brownian or Lévy movement;
  • In unit velocity ratio ( v < = 1 ) , if the prey moves in Lévy, the most suitable movement for MP is Brownian;
  • In high velocity ratio ( v > 1 ) , the best strategy for a predator is not moving at all. In this case, either prey is moving Brownian or Lévy.
Standard MPA Methodology. The standard MPA is an iterative, population-based optimization algorithm. The first step is to generate an initial population of solutions. The population matrix of size n × d is generated as follows:
P = X 1 , 1 X 1 , 2 X 1 , d X 2 , 1 X 2 , 2 X 2 , d X n , 1 X n , 2 X n , d
where n is the size of population, i.e., number of search agents (each predator and prey are searching for food and can be considered as a search agent), and d is the dimension (no. of variables) of each agent. Each variable of initial solution is uniformly distributed over the search space computed as
X i = l b + r a n d × ( u b l b )
where l b denotes the lower bound, u b denotes the upper bound, and r a n d is a uniformly distributed random number from the interval [ 0 , 1 ] . Based on the concept of survival of the fittest, the top predators have the best foraging capabilities. Therefore, the fittest solution is nominated as the best predator and used to construct a matrix called Elite. In an iteration I, the E l i t e matrix is constructed as
E l i t e = X 1 , 1 I X 1 , 2 I X 1 , d I X 2 , 1 I X 2 , 2 I X 2 , d I X n , 1 I X n , 2 I X n , d I = X ¯ X ¯ X ¯
where X ¯ is the top predator vector which is replicated n times to construct the E l i t e matrix. At the end of each iteration, the E l i t e matrix will be updated if the fittest predator of a population is replaced by another better predator. Another matrix named P r e y is generated with the same dimensions as E l i t e . The Prey matrix is computed as
P r e y = X 1 , 1 X 1 , 2 X 1 , d X 2 , 1 X 2 , 2 X 2 , d X n , 1 X n , 2 X n , d
where X i , j denotes the jth dimension of ith prey. During first iteration, the Prey matrix is equal to randomly generated population matrix P. In all subsequent iterations, the Prey is updated and its values are used to compute the E l i t e matrix. The update of the P r e y matrix is carried separately in three phases of MPA optimization. These phases include:
Phase 1: This phase corresponds to the high velocity ratio and happens in the first ( 1 3 ) r d of maximum iterations of algorithm where exploration is more significant. The update rule of this phase is given as:
S t e p s i z e i ¯ = R B ¯ ( E l i t e i ¯ R B ¯ P r e y i ¯ ) , i = 1 , , n
P r e y i ¯ = P r e y i ¯ + P . R ¯ S t e p s i z e i ¯ , i = 1 , , n
where P r e y i ¯ is a vector of Prey matrix, R B ¯ and R ¯ are vectors of dimensions d containing random numbers from Normal and Uniform distribution, respectively, P is constant value equal to 0.5 , and ⨂ shows element-wise multiplication.
Phase 2: This phase corresponds to unit velocity ratio when predator and prey are moving at the same pace. This is the phase which occurs for intermediate ( 1 3 ) r d of iterations, where exploration and exploitation matters. The update rules for this phase are given as follows:
S t e p s i z e i ¯ = R L ¯ ( E l i t e i ¯ R L ¯ P r e y i ¯ ) , i = 1 , , n 2
P r e y i ¯ = P r e y i ¯ + P R ¯ S t e p s i z e i ¯ , i = 1 , n 2
S t e p s i z e i ¯ = R B ¯ ( E l i t e i ¯ R B ¯ P r e y i ¯ ) , i = n 2 + 1 , , n
P r e y i ¯ = E l i t e i ¯ + P . C F S t e p s i z e i ¯ , i = n 2 + 1 , n
 
where R L ¯ is vector of size d containing random numbers based on Lévy distribution, C F = ( 1 I I M a x ) ( 2 I I M a x ) is an adaptive parameter to control the step size for predator movement, I is the current iteration, and I M a x is the maximum number of iterations.
Phase 3: This phase corresponds to low velocity ratio when predator is moving faster than prey. This scenario happens in the last ( 1 3 ) r d iterations of the optimization phase where exploitation matters. The update rules for this phase are given as follows:
S t e p s i z e i ¯ = R L ¯ ( E l i t e i ¯ P r e y i ¯ ) , i = 1 , , n
P r e y i ¯ = E l i t e i ¯ + P . C F S t e p s i z e i ¯ , i = 1 , n
The next step is to model the behavioral change in MPs as a result of environmental effects. These effects are known as fish aggregating devices (FADs). The FADs are known as local optima; therefore, the prey and predators must perform longer jumps during simulation to avoid stagnation in local optima. The update of P r e y matrix considering the FAD effect is mathematically represented as follows:
P r e y i ¯ = P r e y i ¯ + C F [ X m i n ¯ + R ¯ ( X m a x ¯ X m i n ¯ ) ] U ¯ r F A D s P r e y i ¯ + [ F A D s ( 1 r ) + r ] ( P r e y r 1 ¯ P r e y r 2 ¯ ) r > F A D s )
where F A D s = 0.2 is the probability of occurrence of FAD effects, U ¯ is a randomly generated binary vector, r is the uniform number in [ 0 , 1 ] , X m a x and X m i n are the vectors containing lower and upper bounds of dimensions, respectively, and r 1 , r 2 are the random indices of the P r e y matrix.
After the P r e y matrix is updated using Equations (6)–(13), and incorporating the FAD effects of Equation (14), this matrix is evaluated for fitness function. The fitness of each solution of current iteration is compared to its equivalent solution in prior iteration. If the current solution is more fitted, it replaces the previous one. In the next iteration, the best solution of P r e y is used to generate the E l i t e matrix and update the P r e y matrix using Equations (6)–(13).

2.2.5. Entropy-Controlled MPA for Feature Selection

In this work, we proposed a multi-level feature selection algorithm named entropy-controlled marine predators algorithm (ECMPA). The proposed technique is based on two stages of feature selection, the first of which corresponds to feature reduction based on the entropy of the fused feature vector, followed by additional feature reduction based on the MPA. The main computational steps of ECMPA are discussed in Algorithm 1.
Notations: In Algorithm 1, matrices and vectors are represented as double struck characters (e.g., F ) and scalars are represented as normal letters.
The algorithm receives as inputs the fused feature matrix F , the label vector L , the entropy-controlled feature reduction parameter e c , the maximum number of iterations I M a x , and the population size N. Other constant parameters specific to MPA are upper bound u b , lower bound l b , threshold t, constant P, Levy coefficient β , and fish aggregating devices effect F A D S . The matrix F is of dimensions n t × D , where n t is the number of training images and D is the number of features extracted from the feature fusion step. The first level of feature extraction, i.e., entropy-based, is performed by steps 4–7. Step 4 computes the entropy of each column of F and returns a vector E of size 1 × D . Step 5 sorts E in descending order, thus returning the sorted vector E 2 along with indices stored in I . Step 6 extracts the indices of the first e c percentage of features with maximum entropy. Step 7 extracts features in these indices from F and stores them in F 2 . The task for generating initial population matrix P using Equation (1) is performed in Steps 11–15. Step 13 corresponds to Equation (2), where r a n d ( ) computes a random number from uniform distribution from interval [ 0 , 1 ] . In Steps 19–26, the fitness function of each individual of P is computed using C o s t F u n c t i o n and stored in fitness vector F i t . The best fitness is stored in f i t g and best individual is stored in x g b . The E l i t e matrix E is computed in Step 28, by performing the R e p e a t function which concatenates N copies of x g b along the first dimension (i.e., column-wise). Steps 31–37 perform update of P r e y matrix P according to Phase 1 (Equations (5) and (6)). In Step 32, the function r a n d n ( 1 , D ) returns a vector of size 1 × D , containing random numbers from Normal distribution. The Phase 2 of MPA according to update rules of Equations (7)–(11) is performed by Steps 39–57. In Step 43, L e v y ( β , D ) generates a vector of size D containing random numbers from Lévy Distribution. Steps 59-66 perform the P matrix update according to Equations (12) and (13) of Phase 3. In Steps 68–83, the FAD effects are added to the P r e y according to update rules of Equation (14). In each iteration, Steps 85–92 correspond to the memory-saving step where the updated P r e y matrix P is evaluated for C o s t F u n c t i o n and best individual of X g b is obtained. Steps 95–97 correspond to the output step where S F , i.e., a binary vector of size D is obtained. The indices of non-zero entries of S F correspond to the indices of selected features of the fused feature vector.
In Steps 100–115 the execution of C o s t F u n c t i o n is demonstrated, which performs the computation of fitness value of each individual solution. The function accepts as inputs the entropy-reduced feature vector F 2 , the label vector L , and a binary vector a computed by comparing the entries of ith solution of P with the threshold t. The Step 104 extracts all the features of F 2 whose indices correspond to non-zero values of a , and then the function p a r t i t i o n splits the feature matrix F 2 and label vector L into training feature set F t r a i n , testing feature set F t e s t , training label set L t r a i n , and testing label set L t e s t with a splitting ratio h o . In the subsequent steps of the function, model training and prediction is performed using KNN classifier and fitness value (cost) is computed using the classification error metric. An individual with a smaller c o s t value is the fitter individual.
Algorithm 1: ECMPA for feature selection.
Diagnostics 13 00352 i001

2.2.6. Classification

The indices of the reduced feature set extracted from the ECMPA step discussed above are used to perform training and validation using baseline classifiers with multiple feature settings. In this work, we used KNN [38] and SVM [39] classifiers with multiple kernel settings.

3. Results and Discussion

The proposed system was implemented in Matlab 2021 on Windows 10 64-bit using a Core i5, 2.5 GHz CPU, and 8 GB of RAM. The dataset of 5000 thousand images was split into training and testing with an 80 % splitting ratio in order to perform transfer learning using DenseNet201 and DarkNet53 deep models. Table 3 demonstrates the main model training parameters for deep transfer learning. For both networks, a significant level of training and validation accuracy was achieved with five epochs. Figure 4 shows the accuracy and loss function plots for DenseNet201. In order to extract the indices of the most dominant features, the fused feature vector of size 2944 features was then subjected to feature reduction using ECMPA. The reduced set of features is then used to train the KNN and SVM classifiers with multiple kernel settings. In order to perform the classification task, testing images are applied to the trained deep models and a fused feature vector is obtained. The reduced feature vector is generated by using the indices obtained by the ECMPA. This is then classified using the trained KNN and SVM classifiers. Figure 5 demonstrates a set of reduced features extracted from the ECMPA step. Figure 6 demonstrates the results of the proposed WBCs classification system with various kernels of SVM and KNN classifiers. The SVM classifier achieves a 99.9 % accuracy with a reduced feature set consisting of 70 strong features. The confusion matrix of SVM with the quadratic kernel is also demonstrated. The high value of true positive rate (TPR) and low value of false negative rate (FNR) are achieved for all image classes.
In Figure 7, the convergence of the proposed ECMPA is compared with a classical population-based meta-heuristic algorithm, i.e., genetic algorithm (GA). The graph demonstrates that ECMPA achieves a better value of cost function with a smaller number of iterations.
Table 4 shows an accuracy comparison of the proposed approach with some recent works on WBC classification using deep learning networks that use similar datasets. The proposed method shows a comparable or even better accuracy performance with a smaller number of features as compared to other works. This demonstrates the validity of the proposed approach.

Statistical Significance

Obtaining a certain level of confidence in the proposed strategy is the main goal of this statistical investigation. We use the analysis of variance (ANOVA) [44] to compare the means of several distributions in order to determine whether the results are statistically significant. We consider classification accuracy as a performance characteristic for our proposed framework. In order to implement ANOVA, we performed a series of tests to validate the assumption of normality using the Shapiro–Wilk test [45], and homogeneity of variance using the Bartlett’s test [44]. In these testing procedures, we used 1 % level of significance (i.e., α = 0.01 ). The means of classification accuracy values for the selected classifiers, i.e., SVM, KNN, and NNN, as μ 1 , μ 2 , and μ 3 , respectively. For each of the tests mentioned above, the null hypotheses are considered true if the computed Shapiro–Wilk p-values are less than or equal to α ; otherwise, the alternative hypotheses are affirmed as true. The p-values of SVM, KNN, and NNN classifiers are obtained as p 1 = 0.784 , p 2 = 0.7124 , and p 3 = 0.7031 , respectively. The Chi-squared probability from the Bartlett’s test is p c h = 0.85 . From these p-values, we fail to renounce the null hypotheses and confidently claim that our accuracy data are normally distributed with homogeneous variances.
Table 5 presents the statistical results obtained from the ANOVA test including the sum of squared deviation (SS), degree of freedom (df), F-statistics, mean squared error (MSE), and p-value. The obtained p-value is 0.705 , which is greater than α and leads to the conclusion that the means of three classifiers are identical.
Figure 8 shows the confidence interval plots of accuracy values of the three selected classifiers. In the figure, red bars present the average accuracy, whereas the black bars present the 99% confidence limits of each classifier. In addition, the blue bars show lower and upper quantile points obtained by performing the above-mentioned statistical tests. From the figure, we can observe that the SVM classifier achieves a higher average accuracy with relatively smaller confidence interval size as compared to other classifiers. The quantile points of each classifier lie within their respective confidence limits. The higher p-values resulting from these quantile points lead to the acceptance of null hypotheses, which means significant differences in the accuracy distribution of the classifiers.

4. Conclusions

WBCs classification is a vital step in the correct diagnosis of Leukemia. The existing manual methods of WBC classification are labor-intensive and error-prone. Automated WBC classification using computer vision techniques is an emerging paradigm. Modern approaches using deep neural networks achieve a significant level of accuracy for a variety of tasks. However, these neural networks suffer from exorbitant computational complexity, processing power, and memory requirement owing to very large feature sets. Therefore, an efficient feature reduction is essential to make deep neural networks feasible for real-time biomedical applications. This work proposes a complete WBCs classification pipeline that performs transfer learning using deep neural networks followed by an efficient feature reduction algorithm. The proposed feature reduction method is validated using several baseline classifiers with multiple kernel settings. An accuracy of 99.9 % is achieved with a feature reduction of 95 % , which demonstrates the feasibility of the proposed WBCs classification method. While the proposed approach has been applied to an augmented clean dataset containing only WBC subtype images, the ECMPA feature selection algorithm can be applied in any blood cell classification setup with little tuning of parameters. In the future, we plan to extend this work to a more challenging dataset for clinical-grade classification of other cell entities such as platelets and red blood cells, among others. The proposed algorithm can also be tested on the bench mark datasets for other diseases such as skin lesion and brain tumors, among others. In order to address the “curse of dimensionality”, other similar bio-inspired meta-heuristics can be investigated to obtain a trade-off between classification accuracy and computational complexity.

Author Contributions

Conceptualization, R.A. and M.A.; methodology, R.A. and M.A.; software, R.A., M.A. and T.A.; validation, M.A., R.A. and T.A.; resources, N.K.; writing—original draft preparation, R.A. and M.A.; writing—review and editing, M.A., T.A. and R.A. visualization, N.K.; supervision, N.K.; project administration, N.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research recieved no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data for this work shall be available upon request.

Conflicts of Interest

The authors declare no conflict of interest for this work.

Abbreviations

The following abbreviations are used in this manuscript:
WBCWhite blood cell
CNNConvolutional neural network
DNNDeep neural network
SVMsSupport vector machines
MPAMarine predators algorithm
ECMPAEntropy-controlled marine predators algorithm
KNNK-nearest neighbors
TPRTrue positive rate
FNRFalse negative rate
GANGenerative adversarial network
GVFGradient vector flow

References

  1. Kuan, D.H.; Wu, C.C.; Su, W.Y.; Huang, N.T. A microfluidic device for simultaneous extraction of plasma, red blood cells, and on-chip white blood cell trapping. Sci. Rep. 2018, 8, 15345. [Google Scholar] [CrossRef] [Green Version]
  2. Farag, M.R.; Alagawany, M. Erythrocytes as a biological model for screening of xenobiotics toxicity. Chem. Biol. Interact. 2018, 279, 73–83. [Google Scholar] [CrossRef] [PubMed]
  3. Rezatofighi, S.H.; Soltanian-Zadeh, H. Automatic recognition of five types of white blood cells in peripheral blood. Comput. Med. Imaging Graph. 2011, 35, 333–343. [Google Scholar] [CrossRef]
  4. Weatherspoon, D. What to Know about White Blood Cells. Available online: https://www.medicalnewstoday.com/articles/327446#types-and-function (accessed on 22 November 2022).
  5. Mathur, A.; Tripathi, A.S.; Kuse, M. Scalable system for classification of white blood cells from Leishman stained blood stain images. J. Pathol. Inform. 2013, 4, 15. [Google Scholar] [CrossRef]
  6. Khan, M.A.; Sharif, M.; Akram, T.; Damaševičius, R.; Maskeliūnas, R. Skin Lesion Segmentation and Multiclass Classification Using Deep Learning Features and Improved Moth Flame Optimization. Diagnostics 2021, 11, 811. [Google Scholar] [CrossRef] [PubMed]
  7. Bengio, Y.; Courville, A.; Vincent, P. Representation learning: A review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1798–1828. [Google Scholar] [CrossRef] [Green Version]
  8. Donahue, J.; Jia, Y.; Vinyals, O.; Hoffman, J.; Zhang, N.; Tzeng, E.; Darrell, T. Decaf: A deep convolutional activation feature for generic visual recognition. In Proceedings of the International Conference on Machine Learning, PMLR, Bejing, China, 21–26 June 2014; pp. 647–655. [Google Scholar]
  9. Nguyen, L.D.; Lin, D.; Lin, Z.; Cao, J. Deep CNNs for microscopic image classification by exploiting transfer learning and feature concatenation. In Proceedings of the 2018 IEEE International Symposium on Circuits and Systems (ISCAS), Florence, Italy, 27–30 May 2018; pp. 1–5. [Google Scholar]
  10. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  11. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  12. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  13. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  14. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  15. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  16. Shin, H.C.; Roth, H.R.; Gao, M.; Lu, L.; Xu, Z.; Nogues, I.; Yao, J.; Mollura, D.; Summers, R.M. Deep convolutional neural networks for computer-aided detection: CNN architectures, dataset characteristics and transfer learning. IEEE Trans. Med. Imaging 2016, 35, 1285–1298. [Google Scholar] [CrossRef] [Green Version]
  17. Sanei, S.; Lee, T.K. T: Cell recognition based on pca and bayesian classification. In Proceedings of the 4th International Symposium, ICA 2003, Nara, Japan, 1–4 April 2003. [Google Scholar]
  18. Sarrafzadeh, O.; Rabbani, H.; Talebi, A.; Banaem, H.U. Selection of the best features for leukocytes classification in blood smear microscopic images. In Proceedings of the Medical Imaging 2014: Digital Pathology, San Diego, CA, USA, 15–20 February 2014; Volume 9041, pp. 159–166. [Google Scholar]
  19. Ko, B.; Gim, J.; Nam, J. Cell image classification based on ensemble features and random forest. Electron. Lett. 2011, 47, 638–639. [Google Scholar] [CrossRef]
  20. Abdullah, E.; Turan, M.K. Classifying white blood cells using machine learning algorithms. Int. J. Eng. Res. Dev. 2019, 11, 141–152. [Google Scholar]
  21. Alruwaili, M. An intelligent medical imaging approach for various blood structure classifications. Complexity 2021, 2021, 5573300. [Google Scholar] [CrossRef]
  22. Nithyaa, A.N.; Kumar, R.P.; Gokul, M.; Aananthi, C.G. Matlab Based Potent Algorithm for Wbc Cancer Detection and Classification. Biomed. Pharmacol. J. 2021, 14, 2277–2284. [Google Scholar] [CrossRef]
  23. Sharma, S.; Gupta, S.; Gupta, D.; Juneja, S.; Gupta, P.; Dhiman, G.; Kautish, S. Deep learning model for the automatic classification of white blood cells. Comput. Intell. Neurosci. 2022, 2022, 7384131. [Google Scholar] [CrossRef]
  24. Baby, D.; Devaraj, S.J.; Anishin Raj, M.M. Leukocyte classification based on transfer learning of VGG16 features by K-nearest neighbor classifier. In Proceedings of the IEEE 2021 3rd International Conference on Signal Processing and Communication (ICPSC), Coimbatore, India, 13–14 May 2021; pp. 252–256. [Google Scholar]
  25. Almezhghwi, K.; Serte, S. Improved classification of white blood cells with the generative adversarial network and deep convolutional neural network. Comput. Intell. Neurosci. 2020, 2020, 6490479. [Google Scholar] [CrossRef]
  26. Yildirim, M.; Çinar, A. Classification of White Blood Cells by Deep Learning Methods for Diagnosing Disease. Rev. d’Intell. Artif. 2019, 33, 335–340. [Google Scholar] [CrossRef]
  27. Alam, M.M.; Islam, M.T. Machine learning approach of automatic identification and counting of blood cells. Healthc. Technol. Lett. 2019, 6, 103–108. [Google Scholar] [CrossRef]
  28. Wang, Q.; Bi, S.; Sun, M.; Wang, Y.; Wang, D.; Yang, S. Deep learning approach to peripheral leukocyte recognition. PLoS ONE 2019, 14, e0218808. [Google Scholar] [CrossRef] [Green Version]
  29. Sahlol, A.T.; Abdeldaim, A.M.; Hassanien, A.E. Automatic acute lymphoblastic leukemia classification model using social spider optimization algorithm. Soft Comput. 2019, 23, 6345–6360. [Google Scholar] [CrossRef]
  30. Sahlol, A.T.; Kollmannsberger, P.; Ewees, A.A. Efficient classification of white blood cell leukemia with improved swarm optimization of deep features. Sci. Rep. 2020, 10, 2536. [Google Scholar] [CrossRef] [Green Version]
  31. Jung, C.; Abuhamad, M.; Mohaisen, D.; Han, K.; Nyang, D. WBC image classification and generative models based on convolutional neural network. BMC Med. Imaging 2022, 22, 94. [Google Scholar] [CrossRef] [PubMed]
  32. The Catholic University of Korea Institutional Review Board. 2019. Available online: https://bit.ly/2YrlQPl (accessed on 22 November 2022).
  33. ImageNet. Available online: http://www.image-net.org (accessed on 22 November 2022).
  34. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, PMLR, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  35. Awais, M.; Ahmed, A.; Naeem, M.; Iqbal, M.; Qadri, N.; Anpalagan, A. Multiple line outages identification: A customized quantum inspired approach. Electr. Power Syst. Res. 2016, 134, 47–55. [Google Scholar] [CrossRef]
  36. Ahmed, A.; Khan, Q.; Naeem, M.; Iqbal, M.; Anpalagan, A.; Awais, M. An insight to the performance of estimation of distribution algorithm for multiple line outage identification. Swarm Evol. Comput. 2018, 39, 114–122. [Google Scholar] [CrossRef]
  37. Faramarzi, A.; Heidarinejad, M.; Mirjalili, S.; Gandomi, A.H. Marine Predators Algorithm: A nature-inspired metaheuristic. Expert Syst. Appl. 2020, 152, 113377. [Google Scholar] [CrossRef]
  38. Kafeel, A.; Aziz, S.; Awais, M.; Khan, M.A.; Afaq, K.; Idris, S.A.; Alshazly, H.; Mostafa, S.M. An Expert System for Rotating Machine Fault Detection Using Vibration Signal Analysis. Sensors 2021, 21, 7587. [Google Scholar] [CrossRef] [PubMed]
  39. Khan, S.A.; Nazir, M.; Khan, M.A.; Saba, T.; Javed, K.; Rehman, A.; Akram, T.; Awais, M. Lungs nodule detection framework from computed tomography images using support vector machine. Microsc. Res. Tech. 2019, 82, 1256–1266. [Google Scholar] [CrossRef]
  40. Toğaçar, M.; Ergen, B.; Cömert, Z. Classification of white blood cells using deep features obtained from Convolutional Neural Network models based on the combination of feature selection methods. Appl. Soft Comput. 2020, 97, 106810. [Google Scholar] [CrossRef]
  41. Hegde, R.B.; Prasad, K.; Hebbar, H.; Singh, B.M.K. Feature extraction using traditional image processing and convolutional neural network methods to classify white blood cells: A study. Australas. Phys. Eng. Sci. Med. 2019, 42, 627–638. [Google Scholar] [CrossRef]
  42. Wang, J.L.; Li, A.Y.; Huang, M.; Ibrahim, A.K.; Zhuang, H.; Ali, A.M. Classification of white blood cells with patternnet-fused ensemble of convolutional neural networks (pecnn). In Proceedings of the 2018 IEEE International Symposium on Signal Processing and Information Technology (ISSPIT), Louisville, KY, USA, 6–8 December 2018; pp. 325–330. [Google Scholar]
  43. Habibzadeh, M.; Jannesari, M.; Rezaei, Z.; Baharvand, H.; Totonchi, M. Automatic white blood cell classification using pre-trained deep learning models: Resnet and inception. In Proceedings of the Tenth International Conference on Machine Vision (ICMV 2017), Vienna, Austria, 13–15 November 2017; Volume 10696, pp. 274–281. [Google Scholar]
  44. Akram, T.; Laurent, B.; Naqvi, S.R.; Alex, M.M.; Muhammad, N. A deep heterogeneous feature fusion approach for automatic land-use classification. Inf. Sci. 2018, 467, 199–218. [Google Scholar] [CrossRef]
  45. Akram, T.; Naqvi, S.R.; Haider, S.A.; Kamran, M.; Qamar, A. A novel framework for approximation of magneto-resistance curves of a superconducting film using GMDH-type neural networks. Superlattices Microstruct. 2020, 145, 106635. [Google Scholar] [CrossRef]
Figure 1. White blood cell subtypes.
Figure 1. White blood cell subtypes.
Diagnostics 13 00352 g001
Figure 2. Samplesof WBC images of dataset used in this work.
Figure 2. Samplesof WBC images of dataset used in this work.
Diagnostics 13 00352 g002
Figure 3. Pipeline of proposed WBCs classification system.
Figure 3. Pipeline of proposed WBCs classification system.
Diagnostics 13 00352 g003
Figure 4. Training accuracy and loss function plots for DenseNet201 network.
Figure 4. Training accuracy and loss function plots for DenseNet201 network.
Diagnostics 13 00352 g004
Figure 5. Extracted deep features from proposed ECMPA.
Figure 5. Extracted deep features from proposed ECMPA.
Diagnostics 13 00352 g005
Figure 6. Classification results of proposed WBCs classification system. Left: Test accuracy achieved by SVM and KNN classifiers with several kernels. Right: Confusion matrix of SVM with quadratic kernel.
Figure 6. Classification results of proposed WBCs classification system. Left: Test accuracy achieved by SVM and KNN classifiers with several kernels. Right: Confusion matrix of SVM with quadratic kernel.
Diagnostics 13 00352 g006
Figure 7. Convergence performance of ECMPA and GA.
Figure 7. Convergence performance of ECMPA and GA.
Diagnostics 13 00352 g007
Figure 8. Confidence interval for selected classifiers.
Figure 8. Confidence interval for selected classifiers.
Diagnostics 13 00352 g008
Table 1. DarkNet53 layer architecture.
Table 1. DarkNet53 layer architecture.
Layer TypeFiltersFilter SizeStride SizeRepeatOutput Size
Input---- 224 × 256
Convolutional32 3 × 3 11 256 × 256
Convolutional64 3 × 3 21 128 × 128
Convolutional32 1 × 1 11
Convolutional64 3 × 3 1
Residual 128 × 128
Convolutional128 3 × 3 21 64 × 64
Convolutional64 1 × 1 12
Convolutional128 3 × 3 1
Residual 64 × 64
Convolutional256 3 × 3 21 32 × 32
Convolutional128 1 × 1 18
Convolutional256 3 × 3 1
Residual 32 × 32
Convolutional512 3 × 3 21 16 × 16
Convolutional256 1 × 1 18
Convolutional512 3 × 3 1
Residual 16 × 16
Convolutional1024 3 × 3 21 8 × 8
Convolutional512 1 × 1 14
Convolutional1024 3 × 3 1
Residual 8 × 8
GlobalAvgPool
Fully Connected1000
Softmax
Table 2. DenseNet201 layer architecture.
Table 2. DenseNet201 layer architecture.
Layer TypeCompositionRepeatOutSize
Input 224 × 224
ConvolutionConv( 7 × 7 ), stride 2 112 × 112
MaxPool( 3 × 3 ), stride 2 56 × 56
Dense Block 1Conv( 1 × 1 )6
Conv( 3 × 3 ) 56 × 56
Transition Layer 1Conv( 1 × 1 )1 56 × 56
Avg Pool( 2 × 2 ), Stride 2 28 × 28
Dense Block 2Conv( 1 × 1 )12
Conv( 3 × 3 ) 28 × 28
Transition Layer 2Conv( 1 × 1 )1 28 × 28
Avg Pool( 2 × 2 ), Stride 2 14 × 14
Dense Block 3Conv( 1 × 1 )48
Conv( 3 × 3 ) 14 × 14
Transition Layer 3Conv( 1 × 1 )1 14 × 14
Avg Pool( 2 × 2 ), Stride 2 7 × 7
Dense Block 4Conv( 1 × 1 )32
Conv( 3 × 3 ) 7 × 7
Classification Layer 7 × 7 Global Avg. Pool
1000D fully Connected, softmax 1 × 1
Table 3. Model training parameters for transfer learning of DenseNet201 and DarkNet53 models.
Table 3. Model training parameters for transfer learning of DenseNet201 and DarkNet53 models.
PropertyValuePropertyValue
KernelsdgmInitial Learning Rate 1 × 10 4
Execution EnvironmentAutoMiniBatch Size20
MaxEpochs5Validation Frequency30
Dropout rate0.1Stride Size1
Table 4. Performance Comparison of proposed method with some existing works. ×: Not done, N.A: Information not available.
Table 4. Performance Comparison of proposed method with some existing works. ×: Not done, N.A: Information not available.
WorkDeep Learning ModelFeature SelectionFeature Vector SizeClassifierAccuracy %
[40]GoogleNet, ResNet-50Maximal Information Coefficient, Ridge Regression Model755Quadratic Discriminant Analysis97.95
[41]AlexNet×1000CNN98.4
[42]PatternNet fused ensemble of CNNs×N.ACNN99.90
[43]ResNet and InceptionHierarchical ApproachN.AResNet and Inception99.84
This WorkDenseNet201 and DartkNet53ECMPA76SVM and KNN99.6
Table 5. Statistical test results based on ANOVA using accuracy metric.
Table 5. Statistical test results based on ANOVA using accuracy metric.
V-SourceSSdfMSEF-Statisticsp-Value
Between7.0812 × 10 5 23.6333 × 10 5 0.370.705
Within5.9123 × 10 4 69.8222 × 10 5 --
Total6.6232 × 10 4 8---
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ahmad, R.; Awais, M.; Kausar, N.; Akram, T. White Blood Cells Classification Using Entropy-Controlled Deep Features Optimization. Diagnostics 2023, 13, 352. https://doi.org/10.3390/diagnostics13030352

AMA Style

Ahmad R, Awais M, Kausar N, Akram T. White Blood Cells Classification Using Entropy-Controlled Deep Features Optimization. Diagnostics. 2023; 13(3):352. https://doi.org/10.3390/diagnostics13030352

Chicago/Turabian Style

Ahmad, Riaz, Muhammad Awais, Nabeela Kausar, and Tallha Akram. 2023. "White Blood Cells Classification Using Entropy-Controlled Deep Features Optimization" Diagnostics 13, no. 3: 352. https://doi.org/10.3390/diagnostics13030352

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop