Next Article in Journal
Secure Vehicular Platoon Management against Sybil Attacks
Next Article in Special Issue
Classification Strategies for P300-Based BCI-Spellers Adopting the Row Column Paradigm
Previous Article in Journal
Movable Surface Rotation Angle Measurement System Using IMU
Previous Article in Special Issue
SoftMatch: Comparing Scanpaths Using Combinatorial Spatio-Temporal Sequences with Fractal Curves
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning-Based Computer-Aided Diagnosis (CAD): Applications for Medical Image Datasets

1
Department of Modeling and Design of Engineering Systems (MODES), Atilim University, Ankara 06830, Turkey
2
Department of Electrical and Electronics Engineering, Atilim University, Incek, Ankara 06830, Turkey
3
Department of Mechatronics Engineering, Atilim University, Incek, Ankara 06830, Turkey
4
Department of Software Engineering, Atilim University, Incek, Ankara 06830, Turkey
5
Informatics and Digitalization Group, Molde University College—Specialized University in Logistics, 6410 Molde, Norway
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(22), 8999; https://doi.org/10.3390/s22228999
Submission received: 23 October 2022 / Revised: 17 November 2022 / Accepted: 18 November 2022 / Published: 21 November 2022

Abstract

:
Computer-aided diagnosis (CAD) has proved to be an effective and accurate method for diagnostic prediction over the years. This article focuses on the development of an automated CAD system with the intent to perform diagnosis as accurately as possible. Deep learning methods have been able to produce impressive results on medical image datasets. This study employs deep learning methods in conjunction with meta-heuristic algorithms and supervised machine-learning algorithms to perform an accurate diagnosis. Pre-trained convolutional neural networks (CNNs) or auto-encoder are used for feature extraction, whereas feature selection is performed using an ant colony optimization (ACO) algorithm. Ant colony optimization helps to search for the best optimal features while reducing the amount of data. Lastly, diagnosis prediction (classification) is achieved using learnable classifiers. The novel framework for the extraction and selection of features is based on deep learning, auto-encoder, and ACO. The performance of the proposed approach is evaluated using two medical image datasets: chest X-ray (CXR) and magnetic resonance imaging (MRI) for the prediction of the existence of COVID-19 and brain tumors. Accuracy is used as the main measure to compare the performance of the proposed approach with existing state-of-the-art methods. The proposed system achieves an average accuracy of 99.61% and 99.18%, outperforming all other methods in diagnosing the presence of COVID-19 and brain tumors, respectively. Based on the achieved results, it can be claimed that physicians or radiologists can confidently utilize the proposed approach for diagnosing COVID-19 patients and patients with specific brain tumors.

1. Introduction

Ever since the development of the first known “expert system” in medicine, researchers have been striving to explore the possibilities of engaging artificial intelligence to solve medical problems of various natures [1]. The advent of modern computers paved the way for a completely new domain that emerged as a combined struggle of medical personnel and computer scientists, known as computer-aided diagnosis (CAD) [2]. The first commercial CAD system was approved by the US Food and Drug Administration (FDA) for mammography in 1998 [3]. Since then, several CAD systems have been developed for analyzing the chest, brain, heart, colon, and kidney, among many other organs, based upon conventional projection radiography, ultrasound [4], computed tomography (CT), and magnetic resonance imaging.
The modern concept of the CAD is to help physicians, pathologists, or radiologists in decision-making. However, researchers in the 1960s and 1970s had a different perspective [5,6]. The original intention was to replace humans with computers for detecting abnormalities, as many considered them superior to perform specific tasks. However, in the following years, computers could not prove their supremacy, probably due to insufficient computational capability and the unavailability of advanced image-processing and image-interpretation techniques. It helped to change the notion from computers being the diagnosticians to the assistants of a physician for decision-making.
In recent years, due to the availability of better computational tools and processors, some investigators have made impressive progress toward the development of automated computer-aided diagnosis tools [7]. The purpose is to develop error-prone diagnostic methods that can achieve a higher level of diagnosis results than the average level achievable by physicians. A significant portion of the CAD systems depends upon the computational analysis of the images to deduce the conclusion. Modern computational tools, particularly deep learning methods, have been effectively utilized for the analysis to obtain promising results [8]. This study employs more recent deep learning methods to encounter the most explored problems in medicine, COVID-19, and brain tumors.
COVID-19 broke out in late 2019 and has not only claimed many lives all around the world, but also impacted many economically and socially. Detecting COVID-19 at the earlier stages can prevent the loss of many precious lives. The nucleic acid detection kit (TR-PCR) and nucleic-acid sequencing are among the most commonly used diagnostic procedures for detection [9]. However, the TR-PCR approach requires at least 8 h to produce test results, while the sequencing takes much longer. By delaying the prompt diagnosis of the afflicted patients, the transmission of this contagious disease can increase at an astonishing rate. Radiologists have utilized CT and chest X-ray (CXR) for lung diseases equally. Many researchers have utilized deep learning technology to diagnose COVID-19 based on CT and chest X-ray (CXR) [10,11,12]. Due to their portability, fewer storage requirements, and availability, deep learning methods are more inclined towards chest X-ray (CXR) over CT for COVID-19 diagnosis.
Brain tumors are also considered among the most explored diagnostics and have seen unprecedented success due to deep learning methods [13]. A brain tumor caused by abnormal cell development that impairs the nervous system’s activities is classified as either malignant or benign [14]. Radiologists widely use MRI for tumor diagnosis. Among all brain tumors, meningioma, glioma, and pituitary are considered the most common types; about 33% of all brain tumors are gliomas [15]. Brain tumor is considered one of the most challenging medical research problems as it exhibits high shape, size, and intensity variations.
Though deep learning methods have achieved promising results in computer-aided diagnosis in recent years, their performance still has much room for improvement. The most notable challenges are the development of advanced segmentation approaches, advanced feature extraction and selection approaches, and better classification approaches. In this study, the main emphasis was on selecting the most effective and essential features based on the ant colony optimization algorithm. Selecting the most prominent features helped improve results by at least 10 percent for both datasets. The main contributions of this paper are four-fold:
  • The extraction of features from the image dataset using auto-encoder and well-known pre-trained CNNs, e.g., ResNet and AlexNet;
  • The selection of the most notable features using ACO to enhance accuracy;
  • A generic framework is proposed that can work on multiple datasets, such as MRIs and X-rays;
  • The evaluation of the proposed classification model against the current baseline diagnostic models.
The remainder of this paper is organized as follows: the related work is presented in Section 2; the datasets and their pre-processing, methodologies, deep learning techniques, and the proposed feature selector are described in Section 3; and the results are presented and analyzed in Section 4. Then, in Section 5, some discussions are made. Finally, the conclusion and future recommendations are presented in Section 6.

2. Related Works

The adaptation of deep learning methods to perform automated CAD for brain tumors and COVID-19 has been astonishing [16,17]. Deep convolutional neural networks are compelling neural network architectures that can produce promising results when applied for image analysis, pattern recognition, and other purposes. For medical image analysis primarily used for diagnostics, the accuracy of the prediction is vital. The prediction based upon image dataset generally depends upon segmentation and pre-processing, feature extraction and selection, and classification [18].

2.1. Segmentation and Pre-Processing

The first stage in the CAD process is gathering the relevant data and executing their quality assessment [18]. After the initial screening, all corrupted images are removed from the dataset. After the attainment of the dataset, a large group of data is distributed into several subsets based upon some uniformity conditions through a process known as segmentation. It is regarded as one of the essential tasks in medical image processing to enhance the already existing information [19]. The wide variety of segmentation techniques can be broadly classified into classical (pixel-based, region-based, edge-based, texture-based) and intelligent (artificial neural network-based, fuzzy logic-based, decision tree-based, genetic algorithm-based) [20].
The captured chest X-ray (CXR) and MRI images are mostly considered noisy, inconsistent, and incomplete; therefore, pre-processing these images plays a crucial role in acquiring higher accuracy [21]. Different types of linear and non-linear filters are extensively used for denoising. Filtering is also considered a mean for suppressing unwanted information and enhancing the wanted information. According to their architecture, different deep learning methods have specific input image requirements; thus, image resizing, cropping, normalization, and padding are considered the essential steps in pre-processing [22]. The volume, region, and intensity adjustments are also considered effective pre-processing techniques to improve the execution rate.

2.2. Feature Extraction and Selection

Accurate diagnosis is also dependent upon the availability of the large-scale and complex dataset. This also entails the requirement of large memory space and computing power. Feature extraction and selection can be potentially used as a means to remove irrelevant or duplicate information; hence, the time and storage-space requirements can be reduced, which may also lead to the better performance of the CAD systems.
To minimize the information loss and reduce the dimensionality, principal component analysis (PCA) is utilized for the feature selection and extraction [23]. The authors in [24] used a combination of discrete wavelet transform [25] and Gabor filter for feature selection from magnetic resonance images. Some recent work used deep learning techniques for the extraction and selection of features from medical images [22,26,27]. Studies were conducted to extract deep features from brain MRI images using pre-trained networks [28]. Rajpurkar et al. [29] used discriminative chest X-ray (CXR) imaging features to detect pneumonia at different stages through CNN.

2.3. Classification

The diagnostic predictions are performed through the central processing module, usually called a classifier. A common problem that exists today is the determination of an appropriate classifier for a given dataset. It is observed that improper classifier selection may substantially reduce the accuracy. Multiple classifiers exist, including those based on neural networks, decision trees, and support vector machines. Recent trends have shown that deep learning-based methods have received much appreciation from investigators for the improved classification on various datasets. In [30], the gray-level co-occurrence matrix is proposed to classify brain tumors as Meningioma, Glioma, and Pituitary based upon CNN. Mustafa R. Ismael et al. [24] proposed a system that integrates statistical features and neural network methods to classify brain cancers in MRI images. The deep convolutional neural network-based multi-grade is presented in [31] for brain tumor classification.
It is typical for a Rician distribution to control the noise in MRI scans. The bendlets system is a second-order shearlet transform with bent elements and is considered an effective tool for sparsely representing images with curve outlines, such as brain MRI scans. An adaptive denoising approach for microsection pictures with Rician noise is proposed in [32] using the bendlet characteristic. This method allows for identifying the curve’s texture and contour as low-frequency components. The Rician noise is clearly recognized to belong to a high-frequency channel, thus making it simple to eliminate without impairing the clarity of the contour.
Alqahtani, Ali et al. developed a novel deep CNN-based framework for chest X-ray analysis [33] that is computationally light and efficient. They used several machine learning classifiers to improve the framework’s discriminating capabilities and suggested a new COV-Net to learn COVID-specific patterns from chest X-rays. The network can assist in the multi-class categorization of two different infection types by utilizing max-pooling procedures.
In [34], soft sensor techniques are used to assess the neurocognitive dysfunctions unique to neurodevelopmental disorders, ADD/HD, and specific learning impairments. The bendlet transform was proposed on the basis of the shearlet transform and used for image processing [35]. K Gopalakrishnan et al. investigated the capacity of several pre-trained DCNN models—Alexnet, Resnet50, GoogLeNet, VGG-16, Resnet101, VGG-19, Inceptionv3, and InceptionResNetV2—using transfer learning to classify diseased brain images [36].
In [37], the authors attempted to establish an automated method for categorizing fundus images. Through an ablation study, Tong He et al. evaluated a number of these modifications’ effects on the final model’s accuracy empirically [38]. Samir S. Yadav et al. classified pneumonia using a convolutional neural network (CNN)-based technique on a chest X-ray dataset [39].
In [40], the authors used the deep convolutional neural network to detect COVID-19 from the X-ray imaging system. They introduced the new open-access benchmark that included 13,975 chest X-ray (CXR) images across 13,870 patient case datasets. The three-player knowledge transfer and concentration framework counting a pre-trained appearing network that extracts the chest X-ray (CXR) imaging features from a large scale of lung disease CXR images is presented in [41]. The deep learning based on the DarkNet model is used to classify the binary and multi-classes cases of COVID-19 [42]. In this model, YOLO was employed for the real-time object detection. A rapid and valid method [43] diagnoses COVID-19 with an artificial neural network. They used and tested ten various CNNs to distinguish infection of COVID-19 cases from non-COVID-19.
The existing deep learning diagnostic models use CT, CXR, or MRI images for the detection of brain tumors and COVID-19. The performance indicators of these models indicate that there is still massive room for improvement. In this study, the main emphasis is on selecting the most effective and important features based on ant colony optimization (ACO). The obtained features are then processed through state-of-the-art deep learning classifiers to perform an accurate diagnosis. This paper also strives to develop a generic framework that can work on multiple datasets, such as MRI and CXR.

3. Material and Methods

In this study, for detecting brain tumors and COVID-19, the features from the image datasets are obtained either through pre-trained CNN or Auto-encoder. These methods are implemented separately and combined with feature selection methods. The proposed framework tends to work with datasets of different natures. To prove its efficacy, it is tested on two different medical image datasets based on MRI and CXR to detect brain tumors and COVID-19. Both datasets have been used widely by the research community in the recent past.

3.1. COVID-19 Dataset

The coronavirus strain COVID-19 (coronavirus disease 2019) causes the severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2). The first case was reported in Wuhan, China, in late December 2019, before spreading worldwide [44]. The dataset contains 6432 X-ray images of varying sizes [45]. The distribution for the training-testing is defined as 80-20, consisting of three classes: COVID-19, pneumonia, and normal. The actual number of training images for COVID-19, pneumonia, and normal is defined as 460, 3418, and 1266, respectively. For testing, 116 COVID-19, 855 pneumonia, and 317 normal images are selected. Figure 1 shows a sample from each class for the COVID-19 dataset.

3.2. Brain Tumor Dataset

The figshare brain tumor MRI dataset is publicly available and is commonly used for classification evaluation. The involved dataset was acquired from Nanfang Hospital, Guangzhou, China, and Tianjin Medical University General Hospital, China, between 2005 and 2010 [46]. This dataset includes 3064 T1-weighted contrast-enhanced images obtained from 233 patients. Three different types of brain tumor are considered in this dataset: meningioma (708 slices), glioma (1426 slices), and pituitary (930 slices). This study defines the distribution between train and test datasets as 80-20. It means that meningioma has 566 slices for training the model and 142 slices for testing, glioma has 1141 slices for training and 285 slices for the testing, and pituitary has 744 slices to train the model and 186 slices for evaluation. Figure 2 represents the brain dataset samples for the three classes.

3.3. Deep Learning Techniques

3.3.1. Convolutional Neural Network

This study combines pre-trained CNNs with ant colony optimization (ACO) and some classifiers for image classification. In the first stage, the pre-trained model is applied to extract important features from the image. AlexNet is one of the several pre-trained classifiers involved in this study, and is known as a common model in image classification studies. The study aims to extract high-level and sensitive features from input images. The CNN model realizes human vision technology. Nowadays, CNN is widely used in several computer vision problems as it implements the shared weight technique to reduce the computation rather than the fully connected approach. AlexNet consists of several layers: a convolution layer, pooling layer, and a fully connected layer. The extracted features differ from one classifier to another:
  • Alexnet (4096) features.
  • Googlenet (1000) features.
  • Resnet 50 (2048) features.
  • Densenet 201 (1920) features.

3.3.2. Auto-Encoders

An auto-encoder is a neural network that is trained to replicate its input at its output. Auto-encoders can be used as tools to learn deep neural networks. Training an auto-encoder is unsupervised in the sense that no labeled data are needed. The training process is still based on the optimization of a cost function. The cost function measures the error between the input x and its reconstruction at the output x ^ .
An auto-encoder is composed of an encoder and a decoder. The encoder and decoder can have multiple layers, but for simplicity consider that each of them has only one layer [47,48].
If the input to an auto-encoder is a vector x R D x , then the encoder maps the vector x to another vector z R D ( 1 ) , as follows:
z = h ( 1 ) ( W ( 1 ) x + b ( 1 ) )
where the superscript (1) indicates the first layer. h ( 1 ) : R D ( 1 ) R D ( 1 ) is a transfer function for the encoder, W ( 1 ) R D ( 1 ) × D x is a weight matrix, b ( 1 ) R D ( 1 ) is a bias vector, and D x is the dimensional of layer x; R is real values. Then, the decoder maps the encoded representation z back into an estimate of the original input vector, x, as follows:
x ^ = h ( 2 ) ( W ( 2 ) z + b 2 )
where the superscript (2) represents the second layer. h ( 2 ) : R D x R D x is the transfer function for the decoder, W ( 1 ) R D x × D ( 1 ) is a weight matrix, and b ( 2 ) R D x is a bias vector.

3.4. Ant Colony Optimization Algorithm for Feature Selection

One of the interesting behaviors in nature for finding food, which has a high intelligence nature, is the behavior of ants. Ants have a clever way of finding optimal food and reducing their output error in finding the shortest way to it. In this paper, the behavior of ants is first introduced and a modified ant colony optimization for the feature selection method is introduced based on it. In the second phase, the deep learning and auto-encoder-based method are used for classifying the COVID-19 disease and brain tumor data [49].
The ant colony optimization algorithm was proposed in 1990 by Marco Dorigo as an innovative method for solving hybrid optimization problems [49]. This algorithm is derived from the actual behavior of ants to find food along the shortest path [50]. Each ant leaves a chemical material called pheromone along its path of finding food, and other ants choose the shortest path using these previously secreted pheromones. This algorithm is very useful for solving non-deterministic polynomial (NP) problems and is used in problems such as itinerant vendors, scheduling problems, vehicle routing problems, etc. [51].
To solve any NP problem using the ant colony algorithm, the following must be specified:
  • First, you have to turn the problem into a graph, including nodes and edges.
  • See distance nodes ( η ) are raised and specified.
  • A possible solution is created according to the problem.
  • The pheromone update rule is used to determine the effective edges in achieving the best answer.
  • The probabilistic transition rule is used to find the next node [52].
Several different implementations of ant colony algorithms such as Ant System, Max-Min Ant System, and Ant Colony System have been proposed; the main difference of these methods is in the pheromone update formula [52]. The method of implementing the proposed algorithm is described. Feature selection will be based on the ant colony method, with two methods of measuring feature-class (FC) and feature-feature (FF).

3.4.1. Working Method of the Proposed Model

The method of implementing the algorithm is shown in Figure 3.
In this flowchart, Τ i is the pheromone value in ith iteration.
First, a graphical model of all the features in the S dataset is introduced. Attributes are considered nodes and all nodes are interconnected. Then, τ, η, the number of ants, and number of repetitions must be determined [53]. The value of τ is known as the pheromone trail, and at the beginning of the algorithm its values for all attributes are a fixed number of one by default. The value of η is known as heuristic information and is equal to the inverse of the distance between the properties [54]; in this article this distance will be set based on the two methods FC and FF.
After determining the initial values, the algorithm is applicable. In each iteration, the ant is first randomly placed on a node. To determine the next ninety, the law of transfer is used, which is shown in Equation (3):
  P i k ( t ) = |   τ i ( t ) | α · | η i ( t ) | β u j k |   τ i ( t ) | α .   | η i ( t ) | β   i f ( q > q 0 )
j = max u j k   ( τ i ( i ) α . η i ( i ) β )   i f ( q < q 0 )
The values of α and β are determined to make the values of τ and η more effective. j k is a set of traits that the ant has not yet met, and the trait that the ant has seen before is zero. The parameter q 0 is very important in determining the choice of both greedy and probabilistic methods, and the value of q is a random number between zero and one.
After the n ant has completed the node scan, the amount of pheromone obtained from the scan should be updated according to Equation (4):
τ i ( t + 1 ) = ( 1 ρ ) τ i ( t ) + i = 1 n τ i k ( t )
The value of ρ must be determined to reduce the effect. τ i k is the inverse of the error obtained for the Wrapper method and is equal to the average number of nodes selected for the Filter method [53,54].

3.4.2. Criteria for Distance and Similarity of Features

The relationship between two random variables is divided into two categories: linear, and nonlinear. The most famous method for calculating linear variables is the correlation coefficient formula. In [55], the authors used the entropy method and information theory to calculate nonlinear variables. The problem with the correlation coefficient method is that it is inefficient on non-numerical and batch data, while the entropy method works well [55].
Entropy or irregularity criteria are used to measure the uncertainty of a discrete or continuous random variable. The entropy or H(X) of the discrete random variable X = ( x 1 , x 2 , , x n ) is calculated from Equation (5).
H ( X ) = i = 1 n p ( x i ) log ( p ( x i ) )
where p ( x i ) is the probability of   x i occurring on the whole set.
Calculate the entropy of two discrete random variables X and Y with dimensions’ n according to Equation (6).
H ( X , Y ) = i = 1 n j = 1 n p ( x i , y i ) log ( p ( x i , y i ) )
Equation (7) is used to calculate the conditional entropy of X to condition Y.
H ( X | Y ) = i = 1 n j = 1 n p ( x i , y i ) log ( p ( x i | y i ) )
The purpose of the above formulas is to calculate the information factor (IF). To examine how interdependent the two variables are, the IF criterion is used, which is in accordance with Equation (8):
I ( X , Y ) = H ( X ) H ( X | Y )
If the value of IF becomes zero, it means that the two variables are independent, and the higher this value, the more X and Y are dependent [56]. Figure 4 shows the relationship between information coefficient and entropy.
This study uses the normalized form of IG, known as symmetrical uncertainty (SU), which is consistent with Equation (9). The advantage of this formula is the normality of the dependence of the two variables between the range 0 and 1. If the value of SU is close to one, it means that the two variables are dependent, and if it is close to zero, it means that it is independent.
S U ( X , Y ) = 2     I ( X , Y ) H ( X ) + H ( Y )
In this paper, two criteria S U F C and S U F F   are used to calculate η [55].
The S U F C criterion is defined as the dependence of each attribute on the class. The closer this value is to one, the more important that feature will be and it should be selected.
η i = 1 1 S U F C
The criterion S U F F means the dependence of two properties on each other. If its value is close to one, it means that the two properties are very similar, so we will be looking to remove one of the features.
η i = 1 S U F F
In selecting attributes, we seek to preserve class-related attributes and remove duplicate or trivial attributes. The goal is to select features with S U F C higher and S U F F   lower [57].

3.5. Dataset Pre-Processing

Pre-processing image datasets is an essential step that improves the quality of the features, hence diagnosis prediction. The pre-processing steps for CXR and MRI image datasets are detailed in Figure 5. Either auto-encoder or CNN later uses the processed data for the feature extraction. The RGB images are converted to grayscale when processed through an auto-encoder to reduce the amount of processed data by one-third. Furthermore, for auto-encoder images are resized to 64 × 64 and later converted to a single array (vector) without losing any features and to enhance the training model’s quality. The brain tumor images are available as png files with the size of 512 × 512, whereas the chest X-ray (CXR) images have a jpg format of varying sizes. The mismatch among the input images is fixed by resizing them to a fixed 227 × 227 (Alexnet) or 224 × 224 (GoogleNet, ResNet-50, and DenseNet-201) resolution.

3.6. Overall Architecture of the Proposed Framework

In the proposed framework (Figure 6), features can be extracted either using pre-trained CNNs (AlexNet, GoogleNet, ResNet-50, or DenseNet-201) or auto-encoder. Once features are extracted, the meta-heuristic, ant colony optimization (ACO), is utilized to choose the most effective and prominent features [58].
The main goal of this study is to generate smaller subsets of essential features that help to expedite the search process and select the most optimal features. Afterward, learnable classifiers (decision tree, support vector machine, k-nearest neighbor, ensemble, naive Bayes, or discriminant analysis) can be employed for the diagnosis prediction. These classifiers learn from the features taken from the ACO algorithm and classify them based on the labels of input type. Table 1 shows the characteristics of the chosen classifiers. The initial selection of the classifier can be made based on the trade-off among prediction speed, memory usage, and interpretability.

4. Experiments and Results Analysis

In this study, MATLAB2021a is used to execute the proposed method on Intel(R) Core (TM) i7-6500U CPU @ 2.50GHz 2.60 GHz. Many research studies are conducted to investigate the performance of the existing prediction methods. Once done with the pre-processing, the feature selection-extraction is performed, followed by the classification. In the former stage, an auto-encoder/CNN is combined with an ACO algorithm to extract and select the most prominent features from the input training dataset. The second stage employs a range of learnable classifiers to diagnose tumors or COVID-19 accurately.

4.1. Evaluation Metrics

The performance evaluation of the proposed framework is based upon the following performance matrices: sensitivity, specificity, precision, accuracy, F1-score, misclassification rate, and Mathews correlation coefficient (MCC). For the confusion matrix, TP represents the true positive, TN represents the true negative, FP represents the false positive, and FN represents the false negative. All the parameters used for the evaluation are listed in Table 2. Among these, accuracy is usually considered the most reliable measure to compare the performance of different classification methods.

4.2. Classification Using Learnable Classifiers

The performance of the CAD prediction system depends upon the combination of features and the classifier model. In this study, feature extraction-selection is either done using a combination of auto-encoder + ACO or pre-trained CNNs + ACO. For classification, a broad range of learnable classifiers is selected: decision tree, support vector machine, kernel nearest neighbor, ensemble, naive bayes, and discriminant analysis.

4.2.1. Feature Extraction and Feature Selection Using Auto-Encoder and ACO

In the initial stage, image features from the chest X-ray (CXR) images dataset are extracted using an auto-encoder; whereas most optimal features are selected based on ACO. For training and testing, six different classifiers are chosen to evaluate the performance of the proposed framework. The chest X-ray (CXR) image dataset is used to determine the health status of the patients, such as those affected by COVID-19 or pneumonia or considered normal. Table 3 shows the classification results using the evaluation matrices. The accuracy determined using true values of the test images is considered the most important parameter to evaluate the performance. Support vector machines are able to achieve the highest accuracy of 98.68% among all classifiers, followed by ensemble and discriminant as the second-best classifiers to achieve almost 98% accuracy. The misclassification rate also supports the accuracy results, as the highest MR of 1.31% is obtained using SVM. The NB classifier exhibits the worst performance with the lowest accuracy of 89.28% and an MR of 10.71%.
The MRI image dataset is used to classify three types of brain tumors: meningioma, glioma, and pituitary. Once the image features are selected based on the auto-encoder, ACO searches for the best prominent features. Six different classifiers are used to evaluate the performance of the proposed framework on the MRI image dataset for diagnosing tumors (see Table 3). The classification results demonstrate that the highest classification accuracy of 99.18% is obtained through KNN; likewise, the MR is the lowest, i.e., 0.81%. The worst prediction results are obtained through NB, which achieves the lowest accuracy of 78.62% and the MR of 21.37%.

4.2.2. Feature Extraction-Selection Using Pre-Trained CNNs and ACO

For the feature extraction of the chest X-ray (CXR) images dataset, several pre-trained CNNs (AlexNet, GoogleNet, ResNet-50, and DenseNet-201) are considered. Once again, the ACO method is employed to select the most prominent image features. The training and testing of the CXR image dataset for the diagnosis of COVID-19, pneumonia, and normal is performed using six classifiers (see Table 1). Table 4 represents the evaluation matrices obtained through all classifiers against four pre-trained CNNs + ACO. The best overall results are obtained when feature extraction is made through ResNet-50 and classification is done through SVM. The highest accuracy of 99.61% is obtained with SVM and an MR of 0.38. The lowest possible accuracy of 94.95% is obtained for AlexNet evaluated against the NB classifier. It is worth mentioning that all the pre-trained CNNs have been able to achieve decent results.
The MRI image dataset is processed through pre-trained CNNs and ACO to determine the existence of a specific type of brain tumor. The evaluation matrices for classifying brain tumors using six different classifiers are shown in Table 5. It is observable that features extracted using AlexNet, when evaluated against discriminant analysis classifier, produce the best results. The highest classification accuracy of 98.69% is obtained through discriminant analysis, whereas the lowest accuracy of 87.27% is obtained when features using GoogleNet are evaluated against the decision tree classifier.

4.3. Performance Comparison with Existing Methods

The comparison analysis is performed for our proposed method with the most prominent methods related to brain tumors and COVID-19. Table 6 shows the classification accuracy results obtained through the selected state-of-the-art methods and our proposed approach using MRI image-dataset. Classification accuracy is a measure to gauge the amount of certainty to perform the correct diagnosis of brain tumor type. For training-testing, a proportion of 80-20 is defined for all the involved methods. The comparison table shows that the proposed method attains superior results in diagnosing brain tumor type when feature extraction, selection, and classification are performed using auto-encoder, ACO, and KNN, respectively.
The same approach is adapted for the COVID-19 image dataset where the proposed method’s performance is compared against the existing methods in terms of accuracy (Table 7). For training-testing, the classification methods, a proportion of 80-20 is defined for all the involved methods. When combined with SVM classifier, the feature extraction-selection using pre-trained CNN (Resnet-50) and ACO produces the highest accuracy index for diagnosing COVID-19, pneumonia, or normal.
(Table 8) compares the work of (ACO) vs. the work of (GA) as a feature selection, and shows an obvious difference in time and accuracy; GA takes twice as long and is less accurate than ACO.
Ant colony optimization reaches the global minimum point faster than the genetic algorithm (GA) as it avoids being trapped in the local minimum. By simulating the intelligent behavior of ants, ACO tries to find optimal solutions to various optimization problems. It has gained considerable interest worldwide because of its advantages, such as simple implementation, small number of parameters, flexibility, etc.
Ant colony optimization has simplicity, flexibility, robustness, scalability, and self-organization. It has few control parameters as compared to a genetic algorithm (GA). The execution of various tasks can be undertaken by individuals simultaneously. Memory space is less utilized by swarm intelligence (SI) compared to GA.
The benefits of ACOs are numerous and there are many stakeholders who obtain advantages from this model of care. The patient community gains a wide number of advantages including improved outcomes, a better quality of care, greater engagement with providers, and an overall reduction in out-of-pocket costs. They have an advantage over the simulated annealing and genetic algorithm approaches of similar problems when the graph may change dynamically; the ant colony algorithm can be run continuously and adapt to changes in real-time [72,73,74].

5. Discussion

The advancement in the domain of artificial intelligence has also influenced diagnostics in medical science. This progress has enabled computer scientists to build up CAD tools with the assistance of medical personnel. Physicians frequently use such tools worldwide to perform an accurate diagnosis prediction. Although deep learning methods achieved higher accuracy, their success is highly dependent upon the features and their pre-processing. Today, there exist a plethora of diverse methods; the main problem is the selection of an appropriate method. In this study, we attempted to combine the most suitable methods for feature extraction-selection and classification to attain the highest accuracy using image-dataset.
To prove the effectiveness of the proposed method, two different medical image datasets are employed: CXR for the determination of COVID-19 and MRI for diagnosing brain tumors. Many researchers have evaluated the performance of different existing prediction models for COVID-19 and brain tumors. It is observed that they can achieve accuracy to a certain extent, beyond which further improvement is not made. From [9,69], it is witnessed that the auto-encoder extracted many unimportant image features; thus, it is required that the features should be further enhanced and the best ones should be selected. This study used a unique combination of deep learning techniques and a meta-heuristic algorithm to perform feature extraction and selection. The results reported in Table 6 and Table 7 show that the proposed method outperformed other state-of-the-art methods. This shows that our method is extensive and can be used for different types of medical image datasets.
In [31], researchers applied the Deep Convolution Neural Network VGG-19 based on the SoftMax classifier, resulting in 94.58% accuracy. This showed that combining the GLCM and Deep Convolution Neural Network VGG-19 leads to highly accurate results, which the authors in [62] obtained for the related comparison of the brain tumor dataset mentioned in Table 6.
For other Covid-19 dataset comparisons in Table 7, when the feature extraction method is not combined with feature selection methods such as Dense Net, we cannot expect results with high accuracy. However, in [41], our study, when we made a combination of the pre-trained Densenet201 with the ACO algorithm, we achieved 98.99% accuracy.
The method used by GLCM [30,62] has an accuracy is 82.00% and 96.50%. For this type of dataset, a gray level co-occurrence matrix cannot give accurate results from the GLCM features, so GLCM is useful for texture images such as fingerprint and Palmprint images [75]. Also, the method used in [30] used the GLCM, and the result cannot be high, because texture features cannot be used in medical images. However, in [62] they used both GLCM and Pre-Trained VGG-16 CNN, and the combination of the GLCM and Pre-Trained VGG-16 CNN improved the results. That means the use of GLCM only cannot improve the performance of the system.
In [23], texture feature extraction is used; also, to find the best feature, PCA is used. Principal component analysis feature selection cannot find the best feature from the medical data [76]. Also, there are no combination methods for feature extraction with feature selection employed in the other methods mentioned in Table 6 and Table 7. The ineffective features cannot improve the results. Furthermore, the reason for using these related works is using similar methods either for feature extraction or selection with the same dataset that we used.
Sachdeva et al. [23] recommended PCA based on the intensity and texture features. In this method, PCA finds the eigenvalues of the features and this will tend to find the high eigenvalues in the image. However, in the brain tumor images, some regions of the image are not effective on the high eigenvalues, so the result will contain some mistakes and ultimately the result will be not so accurate.
Cheng et al. [59] used the Bag of Words (BoW) to extract the features from the images and SVM are used in the classification of the images. The BoW is used to recognize the text sources [77,78]. This method for finding the features of medical images is not suitable, and also in [59,60,66] researchers used deep learning with the SVM classifier without the aid of any feature selection function in order to select the effective features and achieve quality results. In [22], a modified deep CNN network was assigned for the dataset to extract the most effective features and classify them with the SVM classifier.
The 2D Discrete Wavelet transform (DWT) and 2D Gabor filter were used in [24]. These feature extraction methods are also useful in the face recognition systems used in [79,80,81].
Capsule networks (CapsNets) were used in [61,71] to extract the features from brain tumor images and Covid-19 images. This method cannot give high accuracy to 2D signals such as medical images and was recently used in speech recognition signals [82]. For 1D signals the performance is better than for 2D signals.
Ghassemi et al. [63] combined the Generative adversarial network (GAN) and ConvNet (random split) to find and classify brain tumor images. Recently, GAN methods were used in a face recognition system [83,84]. Face images have high resolution, and there are more objects on the human face than on tumor images. However, combining with ConvNet can produce high accuracy.
The use of Deep Convolutional Auto-encoder alone cannot produce good results [9,69,70]. Auto-encoder produces more unimportant features from images. These features should be enhanced, and the best ones selected.
Umut Özkaya et al. [64] used CNN with data fusion and they obtained 98.27% accuracy. Compared with other methods, this method can be considered as high performance because data fusion creates robust features.
In Covid-19 images, when the feature extraction method is not combined with feature selection methods such as ANN [65], DNN [67], Random Forest [68], Tailored CNN [40], DenseNet [41], DarkNet-19-based CNN [42], and deep learning [43] we cannot expect highly accurate results.
We implemented a powerful method to extract features from medical image datasets in the proposed study; the deep learning algorithms (CNN, Auto-encoder) produced high-quality results when they extracted features. In addition, the meta-heuristic algorithm was utilized to select the effective features; this combination had a favorable impact. Ant colony optimization was used to select important features, to test its performance.

6. Conclusions

Computer-aided diagnostic systems have emerged as an effective tool for performing diagnosis prediction based on medical images. The performance of these systems is challenged by the processing of ‘big data’, which is also vital for accurate diagnosis. This study proposes a novel framework for extracting and selecting features based on deep learning, auto-encoder, and ACO with an intention to select the most prominent image features to reduce the amount of data to be processed. The meta-heuristic algorithm, ACO, is utilized to search for the essential features from the available feature-set to minimize the error rate. The obtained features are then processed through the learnable classifiers to determine the accuracy of the CAD system. The performance of the proposed system is evaluated for two medical image datasets: CXR and MRI. The CXR dataset is used to diagnose the patient’s condition as COVID-19, pneumonia, and normal, whereas, based upon MRI images, the objective is to determine the type of brain tumor (meningioma, glioma, and pituitary). The removal of the minor, redundant, and noisy features produce a significant effect on the accuracy of the overall system. The proposed approach achieves the highest accuracy compared to the other state-of-the-art methods, such as ANN, CNN, CNN with data fusion, Stacked-auto-encoder, Capsule Networks, and DarkNet-19-based CNN, for diagnosing various medical disease datasets. The basic notion of developing the proposed CAD system is assisting the physician. This study’s primary limitation is using the labeled data for supervised learning; hence, this process is not claimed to be completely automated. To exploit the potential of deep learning, the diagnostic prediction must be made using unlabeled data without any human intervention. Further studies will explore the possibility of performing disease prediction without requiring supervision to train the model.

Author Contributions

Conceptualization, Y.A.K., data curation, Y.A.K.; formal analysis, Y.A.K., M.U.K., A.M.; investigation, Y.A.K., A.M.; methodology, Y.A.K., M.U.K., A.M.; supervision, A.M.; validation, Y.A.K., M.U.K. and A.M., visualization, Y.A.K., M.U.K. and A.M., writing—original draft, Y.A.K.; writing—review and editing, Y.A.K., M.U.K. and A.M. All authors have read and agreed to the published version of the manuscript.

Funding

Authors thanks to the Molde University College-Specialized Univ. in Logistics, for the support of Open access fund.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

It is available in the reference.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Briganti, G.; Le Moine, O. Artificial Intelligence in Medicine: Today and Tomorrow. Front. Med. 2020, 7, 27. [Google Scholar] [CrossRef] [PubMed]
  2. Lodwick, G.S. Computer-aided diagnosis in radiology: A research plan. Investig. Radiol. 1966, 1, 72–80. [Google Scholar] [CrossRef] [PubMed]
  3. Giger, M.L.; Chan, H.; Boone, J. Anniversary paper: History and status of CAD and quantitative image analysis: The role of medical physics and AAPM. Med. Phys. 2008, 35, 5799–5820. [Google Scholar] [CrossRef]
  4. Czajkowska, J.; Borak, M. Computer-Aided Diagnosis Methods for High-Frequency Ultrasound Data Analysis: A Review. Sensors 2022, 22, 8326. [Google Scholar] [CrossRef]
  5. Meyers, P.H.; Nice, C.M., Jr.; Becker, H.C.; Nettleton, W.J., Jr.; Sweeney, J.W.; Meckstroth, G.R. Automated computer analysis of radiographic images. Radiology 1964, 83, 1029–1034. [Google Scholar] [CrossRef] [PubMed]
  6. Engle, R.L., Jr. Attempts to use computers as diagnostic aids in medical decision making: A thirty-year experience. Perspect. Biol. Med. 1992, 35, 207–219. [Google Scholar] [CrossRef]
  7. Doi, K. Computer-aided diagnosis in medical imaging: Historical review, current status and future potential. Comput. Med. Imaging Graph. 2007, 31, 198–211. [Google Scholar] [CrossRef] [Green Version]
  8. Chan, H.; Hadjiiski, L.M.; Samala, R.K. Computer-aided diagnosis in the era of deep learning. Med. Phys. 2020, 47, e218–e227. [Google Scholar] [CrossRef]
  9. Li, D.; Fu, Z.; Xu, J. Stacked-autoencoder-based model for COVID-19 diagnosis on CT images. Appl. Intell. 2021, 51, 2805–2817. [Google Scholar] [CrossRef]
  10. Paganin, F.; Seneterre, E.; Chanez, P.; Daures, J.P.; Bruel, J.M.; Michel, F.B.; Bousquet, J. Computed tomography of the lungs in asthma: Influence of disease severity and etiology. Am. J. Respir. Crit. Care Med. 1996, 153, 110–114. [Google Scholar] [CrossRef]
  11. Warrick, J.H.; Bhalla, M.; Schabel, S.I.; Silver, R.M. High resolution computed tomography in early scleroderma lung disease. J. Rheumatol. 1991, 18, 1520–1528. [Google Scholar] [PubMed]
  12. Na’am, J.; Harlan, J.; Nercahyo, G.W.; Arlis, S.; Rani, L.N. Detection of Infiltrate on Infant Chest X-ray. Telkomnika 2017, 15, 1943–1951. [Google Scholar] [CrossRef]
  13. Paul, J.S.; Plassard, A.J.; Landman, B.A.; Fabbri, D. Deep learning for brain tumor classification. In Proceedings of the Medical Imaging 2017: Biomedical Applications in Molecular, Structural, and Functional Imaging, Orlando, FL, USA, 12–14 February 2017; Volume 10137, p. 1013710. [Google Scholar]
  14. Black, P.M. Brain tumors. N. Engl. J. Med. 1991, 324, 1555–1564. [Google Scholar] [CrossRef] [PubMed]
  15. Toğaçar, M.; Cömert, Z.; Ergen, B. Classification of brain MRI using hyper column technique with convolutional neural network and feature selection method. Expert Syst. Appl. 2020, 149, 113274. [Google Scholar] [CrossRef]
  16. Morais, A.; Egger, J.; Alves, V. Automated computer-aided design of cranial implants using a deep volumetric convolutional denoising autoencoder. In Proceedings of the 2019 World Conference on Information Systems and Technologies, Cairo, Egypt, 24–26 March 2019; pp. 151–160. [Google Scholar]
  17. Ali, N.A.; Syafeeza, A.R.; Geok, L.J.; Wong, Y.C.; Hamid, N.A.; Jaafar, A.S. Design of Automated Computer-Aided Classification of Brain Tumor Using Deep Learning. In Intelligent and Interactive Computing; Springer: Berlin/Heidelberg, Germany, 2019; pp. 285–291. [Google Scholar]
  18. Yanase, J.; Triantaphyllou, E. The seven key challenges for the future of computer-aided diagnosis in medicine. Int. J. Med. Inform. 2019, 129, 413–422. [Google Scholar] [CrossRef]
  19. Sonka, M.; Hlavac, V.; Boyle, R. Image Processing, Analysis, and Machine Vision; Cengage Learning: Boston, MA, USA, 2014; ISBN 1285981448. [Google Scholar]
  20. Suetens, P. Fundamentals of Medical Imaging; Cambridge University Press: Cambridge, UK, 2017; ISBN 1107159784. [Google Scholar]
  21. Jeyavathana, R.B.; Balasubramanian, R.; Pandian, A.A. A survey: Analysis on pre-processing and segmentation techniques for medical images. Int. J. Res. Sci. Innov. 2016, 3, 113–120. [Google Scholar]
  22. Deepak, S.; Ameer, P.M. Brain tumor classification using deep CNN features via transfer learning. Comput. Biol. Med. 2019, 111, 103345. [Google Scholar] [CrossRef]
  23. Sachdeva, J.; Kumar, V.; Gupta, I.; Khandelwal, N.; Ahuja, C.K. Segmentation, feature extraction, and multiclass brain tumor classification. J. Digit. Imaging 2013, 26, 1141–1150. [Google Scholar] [CrossRef] [Green Version]
  24. Ismael, M.R.; Abdel-Qader, I. Brain tumor classification via statistical features and back-propagation neural network. In Proceedings of the 2018 IEEE International Conference on Electro/Information Technology (EIT), Rochester, MI, USA, 3–5 May 2018; pp. 252–257. [Google Scholar]
  25. Qin, W.; Zhao, M.; Mei, S.; Cattani, P.; Guercio, V.; Villecco, F. Realization of Single Image Super-Resolution Reconstruction Based on Wavelet Transform and Coupled Dictionary. In Proceedings of the International Conference “New Technologies, Development and Applications”, Sarajevo, Bosnia and Herzegovina, 23–25 June 2022; pp. 449–456. [Google Scholar]
  26. Jogin, M.; Madhulika, M.S.; Divya, G.D.; Meghana, R.K.; Apoorva, S. Feature extraction using convolution neural networks (CNN) and deep learning. In Proceedings of the 2018 3rd IEEE International Conference on Recent Trends in Electronics, Information & Communication Technology (RTEICT), Bangalore, India, 18–19 May 2018; pp. 2319–2323. [Google Scholar]
  27. Bar, Y.; Diamant, I.; Wolf, L.; Lieberman, S.; Konen, E.; Greenspan, H. Chest pathology identification using deep feature selection with non-medical training. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 2018, 6, 259–263. [Google Scholar] [CrossRef]
  28. Ahmed, K.B.; Hall, L.O.; Goldgof, D.B.; Liu, R.; Gatenby, R.A. Fine-tuning convolutional deep features for MRI based brain tumor classification. In Proceedings of the Medical Imaging 2017: Computer-Aided Diagnosis, Orlando, FL, USA, 13–16 February 2017; Volume 10134, p. 101342E. [Google Scholar]
  29. Rajpurkar, P.; Irvin, J.; Zhu, K.; Yang, B.; Mehta, H.; Duan, T.; Ding, D.; Bagul, A.; Langlotz, C.; Shpanskaya, K.; et al. Chexnet: Radiologist-level pneumonia detection on chest x-rays with deep learning. arXiv 2017, arXiv:1711.05225. [Google Scholar]
  30. Widhiarso, W.; Yohannes, Y.; Prakarsah, C. Brain tumor classification using gray level co-occurrence matrix and convolutional neural network. IJEIS Indones. J. Electron. Instrum. Syst. 2018, 8, 179–190. [Google Scholar] [CrossRef]
  31. Sajjad, M.; Khan, S.; Muhammad, K.; Wu, W.; Ullah, A.; Baik, S.W. Multi-grade brain tumor classification using deep CNN with extensive data augmentation. J. Comput. Sci. 2019, 30, 174–182. [Google Scholar] [CrossRef]
  32. Mei, S.; Liu, M.; Kudreyko, A.; Cattani, P.; Baikov, D.; Villecco, F. Bendlet Transform Based Adaptive Denoising Method for Microsection Images. Entropy 2022, 24, 869. [Google Scholar] [CrossRef] [PubMed]
  33. Alqahtani, A.; Zahoor, M.M.; Nasrullah, R.; Fareed, A.; Cheema, A.A.; Shahrose, A.; Irfan, M.; Alqhatani, A.; Alsulami, A.A.; Zaffar, M.; et al. Computer Aided COVID-19 Diagnosis in Pandemic Era Using CNN in Chest X-ray Images. Life 2022, 12, 1709. [Google Scholar] [CrossRef] [PubMed]
  34. Mohai, K.; Kálózi-Szabó, C.; Jakab, Z.; Fecht, S.D.; Domonkos, M.; Botzheim, J. Development of an Adaptive Computer-Aided Soft Sensor Diagnosis System for Assessment of Executive Functions. Sensors 2022, 22, 5880. [Google Scholar] [CrossRef]
  35. Haitao, H.; Cattani, P.; Guercio, V.; Villecco, F. Shearlet Transform and the Application in Image Processing. In Proceedings of the International Conference “New Technologies, Development and Applications”, Sarajevo, Bosnia and Herzegovina, 23–25 June 2022; pp. 464–470. [Google Scholar]
  36. Gopalakrishnan, K.; Khaitan, S.K.; Choudhary, A.; Agrawal, A. Deep convolutional neural networks with transfer learning for computer vision-based data-driven pavement distress detection. Constr. Build. Mater. 2017, 157, 322–330. [Google Scholar] [CrossRef]
  37. Wan, S.; Liang, Y.; Zhang, Y. Deep convolutional neural networks for diabetic retinopathy detection by image classification. Comput. Electr. Eng. 2018, 72, 274–282. [Google Scholar] [CrossRef]
  38. He, T.; Zhang, Z.; Zhang, H.; Zhang, Z.; Xie, J.; Li, M. Bag of tricks for image classification with convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 558–567. [Google Scholar]
  39. Yadav, S.S.; Jadhav, S.M. Deep convolutional neural network based medical image classification for disease diagnosis. J. Big Data 2019, 6, 113. [Google Scholar] [CrossRef] [Green Version]
  40. Wang, L.; Lin, Z.Q.; Wong, A. COVID-Net: A tailored deep convolutional neural network design for detection of COVID-19 cases from chest X-ray images. Sci. Rep. 2020, 10, 19549. [Google Scholar] [CrossRef]
  41. Li, X.; Zhu, D. Covid-xpert: An ai powered population screening of covid-19 cases using chest radiography images. arXiv PPR346252. 2020. Available online: https://europepmc.org/article/ppr/ppr346252 (accessed on 22 October 2022).
  42. Ozturk, T.; Talo, M.; Yildirim, E.A.; Baloglu, U.B.; Yildirim, O.; Acharya, U.R. Automated detection of COVID-19 cases using deep neural networks with X-ray images. Comput. Biol. Med. 2020, 121, 103792. [Google Scholar] [CrossRef] [PubMed]
  43. Ardakani, A.A.; Kanafi, A.R.; Acharya, U.R.; Khadem, N.; Mohammadi, A. Application of deep learning technique to manage COVID-19 in routine clinical practice using CT images: Results of 10 convolutional neural networks. Comput. Biol. Med. 2020, 121, 103795. [Google Scholar] [CrossRef] [PubMed]
  44. Zhu, H.; Wei, L.; Niu, P. The novel coronavirus outbreak in Wuhan, China. Glob. Health Res. Policy 2020, 5, 6. [Google Scholar] [CrossRef] [PubMed]
  45. Li, A.C.; Lee, D.T.; Misquitta, K.K.; Uno, K.; Wald, S. COVID-19 detection from chest radiographs using machine learning and convolutional neural networks. medRxiv 2020. [Google Scholar] [CrossRef]
  46. Cheng, J.; Yang, W.; Huang, M.; Huang, W.; Jiang, J.; Zhou, Y.; Yang, R.; Zhao, J.; Feng, Y.; Feng, Q. Retrieval of brain tumors by adaptive spatial pooling and fisher vector representation. PLoS ONE 2016, 11, e0157112. [Google Scholar] [CrossRef] [Green Version]
  47. Shrestha, A.; Mahmood, A. Review of deep learning algorithms and architectures. IEEE Access 2019, 7, 53040–53065. [Google Scholar] [CrossRef]
  48. Dong, G.; Liao, G.; Liu, H.; Kuang, G. A review of the autoencoder and its variants: A comparative perspective from target recognition in synthetic-aperture radar images. IEEE Geosci. Remote Sens. Mag. 2018, 6, 44–68. [Google Scholar] [CrossRef]
  49. Dorigo, M.; Birattari, M.; Stutzle, T. Ant colony optimization. IEEE Comput. Intell. Mag. 2006, 1, 28–39. [Google Scholar] [CrossRef]
  50. Dorigo, M.; Socha, K. An introduction to ant colony optimization. In Handbook of Approximation Algorithms and Metaheuristics, 2nd ed.; Chapman and Hall/CRC: Boca Raton, FL, USA, 2018; pp. 395–408. ISBN 1351236423. [Google Scholar]
  51. Dorigo, M.; Blum, C. Ant colony optimization theory: A survey. Theor. Comput. Sci. 2005, 344, 243–278. [Google Scholar] [CrossRef]
  52. Kanan, H.R.; Faez, K.; Taheri, S.M. Feature selection using ant colony optimization (ACO): A new method and comparative study in the application of face recognition system. In Proceedings of the Industrial Conference on Data Mining, Leipzig, Germany, 14–18 July 2007; pp. 63–76. [Google Scholar]
  53. Tabakhi, S.; Moradi, P.; Akhlaghian, F. An unsupervised feature selection algorithm based on ant colony optimization. Eng. Appl. Artif. Intell. 2014, 32, 112–123. [Google Scholar] [CrossRef]
  54. Sivagaminathan, R.K.; Ramakrishnan, S. A hybrid approach for feature subset selection using neural networks and ant colony optimization. Expert Syst. Appl. 2007, 33, 49–60. [Google Scholar] [CrossRef]
  55. Yu, L.; Liu, H. Efficient feature selection via analysis of relevance and redundancy. J. Mach. Learn. Res. 2004, 5, 1205–1224. [Google Scholar]
  56. Bennasar, M.; Hicks, Y.; Setchi, R. Feature selection using joint mutual information maximisation. Expert Syst. Appl. 2015, 42, 8520–8532. [Google Scholar] [CrossRef] [Green Version]
  57. Hoque, N.; Bhattacharyya, D.K.; Kalita, J.K. MIFS-ND: A mutual information-based feature selection method. Expert Syst. Appl. 2014, 41, 6371–6385. [Google Scholar] [CrossRef]
  58. Kabir, M.; Shahjahan, M.; Murase, K.; Barbosa, H.J.C. Ant colony optimization toward feature selection. In Ant Colony Optimization—Techniques and Applications; IntechOpen: London, UK, 2013; pp. 1–43. [Google Scholar]
  59. Cheng, J.; Huang, W.; Cao, S.; Yang, R.; Yang, W.; Yun, Z.; Wang, Z.; Feng, Q. Enhanced performance of brain tumor classification via tumor region augmentation and partition. PLoS ONE 2015, 10, e0140381. [Google Scholar] [CrossRef]
  60. Thejaswini, P.; Bhat, M.B.; Prakash, M.K. Detection and classification of tumour in brain MRI. Int. J. Eng. Manufact.(IJEM) 2019, 9, 11–20. [Google Scholar]
  61. Afshar, P.; Mohammadi, A.; Plataniotis, K.N. Brain tumor type classification via capsule networks. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 3129–3133. [Google Scholar]
  62. Belaid, O.N.; Loudini, M. Classification of Brain Tumor by Combination of Pre-Trained VGG16 CNN. J. Inf. Technol. Manag. 2020, 12, 13–25. [Google Scholar]
  63. Ghassemi, N.; Shoeibi, A.; Rouhani, M. Deep neural network with generative adversarial networks pre-training for brain tumor classification based on MR images. Biomed. Signal Process. Control 2020, 57, 101678. [Google Scholar] [CrossRef]
  64. Özkaya, U.; Öztürk, Ş.; Barstugan, M. Coronavirus (COVID-19) classification using deep features fusion and ranking technique. In Big Data Analytics and Artificial Intelligence against COVID-19: Innovation Vision and Approach; Springer: Berlin/Heidelberg, Germany, 2020; pp. 281–295. [Google Scholar]
  65. Mollalo, A.; Rivera, K.M.; Vahedi, B. Artificial neural network modeling of novel coronavirus (COVID-19) incidence rates across the continental United States. Int. J. Environ. Res. Public Health 2020, 17, 4204. [Google Scholar] [CrossRef]
  66. Sethy, P.K.; Behera, S.K. Detection of coronavirus disease (COVID-19) based on deep features. Preprints 2020. [Google Scholar] [CrossRef] [Green Version]
  67. Gozes, O.; Frid-Adar, M.; Sagie, N.; Zhang, H.; Ji, W.; Greenspan, H. Coronavirus detection and analysis on chest ct with deep learning. arXiv 2020, arXiv:2004.02640. [Google Scholar]
  68. Yeşilkanat, C.M. Spatio-temporal estimation of the daily cases of COVID-19 in worldwide using random forest machine learning algorithm. Chaos Solitons Fractals 2020, 140, 110210. [Google Scholar] [CrossRef] [PubMed]
  69. Khoshbakhtian, F.; Ashraf, A.B.; Khan, S.S. Covidomaly: A deep convolutional autoencoder approach for detecting early cases of covid-19. arXiv 2020, arXiv:2010.02814. [Google Scholar]
  70. Khozeimeh, F.; Sharifrazi, D.; Izadi, N.H.; Joloudari, J.H.; Shoeibi, A.; Alizadehsani, R.; Gorriz, J.M.; Hussain, S.; Sani, Z.A.; Moosaei, H. CNN AE: Convolution Neural Network combined with Autoencoder approach to detect survival chance of COVID 19 patients. arXiv 2021, arXiv:2104.08954. [Google Scholar]
  71. Afshar, P.; Heidarian, S.; Naderkhani, F.; Oikonomou, A.; Plataniotis, K.N.; Mohammadi, A. Covid-caps: A capsule network-based framework for identification of covid-19 cases from x-ray images. Pattern Recognit. Lett. 2020, 138, 638–643. [Google Scholar] [CrossRef]
  72. de Oliveira, S.M.; Bezerra, L.C.T.; Stützle, T.; Dorigo, M.; Wanner, E.F.; de Souza, S.R. A computational study on ant colony optimization for the traveling salesman problem with dynamic demands. Comput. Oper. Res. 2021, 135, 105359. [Google Scholar] [CrossRef]
  73. Ning, J.; Zhang, Q.; Zhang, C.; Zhang, B. A best-path-updating information-guided ant colony optimization algorithm. Inf. Sci. 2018, 433, 142–162. [Google Scholar] [CrossRef]
  74. Shyu, S.J.; Yin, P.-Y.; Lin, B.M.T. An ant colony optimization algorithm for the minimum weight vertex cover problem. Ann. Oper. Res. 2004, 131, 283–304. [Google Scholar] [CrossRef]
  75. Latha, Y.L.M.; Prasad, M.V.N.K. GLCM based texture features for palmprint identification system. In Computational Intelligence in Data Mining-Volume 1; Springer: Berlin/Heidelberg, Germany, 2015; pp. 155–163. [Google Scholar]
  76. Gárate-Escamila, A.K.; El Hassani, A.H.; Andrès, E. Classification models for heart disease prediction using feature selection and PCA. Informatics Med. Unlocked 2020, 19, 100330. [Google Scholar] [CrossRef]
  77. Huang, C.-R.; Lee, L.-H. Contrastive approach towards text source classification based on top-bag-of-word similarity. In Proceedings of the 22nd Pacific Asia Conference on Language, Information and Computation, Cebu City, Philippines, 20–22 November 2008; pp. 404–410. [Google Scholar]
  78. Rui, W.; Xing, K.; Jia, Y. BOWL: Bag of word clusters text representation using word embeddings. In Proceedings of the International Conference on Knowledge Science, Engineering and Management, Passau, Germany, 5–7 October 2016; pp. 3–14. [Google Scholar]
  79. Ahmed, S.; Frikha, M.; Hussein, T.D.H.; Rahebi, J. Face Recognition System using Histograms of Oriented Gradients and Convolutional Neural Network based on with Particle Swarm Optimization. In Proceedings of the 2021 International Conference on Electrical, Communication, and Computer Engineering (ICECCE), Kuala Lumpur, Malaysia, 12–13 June 2021; pp. 1–5. [Google Scholar]
  80. Ahmed, S.; Frikha, M.; Hussein, T.D.H.; Rahebi, J. Optimum Feature Selection with Particle Swarm Optimization to Face Recognition System Using Gabor Wavelet Transform and Deep Learning. Biomed Res. Int. 2021, 2021, 6621540. [Google Scholar] [CrossRef]
  81. Allagwail, S.; Gedik, O.S.; Rahebi, J. Face recognition with symmetrical face training samples based on local binary patterns and the Gabor filter. Symmetry 2019, 11, 157. [Google Scholar] [CrossRef] [Green Version]
  82. Wu, X.; Liu, S.; Cao, Y.; Li, X.; Yu, J.; Dai, D.; Ma, X.; Hu, S.; Wu, Z.; Liu, X.; et al. Speech emotion recognition using capsule networks. In Proceedings of the ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 6695–6699. [Google Scholar]
  83. Alashik, K.M.; Yildirim, R. Human Identity Verification From Biometric Dorsal Hand Vein Images Using the DL-GAN Method. IEEE Access 2021, 9, 74194–74208. [Google Scholar] [CrossRef]
  84. Hussin, S.H.S.; Yildirim, R. StyleGAN-LSRO Method for Person Re-Identification. IEEE Access 2021, 9, 13857–13869. [Google Scholar] [CrossRef]
Figure 1. Three classes of COVID-19 dataset: (left) COVID-19, (middle) Pneumonia, and (right) Normal.
Figure 1. Three classes of COVID-19 dataset: (left) COVID-19, (middle) Pneumonia, and (right) Normal.
Sensors 22 08999 g001
Figure 2. Three classes of brain tumor dataset: (right) Meningioma, (middle) Glioma, and (left) Pituitary Brain Tumors.
Figure 2. Three classes of brain tumor dataset: (right) Meningioma, (middle) Glioma, and (left) Pituitary Brain Tumors.
Sensors 22 08999 g002
Figure 3. Working method of the proposed feature selection model.
Figure 3. Working method of the proposed feature selection model.
Sensors 22 08999 g003
Figure 4. Relationship between information coefficient and entropy.
Figure 4. Relationship between information coefficient and entropy.
Sensors 22 08999 g004
Figure 5. Image dataset pre-processing steps.
Figure 5. Image dataset pre-processing steps.
Sensors 22 08999 g005
Figure 6. Overall framework of the proposed classification system.
Figure 6. Overall framework of the proposed classification system.
Sensors 22 08999 g006
Table 1. The characteristics of classifier types.
Table 1. The characteristics of classifier types.
ClassifierPrediction SpeedMemory UsageInterpretability
Decision Tree (DT)FastSmallEasy
Support Vector Machines (SVM)Medium for linear
Slow for others
Medium for linear
All others: medium for multiclass, large for binary
Easy for Linear SVM
Hard for all other kernel types
Kernel Nearest Neighbor (KNN)Slow for cubic
Medium for others
MediumHard
EnsembleFast to medium depending on choice of algorithmLow to high depending on choice of algorithmHard
Naïve BayesMedium for simple distributions
Slow for kernel distributions or high-dimensional data
Small for simple distributions
Medium for kernel distributions or high-dimensional data
Easy
Discriminant
Analysis
FastSmall for linear, large for quadraticEasy
Table 2. Evaluation metrics.
Table 2. Evaluation metrics.
ParameterAbbreviationMathematic Expression
Sensitivity, Recall, hit rate, True Positive Rate(TPR) T P T P + F N = 1 F N R
Specificity, Selectivity, True Negative Rate(TNR) T N T N + F P = 1 F P R
Precision, Positive Predictive Value(PPV) T P T P + F P = 1 F D R
Negative Predictive Value(NPV) T N T N + F N   = 1 F O R
Accuracy(ACC) T P + T N ( T P + T N + F P + F N )
F1 Score(F1 Score) 2     ( P P V     T P R )   ( P P V     T P R ) = 2 T P   2 T P + F P + F N  
Misclassification rate(MR) F P + F N T P + T N + F P + F N  
Matthews correlation coefficient (MCC)(MCC) T P     T N F P     F N ( T P + F P ) ( T P + F N ) ( T N + F P ) ( T N + F N )
Table 3. Auto-encoder with ACO as feature selection-extraction method.
Table 3. Auto-encoder with ACO as feature selection-extraction method.
MethodDatasetACCTPRTNRPPVNPVF1-ScoreMRMCC
DECISION TREECOVID-1997.7488.4998.6386.2098.8987.332.250.86
Brain Tumor92.8288.1794.8488.1794.8488.177.170.83
SVMCOVID-1998.6898.0598.7387.0699.8292.231.310.92
Brain Tumor96.5793.1998.1095.6996.9594.423.420.92
KNNCOVID-1997.7489.1898.5585.3498.9787.222.250.86
Brain Tumor99.1897.8899.7699.4699.0698.660.810.98
ENSEMBLECOVID-1998.2195.1498.4884.4899.5789.491.780.89
Brain Tumor95.7593.9596.5191.9397.4292.934.240.90
NAIVE BAYESCOVID-1989.2845.4599.4294.8288.7361.4510.710.61
Brain Tumor78.6259.3296.5494.0871.8972.7621.370.61
DISCRIMINANTCOVID-1998.4498.0098.4884.4899.8290.741.550.90
Brain Tumor97.0694.6898.1195.6997.6595.182.930.93
Table 4. Several pre-trained CNN with ACO as a feature selection method for COVID-19 dataset.
Table 4. Several pre-trained CNN with ACO as a feature selection method for COVID-19 dataset.
Pretrained CNN with ACO (COVID-19 Dataset)
ClassifiersACCTPRTNRPPVNPVF1-SCOREMRMCC
Pre-trained CNN (AlexNet) + ACO
DECISION TREE96.1177.9697.9479.3197.7878.633.880.77
SVM98.9998.1399.0690.5199.8294.171.000.94
KNN98.6898.8698.5785.3499.9892.091.310.92
ENSEMBLE98.5298.9898.4884.4899.9191.161.470.91
NAÏVE BAYES94.9565.0899.4694.8294.9677.195.040.76
DISCRIMINANT99.5399.6299.4994.8299.8297.340.460.97
Pre-trained CNN (GoogleNet) + ACO
DECISION TREE97.0484.8298.2181.8998.5483.332.950.82
SVM98.9198.1198.9889.6599.8293.691.080.93
KNN98.6899.0098.6586.2099.9192.161.310.92
ENSEMBLE97.6795.7497.8277.5899.6585.712.320.85
NAÏVE BAYES98.4491.3799.1491.3799.1491.371.550.91
DISCRIMINANT98.8398.0998.9088.7999.8293.211.160.93
Pre-trained CNN (ResNet 50) + ACO
DECISION TREE97.5191.1798.0680.1799.2385.322.480.84
SVM99.6199.8999.5795.6899.9997.790.380.98
KNN98.9199.0398.9088.7999.9193.631.080.93
ENSEMBLE99.1499.2399.0790.5199.9795.020.850.95
NAÏVE BAYES99.1492.0099.9199.1399.1495.430.850.95
DISCRIMINANT99.2199.3298.8999.0199.2196.200.790.95
Pre-trained CNN (DenseNet 201) + ACO
DECISION TREE97.5992.9297.9879.3199.4085.582.400.85
SVM98.9999.0498.9889.6599.9194.111.000.94
KNN98.7599.3798.6586.2099.9092.591.240.92
ENSEMBLE98.8399.5698.7387.0699.8693.081.160.93
NAÏVE BAYES99.1494.1199.6596.5599.4095.310.850.95
DISCRIMINANT98.9999.5798.8892.3499.8093.121.010.94
Table 5. Several pre-trained CNN with ACO as a feature selection method for brain tumor dataset.
Table 5. Several pre-trained CNN with ACO as a feature selection method for brain tumor dataset.
Pretrained CNN with ACO (Brain Tumor Dataset)
ClassifiersACCTPRTNRPPVNPVF1-SCOREMRMCC
Pre-trained CNN (AlexNet) + ACO
DECISION TREE91.6886.0994.1386.5593.9186.328.310.80
SVM98.5396.3399.5298.9298.3697.611.460.97
KNN98.2096.2999.0597.8498.3697.061.790.96
ENSEMBLE96.7393.6898.1095.6997.1894.683.260.92
NAÏVE BAYES89.5574.7999.4598.9285.4885.1810.440.79
DISCRIMINANT98.6995.8799.9299.9598.1297.891.300.97
Pre-trained CNN (GoogleNet) + ACO
DECISION TREE87.2781.7689.3974.7392.7478.0812.720.69
SVM95.2694.8595.4389.2497.8991.964.730.89
KNN94.9490.1597.1493.5495.5591.825.050.88
ENSEMBLE93.8091.5794.7187.6396.4889.566.190.85
NAÏVE BAYES91.1982.6795.3789.7891.8086.088.800.80
DISCRIMINANT96.7393.2298.3396.2396.9594.703.260.92
Pre-trained CNN (ResNet 50) + ACO
DECISION TREE88.0979.5891.9481.7290.8680.6311.900.72
SVM97.5598.3097.2493.5499.2995.862.440.94
KNN97.8795.7698.8297.3198.1296.532.120.95
ENSEMBLE96.2495.5396.5491.9398.1293.693.750.91
NAÏVE BAYES91.0279.1197.9395.6988.9986.618.970.81
DISCRIMINANT96.2494.5396.9793.0197.6593.763.750.91
Pre-trained CNN (DenseNet 201) + ACO
DECISION TREE91.5186.4193.7085.4894.1485.948.480.80
SVM98.2096.2999.0597.8498.3697.061.790.96
KNN97.7195.2698.8197.3197.8996.272.280.95
ENSEMBLE97.2294.2498.5796.7797.4295.492.770.94
NAÏVE BAYES93.9687.8196.8793.0194.3790.336.030.86
DISCRIMINANT95.2293.8496.0588.5490.4191.454.780.89
Table 6. Brain tumor classification results comparison.
Table 6. Brain tumor classification results comparison.
MethodAcc (%)
GLCM + CNN [30]82.00
Intensity and texture features + PCA + ANN [23]91.00
BOW + SVM [59]91.28
2D Discrete Wavelet transform (DWT) and 2D Gabor filter [24]95.66
SVM+ANN [60]91.40
Deep CNN+SVM [22]98.00
Deep Convolution Neural Network VGG 19 + SoftMax classifier [31]94.58
Capsule networks (CapsNets) [61]86.56
GLCM + Pre-Trained VGG16 CNN [62]96.50
GAN + ConvNet (random split) [63]95.60
Auto-encoder + KNN (without ACO)97.22
Proposed Method (Auto-encoder + ACO + KNN)99.18
Table 7. COVID-19 classification results comparison. The results that are mentioned in the table are obtained from the literature and the authors used the same dataset that we used for the Brain tumor.
Table 7. COVID-19 classification results comparison. The results that are mentioned in the table are obtained from the literature and the authors used the same dataset that we used for the Brain tumor.
MethodAcc (%)
CNN with data fusion [64]98.27
ANN [65]83.98
DNN + SVM [66]95.33
DNN [67]94.80
Random Forest [68]95.90
Stacked-auto-encoder [9]94.70
Deep Convolutional Auto-encoder [69]76.52
CNN + Auto-encoder [70]96.05
Tailored CNN [40]92.30
Dense Net [41]88.90
Capsule Networks [71]95.70
DarkNet-19 based CNN [42]87.02
Deep Learning [42]98.08
Deep Learning [43]86.27
CNN + SVM (without ACO)96.45
Proposed Method (CNN + ACO + SVM)99.61
Table 8. Comparison between the work of ant colony optimization (ACO) vs. the work of genetic algorithm (GA) as a feature selection (accuracy and time).
Table 8. Comparison between the work of ant colony optimization (ACO) vs. the work of genetic algorithm (GA) as a feature selection (accuracy and time).
DatasetCombined MethodsACOGA
ClassifiersAccuracyTime (h)ClassifiersAccuracyTime (h)
Covid-19Auto-encoderSVM98.68%0:27:00KNN97.98%1:00:00
CNN (AlexNet)DISCRIMINANT99.53%1:17:00KNN98.60%2:07:00
CNN (GoogleNet)SVM98.91%0:21:00NAÏVE BAYES98.13%0:40:00
CNN (ResNet 50)SVM99.61%0:43:00KNN98.60%1:05:00
CNN (DenseNet 201)NAÏVE BAYES99.14%0:39:00KNN98.75%1:04:00
Brain TumorAuto-encoderKNN99.18%0:11:00ENSEMBLE96.24%0:16:00
CNN (AlexNet)DISCRIMINANT98.69%0:12:00ENSEMBLE94.61%0:30:00
CNN (GoogleNet)DISCRIMINANT96.73%0:05:00KNN93.96%0:09:00
CNN (ResNet 50)KNN97.87%0:11:00ENSEMBLE97.06%0:16:00
CNN (DenseNet 201)SVM98.20%0:09:00ENSEMBLE96.57%0:18:00
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kadhim, Y.A.; Khan, M.U.; Mishra, A. Deep Learning-Based Computer-Aided Diagnosis (CAD): Applications for Medical Image Datasets. Sensors 2022, 22, 8999. https://doi.org/10.3390/s22228999

AMA Style

Kadhim YA, Khan MU, Mishra A. Deep Learning-Based Computer-Aided Diagnosis (CAD): Applications for Medical Image Datasets. Sensors. 2022; 22(22):8999. https://doi.org/10.3390/s22228999

Chicago/Turabian Style

Kadhim, Yezi Ali, Muhammad Umer Khan, and Alok Mishra. 2022. "Deep Learning-Based Computer-Aided Diagnosis (CAD): Applications for Medical Image Datasets" Sensors 22, no. 22: 8999. https://doi.org/10.3390/s22228999

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop