Next Article in Journal
Inferior Vena Cava (IVC) Resections without Reconstruction in Renal Tumors: Two Case Reports
Next Article in Special Issue
Machine Learning Assisting the Prediction of Clinical Outcomes following Nucleoplasty for Lumbar Degenerative Disc Disease
Previous Article in Journal
Analysis of Temperatures Generated during Conventional Laser Irradiation of Root Canals—A Finite Element Study
Previous Article in Special Issue
Classification of Monkeypox Images Using LIME-Enabled Investigation of Deep Convolutional Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid Models for Endoscopy Image Analysis for Early Detection of Gastrointestinal Diseases Based on Fused Features

by
Ibrahim Abdulrab Ahmed
1,*,
Ebrahim Mohammed Senan
2,* and
Hamzeh Salameh Ahmad Shatnawi
1
1
Computer Department, Applied College, Najran University, Najran 66462, Saudi Arabia
2
Department of Artificial Intelligence, Faculty of Computer Science and Information Technology, Alrazi University, Sana’a, Yemen
*
Authors to whom correspondence should be addressed.
Diagnostics 2023, 13(10), 1758; https://doi.org/10.3390/diagnostics13101758
Submission received: 13 March 2023 / Revised: 12 May 2023 / Accepted: 13 May 2023 / Published: 16 May 2023
(This article belongs to the Special Issue Medical Image Processing and Analysis)

Abstract

:
The gastrointestinal system contains the upper and lower gastrointestinal tracts. The main tasks of the gastrointestinal system are to break down food and convert it into essential elements that the body can benefit from and expel waste in the form of feces. If any organ is affected, it does not work well, which affects the body. Many gastrointestinal diseases, such as infections, ulcers, and benign and malignant tumors, threaten human life. Endoscopy techniques are the gold standard for detecting infected parts within the organs of the gastrointestinal tract. Endoscopy techniques produce videos that are converted into thousands of frames that show the disease’s characteristics in only some frames. Therefore, this represents a challenge for doctors because it is a tedious task that requires time, effort, and experience. Computer-assisted automated diagnostic techniques help achieve effective diagnosis to help doctors identify the disease and give the patient the appropriate treatment. In this study, many efficient methodologies for analyzing endoscopy images for diagnosing gastrointestinal diseases were developed for the Kvasir dataset. The Kvasir dataset was classified by three pre-trained models: GoogLeNet, MobileNet, and DenseNet121. The images were optimized, and the gradient vector flow (GVF) algorithm was applied to segment the regions of interest (ROIs), isolating them from healthy regions and saving the endoscopy images as Kvasir-ROI. The Kvasir-ROI dataset was classified by the three pre-trained GoogLeNet, MobileNet, and DenseNet121 models. Hybrid methodologies (CNN–FFNN and CNN–XGBoost) were developed based on the GVF algorithm and achieved promising results for diagnosing disease based on endoscopy images of gastroenterology. The last methodology is based on fused CNN models and their classification by FFNN and XGBoost networks. The hybrid methodology based on the fused CNN features, called GoogLeNet–MobileNet–DenseNet121–XGBoost, achieved an AUC of 97.54%, accuracy of 97.25%, sensitivity of 96.86%, precision of 97.25%, and specificity of 99.48%.

1. Introduction

The gastrointestinal system consists of internal body organs. The system is divided into the upper system, which includes the mouth cavity, esophagus, and stomach, and the lower system, which contains the small intestine, colon, and rectum. The primary task of the gastrointestinal system is to break down food into essential elements for the body and to expel waste in the form of feces [1]. If any organ is disturbed, its functions do not work well, which affects the body. Many gastrointestinal disorders exist, such as infections, ulcers, polyps, and malignant tumors [2]. Cancers of the stomach, esophagus, colon, and rectum constitute about 3 million new cancer cases per year, and about 2 million people die annually from these diseases [3]. Malignant tumors in the upper and lower gastrointestinal tract threaten human life. For the upper gastrointestinal tract, esophageal cancer is the least prevalent, whereas stomach cancer is the fifth most common type of cancer and the third most common cause of death [4]. In contrast, malignant tumors of the lower gastrointestinal tract are the second most common cause of death and the third most common type of cancer [5]. There are many imaging techniques for the organs of the gastrointestinal tract. Wireless capsule endoscopy (WCE) [6] is a wireless capsule-sized camera that is swallowed through the mouth and takes thousands of images, starting from the pharynx until it reaches the rectum. This technique is good because it reaches all places of the gastrointestinal tract, such as the small intestine, but it takes two days to pass to the end of the gastrointestinal system. However, there are some problems with WCE, the device can be difficult to swallow and pass normally and can get stuck in the narrow places in the intestines and cause an obstruction. Thus, endoscopy is the gold standard for detecting diseases of the gastrointestinal system [7]. An upper GI endoscopy examines the stomach, esophagus, and the beginning of the small intestine. In contrast, a lower gastrointestinal endoscopy covers the colon and rectum. Endoscopy is a video to discover the gastrointestinal internal parts using high-resolution digital endoscopes [8]. Endoscopic examinations require expensive equipment and a skilled specialist. Detection by endoscopy is imperative for colorectal cancer to be removed. Therefore, detecting benign tumors is necessary to remove them before they develop into a malignant type. However, the diagnostic ability of endoscopy varies according to the experience of doctors, which may result in colon and rectal cancer. Accurate diagnosis of diseases is vital to avoid malignancy and to survive. Esophageal infections are a normal and pathological condition, and early detection of esophagitis is necessary before complications lead to ulcers, stenosis, and bleeding [9]. If left untreated it develops into cancer. Therefore, a distinction must be made between the esophagus and the z-line. Endoscopy technology produces a video that converts to thousands of frames, increasing the burden of the doctors to keep track of all the video frames. The frames appear healthy; the disease (lesion, ulcer, or bleeding) does not occur except in a few frames. Therefore, this technique requires examining all frames, which takes a long time and is a tedious process [10]. Computer-aided diagnostic methods are used to support clinicians and experts in addressing the challenge of manual insufficiency. With the advancement in AI technologies, many researchers have focused on its application in healthcare, including biomedical image processing. Artificial intelligence techniques are used to detect gastrointestinal diseases and develop systems that can detect infections, ulcers, bleeding, polyps, colon cancer, and stomach cancer. The application of automated techniques is expected to improve the efficiency of diagnosis, reduce the gap between the number of doctors and patients, and reduce diagnostic costs. Many researchers have proposed deep learning techniques for analyzing and diagnosing medical images, including CNN models. CNN models are a good research topic in medical imaging diagnostics to help radiologists and medical specialists provide high-quality healthcare. CNN is an integrated model with dozens of layers to analyze thousands of endoscopy images and extract accurate feature maps to detect infections, ulcers, bleeding, polyps, and colorectal cancer. In this study, several effective systems for diagnosing endoscopy images of the gastrointestinal tract were developed. Due to the similarity of the clinical signs between infections and ulcers and the need for early detection of polyps, this study focused on developing hybrid techniques between deep and machine learning. CNN models have superior capabilities to extract features that are not visible to the naked eye, thus extracting features from many CNNs and integrating feature maps vectors CNN.
Artificial intelligence (AI) has made significant contributions to medical image processing, revolutionizing the field with its ability to analyze and interpret complex medical images. Some of the most important applications of AI in this domain include:
Classification of medical images: AI algorithms can analyze medical images such as endoscopy, X-rays, CT scans, and MRI scans to classify them into different categories, which helps in detecting diseases and abnormalities. For example, AI can identify cancerous tumors or diagnose certain conditions such as gastrointestinal diseases, pneumonia or retinal diseases, heart diseases, and skin diseases.
Computer-Aided Detection/Diagnosis (CAD): AI-based CAD systems can act as a “second opinion” by flagging potential abnormalities in medical images. These systems assist radiologists in the detection of tumors, lesions, or other abnormalities that may be missed or difficult to spot. CAD systems help reduce human error and improve diagnostic accuracy.
Segmentation and Annotation: AI techniques can segment medical images, separating different structures or organs for further analysis. This enables precise measurements and quantitative analysis. AI can also assist in annotating medical images, marking specific regions of interest for radiologists or clinicians.
Disease Detection and Diagnosis: AI algorithms assist in detecting diseases and conditions by analyzing medical images. They can learn patterns and features associated with various diseases, aiding in early detection and accurate diagnosis. For example, AI could help identify signs of gastrointestinal disease in endoscopic images, Alzheimer’s disease in brain scans, or detect diabetic retinopathy in retinal images.
Image Reconstruction and Enhancement: AI can reconstruct or enhance medical images to improve their quality or extract additional information. This can be particularly useful when dealing with noisy or low-resolution images. AI algorithms can fill in missing data or enhance image details, assisting radiologists, endoscopists, and physicians in making informed decisions.
Prognosis and Predictive Analytics: AI can analyze medical images to predict patient outcomes and provide prognostic information. By identifying patterns and features in images, AI algorithms can estimate disease progression, survival rates, or response to treatment, aiding in personalized patient care.
These applications of AI in medical image processing have the potential to enhance diagnostic accuracy, improve patient outcomes, and streamline healthcare workflows by reducing the time and effort required for manual analysis. However, it is important to note that AI systems should always be used in conjunction with human expertise and not as a replacement for clinical decision making.
The main contribution of this study is as follows:
  • Classification of endoscopy images of gastrointestinal diseases by a hybrid methodology CNN-FFNN and CNN-XGBoost based on the GVF segmentation algorithm.
  • Fusion of features of CNN models and their classification by FFNN and XGBoost based on the GVF segmentation algorithm.
The rest of this paper is organized as follows: Section 2 discusses relevant techniques and results from the literature. Section 3 explains the methodologies and tools applied to analyze endoscopy images from gastroenterology. Section 4 presents the results of the proposed methodologies. Section 5 discusses the performance of the methodologies and compares their results. Section 6 concludes the study.

2. Related Work

Öztürk et al. [11] developed three CNN models for classifying gastroenterological endoscopy images based on combining LSTM layers. The features are extracted from each CNN model and sent to an LSTM layer to evaluate the contribution to a dataset with various image counts. Imran et al. [12] propose a DCNN architecture with multiple pathways and layers to improve the efficiency of detecting endoscopy image abnormalities of the human gastrointestinal tract. The network works on two tracks with different resolutions, each with four units. Different degrees of image resolution may be useful in detecting gastrointestinal tract lesions. Yoshiok et al. [13] reported comparison of the performance of four CNN models to the analysis of endoscopy images to detect of esophagitis. The GoogLeNet achieved an F1-score that was better than the other models, whereas MobileNet V3 achieved a better average positive rate. Zahra et al. [14] presented a system for detecting abnormalities in WCE images of the gastrointestinal tract. The images were enhanced, the region of interest was extracted by the thresholding method, and the color, shape, and texture features were extracted. Finally, the features were fed to SVM to classify the extracted features. Shima et al. [15] described a method for detecting the anatomical features of endoscopy images of the gastrointestinal tract by supervised CNN and their comparison with semi-supervised CNN. The results showed that the supervised CNN achieved better results than the semi-supervised CNN. Ibtesam et al. [16] studied the retraining of five pre-trained CNN models to classify a Kvasir dataset containing eight classes, including an anatomical parameter, disease condition, and medical procedure. Inception-v3 and ResNet achieved 90% and 97.6% accuracy. Xing et al. [17] reported on a hybrid learning framework for tagging in medical endoscopy images and its fusion with deep learning. Experiments demonstrated the superiority of the framework for classifying the Hyper-Kvasir dataset with an accuracy of 95%. Ranit et al. [18] proposed a lightweight Mobile-PolypNet for the segmentation of colorectal polyps. The network uses a set of blocks of the autoencoder to improve the system’s efficiency. The network was trained and tested on the Kvasir dataset, achieving a DICE score of 93.5%. Subhashree et al. [19] developed an intelligent system for detecting abnormalities in GI endoscopy images based on CNN and time–frequency analysis. The images were improved, and separate waveform coefficients were extracted and submitted to two CNN models for training and testing at two levels. The system achieved an accuracy of 93.75% in the first level. Ali et al. [20] used pre-trained CNNs with filters to improve endoscopy images to detect GI disease. The preprocessing stage was combined with the classification methods of the Kvasir dataset, which achieved an accuracy of 90.17%. Debesh et al. [21] presented a classic U-Net architecture for hashing the Kvasir-Instrument dataset, which achieved a Jaccard index of 85.78% and a dice factor of 91.58%. Ramzan et al. [22] described a new framework for image preprocessing, integrating texture features, LBP, and deep learning serially to improve diagnostic imaging of gastrointestinal diseases. The improved features were selected by PCA and entropy. Yogapriya et al. [23] developed a model that integrates traditional algorithms and data augmentation with CNNs to classify gastrointestinal diseases. The model with the VGG16 network achieved better results than the rest of the CNN networks, achieving an accuracy of 96.33% and an F1 score of 96.5%. Muhammad et al. [24] reported on two deep learning models based on the hybrid crow–moth method for classifying gastrointestinal disease datasets. The features were extracted using average layers and then combined with crow–moth features. Muhammad et al. [25] used contrast enhancement technology to improve images and segmentation of the region of interest by saliency map, feeding regions of interest to MobileNet-V2 for training and feature extraction. The authors applied a hybrid whale method to select the best features.
The researchers devoted their efforts to developing a modern technique to obtain good results for diagnosing abnormalities in diseases of the gastrointestinal tract. The characteristics of gastrointestinal disorders are similar in the early stages, so this study focused on extracting the attributes from the region of interest from many methods and integrating them to achieve satisfactory results for the early detection of gastrointestinal diseases.

3. Materials and Methods

3.1. Description of the Kvasir Dataset

In this study, the performances of the systems were evaluated on endoscopy images from the available Kvasir dataset. Endoscopy images were obtained by the Vestre Viken Health Trust (VV) of four hospitals and the Cancer Registry of Norway (CRN), Norway. CRN provides knowledge about cancer to prevent its spread and is part of a regional body for Norway under the supervision of the Oslo Hospital. The Kvasir dataset images were classified into eight types divided into three anatomical landmarks (ALs), three types of pathological findings (PFs), and two types of polyps removal (PR). Endoscopy experts have classified and annotated all images. The dataset consists of 8000 images distributed evenly among eight classes. LA include three types: cecum and Z-line pylorus. LA within the gastrointestinal tract can be identified endoscopically as marker of inflammation and ulcers. In contrast, PFs include esophagitis, ulcerative colitis, and polyps. PFs are abnormal signs of the gastrointestinal tract that are identified endoscopically [26]. These clinical characteristics may be signs of an ongoing disease or an early cancer stage. PR includes two types: dyed and lifted polyps and dyed resection margins, which are polyps that may be a harbinger of cancer. All images of the Kvasir dataset are in RGB color with resolutions ranging from 720 × 576 to 1920 × 1072.

3.2. Enhancement of the Images of the Kvasir Dataset

When acquiring images through the endoscope, some artifacts negatively affect the image enhancement stages and lead to incorrect results. The artifacts comprise air bubbles, food residues in the upper digestive tract, stool residues in the lower digestive tract, mucous membranes, and fluids [27]. All these artifacts are considered features in the next stages, and they are artifacts and have nothing to do with the disease. Additionally, the low contrast between the borders of the disease and the healthy part presents a challenge to extract and analyze the diseased part by extracting its features. Thus, removing artifacts is essential for proper feature extraction, and increased low variance is necessary for efficient segmentation of a region of interest. This study calculated the mean achromatic area for each RGB channel. After that, an averaging filter was applied to improve the images and remove artifacts, followed by increasing the contrast of the low area between the infected and the healthy regions by a contrast-limited adaptive histogram equalization (CLAHE) method [28].
First, to refine the endoscopy images of the Kvasir dataset, each image was passed through an averaging filter with an operator of size 6 × 6. The image is processed pixel by pixel, called the target pixel, based on the value of 35 adjacent pixels. The filter continues to run until the last pixel in the image is processed, as in Equation (1) [29]:
F x = 1 N i = 0 N 1 a x i
where F(x) means the output, N means the number of pixels, a(x) means the input, and a(xi) means the prior input.
In order to address the low contrast between the infected and healthy areas, images are passed through a CLAHE method optimization method [30]. The method works to light up the dark areas by distributing bright pixels to the dark areas. The method processes pixels based on the value of adjacent pixels. When processing each pixel, the method compares it with the value of neighboring pixels as follows: if the value of the neighboring pixels is greater than the target pixel, then the contrast is increased, whereas the contrast decreases when the target pixel is larger than the adjacent ones. The method continues for each pixel of the image until the contrast of the affected areas is improved. Figure 1b shows a set of images from the Kvasir dataset after optimization.

3.3. Gradient Vector Flow Method

The GVF method is an extended method of the active contour method, which is defined by the following steps: first step, detect the edges of the affected regions F(x, y) from endoscopy images I(x, y). Equation (2) illustrates the mechanism for detecting the edges of grayscale images [31]. The GVF algorithm was used because it spatially extends the gradient vectors of the edge map, which produces a vector that contains information about the edges of the lesion throughout the image [32].
f x , y = | G σ x , y I x , y | 2
where ∇ refers to gradient operator and (x, y) refer to a 2D Gaussian function with σ standard deviation.
The flow minimizes the energy functionalization in the gradient vector by the GVF method. The functional energy depends on the smoothing and the level of edge noise present in the image. The image edge noise level is measured through the parameter μ. If the edge noise is high, the parameter μ should be increased, which leads to a decrease in the noise around the edges and a weakening of the contour, as in Equation (3) [33].
ε = μ u x 2 + v x 2 + v y 2 + f | 2   g f | 2 d x d y
where g refers to a gradient vector flow that can be derived from Euler equations that encourage edge map gradients that make g f smaller. When the gradients are large, g f is multiplied by the square of the gradient length subscripts that indicate partisan derivatives; f = f x , f y refers to a gradient map that shows edges. The affected area is segmented with specific parameters using gradient vector flow models. These models mark the edges of the affected regions and separate them from the rest of the image. The aim of applying this algorithm is to segment the infected areas and separate them from the healthy areas for further analysis. This method is an important contribution to this study as the affected areas of all the endoscopy images from the Kvasir dataset are segmented and saved in a new folder called Kvasir-ROI. CNN models receive endoscopy images of the gastrointestinal tract to the new Kvasir-ROI dataset. Thus, the CNN models work on analyzing endoscopy images of infected areas only instead of feeding them complete images including the healthy regions. Figure 2 shows a dataset sample that underwent segmentation and selected only affected areas.

3.4. CNN Models for the Extract Feature

CNN networks are characterized by their superior ability to process, interpret, and analyze the input data with high efficiency. The field of health care, especially the processing of biomedical images, has enjoyed a large share of CNN networks for analyzing medical images and diagnosing and predicting diseases and cancers at an early stage [34]. CNN is characterized by dozens of convolutional, pooling, fully connected, and other auxiliary layers. Convolution is a commutative operation involving two real-value functions. The input to a convolutional layer is a multidimensional array, and the filter is also a multidimensional array of parameters tuned by the learning process [35]. The convolutional layers receive endoscope images with an RGB color space of size m × n × j, where m, n, and j represent the width, height, and number of color channels of an image. Each CNN has a different number of convolutional layers. Each layer has a different filter corresponding to the image size regarding width, height, and number of filter channels. The filter f (t) wraps around a region of the image x(t) as in Equation (4). Each filter cell will be multiplied by the corresponding cell in the image to produce a matrix with smaller dimensions than the original image [36].
y t = x f t = x a f t a   d a
where f (t), x(t), and y(t) denote the filter, input, and the image output, respectively.
Thus, a feature map is created by wrapping filters around the image. The number of feature maps equals the number of filters in the layer, and the layer’s output becomes the input for the next layer [37]. The convolutional layers are followed by a ReLU helper layer whose task is to pass only positive values, while converting negative values to zero, as in Equation (5).
ReLU x = x , x 0 0 , x < 0
The output of convolutional layers is millions of neurons, connections, and parameters; therefore, networks will need complex calculations and a lot of training time. Thus, pooling layers will reduce the high aspect ratio, which reduces the image dimensions by pooling the output of a group of neurons in one layer to a single neuron in the next layer. There are two types of pooling layers. Either maximum or average grouping performs the calculations [38]. The Max pooling method uses the greatest value from a group of neurons and represents it with one neuron in the next layer, as in Equation (6), whereas the average pooling method calculates the average for a group of selected neurons and represents it with one neuron in the next layer, as in Equation (7) [39].
z i ;   j = m a x m , n = 1 . k   f i 1 p + m ;     j 1 p + n
z i ;   j = 1 k 2 m , n = 1 . k f i 1 p + m ;     j 1 p + n
where f, p, k, and m, n are the wrapped filter, p-step filter, number of pixels, and matrix sites.
The fully connected layers (FCLs) are the classification layers in CNNs. In these layers each neuron is connected to all neurons in the previous layer. The FCL transforms map features from high-dimensional to vector levels. The SoftMax function is non-linear and useful for handling classification problems. SoftMax is used with multiple classes; it divides the output between 0 and 1 according to the classes and then labels each image to its appropriate class [40].
Finally, in the last convolutional layer of the GoogLeNet, MobileNet, and DenseNet121 models are high-level features with sizes: (7, 7, 512), (7, 7, 1024), and (16, 32, 512), respectively. The global average pooling layer represents the high levels of features in feature vectors with sizes of 4069, 1024, and 1024 for the GoogLeNet, MobileNet, and DenseNet121 models, respectively. Therefore, the size of the gastrointestinal disease dataset is represented by feature matrices with sizes 8000 × 4096, 8000 × 1024, and 8000 × 8000 × 1024 for the GoogLeNet, MobileNet and DenseNet121 models, respectively.

3.5. Training and Performance Evaluation

The classification phase, including training and testing of the systems, is based on the previous phases of medical image processing. After enhancing the images, the ROI was segmented and the images were sent to CNN models to extract the features. Because of the high dimensions of the features, the CNN feature maps were passed to the PCA method [41] to select the representative features and delete the redundant and unimportant ones. [42] FFNN and XGBoost receive high-representation, low-dimensional features [43]. The classification phase includes an inductive phase called the training phase to build a taxonomic model. The deductive stage, the most important stage for the system’s efficiency, is called the testing stage.

3.5.1. FFNN Network

FFNN has the ability to handle computational complexities and different parameters and perform classifications efficiently according to the training data. Whenever the training data is well trained, this leads to good results for the test data [44]. Training the network well depends on its internal structure, which consists of three basic layers: the input, hidden, and output layers. Increasing the hidden layers and neurons does not necessarily lead to satisfactory results; therefore, the network efficiency may improve when reducing the hidden layers and interconnected neurons. The number of hidden cells varies from one network to another. Thus, no fixed mechanism exists for identifying hidden layers and their associated neurons to obtain accurate classification results [45]. Thus, the number of hidden layers is adjusted through trial and error. The learning method is based on adjusting the weights continuously, returning from the output layers to the hidden layers and updating the weights. Each time the weights are updated, the minimum square error (MSE) is calculated between the actual x j and expected y j values, as in Equation (8). The network continues until no change appears with the least error value and the weights are fixed at the minimum error rate. For the network mechanism, the input layer receives features extracted from the CNN models (GoogLeNet, MobileNet and DenseNet121). The number of entries for FFNN is 670, 495, and 520, based on the features of GoogLeNet, MobileNet, and DenseNet121, respectively. Features move to hidden layers, where they are analyzed and interpreted through 20 hidden layers. Finally, the output layer shows the FFNN performance results [46]. The output layer consists of eight neurons, each representing a class in the dataset, where each image is mapped to an appropriate neuron (class). Figure 3 shows the basic structure of FFNN for receiving, analyzing, and interpreting features of the GoogLeNet, MobileNet, and DenseNet121 models.
MSE = 1 k j = 1 k   ( x j y j ) 2
where k is the number of data points.

3.5.2. XGBoost Algorithm

XGBoost is a machine learning algorithm that works with the ensemble learning approach. A single model may not achieve good results in some algorithms, so the ensemble learning model offers a powerful approach to predicting multiple learners; the output is a single model. The algorithm sequentially generates many decision trees with different weights so that the algorithm merges weak learners with strong learners to produce an effective model; this process is called boosting. The boosting mechanism builds trees sequentially so that each subsequent tree takes advantage of the previous tree’s errors and reduces them [47]. Thus, any new tree is an updated version of the previous one with fewer errors than the previous one. In boosting, basic learners are weak learners who contribute some information vital to prediction. Thus, boosting can produce a strong learner by combining weak learners. When training the network, the weights of the variables that predict the decision tree are incorrectly increased. The variables are fed to the next tree to solve the errors in the previous tree, and so the network continues until the prediction of an effective and accurate dataset.
The first strategy is to classify endoscopy images from the Kvasir dataset by three pre-trained GoogLeNet, MobileNet, and DenseNet121 models. The second strategy is to analyze endoscopy images from the Kvasir gastroenterology dataset using the GoogLeNet, MobileNet, and DenseNet121 models based on the GVF segmentation algorithm. In this method, the images are improved and the infected areas are segmented and isolated from the healthy areas so that the CNN models receive the photos from the dataset after segmentation of the infected areas.
The third strategy for analyzing endoscopy images from the Kvasir gastrointestinal disease dataset is by a hybrid technique; CNN–FFNN and CNN–RF based on the GVF and PCA algorithms goes through several stages (as shown in Figure 4): First, passing the endoscopy images to the averaging and CLAHE techniques to improve the images. Second, applying the GVF algorithm to segment the region of interest for further analysis in the following stages. Third, applying the GoogLeNet, MobileNet, and DenseNet121 models to analyze regions of interest by convolutional, auxiliary, and pooling layers. Feature maps of 8000 × 4096, 8000 × 1024, and 8000 × 8000 × 1024 are produced for the GoogLeNet, MobileNet, and DenseNet121 models, respectively. The fourth stage is to obtain more representative features for the affected and diseased regions and delete the repetitive features using the PCA method [48]. The PCA method produces reduced feature vectors organized into feature matrices of sizes 8000 × 670, 8000 × 495, 8000 × 520, and 8000 × 520 for the GoogLeNet, MobileNet, and DenseNet121 models, respectively. The fifth stage is the training phase and performance evaluation by FFNN and XGBoost. FFNN and XGBoost receive critical feature vectors after the PCA method [49] and divide them into two parts: 80% for training and validating a model and 20% for measuring its performance.
The fourth strategy for analyzing endoscopy images to detect gastrointestinal diseases is by a hybrid model based on fused CNN features and the GVF and PCA algorithms, as shown in Figure 5. The first four steps of the third strategy are the same in this strategy. Fifthly, the fusion between the CNN features are as follows: GoogLeNet–MobileNet, MobileNet–DenseNet121, GoogLeNet–DenseNet121, and GoogLeNet–MobileNet–DenseNet121. So, the feature matrices in this strategy become 8000 × 1165, 8000 × 1015, 8000 × 1165, and 8000 × 1685 for the fused models GoogLeNet–MobileNet, MobileNet–DenseNet121, GoogLeNet–DenseNet121, and GoogLeNet–MobileNet–DenseNet121, respectively. Sixth, the training phase and performance evaluation by FFNN and XGBoost. FFNN and XGBoost receive essential feature vectors after the PCA method and divide them into 80% for training and validating a model and 20% for measuring its performance.

4. Results of Systems

4.1. Split of the Kvasir Dataset

The model performance was measured on endoscopy images from the Kvasir gastroenterology dataset. The Kvasir dataset consists of 8000 images distributed evenly across eight classes. During the training phase, the dataset was divided into 640 images (64%) for systems training, 160 images (16%) for validation, and 200 images (20%) for testing.

4.2. Metrics of Models Evaluation

When implementing systems and testing their performance, a confusion matrix can be produced, which represents the gold standard for systems performance. A confusion matrix is in a form similar to a quadrilateral matrix. The columns represent the number of target samples, which corresponds to the number of output samples by the rows. All samples in the confusion matrix are correctly classified samples located in the main diagonal and are called true positives (TPs). At the same time, the other cells represent false positives (FPs) or false negative (FN) samples and are incorrectly classified. The results of the systems are measured through Equations (9)–(13) [50].
AUC = TP Rate FP Rate 100 %
Accuracy = TN + TP TN + TP + FN + FP 100 %
Sensitivity = TP TP + FN 100 %
Precision = TP TP + FP 100 %
Specificity = TN TN + FP 100 %

4.3. Results of CNN Models

This section discusses the performance of the pre-trained GoogLeNet, MobileNet, and DenseNet121 models. The pre-trained CNN models are based on an ImageNet dataset with over a million images of over a thousand classes. This huge dataset lacks medical images, so pre-trained CNN models do not reach good results for classifying some medical datasets. In this study, the experience gained from the CNN models to classify the ImageNet dataset has been transferred to perform the new task of classifying the Kvasir gastroenterology dataset. The input layers of the GoogLeNet, MobileNet, and DenseNet121 models receive endoscopy images of the Kvasir gastroenterology dataset and pass them to convolutional layers to represent the images in high-level feature maps. Pooling layers reduces high dimensionality and feature manipulation through helper layers. Fully connected layers receive feature maps and convert them from 2D to 1D. Finally, the SoftMax function adjusts each carrier to its appropriate class.
Table 1 summarize the results of the pre-trained GoogLeNet, MobileNet, and DenseNet121 models for endoscopy image classification of the Kvasir gastroenterology dataset. GoogLeNet achieved an AUC of 91.46%, accuracy of 88.2%, sensitivity of 88.05%, precision of 88.34%, and specificity of 98.19%. The MobileNet achieved an AUC of 90.99%, accuracy of 86.4%, sensitivity of 86.34%, precision of 86.34%, and specificity of 97.96%. The DenseNet121 achieved an AUC of 89.13%, accuracy of 85.3%, sensitivity of 85.23%, precision of 85.29%, and specificity of 97.85%.

4.4. Results of CNN Models Based on the GVF Algorithm

This section discusses the performance of the GoogLeNet, MobileNet, and Dense-Net121 models based on the GVF hashing algorithm. In this section, the endoscopic images are improved by an average filter and then passed to the CLAHE method to highlight the edges of the lesion. Then, infected regions are divided and separated from healthy regions and saved in a folder named Kvasir-ROI. The models receive pre-trained GoogLeNet, MobileNet, and DenseNet121. Regions of interest are analyzed through convolutional layers and pooling and saved in deep feature maps. A fully connected layer categorizes the feature maps.
Table 2 summarize the results of the pre-trained GoogLeNet, MobileNet, and DenseNet121 models based on the GVF algorithm for endoscopy image classification of the Kvasir gastroenterology dataset. The GoogLeNet–GVF achieved an AUC of 91.8%, accuracy of 90.8%, sensitivity of 91.09%, precision of 90.81%, and specificity of 98.49%. The MobileNet–GVF achieved an AUC of 91.84%, accuracy of 91.5%, sensitivity of 91.56%, precision of 91.55%, and specificity of 98.83%. The DenseNet121–GVF achieved an AUC of 90.89%, accuracy of 90.1%, sensitivity of 89.83%, precision of 90.03%, and specificity of 98.35%.

4.5. Results of Hybrid Models of CNN, FFNN, and XGBoost

The section summarizes the results of the hybrid methodology between CNN models (GoogLeNet, MobileNet, and DenseNet121) and the FFNN or XGBoost networks for endoscopy image classification of the Kvasir-ROI gastroenterology dataset.
The mechanism of the hybrid method is to improve the images, segment the affected regions, and extract the feature map through the convolutional layers of the CNN models. The output of the CNN models was fed to PCA to reduce the features, save the most important features, and delete the unnecessary and redundant ones. FFNN and XGBoost receive the most important features and segment them during the training and testing phases. The CNN–FFNN and CNN–XGBoost models can analyze endoscopy images of gastrointestinal diseases and distinguish between the types of gastrointestinal diseases with high accuracy.
Table 3 presents the performance of the hybrid CNN–FFNN for endoscopy image classification of the Kvasir gastroenterology dataset. The GoogLeNet–FFNN achieved an AUC of 93.46%, accuracy of 93.8%, sensitivity of 93.99%, precision of 93.88%, and specificity of 98.91%. The MobileNet–FFNN achieved an AUC of 95.08%, accuracy of 94.4%, sensitivity of 94.45%, precision of 94.49%, and specificity of 98.98%. The DenseNet121–FFNN achieved an AUC of 95.93%, accuracy of 94.8%, sensitivity of 94.76%, precision of 94.84%, and specificity of 99.05%.
Table 4 presents the performance of the hybrid CNN–XGBoost for endoscopy image classification of the Kvasir gastroenterology dataset. The GoogLeNet–XGBoost achieved an AUC of 95.46%, accuracy of 94.3%, sensitivity of 94.29%, precision of 94.26%, and specificity of 99.14%. The MobileNet–XGBoost achieved an AUC of 94.36%, accuracy of 93.8%, sensitivity of 94.06%, precision of 93.83%, and specificity of 99.24%. The DenseNet121–XGBoost achieved an AUC of 94.91%, accuracy of 94.20%, sensitivity of 94.29%, precision of 94.2%, and specificity of 98.95%.
A confusion matrix is the gold standard for presenting system performance measures. This section shows the performance of hybrid CNN–FFNN networks for classifying endoscopy images of the Kvasir-ROI gastroenterology dataset. Figure 6 shows the performance of the GoogLeNet–FFNN, MobileNet–FFNN, and DenseNet121–FFNN hybrid networks through confusion matrices. The hybrid GoogLeNet–FFNN achieved accuracy for each type in the Kvasir gastroenterology dataset as follows: for dyed-lifted-polyps, 92.5%; for dyed resection margins, 92.5%; for esophagitis, 94%, for normal cecum, 98.5%; for normal pylorus, 96.5%; for normal z-line, 92.5%, for polyps, 93%; for ulcerative colitis, 91%. The hybrid MobileNet–FFNN achieved accuracy for each type in the Kvasir gastroenterology dataset as follows: for dyed-lifted-polyps, 94.5%; for dyed resection margins, 93%; for esophagitis, 98.5%; for normal cecum, 97%; for normal pylorus, 95%; for normal z-line, 93.5%; for polyps, 92%; for ulcerative colitis, 92%. The hybrid DenseNet121–FFNN achieved accuracy for each type in the Kvasir gastroenterology dataset as follows: for dyed-lifted-polyps, 93%; for dyed resection margins, 94.5%; for esophagitis, 97%; for normal cecum, 99%; for normal pylorus, 95.5%; for normal z-line, 96%; for polyps, 94%; for ulcerative colitis, 89%.
This section shows the performance of hybrid CNN–XGBoost and CNN–XGBoost networks for classifying endoscopy images from the Kvasir-ROI gastroenterology dataset. Figure 7 shows the performance of the GoogLeNet–XGBoost, MobileNet–XGBoost, and DenseNet121–XGBoost hybrid networks through the confusion matrix. The hybrid GoogLeNet–XGBoost achieved accuracy for each type in the Kvasir gastroenterology dataset as follows: for dyed lifted polyps, 93%; for dyed resection margins, 96%; for esophagitis, 96%; for normal cecum, 97.5%; for normal pylorus, 95.5%; for normal z-line, 93%, for polyps, 90.5%, for ulcerative colitis, 92.5%. The hybrid MobileNet–XGBoost achieved accuracy for each type in the Kvasir gastroenterology dataset as follows: for dyed-lifted-polyps, 93.5%; for dyed resection margins, 91.5%; for esophagitis, 94%; for normal cecum, 96%; for normal pylorus, 97.5%; for normal z-line, 94%; for polyps, 91%; for ulcerative colitis, 93%. The hybrid DenseNet121–XGBoost achieved accuracy for each type in the Kvasir gastroenterology dataset as follows: for dyed lifted polyps, 94.5%; for dyed resection margins, 95%; for esophagitis, 96.5%; for normal cecum, 99%; for normal pylorus, 95%; for normal z-line, 94%; for polyps, 87.5%; for ulcerative colitis, 92%.

4.6. Results of Hybrid Methodology Based on Fused Deep Learning Features

The section summarizes the results of the hybrid methodology between the CNN models (GoogLeNet, MobileNet, and DenseNet121) and the FFNN or XGBoost networks based on integrating the features of the CNN models for endoscopy image classification of the Kvasir-ROI gastroenterology dataset. The hybrid methodology aims to optimize the images, segment the affected regions and isolate them from the healthy ones, and extract the feature map through the convolutional layers of the CNN models. The features of the CNN models are sequentially combined as follows: GoogLeNet–MobileNet, MobileNet–DenseNet121, GoogLeNet–DenseNet121, and GoogLeNet–MobileNet–DenseNet121. The features of the hybrid CNN models were then fed into PCA to reduce the features, save the most important features, and delete the unnecessary and redundant features. FFNN and XGBoost receive and breakdown the most important features during the training and testing phases. The CNN-FFNN and CNN-XGBoost, based on features of CNN fusion, analyze endoscopy images of gastroenterology and accurately distinguish between gastroenterological disease types.
Table 5 presents the performance of the hybrid CNN–FFNN based on features of CNN fusion for endoscopy image classification of the Kvasir-ROI gastroenterology dataset. The GoogLeNet–MobileNet–FFNN achieved an AUC of 94.49%, accuracy of 96.06%, sensitivity of 95.75%, precision of 96.09%, and specificity of 99.33%. The MobileNet–DenseNet121–FFNN achieved an AUC of 96.4%, accuracy of 95.8%, sensitivity of 95.64%, precision of 95.84%, and specificity of 99.28%. The GoogLeNet–DenseNet121–FFNN achieved an AUC of 97.21%, accuracy of 96.5%, sensitivity of 96.29%, precision of 96.53%, and specificity of 99.29%. The GoogLeNet–MobileNet–DenseNet121–FFNN achieved an AUC of 97.08%, accuracy of 96.9%, sensitivity of 96.68%, precision of 96.9%, and specificity of 99.3%.
Table 6 presents the performance of the hybrid CNN–XGBoost based on features of CNN fusion for endoscopy image classification of the Kvasir-ROI gastroenterology dataset as follows: The GoogLeNet–MobileNet–XGBoost achieved an AUC of 96.6%, accuracy of 96.8%, sensitivity of 96.43%, precision of 96.78%, and specificity of 99.4%. The MobileNet–DenseNet121–XGBoost achieved an AUC of 95.5%, accuracy of 95.9%, sensitivity of 95.78%, precision of 95.9%, and specificity of 99.23%. The GoogLeNet–DenseNet121–XGBoost achieved an AUC of 95.8%, accuracy of 95.3%, sensitivity of 95.15%, precision of 95.26%, and specificity of 99.15%. The GoogLeNet–MobileNet–DenseNet121–XGBoost achieved an AUC of 97.54%, accuracy of 97.25%, sensitivity of 96.86%, precision of 97.25%, and specificity of 99.48%.
This section shows the performance of hybrid CNN–FFNN and CNN–XGBoost based on features of CNN fusion for classifying endoscopy images from the Kvasir-ROI gastroenterology dataset. Figure 8 shows the performance of the CNN–FFNN based on the features of CNN fusion through the confusion matrix. The hybrid GoogLeNet–MobileNet–FFNN achieved accuracy for each type in the Kvasir gastroenterology dataset as follows: for dyed lifted polyps, 96%; for dyed resection margins, 94.5%; for esophagitis, 99%; for normal cecum, 99.5%; for normal pylorus, 95.5%; for normal z-line, 95.5%; for polyps, 95.5%; for ulcerative colitis, 93%. The hybrid MobileNet–DenseNet121–FFNN achieved accuracy for each type in the Kvasir gastroenterology dataset as follows: for dyed lifted polyps, 95.5%; for dyed resection margins, 95%; for esophagitis, 98.5%; for normal cecum, 97.5%; for normal pylorus, 96.5%; for normal z-line, 95%; for polyps, 93%; for ulcerative colitis, 95.5%. The hybrid GoogLeNet–DenseNet121–FFNN achieved accuracy for each type in the Kvasir gastroenterology dataset as follows: for dyed lifted polyps, 95%; for dyed resection margin, 97.5%; for esophagitis, 97%; for normal cecum, 99.5%; for normal pylorus, 96%; for normal z-line, 97.5%; for polyps, 94%; for ulcerative colitis, 95.5%. The hybrid GoogLeNet–MobileNet–DenseNet121–FFNN achieved accuracy for each type in the Kvasir gastroenterology dataset as follows: for dyed lifted polyps, 97%; for dyed resection margins, 97.5%; for esophagitis, 96.5%; for normal cecum, 99%; for normal pylorus, 97%; for normal z-line, 95.5%; for polyps, 97%; for ulcerative colitis, 95.5%.
It is noted from Figure 9 that there are cases where the system failed and misclassified the images as follows: First, for the class “dyed-lifted-polyps”, 192 images were classified correctly, whereas the failure cases in this class were 6 images that were classified as “dyed-resection-margins” and 2 images that were classified as “polyps”. Second, for the class “dyed-resection-margins”, 189 images were classified correctly, whereas the failure cases in this class were 4 images that were classified as “dyed-lifted-polyps”, 2 images that were classified as “normal-pylorus”, 1 image that was classified as “normal-z-line”, 2 images that were classified as “polyps”, and 2 images that were classified as “ulcerative-colitis”. Thirdly, for the “esophagitis” class 198 images were classified correctly, whereas the failure cases in this class were 2 images classified as “normal-z-line”. Fourth, for the “normal-cecum” class, 199 images were classified correctly, whereas the failure case in this class was one image that was classified as “ulcerative-colitis”. Fifth, for the “normal-pylorus class”, 191 images were classified correctly, whereas the failure cases in this class were 4 images that were classified as “dyed-lifted-polyps” and 5 images that were classified as “normal-z-line”. Sixth, for the “normal-z-line class”, 191 images were classified correctly, whereas the failure cases in this class were 2 images classified as “dyed-resection-margins”, 5 images classified as “esophagitis”, and 2 images classified as “normal-pylorus”. Seventh, for the “polyps” class, 191 images were classified correctly, whereas the failure cases in this class were 3 images classified as “dyed-lifted-polyps”, 1 image classified as “dyed-resection-margins”, 2 images classified as “normal-cecum”, 2 images classified as “normal-pylorus”, and 1 image classified as “ulcerative-colitis”. Eighth, for the “ulcerative-colitis” class, 186 images were classified correctly, whereas the failure cases in this class were 1 image classified as “dyed-resection-margins”, 5 images classified as “normal-cecum”, and 8 images classified as “polyps”.
Figure 8 shows the performance of the CNN–XGBoost based on features of CNN fusion through the confusion matrix. The hybrid GoogLeNet–MobileNet–XGBoost achieved accuracy for each type in the Kvasir gastroenterology dataset as follows: for dyed lifted polyps, 94.5%; for dyed resection margins, 98%; for esophagitis, 97.5%; for normal cecum, 100%; for normal pylorus, 96%; for normal z-line, 97.5%; for polyps, 98.5%; for ulcerative colitis, 97.5%. The hybrid MobileNet–DenseNet121–XGBoost achieved accuracy for each type in the Kvasir gastroenterology dataset as follows: for dyed lifted polyps, 94.5%; for dyed resection margins, 96%; for esophagitis, 97%; for normal cecum, 98.5%; for normal-pylorus, 94.5%; for normal z-line, 98%; for polyps, 93.5%; for ulcerative colitis, 95%. The hybrid GoogLeNet–DenseNet121–XGBoost achieved accuracy for each type in the Kvasir gastroenterology dataset as follows: for dyed lifted polyps, 94%; for dyed resection margins, 93%; for esophagitis, 99%; for normal cecum, 98%; for normal pylorus, 94.5%; for normal z-line, 95.5%; for polyps, 95.5%; for ulcerative colitis, 92.5%. The hybrid GoogLeNet–MobileNet–DenseNet121–XGBoost achieved accuracy for each type in the Kvasir gastroenterology dataset as follows: for dyed lifted polyps, 95%; for dyed resection margins, 98.5%; for esophagitis, 98.5%; for normal cecum, 96.5%; for normal pylorus, 97.5%; for normal z-line, 98.5%; for polyps, 97.5%; for ulcerative colitis, 96%.

5. Discussion of the Performance of the Proposed Methodologies

The gastrointestinal tract suffers from many diseases, such as infections, ulcers, benign tumors, and malignant tumors, which lead to its inability to carry out its tasks well. Additionally, malignant tumors threaten human life and lead to death if discovered late. Polyps may develop into malignancy if not treated and removed early. Manual diagnosis faces many challenges, such as distinguishing between malignant and benign tumors in the early stages and distinguishing between infections and ulcers. Computer-assisted automated diagnostic systems help differentiate between different types of gastrointestinal diseases.
In this study, several hybrid systems (CNN–FFNN and CNN–XGBoost) were developed based on the GVF segmentation algorithm and fused CNN features.
The first methodology was to diagnose endoscopy images from the Kvasir gastroenterology dataset using pre-trained GoogLeNet, MobileNet, and DenseNet121 models, which reached accuracies of 88.2%, 86.4%, and 85.3%, respectively.
The second methodology for diagnosing endoscopy images from the Kvasir gastroenterology dataset was by using pre-trained GoogLeNet, MobileNet, and DenseNet121 models based on the GVF segmentation algorithm and reached accuracies of 90.8%, 91.5%, and 90.1%, respectively.
The third methodology for analyzing endoscopy images for the early diagnosis of gastrointestinal diseases was by hybrid systems (CNN–FFNN and CNN–XGBoost) based on the GVF segmentation algorithm. The GoogLeNet–FFNN, MobileNet–FFNN, and DenseNet121–FFNN hybrid systems achieved accuracies of 93.8%, 94.4%, and 94.8%, respectively, whereas the hybrid systems GoogLeNet–XGBoost, MobileNet–XGBoost, and DenseNet121–XGBoost achieved accuracies of 94.3%, 93.8%, and 94.2%, respectively.
The fourth methodology for the early diagnosis of gastrointestinal diseases was by hybrid systems (CNN–FFNN and CNN–XGBoost) based on the GVF segmentation algorithm and fused CNN features. The hybrid systems GoogLeNet–MobileNet–FFNN, MobileNet–DenseNet121–FFNN, GoogLeNet–DenseNet121–FFNN, and GoogLeNet–MobileNet–DenseNet121–FFNN reached accuracies of 96.06%, 95.8%, 96.5%, and 96.9%, respectively, whereas the hybrid systems GoogLeNet–MobileNet–XGBoost, MobileNet–DenseNet121–XGBoost, GoogLeNet–DenseNet121–XGBoost, and GoogLeNet–MobileNet–DenseNet121–XGBoost reached accuracies of 96.8%, 95.9%, 95.3%, and 97.25%, respectively.
Table 7 and Figure 10 present the results of all proposed methodologies for analyzing endoscopy images for gastroenterology from the Kvasir dataset. The table summarizes the overall accuracy and accuracies for each class in the Kvasir dataset that each methodology yielded. Endoscopic images from the Kvasir gastroenterology dataset were diagnosed using pre-trained GoogLeNet, MobileNet, and DenseNet121 models, which achieved good, but not superior, results. The results of the pre-trained models were improved through the GVF segmentation algorithm, where the images were enhanced and the infected regions were segmented and isolated from the healthy ones. Thus, the pre-trained models received the Kvasir-ROI dataset and an improvement in the results was noted. A hybrid CNN–FFNN or CNN–XGBoost methodology was applied based on the GVF segmentation algorithm, and the results improved significantly. To achieve superior results for the classification of the Kvasir gastroenterology dataset, a hybrid CNN–FFNN or CNN–XGBoost methodology was applied based on the GVF segmentation algorithm and fused CNN features.
The table shows the results for each class across each system and the improvement of the results for each category was as follows: For the class “dyed_lifted_polyps”, the GoogLeNet model achieved an accuracy of 79%, whereas the accuracy was improved to 95.5% by the MobileNet–DenseNet121–FFNN model. For the class “dyed_resection_margins”, the DenseNet121 model achieved an accuracy of 75.5%, whereas the accuracy was improved to 98.5% by the GoogLeNet–MobileNet–DenseNet121–XGBoost model. For the class “esophagitis”, the DenseNet121 model achieved an accuracy of 70.5%, whereas the accuracy was improved to 98.5% by three hybrid models. For the “normal_cecum” class, the MobileNet model achieved an accuracy of 91.5%, whereas the accuracy was improved to 99.5% by the GoogLeNet–MobileNet–FFNN model. For the “normal_pylorus” class, the MobileNet model achieved an accuracy of 86%, whereas the accuracy was improved to 97.5% by the GoogLeNet–MobileNet–DenseNet121–XGBoost model. For the “normal_z_line” class, the DenseNet121 model achieved an accuracy of 76%, whereas the accuracy was improved to 98.5% by the GoogLeNet–MobileNet–XGBoost model. For the “polyps” class, the GoogLeNet model achieved an accuracy of 79%, whereas the accuracy was improved to 98.5% by the GoogLeNet–MobileNet–XGBoost model. For the “ulcerative_colitis” class, the MobileNet model achieved an accuracy of 82.5%, whereas the accuracy was improved to 95.5% by three hybrid models.
It is concluded that the proposed systems are superior to the previous studies by all measures of accuracy, sensitivity, AUC, and specificity thanks to the use of the method of combining CNN features.
Through the results, the following can be concluded: The results of the CNN models based on the GVF algorithm were better than feeding in the images directly without applying the GVF algorithm. The technical results of a hybrid between CNN models (GoogLeNet, MobileNet, and DenseNet121) and FFNN and XGBoost improved the results more than applying CNN models. Finally, it is noted that the performance of the FFNN and XGBoost networks with the hybrid features of the CNN models is clearly better than the previous methods.

6. Conclusions

It is difficult to distinguish between the types of diseases of gastrointestinal tract and the presence of anatomical landmarks, pathological findings, and polyp removal, especially in the early stages. Therefore, many methodologies have been developed to help doctors and specialists diagnose disease from endoscopy images of gastrointestinal tract. The Kvasir dataset was classified by three pre-trained CNN models. In the second methodology, endoscopy images were optimized, and the affected areas were segmented and classified by three pre-trained CNN models. The third methodology analyzed endoscopy images from the Kvasir dataset by a hybrid CNN–FFNN or CNN–XGBoost methodology based on the GVF segmentation algorithm. The fourth methodology analyzed endoscopy images from the Kvasir dataset by a hybrid CNN–FFNN or CNN–XGBoost methodology based on the GVF segmentation algorithm and fused CNN features. The GoogLeNet–MobileNet–DenseNet121–XGBoost hybrid methodology based on the GVF algorithm achieved an AUC of 97.54%, accuracy of 97.25%, sensitivity of 96.86%, precision of 97.25%, and specificity of 99.48%.
These systems can be generalized to help doctors and endoscopy specialists in the early diagnosis of tumors of the gastrointestinal tract in order for the patient to receive appropriate treatment.
The most important limitation we encountered was the similarity in the biological characteristics between some types of gastrointestinal diseases, which were overcome by combining the features of the CNN models.
The future work is to extract the handcrafted features, combine them with combined CNN features, and generalize them to the proposed systems on a new dataset.
Additionally, in future work PCA will be used before and after combining the features and comparing the results.

Author Contributions

Conceptualization, I.A.A., E.M.S. and H.S.A.S.; methodology, I.A.A., E.M.S. and H.S.A.S.; software, E.M.S. and I.A.A.; validation, H.S.A.S., E.M.S. and I.A.A.; formal analysis, I.A.A., E.M.S. and H.S.A.S.; investigation, I.A.A., E.M.S. and H.S.A.S.; resources, H.S.A.S., E.M.S. and I.A.A.; data curation, I.A.A. and E.M.S.; writing—original draft preparation, E.M.S.; writing—review and editing, I.A.A. and H.S.A.S.; visualization, H.S.A.S., E.M.S. and I.A.A.; supervision, I.A.A. and E.M.S.; project administration, I.A.A., E.M.S. and H.S.A.S.; funding acquisition, I.A.A. and H.S.A.S.; All authors have read and agreed to the published version of the manuscript.

Funding

This research has been funded by the Deanship of Scientific Research at Najran University, Kingdom of Saudi Arabia, through a grant code (NU/DRP/SERC/12/8).

Data Availability Statement

In this study, data supporting the performance of the proposed methodologies were collected from the publicly available Kvasir gastroenterology dataset at the link: https://datasets.simula.no/kvasir/#download (accessed date: 2 October 2022).

Acknowledgments

The authors are thankful to the Deanship of Scientific Research at Najran University for funding this work, under the General Research Funding program grant code (NU/DRP/SERC/12/8).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ji, H.; Hu, J.; Zuo, S.; Zhang, S.; Li, M.; Nie, S. In vitro gastrointestinal digestion and fermentation models and their applications in food carbohydrates. Crit. Rev. Food Sci. Nutr. 2021, 62, 5349–5371. [Google Scholar] [CrossRef] [PubMed]
  2. Naz, J.; Sharif, M.; Yasmin, M.; Raza, M.; Khan, M.A. Detection and Classification of Gastrointestinal Diseases using Machine Learning. Curr. Med. Imaging 2020, 17, 479–490. [Google Scholar] [CrossRef] [PubMed]
  3. Sung, H.; Ferlay, J.; Siegel, R.L.; Laversanne, M.; Soerjomataram, I.; Jemal, A.; Bray, F. Global Cancer Statistics 2020: GLOBOCAN Estimates of Incidence and Mortality Worldwide for 36 Cancers in 185 Countries. CA Cancer J. Clin. 2021, 71, 209–249. [Google Scholar] [CrossRef] [PubMed]
  4. Xia, J.Y.; Aadam, A.A. Advances in screening and detection of gastric cancer. J. Surg. Oncol. 2022, 125, 1104–1109. [Google Scholar] [CrossRef]
  5. Xie, Y.; Shi, L.; He, X.; Luo, Y. Gastrointestinal cancers in China, the USA, and Europe. Gastroenterol. Rep. 2021, 9, 91–104. [Google Scholar] [CrossRef]
  6. Rustam, F.; Siddique, M.A.; Siddiqui, H.U.R.; Ullah, S.; Mehmood, A.; Ashraf, I.; Choi, G.S. Wireless Capsule Endoscopy Bleeding Images Classification Using CNN Based Model. IEEE Access 2021, 9, 33675–33688. [Google Scholar] [CrossRef]
  7. Finocchiaro, M.; Valdivia, P.C.; Hernansanz, A.; Marino, N.; Amram, D.; Casals, A.; Menciassi, A.; Marlicz, W.; Ciuti, G.; Koulaouzidis, A. Training Simulators for Gastrointestinal Endoscopy: Current and Future Perspectives. Cancers 2021, 13, 1427. [Google Scholar] [CrossRef]
  8. Kim, M.; Lee, K.W.; Kim, K.; Gulenko, O.; Lee, C.; Keum, B.; Chun, H.J.; Choi, H.S.; Kim, C.U.; Yang, J.-M. Intra-instrument channel workable, optical-resolution photoacoustic and ultrasonic mini-probe system for gastrointestinal endoscopy. Photoacoustics 2022, 26, 100346. [Google Scholar] [CrossRef]
  9. Gill, I.; Antonios, B.; Imam, Z.; Ghaith, G. Self-Expanding Metal Stent (SEMS) Placement to Treat Bleeding from Late Radiation Esophagitis. Case Rep. Gastrointest. Med. 2021, 2021, 6678139. [Google Scholar] [CrossRef]
  10. Tang, Y.; Anandasabapathy, S.; Richards-Kortum, R. Advances in optical gastrointestinal endoscopy: A technical review. Mol. Oncol. 2021, 15, 2580–2599. [Google Scholar] [CrossRef]
  11. Öztürk, S.; Özkaya, U. Residual LSTM layered CNN for classification of gastrointestinal tract diseases. J. Biomed. Inform. 2021, 113, 103638. [Google Scholar] [CrossRef]
  12. Iqbal, I.; Walayat, K.; Kakar, M.U.; Ma, J. Automated identification of human gastrointestinal tract abnormalities based on deep convolutional neural network with endoscopic images. Intell. Syst. Appl. 2022, 16, 200149. [Google Scholar] [CrossRef]
  13. Yoshiok, K.; Tanioka, K.; Hiwa, S.; Hiroyasu, T. Deep-learning models in medical image analysis: Detection of esophagitis from the Kvasir Dataset. arXiv 2023. [Google Scholar] [CrossRef]
  14. Amiri, Z.; Hassanpour, H.; Beghdadi, A. A Computer-Aided Method for Digestive System Abnormality Detection in WCE Images. J. Healthc. Eng. 2021, 2021, 7863113. [Google Scholar] [CrossRef]
  15. Nezhad, S.A.; Khatibi, T.; Sohrabi, M. Proposing Novel Data Analytics Method for Anatomical Landmark Identification from Endoscopic Video Frames. J. Health Eng. 2022, 2022, 8151177. [Google Scholar] [CrossRef]
  16. Dheir, I.M.; Abu-Naser, S.S. Classification of Anomalies in Gastrointestinal Tract Using Deep Learning. Int. J. Acad. Eng. Res. 2022, 6, 15–28. [Google Scholar]
  17. Wu, X.; Chen, C.; Zhong, M.; Wang, J. HAL: Hybrid active learning for efficient labeling in medical domain. Neurocomputing 2021, 456, 563–572. [Google Scholar] [CrossRef]
  18. Karmakar, R.; Nooshabadi, S. Mobile-PolypNet: Lightweight Colon Polyp Segmentation Network for Low-Resource Settings. J. Imaging 2022, 8, 169. [Google Scholar] [CrossRef]
  19. Mohapatra, S.; Nayak, J.; Mishra, M.; Pati, G.K.; Naik, B.; Swarnkar, T. Wavelet Transform and Deep Convolutional Neural Network-Based Smart Healthcare System for Gastrointestinal Disease Detection. Interdiscip. Sci. Comput. Life Sci. 2021, 13, 212–228. [Google Scholar] [CrossRef]
  20. Ahmed, A. Classification of Gastrointestinal Images Based on Transfer Learning and Denoising Convolutional Neural Networks. Lect. Notes Netw. Syst. 2022, 288, 631–639. [Google Scholar] [CrossRef]
  21. Jha, D.; Ali, S.; Emanuelsen, K.; Hicks, S.A.; Thambawita, V.; Garcia-Ceja, E.; Halvorsen, P. Kvasir-instrument: Diagnostic and therapeutic tool segmentation dataset in gastrointestinal endoscopy. In MultiMedia Modeling, Proceedings of the 27th International Conference, MMM 2021, Prague, Czech Republic, 22–24 June 2021; Part II 27, Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer International Publishing: Cham, Switzerland, 2021; Volume 12573, pp. 218–229. [Google Scholar]
  22. Ramzan, M.; Raza, M.; Sharif, M.; Khan, M.A.; Nam, Y. Gastrointestinal Tract Infections Classification Using Deep Learning. Comput. Mater. Contin. 2021, 69, 3239–3257. [Google Scholar] [CrossRef]
  23. Yogapriya, J.; Chandran, V.; Sumithra, M.G.; Anitha, P.; Jenopaul, P.; Dhas, C.S.G. Gastrointestinal Tract Disease Classification from Wireless Endoscopy Images Using Pretrained Deep Learning Model. Comput. Math. Methods Med. 2021, 2021, 5940433. [Google Scholar] [CrossRef] [PubMed]
  24. Khan, M.A.; Muhammad, K.; Wang, S.H.; Alsubai, S.; Binbusayyis, A.; Alqahtani, A.; Thinnukool, O. Gastrointestinal diseases recognition: A framework of deep neural network and improved moth-crow optimization with dcca fusion. Hum.-Cent. Comput. Inf. Sci. 2022, 12, 25. [Google Scholar] [CrossRef]
  25. Khan, M.A.; Sahar, N.; Khan, W.Z.; Alhaisoni, M.; Tariq, U.; Zayyan, M.H.; Kim, Y.J.; Chang, B. GestroNet: A Framework of Saliency Estimation and Optimal Deep Learning Features Based Gastrointestinal Diseases Detection and Classification. Diagnostics 2022, 12, 2718. [Google Scholar] [CrossRef] [PubMed]
  26. Pogorelov, K.; Randel, K.R.; Griwodz, C.; Eskeland, S.L.; de Lange, T.; Johansen, D.; Spampinato, C.; Dang-Nguyen, D.-T.; Lux, M.; Schmidt, P.T.; et al. KVASIR: A multi-class image dataset for computer-aided gastrointestinal disease detection. In Proceedings of the 8th ACM Multimedia Systems Conference, MMSys 2017, Taipei, Taiwan, 20–23 June 2017; pp. 164–169. [Google Scholar] [CrossRef]
  27. Ardestani, M.M.; Yan, H. Noise Reduction in Human Motion-Captured Signals for Computer Animation based on B-Spline Filtering. Sensors 2022, 22, 4629. [Google Scholar] [CrossRef]
  28. Jebadass, J.R.; Balasubramaniam, P. Low contrast enhancement technique for color images using interval-valued intuitionistic fuzzy sets with contrast limited adaptive histogram equalization. Soft Comput. 2022, 26, 4949–4960. [Google Scholar] [CrossRef]
  29. Senan, E.M.; Jadhav, M.E. Techniques for the Detection of Skin Lesions in PH2 Dermoscopy Images Using Local Binary Pattern (LBP). In Proceedings of the International Conference on Recent Trends in Image Processing and Pattern Recognition, Aurangabad, India, 3–4 January 2020; Springer: Singapore, 2020; pp. 14–25. [Google Scholar] [CrossRef]
  30. Hussein, F.; Mughaid, A.; AlZu’bi, S.; El-Salhi, S.M.; Abuhaija, B.; Abualigah, L.; Gandomi, A.H. Hybrid CLAHE-CNN Deep Neural Networks for Classifying Lung Diseases from X-ray Acquisitions. Electronics 2022, 11, 3075. [Google Scholar] [CrossRef]
  31. Mohammed, B.A.; Senan, E.M.; Al-Mekhlafi, Z.G.; Alazmi, M.; Alayba, A.M.; Alanazi, A.A.; Alreshidi, A.; Alshahrani, M. Hybrid Techniques for Diagnosis with WSIs for Early Detection of Cervical Cancer Based on Fusion Features. Appl. Sci. 2022, 12, 8836. [Google Scholar] [CrossRef]
  32. Qian, Q.; Cheng, K.; Qian, W.; Deng, Q.; Wang, Y. Image Segmentation Using Active Contours with Hessian-Based Gradient Vector Flow External Force. Sensors 2022, 22, 4956. [Google Scholar] [CrossRef]
  33. Ahmed, I.A.; Senan, E.M.; Rassem, T.H.; Ali, M.A.H.; Shatnawi, H.S.A.; Alwazer, S.M.; Alshahrani, M. Eye Tracking-Based Diagnosis and Early Detection of Autism Spectrum Disorder Using Machine Learning and Deep Learning Techniques. Electronics 2022, 11, 530. [Google Scholar] [CrossRef]
  34. Khalid, A.; Senan, E.M.; Al-Wagih, K.; Al-Azzam, M.M.A.; Alkhraisha, Z.M. Automatic Analysis of MRI Images for Early Prediction of Alzheimer’s Disease Stages Based on Hybrid Features of CNN and Handcrafted Features. Diagnostics 2023, 13, 1654. [Google Scholar] [CrossRef]
  35. Shamsan, A.; Senan, E.M.; Shatnawi, H.S.A. Automatic Classification of Colour Fundus Images for Prediction Eye Disease Types Based on Hybrid Features. Diagnostics 2023, 13, 1706. [Google Scholar] [CrossRef]
  36. Fati, S.M.; Senan, E.M.; ElHakim, N. Deep and Hybrid Learning Technique for Early Detection of Tuberculosis Based on X-ray Images Using Feature Fusion. Appl. Sci. 2022, 12, 7092. [Google Scholar] [CrossRef]
  37. Nogueira-Rodríguez, A.; Reboiro-Jato, M.; Glez-Peña, D.; López-Fernández, H. Performance of Convolutional Neural Networks for Polyp Localization on Public Colonoscopy Image Datasets. Diagnostics 2022, 12, 898. [Google Scholar] [CrossRef]
  38. Mohammed, B.A.; Senan, E.M.; Rassem, T.H.; Makbol, N.M.; Alanazi, A.A.; Al-Mekhlafi, Z.G.; Almurayziq, T.S.; Ghaleb, F.A. Multi-Method Analysis of Medical Records and MRI Images for Early Diagnosis of Dementia and Alzheimer’s Disease Based on Deep Learning and Hybrid Methods. Electronics 2021, 10, 2860. [Google Scholar] [CrossRef]
  39. Amin, M.S.; Shah, J.H.; Yasmin, M.; Ansari, G.J.; Khan, M.A.; Tariq, U.; Kim, Y.J.; Chang, B. A Two Stream Fusion Assisted Deep Learning Framework for Stomach Diseases Classification. CMC-Comput. Mater. Contin. 2022, 73, 4423–4439. [Google Scholar] [CrossRef]
  40. Mohammed, B.A.; Senan, E.M.; Al-Mekhlafi, Z.G.; Rassem, T.H.; Makbol, N.M.; Alanazi, A.A.; Almurayziq, T.S.; Ghaleb, F.A.; Sallam, A.A. Multi-Method Diagnosis of CT Images for Rapid Detection of Intracranial Hemorrhages Based on Deep and Hybrid Learning. Electronics 2022, 11, 2460. [Google Scholar] [CrossRef]
  41. Zhao, B.; Dong, X.; Guo, Y.; Jia, X.; Huang, Y. PCA Dimensionality Reduction Method for Image Classification. Neural Process. Lett. 2022, 54, 347–368. [Google Scholar] [CrossRef]
  42. Khan, M.A.; Sarfraz, M.S.; Alhaisoni, M.; Albesher, A.A.; Wang, S.; Ashraf, I. StomachNet: Optimal Deep Learning Features Fusion for Stomach Abnormalities Classification. IEEE Access 2020, 8, 197969–197981. [Google Scholar] [CrossRef]
  43. Mohammad, F.; Al-Razgan, M. Deep Feature Fusion and Optimization-Based Approach for Stomach Disease Classification. Sensors 2022, 22, 2801. [Google Scholar] [CrossRef]
  44. Mohammed, F.G.; Thamir, N.N. Esophageal Cancer Detection Using Feed-Forward Neural Network. Webology 2022, 19, 6121–6145. [Google Scholar]
  45. Senan, E.M.; Jadhav, M.E.; Rassem, T.H.; Aljaloud, A.S.; Mohammed, B.A.; Al-Mekhlafi, Z.G. Early Diagnosis of Brain Tumour MRI Images Using Hybrid Techniques between Deep and Machine Learning. Comput. Math. Methods Med. 2022, 2022, 8330833. [Google Scholar] [CrossRef] [PubMed]
  46. Yin, M.; Lin, J.; Liu, L.; Gao, J.; Xu, W.; Yu, C.; Qu, S.; Liu, X.; Qian, L.; Xu, C.; et al. Development of a Deep Learning Model for Malignant Small Bowel Tumors Survival: A SEER-Based Study. Diagnostics 2022, 12, 1247. [Google Scholar] [CrossRef] [PubMed]
  47. Lin, X.; Chen, L.; Zhang, D.; Luo, S.; Sheng, Y.; Liu, X.; Liu, Q.; Li, J.; Shi, B.; Peng, G.; et al. Prediction of Surgical Approach in Mitral Valve Disease by XGBoost Algorithm Based on Echocardiographic Features. J. Clin. Med. 2023, 12, 1193. [Google Scholar] [CrossRef]
  48. Liu, Y.; Durlofsky, L.J. 3D CNN-PCA: A deep-learning-based parameterization for complex geomodels. Comput. Geosci. 2021, 148, 104676. [Google Scholar] [CrossRef]
  49. Nahiduzzaman; Goni, O.F.; Anower, S.; Islam, R.; Ahsan, M.; Haider, J.; Gurusamy, S.; Hassan, R.; Islam, R. A Novel Method for Multivariant Pneumonia Classification Based on Hybrid CNN-PCA Based Feature Extraction Using Extreme Learning Machine With CXR Images. IEEE Access 2021, 9, 147512–147526. [Google Scholar] [CrossRef]
  50. Khan, M.A.; Majid, A.; Hussain, N.; Alhaisoni, M.; Zhang, Y.-D.; Kadry, S.; Nam, Y. Multiclass Stomach Diseases Classification Using Deep Learning Features Optimization. Comput. Mater. Contin. 2021, 67, 3381–3399. [Google Scholar] [CrossRef]
Figure 1. A sample from the Kvasir gastroenterology dataset classes: (a) before improving and (b) after improvement.
Figure 1. A sample from the Kvasir gastroenterology dataset classes: (a) before improving and (b) after improvement.
Diagnostics 13 01758 g001
Figure 2. A sample of all classes from the Kvasir dataset after applying the GVF segmentation algorithm. (a) Original image; (b) image segmentation; (c) region of interest.
Figure 2. A sample of all classes from the Kvasir dataset after applying the GVF segmentation algorithm. (a) Original image; (b) image segmentation; (c) region of interest.
Diagnostics 13 01758 g002
Figure 3. FFNN architecture for endoscopy image analysis of the Kvasir gastroenterology dataset classification.
Figure 3. FFNN architecture for endoscopy image analysis of the Kvasir gastroenterology dataset classification.
Diagnostics 13 01758 g003
Figure 4. Hybrid methodology: CNN–FFNN and CNN–XGBoost based on the GVF segmentation algorithm for endoscopy image analysis to classify the Kvasir gastroenterology dataset.
Figure 4. Hybrid methodology: CNN–FFNN and CNN–XGBoost based on the GVF segmentation algorithm for endoscopy image analysis to classify the Kvasir gastroenterology dataset.
Diagnostics 13 01758 g004
Figure 5. Hybrid methodology: CNN–FFNN and CNN–XGBoost based on fused CNN features and the GVF segmentation algorithm for endoscopy image analysis to classify the Kvasir gastroenterology dataset.
Figure 5. Hybrid methodology: CNN–FFNN and CNN–XGBoost based on fused CNN features and the GVF segmentation algorithm for endoscopy image analysis to classify the Kvasir gastroenterology dataset.
Diagnostics 13 01758 g005
Figure 6. Confusion matrices displaying CNN–FFNN hybrid systems’ performance based on ROIs using the GVF algorithm for endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases: (a) GoogLeNet–FFNN, (b) MobileNet–FFNN, and (c) DenseNet121–FFNN.
Figure 6. Confusion matrices displaying CNN–FFNN hybrid systems’ performance based on ROIs using the GVF algorithm for endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases: (a) GoogLeNet–FFNN, (b) MobileNet–FFNN, and (c) DenseNet121–FFNN.
Diagnostics 13 01758 g006
Figure 7. Confusion matrices displaying CNN–XGBoost hybrid systems’ performances based on ROIs using the GVF algorithm for endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases: (a) GoogLeNet–XGBoost, (b) MobileNet–XGBoost, and (c) DenseNet121–XGBoost.
Figure 7. Confusion matrices displaying CNN–XGBoost hybrid systems’ performances based on ROIs using the GVF algorithm for endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases: (a) GoogLeNet–XGBoost, (b) MobileNet–XGBoost, and (c) DenseNet121–XGBoost.
Diagnostics 13 01758 g007
Figure 8. Confusion matrices displaying CNN–FFNN hybrid systems’ performances based on ROIs and feature CNN fused for the endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases: (a) GoogLeNet–MobileNet–FFNN, (b) MobileNet–DenseNet121–FFNN, (c) GoogLeNet–DenseNet121–FFNN, and (d) GoogLeNet–MobileNet–DenseNet121–FFNN.
Figure 8. Confusion matrices displaying CNN–FFNN hybrid systems’ performances based on ROIs and feature CNN fused for the endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases: (a) GoogLeNet–MobileNet–FFNN, (b) MobileNet–DenseNet121–FFNN, (c) GoogLeNet–DenseNet121–FFNN, and (d) GoogLeNet–MobileNet–DenseNet121–FFNN.
Diagnostics 13 01758 g008
Figure 9. Confusion matrices displaying CNN–XGBoost hybrid systems’ performances based on ROIs and feature CNN fused for the endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases: (a) GoogLeNet–MobileNet–XGBoost, (b) MobileNet–DenseNet121–XGBoost, (c) GoogLeNet–DenseNet121–XGBoost, and (d) GoogLeNet–MobileNet–DenseNet121–XGBoost.
Figure 9. Confusion matrices displaying CNN–XGBoost hybrid systems’ performances based on ROIs and feature CNN fused for the endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases: (a) GoogLeNet–MobileNet–XGBoost, (b) MobileNet–DenseNet121–XGBoost, (c) GoogLeNet–DenseNet121–XGBoost, and (d) GoogLeNet–MobileNet–DenseNet121–XGBoost.
Diagnostics 13 01758 g009
Figure 10. Performance display of all proposed methodologies for analyzing endoscopy images for diagnosing gastroenterological status.
Figure 10. Performance display of all proposed methodologies for analyzing endoscopy images for diagnosing gastroenterological status.
Diagnostics 13 01758 g010
Table 1. Results of pre-trained deep learning models for endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases.
Table 1. Results of pre-trained deep learning models for endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases.
ModelsType of DiseaseAUC %Accuracy %Sensitivity %Precision %Specificity %
GoogLeNetdyed_lifted_polyps88.77979.291.399.2
dyed_resection_margins92.591.592.483.296.5
esophagitis90.283.583.182.797.4
normal_cecum94.598.597.691.698.7
normal_pylorus93.898.597.892.999.3
normal_z_line91.68181.281.896.9
polyps86.27978.99498.6
ulcerative_colitis94.294.594.289.298.9
Average ratio91.4688.2088.0588.3498.19
MobileNetdyed_lifted_polyps92.69191.288.898.1
dyed_resection_margins90.48787.387.498.3
esophagitis92.190.590.889.297.6
normal_cecum93.291.591.591.599.4
normal_pylorus95.49685.786.498.8
normal_z_line88.181.580.983.697.5
polyps86.481.581.276.996.3
ulcerative_colitis89.782.582.188.297.7
Average ratio90.9986.4086.3486.5097.96
DenseNet121dyed_lifted_polyps82.473.572.975.497.4
dyed_resection_margins86.275.576.277.896.6
esophagitis81.27878.478.496.9
normal_cecum94.69392.895.498.6
normal_pylorus95.29998.797.199.6
normal_z_line83.47978.877.596.7
polyps94.99291.784.497.8
ulcerative_colitis95.19292.396.399.2
Average ratio89.1385.3085.2385.2997.85
Table 2. Results of CNN models based on the GVF Algorithm for endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases.
Table 2. Results of CNN models based on the GVF Algorithm for endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases.
ModelsType of DiseaseAUC %Accuracy %Sensitivity %Precision %Specificity %
GoogLeNetdyed_lifted_polyps91.289.590.285.298.2
dyed_resection_margins88.38180.887.197.8
esophagitis95.193.593.795.499.3
normal_cecum92.393.594.394.499.4
normal_pylorus93.195.595.295.598.7
normal_z_line92.994.593.892.699.1
polyps89.187.588.487.997.6
ulcerative_colitis92.491.592.388.497.8
Average ratio91.8090.8091.0990.8198.49
MobileNetdyed_lifted_polyps91.29090.490.998.5
dyed_resection_margins89.38786.989.799.1
esophagitis92.49695.693.298.6
normal_cecum93.19595.19599.1
normal_pylorus92.69493.893.599.4
normal_z_line93.49191.193.898.8
polyps90.99090.484.197.9
ulcerative_colitis91.88989.292.299.2
Average ratio91.8491.5091.5691.5598.83
DenseNet121dyed_lifted_polyps91.392.593.289.898.2
dyed_resection_margins94.297.596.895.199.1
esophagitis80.270.57070.595.6
normal_cecum94.396.596.295.598.7
normal_pylorus95.29797.397.599.5
normal_z_line83.47675.678.496.7
polyps94.996.595.89699.2
ulcerative_colitis93.69493.797.499.8
Average ratio90.8990.1089.8390.0398.35
Table 3. Results of CNN–FFNN hybrid system based on ROI using the GVF algorithm for endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases.
Table 3. Results of CNN–FFNN hybrid system based on ROI using the GVF algorithm for endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases.
Hybrid ModelsType of LesionAUC %Accuracy %Sensitivity %Precision %Specificity %
GoogLeNet–FFNNdyed_lifted_polyps92.592.593.290.799.2
dyed_resection_margins91.792.592.888.597.6
esophagitis93.69494.494.598.7
normal_cecum96.89897.897.599.8
normal_pylorus95.796.596.39899.5
normal_z_line92.492.592.992.598.6
polyps92.99393.393.599.1
ulcerative_colitis92.19191.295.898.8
Average ratio93.4693.8093.9993.8898.91
MobileNet–FFNNdyed_lifted_polyps95.494.594.29598.7
dyed_resection_margins94.69393.491.699.4
esophagitis97.598.598.196.699.1
normal_cecum96.89797.297.599.5
normal_pylorus96.39594.893.698.5
normal_z_line94.193.593.996.999.6
polyps92.89292.488.997.8
ulcerative_colitis93.19291.695.899.2
Average ratio95.0894.4094.4594.4998.98
DenseNet121–FFNNdyed_lifted_polyps94.19392.794.998.7
dyed_resection_margins96.394.594.288.797.6
esophagitis97.69797.398.599.5
normal_cecum98.29999.197.599.7
normal_pylorus96.195.595.29598.6
normal_z_line97.29696.494.699.4
polyps95.89493.992.299.1
ulcerative_colitis92.18989.397.399.8
Average ratio95.9394.8094.7694.8499.05
Table 4. Results of CNN–XGBoost hybrid system based on ROI using the GVF algorithm for endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases.
Table 4. Results of CNN–XGBoost hybrid system based on ROI using the GVF algorithm for endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases.
Hybrid ModelsType of DiseaseAUC %Accuracy %Sensitivity %Precision %Specificity %
GoogLeNet–XGBoostdyed_lifted_polyps93.89393.292.599.2
dyed_resection_margins96.59696.492.899.4
esophagitis97.39695.894.698.7
normal_cecum98.697.597.297.599.5
normal_pylorus96.295.594.797.499.8
normal_z_line94.99393.494.998.6
polyps92.390.590.990.599.4
ulcerative_colitis94.192.592.793.998.5
Average ratio95.4694.3094.2994.2699.14
MobileNet– XGBoostdyed_lifted_polyps94.193.594.192.699.2
dyed_resection_margins92.491.592.490.198.8
esophagitis93.99494.495.499.3
normal_cecum95.79695.69799.5
normal_pylorus98.497.597.39799.7
normal_z_line94.79494.293.198.9
polyps92.19191.191.999.1
ulcerative_colitis93.69393.493.599.4
Average ratio94.3693.8094.0693.8399.24
DenseNet121–XGBoostdyed_lifted_polyps93.894.594.29598.5
dyed_resection_margins94.29595.492.799.1
esophagitis96.996.595.996.598.8
normal_cecum98.29999.197.199.5
normal_pylorus95.39595.495.598.6
normal_z_line94.99493.895.498.8
polyps92.887.588.489.799.4
ulcerative_colitis93.29292.191.598.9
Average ratio94.9194.2094.2994.2098.95
Table 5. Results of CNN–FFNN hybrid system based on feature CNN fused for endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases.
Table 5. Results of CNN–FFNN hybrid system based on feature CNN fused for endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases.
Fusion
Features
ClassifierType of DiseaseAUC %Accuracy %Sensitivity %Precision %Specificity %
GoogLeNet–MobileNetFFNNdyed_lifted_polyps95.99696.294.699.2
dyed_resection_margins94.294.594.49599.4
esophagitis98.79998.697.599.8
normal_cecum99.199.599.196.698.8
normal_pylorus96.495.594.69799.5
normal_z_line96.795.594.99698.6
polyps96.195.595.494.199.4
ulcerative_colitis94.89392.897.999.9
Average ratio96.4996.0695.7596.0999.33
MobileNet–DenseNet121FFNNdyed_lifted_polyps96.295.595.295.598.5
dyed_resection_margins95.79594.89599.2
esophagitis98.598.598.49899.9
normal_cecum98.197.597.198.599.6
normal_pylorus96.496.595.893.798.7
normal_z_line95.99595.497.499.8
polyps94.19393.292.199.1
ulcerative_colitis96.395.595.296.599.4
Average ratio96.4095.8095.6495.8499.28
GoogLeNet–DenseNet121FFNNdyed_lifted_polyps96.29595.29698.5
dyed_resection_margins97.997.597.194.799.2
esophagitis96.89796.898.599.8
normal_cecum99.199.599.397.599.7
normal_pylorus97.29695.997.599.5
normal_z_line98.497.596.695.199.2
polyps95.69494.394.498.6
ulcerative_colitis96.595.595.198.599.8
Average ratio97.2196.5096.2996.5399.29
GoogLeNet–MobileNet–DenseNet121FFNNdyed_lifted_polyps97.89797.296.599.2
dyed_resection_margins98.297.597.494.798.7
esophagitis97.196.595.897.599.5
normal_cecum98.99999.198.599.7
normal_pylorus97.59796.696.598.8
normal_z_line95.195.594.895.599.1
polyps96.29796.99799.6
ulcerative_colitis95.895.595.69999.8
Average ratio97.0896.9096.6896.9099.30
Table 6. Results of CNN–XGBoost hybrid system based on feature CNN fused for endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases.
Table 6. Results of CNN–XGBoost hybrid system based on feature CNN fused for endoscopy image analysis of the Kvasir dataset to discriminate gastrointestinal diseases.
Fusion FeaturesClassifierType of DiseaseAUC %Accuracy %Sensitivity %Precision %Specificity %
GoogLeNet–MobileNetXGBoostdyed_lifted_polyps95.194.594.297.499.5
dyed_resection_margins97.89898.197.599.8
esophagitis96.897.596.796.198.5
normal_cecum95.79696.497.5100
normal_pylorus96.297.597.398.599.6
normal_z_line98.898.597.89799.8
polyps97.297.596.594.798.8
ulcerative_colitis95.294.594.495.599.2
Average ratio96.6096.8096.4396.7899.40
MobileNet–DenseNet121XGBoostdyed_lifted_polyps93.294.593.89599.2
dyed_resection_margins95.19696.293.798.7
esophagitis98.19796.99999.6
normal_cecum97.298.598.496.198.9
normal_pylorus93.894.593.796.499.4
normal_z_line96.89897.794.798.8
polyps95.693.594.195.499.4
ulcerative_colitis94.29595.496.999.8
Average ratio95.595.995.7895.999.23
GoogLeNet–DenseNet121XGBoostdyed_lifted_polyps94.29494.29499.4
dyed_resection_margins93.99393.490.798.8
esophagitis99.49998.797.599.6
normal_cecum97.89897.697.5100
normal_pylorus95.494.594.395.598.5
normal_z_line95.995.594.99699.2
polyps96.195.595.395.599.4
ulcerative_colitis93.792.592.895.498.3
Average ratio95.8095.3095.1595.2699.15
GoogLeNet–MobileNet–DenseNet121XGBoostdyed_lifted_polyps94.69594.697.499.5
dyed_resection_margins98.698.597.89799.7
esophagitis98.498.598.29799.6
normal_cecum97.196.595.998.5100
normal_pylorus98.197.597.498.599.7
normal_z_line99.298.598.297.599.9
polyps98.797.597.395.698.8
ulcerative_colitis95.69695.596.598.6
Average ratio97.5497.2596.8697.2599.48
Table 7. Results of proposed methodologies for analyzing endoscopy images for early diagnosis of gastroenterological status from the Kvasir dataset.
Table 7. Results of proposed methodologies for analyzing endoscopy images for early diagnosis of gastroenterological status from the Kvasir dataset.
TechniquesFeaturesDyed Lifted
Polyps
Dyed
Resection
Margins
EsophagitisNormal
Cecum
Normal
Pylorus
Normal z LinePolypsUlcerative ColitisAccuracy
pre-trainedGoogLeNet7991.583.598.598.5817994.588.2
MobileNet918790.591.58681.581.582.5
DenseNet12173.575.578939979929283.5
Based on GVFGoogLeNet89.58193.593.595.594.587.591.590.8
MobileNet908796959491908991.5
DenseNet12192.597.570.596.5977696.59490.1
HybridFFNNGoogLeNet92.592.59498.596.592.5939193.8
MobileNet94.59398.5979593.5929294.4
DenseNet1219394.5979995.596948994.8
HybridXGBoostGoogLeNet93969697.595.59390.592.594.3
MobileNet93.591.5949697.594919393.8
DenseNet12194.59596.599959487.59294.2
HybridFFNNGoogLeNet–MobileNet9694.59999.595.595.595.59396.1
MobileNet–DenseNet12195.59598.597.596.5959395.595.8
GoogLeNet–DenseNet1219597.59799.59697.59495.596.5
GoogLeNet–MobileNet–DenseNet1219797.596.5999795.59795.596.9
HybridXGBoostGoogLeNet–MobileNet94.59897.59697.598.597.594.596.8
MobileNet–DenseNet12194.5969798.594.59893.59595.9
GoogLeNet–DenseNet1219493999894.595.595.592.595.3
GoogLeNet–MobileNet–DenseNet1219598.598.596.597.598.597.59697.3
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ahmed, I.A.; Senan, E.M.; Shatnawi, H.S.A. Hybrid Models for Endoscopy Image Analysis for Early Detection of Gastrointestinal Diseases Based on Fused Features. Diagnostics 2023, 13, 1758. https://doi.org/10.3390/diagnostics13101758

AMA Style

Ahmed IA, Senan EM, Shatnawi HSA. Hybrid Models for Endoscopy Image Analysis for Early Detection of Gastrointestinal Diseases Based on Fused Features. Diagnostics. 2023; 13(10):1758. https://doi.org/10.3390/diagnostics13101758

Chicago/Turabian Style

Ahmed, Ibrahim Abdulrab, Ebrahim Mohammed Senan, and Hamzeh Salameh Ahmad Shatnawi. 2023. "Hybrid Models for Endoscopy Image Analysis for Early Detection of Gastrointestinal Diseases Based on Fused Features" Diagnostics 13, no. 10: 1758. https://doi.org/10.3390/diagnostics13101758

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop