Next Article in Journal
ITGAM rs1143679 Variant in Systemic Lupus Erythematosus Is Associated with Increased Serum Calcification Propensity
Previous Article in Journal
Genomic Dynamics and Functional Insights under Salt Stress in Gossypium hirsutum L.
Previous Article in Special Issue
Unraveling the Dysbiosis of Vaginal Microbiome to Understand Cervical Cancer Disease Etiology—An Explainable AI Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

EDLM: Ensemble Deep Learning Model to Detect Mutation for the Early Detection of Cholangiocarcinoma

1
Department of Computer Science, Bahria University, Islamabad 04408, Pakistan
2
Department of Computer, College of Science and Arts in Ar Rass, Qassim University, Ar Rass 51921, Qassim, Saudi Arabia
3
Department of Computer Science, Bahria University, Lahore 54782, Pakistan
4
Department of Computer Science, University of Management and Technology, Lahore 54782, Pakistan
*
Author to whom correspondence should be addressed.
Genes 2023, 14(5), 1104; https://doi.org/10.3390/genes14051104
Submission received: 3 April 2023 / Revised: 11 May 2023 / Accepted: 14 May 2023 / Published: 18 May 2023
(This article belongs to the Special Issue Bioinformatics of Sequencing Data: A Machine Learning Approach)

Abstract

:
The most common cause of mortality and disability globally right now is cholangiocarcinoma, one of the worst forms of cancer that may affect people. When cholangiocarcinoma develops, the DNA of the bile duct cells is altered. Cholangiocarcinoma claims the lives of about 7000 individuals annually. Women pass away less often than men. Asians have the greatest fatality rate. Following Whites (20%) and Asians (22%), African Americans (45%) saw the greatest increase in cholangiocarcinoma mortality between 2021 and 2022. For instance, 60–70% of cholangiocarcinoma patients have local infiltration or distant metastases, which makes them unable to receive a curative surgical procedure. Across the board, the median survival time is less than a year. Many researchers work hard to detect cholangiocarcinoma, but this is after the appearance of symptoms, which is late detection. If cholangiocarcinoma progression is detected at an earlier stage, then it will help doctors and patients in treatment. Therefore, an ensemble deep learning model (EDLM), which consists of three deep learning algorithms—long short-term model (LSTM), gated recurrent units (GRUs), and bi-directional LSTM (BLSTM)—is developed for the early identification of cholangiocarcinoma. Several tests are presented, such as a 10-fold cross-validation test (10-FCVT), an independent set test (IST), and a self-consistency test (SCT). Several statistical techniques are used to evaluate the proposed model, such as accuracy (Acc), sensitivity (Sn), specificity (Sp), and Matthew’s correlation coefficient (MCC). There are 672 mutations in 45 distinct cholangiocarcinoma genes among the 516 human samples included in the proposed study. The IST has the highest Acc at 98%, outperforming all other validation approaches.

1. Introduction

With the continuous expansion of medical technology, the time of “big data” has arrived. Artificial intelligence (AI) and many AI technologies are being utilized in the medical services industry to unlock the unlimited potential of big data [1]. Cholangiocarcinoma is one of the deadliest forms of cancer in people and is now the number one killer and disability in the world [2]. Cholangiocarcinoma develops when the DNA of bile duct cells is altered. The DNA of a cell conveys guidelines that direct the cell’s activities. Due to these modifications, cells grow uncontrolled and aggregate into masses known as tumors, which can infiltrate and damage healthy bodily parts [3].
The tumor suppressor gene TP53 is mostly responsible for the alterations in cholangiocarcinoma. Additionally, bile duct cancer may be influenced by the genes KRAS, HER2, and ALK. Some of the genetic modifications that lead to bile duct cancer may be influenced by inflammation [3].
The process of the development of Cholangiocarcinoma is explained with the help of Figure 1. Cholangiocarcinoma is divided into three most common categories: Extrahepatic cholangiocarcinoma is a disease of the extrahepatic bile channels [4]. The cancer may end up in the liver or the small intestine. Cholangiocarcinoma, which begins outside the liver but in the region where the bile ducts and main blood arteries join the liver, is a subclass of extrahepatic cholangiocarcinoma [5]. The second one is intrahepatic cholangiocarcinoma, a cancer of the bile channels in the liver, and the third one is gallbladder cholangiocarcinoma, which starts in the gallbladder.
Cholangiocarcinoma kills about 7000 people a year. Women die less often than men. Asians have the highest death rate. Between 2021 and 2022, African Americans had the largest increase in cholangiocarcinoma deaths (45%), Whites (20%) and Asians (22%) were next [6]. A total of 60–70% of patients with cholangiocarcinoma are diagnosed with local infiltration or distant metastases, thus losing the possibility of a curative surgical intervention. Less than a year is the median survival time across the board [7].
Medical imaging, which includes computed tomography (CT), magnetic resonance imaging (MRI), and ultrasound, is the most effective and common non-invasive diagnostic method for cholangiocarcinoma identification (US) [8]. Cholangiocarcinoma diagnosis could be aided by artificial intelligence (AI). Given the rarity of the condition, the heterogeneity of the tumor’s anatomic location and risk factors, and the importance of AI in cholangiocarcinoma detection [9], several machine learning (ML) AI methods, such as logistic regression, support vector machines (SVMs), artificial neural networks (ANNs), and convolutional neural networks (CNNs), have been used to identify cholangiocarcinoma [10]. This study uses a machine learning framework to pinpoint cholangiocarcinoma.
A rare kind of cancer called cholangiocarcinoma starts in the bile ducts. The small intestine receives bile (a digestive liquid) from the liver and gallbladder through bile ducts, which are tiny tubes. Most cholangiocarcinoma cases are discovered after the disease has progressed outside the bile ducts [11]. Treatment is tough, and the chances of recovery are often slim. The exact etiology of cholangiocarcinoma is not known by experts. Risk factors suggest that diseases that cause chronic (permanent) irritation of the bile ducts may contribute to the growth of this cancer [12]. DNA changes resulting from constant damage, such as inflammation, can alter some cells’ growth, division, and behavior. The following pie chart shows that more than 70% of liver tumors are caused by hepatocellular carcinoma (HCC), and 8% of liver cancers are cholangiocarcinomas (CCAs), the second most common primary malignancy [13].
The ratio of primary tumors is shown in Figure 2. Most patients with cholangiocarcinoma are over 65 years old. Effective treatment can be difficult because the disease is often not diagnosed until it is at an advanced stage. Depending on where the cancer is and how it develops, affected people can live months to years after diagnosis [13]. In medicine, many machine learning and deep learning technologies are used to detect and prevent various diseases. An artificial neural network (ANN) was demonstrated by Matake et al. in 2006, which can distinguish between the four liver masses (HCC, CCA, hemangioma, and metastasis), and reported an average AUC of 0.961 [14].
In 2009, Logeswaran used a popular ANN, a multilayer perceptron, to differentiate images with and without CCA [15]. As a result, the Acc of the test for distinguishing healthy images from tumor images was 94%, and the Acc of the multi-disease test was 88%. To assess the Sp of several serum indicators to enhance CCA diagnosis, Pattanapairoj et al. constructed a classification model in 2015 using both C4.5 (the technique used to build classification models for decision trees in logical form) and ANN, with an AUC of 0.961 [16]. Shao et al. created an ANN model in 2018, which is crucial for patients with inoperable CCA when choosing a course of therapy, with an AUC of 0.964 [17].
Another study by Peng et al., in 2019, offered a novel method of precision treatment for CCA patients, using radiographic signatures of 128 CCA patients based on US scans to noninvasively characterize the biological activity of CCA, with an AUC of 0.930 [18] as shown Table 1. Additionally, Yang et al. investigated the MRI radiomics model’s diagnostic efficacy using random forests in 2020, reporting an AUC of 0.90 [19].
The AUC values were the best results in the above studies. ANN—artificial neural network; MLP—multi-layer perceptron; C4.5—an algorithm used to construct a decision tree classification model in logical form; BP-ANN—back propagation artificial neural network; LASSO—least absolute shrinkage and selection operator; and SVM—support vector machine [13].
Artificial intelligence can automatically offer a quantitative and impartial evaluation of a tumor by recognizing intricate patterns in picture data [20]. The present status of the artwork has a great deal of limitations and cutoff points. There is not yet a clear benchmark dataset for cholangiocarcinoma transformations and specific successions. Additionally, assessment techniques come up short on essential thoroughness and conviction. Subsequently, apparently, there is sufficient space for the models’ precision to be moved along [21]. The latest and most summed-up dataset, as portrayed in the data acquisition framework section, was collected for this study while keeping these requirements in mind. Three deep learning RNN algorithms were utilized in this study: LSTM, GRU, and BLSTM.
Many machine and deep learning algorithms have been utilized for cancer detection, primarily focusing on image-based approaches that identify cancer after symptoms have already appeared. However, early detection is crucial to enhance treatment outcomes. This study aims to achieve early detection by detecting cancer through mutation detection in gene sequences. The dataset used in this study is not publicly available as it is formulated from multiple renowned databases in this study. To enable efficient mutation detection, extensive feature extraction techniques are employed. Moreover, the training process is enhanced by ensemble multiple deep learning algorithms. The performance of the proposed models is evaluated using various testing techniques to ensure their effectiveness.

2. Materials and Methods

This section talks about the inside activity and full clarification of the dataset collection, feature extraction, and classification strategies. The proposed methodology consists of dataset curation, extensive feature extraction, EDLM for accurate classification, testing, and evaluation. The proposed EDLM consists of three deep learning models—LSTM, GRU, and BLSTM. The whole process is explained with the help of Figure 3.

2.1. Data Acquisition Framework

The dataset is the main part of this study. A complete data collection framework was built to train, test, and evaluate the EDLM. The most common way of obtaining reliable and exact information for a study is known as information assortment. Information procurement is the most common way of gathering information to direct research and illustrating how the information is accumulated from a solid source [22].
In human genes, two types of mutations occur: driver mutations, passenger mutations. The type of mutation in cells that causes cancer is known as a driver mutation. Cells grow abnormally because of driver mutation [23]. The dataset was developed so that it contains both normal and mutated sequences. The normal gene sequences were gathered from asia.ensembl.org (accessed on 13 November 2022) [24] using web scrapping code (WSC) developed in Python. CD-HIT was used to reduce sequence redundancy and improve performance. These gene sequences are available, but a generalized dataset related to mutated sequences is not available. Therefore, mutation information was obtained from IntOGen.org [25] using WSC written in Python. The mutation information contains the address of each element in the normal gene sequence. It has both nucleotides before mutation and after mutation. Therefore, a code was written in Python named Generated Mutated sequences (GMSs). which incorporated these changes in the normal gene sequences and built mutated sequences. All the normal gene sequences of all genes were combined in one file, and all the mutated gene sequences of all genes were combined in another file. Thus, the final dataset was formulated to have both normal gene sequences and mutated gene sequences.
The 516 human samples included in the proposed study contain 672 mutations related to 45 different cholangiocarcinoma genes. In Table 2, cholangiocarcinoma genes are listed.
A word cloud visualization generated with Python to highlight the nucleotides and the frequency and significance of each nucleotide in all gene sequences connected to cholangiocarcinoma are represented with the size of each gene in Figure 4.

2.2. Feature Extraction

Feature extraction effectively reduces the amount of data that must be processed while accurately and comprehensively defining the initial data set by combining variables with selections and/or characteristics. It enhances the performance and Acc of training models [26].
Feature extraction techniques are used to extract key characteristics from the raw data source. The process of collecting data in numerous phases to extract critical characteristics required in model training is known as feature extraction [27]. This is the most crucial phase in the preparation of machine learning and deep learning algorithms. Attribute extraction finds data patterns, which are then employed in data training and testing procedures.
This study calculates statistical moments such as Hahn, raw, and central moments. These feature extraction approaches are used to extract key features from data from mutant gene sequences and normal gene sequences [27,28]. All feature extraction techniques are listed in Figure 5.

2.2.1. Hahn Moments Calculation

Hahn moments are used to compute statistical parameters. Hahn moments are the most important concept in pattern recognition. They compute the mean and variance for the data collection. The Hahn polynomial is used to calculate Hahn moments. The size and placement of these Hahn points remain constant. These details are significant because they are sensitive to biological sequence information and can extract hidden properties from gene sequences [29].
Hahn moments need the use of two-dimensional data. As a result, the genomic sequences are transformed into a two-dimensional matrix A of size x x , as shown in the following equation:
A = A 11 A 12 A 1 n A 21 A 22 A 2 n A n 1 A n 2 A n n
The gene sequence is defined by A′ in this case. The value of A′ is used to compute Hahn moments.
Each element A′ represents a genomic sequence residue. The computation of statistical moments of the third order can be found in [30]. Hahn moments are orthogonal since they take in a square matrix. For the benchmark dataset, the Hahn polynomial can be calculated using the following equations:
h m a , b ( C , D ) = ( D + V 1 ) m ( D 1 ) m z = 0 m   ( 1 ) z ( m ) z ( C ) z ( 2 D + a + b m 1 ) z ( D + b 1 ) z ( D 1 ) z 1 z !
All integers a and b must be positive and are predefined constants. The size of the data array is G , and the instant’s order is
A p , q = j = 0 G 1   i = 0 G 1   δ p , q h p a , b ( j , G ) h q a , b ( j , G ) , m , n = 0,1 , 2 , , G 1
where a , b are predetermined constants, is any member of the square matrix and represents the current order.
If A is an integer, then [ 0 , G 1 ] ( G is the provided positive integer). The polynomial’s shape can be altered using these moveable variables [31]. Pochhammer’s symbol is ( a ) k = a ( a + 1 ) ( a + k 1 ) = r ( a + k ) r ( a ) . Equations (2) and (3) may be used to easily determine the normalized Hahn moments of any order. The following numbers stand for the Hahn-moment-based distinctive features: H 00 , H 01 , H 10 , H 11 , H 02 , H 12 , H 21 , H 30 , H 03 . For each gene sequence, 10 raw, 10 central, and 10 Hahn moments are calculated. These moments are united into the miscellaneous super feature vectors and are up to 3rd order.

2.2.2. Central Moments Calculation

Utilizing mean and variance, the central moment of feature extraction is utilized to extract key features. It is the region closest to the mean of the randomly chosen variable in the probability distribution [32]. The following equation illustrates the generic formula for calculating the central moments for the cholangiocarcinoma dataset:
C a , b = g = 1 m   h = 1 m   ( e x ) a ( h y ) b δ g h
C 00 , C 01 , C 10 , C 11 , C 02 , C 12 , C 21 , C 30 ,   and   C 03 are the designations for the distinctive qualities from central moments up to the 3rd order.

2.2.3. Raw Moments Calculation

The raw moments are used for statistical computations. Imputation is the process of maintaining facts by substituting the best available replacement values for missing data values in a data collection. The following equation shows the initial moments for the 2D data of order x + y [33]:
R x , y = g = 1 m   h = 1 m   g x h y δ g h
Up to order three, raw moments are computed. This provides detailed information on important sequence elements, including R 00 , R 01 , R 10 , R 02 , R 20 , R 03 ,   and   R 30 .

2.2.4. Position Relative Incidence Matrix (PRIM)

The location of each gene in the cholangiocarcinoma gene sequence is determined using PRIM [31]. The following equation displays a PRIM-generated matrix with a dimension of 30 × 30:
P P R I M = P 1 1 P 1 2 P 1 j P 1 m P 2 1 P 2 2 P 2 j P 2 m P n 1 P n 2 P n j P n m P m 1 P m 2 P m j P m m

2.2.5. Reverse Position Relative Incidence Matrix (RPRIM)

The RPRIM operates similarly to PRIM but in the other direction [31]. The following equation describes, in detail, how to calculate RPRIM for the cholangiocarcinoma dataset:
R P R I M = R 1 1 R 1 2 R 1 j R 1 m R 2 1 R 2 2 R 2 j R 2 m R n 1 R n 2 R n j R n m R m 1 R m 2 R m j R m m

2.2.6. Accumulative Absolute Position Incidence Vector (AAPIV)

A frequency matrix provides information about the frequency of a gene in a sequence of genes. AAPIV describes the different nucleotide configurations that occur in gene sequences. AAPIV is used to study the nucleotide sequences of cancer gene sequences against each other. The following equation serves as an illustration of the relative gene sequences of cholangiocarcinoma [33]:
K = λ 1 , λ 2 , , λ n
where n is the number of total nucleotides in the gene sequence, which may be estimated using equation regardless of its component:
λ i = j = 1 m   β j
where β j represents where the i t h nucleotide is located.

2.2.7. Reverse Accumulative Absolute Position Incidence Vector (RAAPIV)

RAAPIV functions similarly to AAPIV but in the other direction. Here, the RAAPIV equation is presented:
λ = n 1 , n 2 , , n m

2.2.8. Frequency Vector Calculation

A dataset consists of tens of thousands of data records, each with a unique set of properties. The sequence of genes that come together to produce a gene sequence is represented by a frequency matrix. This is represented by the following equation:
α = ε 1 , ε 2 , , ε n
The frequency of each gene in the cholangiocarcinoma gene sequence may be seen below. The following equation can be used to determine the frequency vector:
F V = f 1 , f 2 , , f n
The frequency of each gene in the gene sequence is shown here by the numbers f1 to fn.

3. Proposed Deep Learning Algorithms

In this study, an ensemble deep learning model (EDLM) is proposed, which consists of three deep learning models—LSTM, GRU, and BLSTM—for the early diagnosis of cholangiocarcinoma. Several malignancies, including cholangiocarcinoma, are recognized, detected, predicted, and diagnosed using the proposed EDLM. An input layer, an output layer, a pooling layer, a dense layer, and a dropout layer are just a few of the layers that make up a deep neural network model. Fully connected layers are then placed on top of everything else [34]. Each layer receives input from the layer before it and analyses the features. Algorithms with intrinsic learning characteristics inside these layers can educate themselves using several learning techniques.
LSTM, GRU, and BLSTM are three types of deep learning RNN algorithms [35] used in this study. For the detection of cholangiocarcinoma, these algorithms employ three assessment methods: An SCT, IST, and 10-FCVT.

3.1. Long Short-Term Memory (LSTM)

In challenges involving sequence prediction, LSTM networks, a kind of recurrent neural network, may discover order dependency [36]. The efficiency of the RNN declines with slot length. In essence, LSTMs are capable of long-term information storage. The information’s length is modified to 64. The addition of a 128-neuron LSTM layer is also mentioned in [37]. The dense layer links the feedback from all the layers and transmits it to the result layer. A 20% dropout capacity is added in the dropout layer to prevent the model from overfitting. In order to combat model overfitting, this model features two dropout layers. In the LSTM layer, Stochastic Gradient Descent (SGD) is utilized as an enhancer. A sigmoid capability is utilized as an initiation capability, the details of which are also available in [38]. These details are also shown with the help of Figure 6.
Which input value is used to change the memory depends on the input gate. Whether 0 or 1 values are allowed is determined by the sigmoid function [39]. The tanh function also assigns a weight to the provided data, defining its value on a scale from −1 to 1.
i t = σ W i h t 1 , x t + b i c t = t a n h W c h t 1 , x t + b c
The Forget gate identifies the details in the block that should be erased. A sigmoid function determines it. It examines the previous state (ht − 1) and the content input (xt) for each number in the cell state, ct − 1, and returns a value between 0 (omit this) and 1 [39].
f t = σ W f . h t 1 , x t + b f
The block’s input and memory are both used to calculate the output. This is determined by the sigmoid function whether 0 or 1 values are acceptable. Additionally, the tanh function defines which values between 0 and 1 can pass [40]. Additionally, the tanh function gives the provided values weight by determining their significance on a scale from −1 to 1 and multiplying it by the sigmoid output.
o t = σ W o h t 1 , x t + b o h t = o t t a n h c t

3.2. Gated Recurrent Unit (GRU)

The second deep learning technique used in the suggested study is the GRU strategy. GRU accomplishes comparable tasks as LSTM but has fewer gates. GRU outperforms LSTM in terms of outcomes [40] since it uses fewer gates and parameters. The reset gate and the update gate are the only gates used by GRU in the cell. The update gate in the GRU controls how much past data are used, whereas the reset gate controls how much past data are disregarded [41]. The GRU cell structure utilized to identify cholangiocarcinomas is shown in the Figure 7.
The following equations show the work process of GRU:
r t = σ x t U r + B t 1 W r z t = σ x t U z + B t 1 W z h t = t a n h r t B t 1 U + x t W y t = 1 z t B t + z t B t 1
In the proposed model, the input is transformed into a vector with a fixed word length of 64 by a single embedding layer. A GRU layer with 256 neurons and a fundamental RNN layer with 128 neurons make up the second layer. Two dropout layers are added at 30% to avoid overfitting. A substantial layer of 10 neurons is introduced at the end. Stochastic Gradient Descent (SGD) is used as an optimizer in the GRU layer, as explained in [42]. As an activation function, the sigmoid function is used. Sparse categorical cross entropy (SCCE) is used to reduce the loss experienced during training the proposed model.

3.3. Bi-Directional LSTM (BLSTM)

The BLSTM extends the regular LSTM. The model incorporates two parallel LSTM layers to produce a forward and backward loop, as seen in the picture below [42]. The network is supposed to produce predictions by using past and future information from forward and backward sequences. In this scenario, the present information is dependent on previous information and is also related to future information [43]. The red and green arrows in the Figure 8 and the equations below reflect the forward and backward consequences, respectively.
h t = g ( U h x t + W h h t 1 + b h ) h t = g U h x t + W h h t 1 + b h y t = g V h h t + V h h t + b y
The following Figure 8 explains the BLSTM cell structure used in the identification of cholangiocarcinoma [44,45].

4. Ensemble Deep Learning Models (EDLM)

EDLM is a cycle where different assorted models are made to foresee a result, either by utilizing an assortment of modeling algorithms or by utilizing an assortment of training data sets. The model then, at that point, joins each base model’s conjecture, yielding a solitary last expectation for the concealed information. The purpose of using EDLM is to reduce prediction generalization errors. When using the EDLM, the prediction error of a model that is different and independent from the base model is reduced. Technology seeks public wisdom when making predictions. An ensemble model has multiple base models, but it works and functions as a single model [46].
With the use of a group learning strategy, this study focuses on the demonstration of unique deep learning models, including LSTM, GRU, and BLSTM. Three groups, such as the training set, validation set, and test set, are created from the extensive feature-extracted dataset. V denotes the validation set, whereas T denotes the test set [47]. Every single LSTM, GRU, and BLSTM deep learning model receives the training set as a contribution. In order to get scan ranges and the optimal attributes for suggested ensemble learning model bounds, the matrix inquiry improvement approach is also used. As shown in Figure 9, EDLM is created from each unique deep learning model under the names training model1, training model2, and training model3; these models are represented as LSTM, GRU, and BLSTM, respectively. Similar studies have also been conducted using machine and deep learning techniques [47,48,49,50,51,52,53,54,55,56,57,58,59,60,61].
Testing and validation sets are utilized to assess each training model. At last, as displayed in the outcomes area, an EDLM produces last superior outcomes.
g p , i = n = 1 N   w n f n , i
The singular profound learning model is provided loads to fabricate group learning forecast in the situation. Here, n signifies the weight given to every exceptional profound learning model, means each model’s forecast, and is the perception [62].
The numerical recipes used to decide the results of the calculations are recorded below. The recipes to determine the responsiveness, particularity, exactness, and Matthew’s correlation coefficient (MCC) are portrayed in the equations below, in a specific order:
S n = T P / ( T P + F N ) S p = T N / ( T N + F P ) A c c = ( T P + T N ) / ( T P + F P + F N + T N )
M C C = ( T P X T N ) ( F P X F N ) ( T P + F P ) ( T P + F N ) ( T N + F P ) ( T N + F N )
In these equations, T N , T P , F N , a n d   F P are represented by every true negative value, all the dataset’s true positive values, false negative values, and false positive values, respectively.
Sn in the equations above relates to the capacity to forecast the count that can accurately detect cholangiocarcinoma. The capacity to anticipate the count that will really reveal the lack of cholangiocarcinoma is referred to as Sp. All individuals in TP + FN have the specified condition [63], while TN + FP are the subjects devoid of the mentioned circumstances. Total participants with good outcomes are denoted by TP + FP, whereas patients with negative results are denoted by TN + FN [64,65].

5. Results

To extract the key aspects of the balanced data, the cholangiocarcinoma dataset is pre-processed. To the retrieved data, extensive feature extraction techniques are applied. Then, the proposed EDLM deep learning methods are employed. IST, SCT, and 10-FCVT were used to gauge the effectiveness of the deep learning algorithm. The outcomes of various validation procedures are as follows.

5.1. Self-Consistency Test (SCT)

SCT is an iterative testing process; when results are satisfied, it stops. A total of 100% of the information collected during an SCT is used for testing and training. The complete dataset is used in SCT for both training and testing. Very minimal loss occurs in BLSTM. In contrast, the SCT showed that LSTM, GRU, and BLSTM had extremely excellent Acc. The ROC curve using SCT is shown in Figure 10.
The outcome, defined by the ROC curve as being between 0.99 and 1.0, should be regarded as excellent. The decision boundary results using SCT are shown in Figure 11.

5.2. Independent Set Test (IST)

The proposed study was also validated via IST. The values from the confusion matrix are used to determine the model’s Acc. This test is related to the primary performance measuring method for the proposed model. In this case, the algorithm is trained on 80% of the dataset and tested on 20% of the dataset. The ROC curve of EDLM in IST is shown in Figure 12.
The decision boundary result using IST are shown in Figure 13.

5.3. 10-Fold Cross-Validation (10-FCVT)

The data are uniformly subsampled into 10 groups for the 10-FCVT. By dividing the training set into 10 divisions, treating each part in the validation set, and training the rest 9-fold, you may choose the model’s hyperparameters and architecture. This process is repeated 10 times, and then, the average value is calculated. Figure 14 shows the ROC curve of EDLM in 10-FCVT.
A decision boundary visualization of each fold obtained through EDLM in 10-FCVT is shown in Figure 15.

5.4. Results Comparison

The result of the EDLM is contrasted with those of its own distinctive algorithms, including LSTM, GRU, and BLSTM, in Table 3. Many measurements are used for comparison. A comparison is made using all three types of tests, such as SCT, IST, and 10-FCVT. Table 3 demonstrates how the proposed EDLM improves the identification precision of the stand-alone deep learning techniques, including LSTM, GRU, and BLSTM. In Table 3, it is clearly mentioned that BLSTM performs very well in SCT and IST, and on the other side, EDLM performs very well in 10-FCVT, which is the best representation of the whole dataset. The statistical tools that evaluate the models are Acc, Sn, Sp, and MCC. The Acc, Sn, Sp, and MCC of BLSTM in SCT are 99%, 100%, 98%, and 0.98, respectively. The Acc, Sn, Sp, and MCC of BLSTM in IST are 98%, 100%, 96%, and 0.95, respectively. The Acc, Sn, Sp, and MCC of EDLM in 10-FCVT are 92%, 94%, 93%, and 0.86, respectively.
Table 4 provides a comprehensive overview of the performance comparison between the previous approaches and proposed models. The BLSTM model stands out with its exceptional AUC value of 0.99, indicating its superior predictive capabilities. Additionally, the proposed algorithms, as a whole, exhibit better accuracies at 98% than those achieved by previous methods, further emphasizing their potential for enhancing predictive modeling tasks. These findings highlight the importance of continued research and development in the field, as advancements in machine learning algorithms have the potential to revolutionize various domains.
The proposed ELDM can also be applied to other types of cancerous datasets. The results produced by the proposed model on a prostate cancer dataset are shown in Table 5. It seems that the proposed model is very effective for the detection of mutations to detect cancer. The model is also validated on other types of cancer datasets, and therefore, this demonstrates the generalizability of the proposed model.

6. Discussion

Cholangiocarcinoma (CCA), one of the deadliest types of cancer in humans, is currently the leading cause of death and disability worldwide. In this study, an EDLM composed of three deep learning models—LSTM, GRU and BLSTM—is proposed. The proposed system is a viable in silico strategy for tracking down mutations in cholangiocarcinoma. When contrasted with the present status of art, the recommended system is a computationally clever indicator. A complete data collection framework is developed in Python to scrap the data from well-known databases and develop mutated gene sequences. An extensive feature extraction framework is developed to extract useful features from the gene sequences and prepare the dataset for training and testing. The EDLM based on ensemble learning of LSTM, GRU, and BLSTM is developed to learn the hidden features of the prepared dataset and identify mutations in cholangiocarcinoma for early detection. Multiple testing techniques, such as SCT, IST, and 10-FCVT, are used to test the proposed model. Multiple statistical tools, such as Acc, Sn, Sp, and MCC, are used for the evaluation of the proposed EDLM, LSTM, GRU, and BLSTM. The performance of EDLM in terms of ROC curve in SCT, IST, and 10-FCVT is shown with the help of Figure 10, Figure 12 and Figure 14 respectively. The performance of EDLM in terms of decision boundary in SCT, IST, and 10-FCVT is shown with the help of Figure 11, Figure 13 and Figure 15 respectively. The evaluation results are shown in Table 3. The best performance in SCT is shown by BLSTM. The best performance in IST is also shown by BLSTM. The best performance in 10-FCVT is shown by EDLM, as discussed in Table 3.

7. Conclusions

Cholangiocarcinoma (CCA), one of the deadliest types of cancer in humans, is currently the leading cause of death and disability worldwide. The best results for early cholangiocarcinoma cancer diagnosis using EDLM are shown in this study. The proposed EDLM consists of three different deep learning algorithms—LSTM, GRU, and BLSTM— that are used to discover mutations in cholangiocarcinoma. All algorithms have above 95% Acc, as shown in Table 3, with an AUC value of 99. The diagnosis of cholangiocarcinoma made using these results is the most precise to date. Table 3 shows the results of the IST, SCT, and 10-FCVT in terms of Acc, Sn, Sp, and MCC. Right now, these are the best techniques for cholangiocarcinoma early diagnosis. Future studies will build on this effort to identify other diseases, and an ensemble deep learning model of other types will also be made.

Author Contributions

Methodology, A.A.S.; software, A.A.S.; validation, F.A.; data curation, T.A.; writing—original draft, A.A.S. and A.F.; writing—review and editing, F.A., T.A., and Y.D.K.; visualization, T.A.; supervision, Y.D.K.; project administration, F.A. and Y.D.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Informed Consent Statement

Not applicable.

Acknowledgments

The researchers would like to thank the Deanship of Scientific Research, Qassim University, for funding the publication of this project.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hulsen, T.; Jamuar, S.S.; Moody, A.R.; Karnes, J.H.; Varga, O.; Hedensted, S.; Spreafico, R.; Hafler, D.A.; McKinney, E.F. From big data to precision medicine. Front. Med. 2019, 6, 34. [Google Scholar] [CrossRef] [PubMed]
  2. Haghbin, H.; Aziz, M. Artificial intelligence and cholangiocarcinoma: Updates and prospects. World J. Clin. Oncol. 2022, 13, 125–134. [Google Scholar] [CrossRef] [PubMed]
  3. Sirica, A.E.; Gores, G.J.; Groopman, J.D.; Selaru, F.M.; Strazzabosco, M.; Wang, X.W.; Zhu, A.X. Intrahepatic Cholangiocarcinoma: Continuing Challenges and Translational Advances. Hepatology 2019, 69, 1803–1815. [Google Scholar] [CrossRef] [PubMed]
  4. Patel, T. Cholangiocarcinoma-controversies and challenges. Nat. Rev. Gastroenterol. Hepatol. 2011, 8, 189–200. [Google Scholar] [CrossRef]
  5. Yao, X.; Huang, X.; Yang, C.; Hu, A.; Zhou, G.; Ju, M.; Lei, J.; Shu, J. A novel approach to assessing differentiation degree and lymph node metastasis of extrahepatic cholangiocarcinoma: Prediction using a radiomics-based particle swarm optimization and support vector machine model. JMIR Med. Inform. 2020, 8, e23578. [Google Scholar] [CrossRef]
  6. Petrick, J.L.; Yang, B.; Altekruse, S.F.; Van Dyke, A.L.; Koshiol, J.; Graubard, B.I.; McGlynn, K.A. Risk factors for intrahepatic and extrahepatic cholangiocarcinoma in the United States: A population-based study in SEER-Medicare. PLoS ONE 2017, 12, e0186643. [Google Scholar] [CrossRef]
  7. Horgan, A.M.; Amir, E.; Walter, T.; Knox, J.J. Adjuvant therapy in the treatment of biliary tract cancer: A systematic review and meta-analysis. J. Clin. Oncol. 2012, 30, 1934–1940. [Google Scholar] [CrossRef]
  8. Malaguarnera, G.; Paladina, I.; Giordano, M.; Malaguarnera, M.; Bertino, G.; Berretta, M. Serum markers of intrahepatic cholangiocarcinoma. Dis. Markers 2013, 34, 219–228. [Google Scholar] [CrossRef]
  9. Bi, Q.; Goodman, K.E.; Kaminsky, J.; Lessler, J. What is machine learning? A primer for the epidemiologist. Am. J. Epidemiol. 2019, 188, 2222–2239. [Google Scholar] [CrossRef]
  10. Saha, S.K.; Zhu, A.X.; Fuchs, C.S.; Brooks, G.A. Forty-Year Trends in Cholangiocarcinoma Incidence in the U.S.: Intrahepatic Disease on the Rise. Oncologist 2016, 21, 594–599. [Google Scholar] [CrossRef]
  11. Khan, A.S.; Dageforde, L.A. Cholangiocarcinoma. Surg. Clin. N. Am. 2019, 99, 315–335. [Google Scholar] [CrossRef] [PubMed]
  12. Tyson, G.L.; El-Serag, H.B. Risk factors for cholangiocarcinoma. Hepatology 2011, 54, 173–184. [Google Scholar] [CrossRef] [PubMed]
  13. Beretta, G.D.; Robertolabianca, B.; Zampino, M.G.; Gemmagatta, C.; Volkerheinemann, D. Cholangiocarcinoma. Crit. Rev. Oncol. Hematol. 2009, 69, 259–270. [Google Scholar] [CrossRef]
  14. Matake, K.; Yoshimitsu, K.; Kumazawa, S.; Higashida, Y.; Irie, H.; Asayama, Y.; Nakayama, T.; Kakihara, D.; Katsuragawa, S.; Doi, K.; et al. Usefulness of Artificial Neural Network for Differential Diagnosis of Hepatic Masses on CT Images. Acad. Radiol. 2006, 13, 951–962. [Google Scholar] [CrossRef] [PubMed]
  15. Logeswaran, R. Cholangiocarcinoma-An automated preliminary detection system using MLP. J. Med. Syst. 2009, 33, 413–421. [Google Scholar] [CrossRef] [PubMed]
  16. Pattanapairoj, S.; Silsirivanit, A.; Muisuk, K.; Seubwai, W.; Cha’On, U.; Vaeteewoottacharn, K.; Sawanyawisuth, K.; Chetchotsak, D.; Wongkham, S. Improve Discrimination Power of Serum Markers for Diagnosis of Cholangiocarcinoma Using Data Mining-Based Approach; Elsevier: Amsterdam, The Netherlands, 2015; Available online: https://www.sciencedirect.com/science/article/pii/S0009912015001204 (accessed on 16 October 2022).
  17. Shao, F.; Huang, Q.; Wang, C.; Qiu, L.; Hu, Y.G.; Zha, S.Y. Artificial neural networking model for the prediction of early occlusion of bilateral plastic stent placement for inoperable hilar cholangiocarcinoma. Surg. Laparosc. Endosc. Percutaneous Tech. 2018, 28, e54–e58. Available online: https://www.ingentaconnect.com/content/wk/slept/2018/00000028/00000002/art00004 (accessed on 16 October 2022). [CrossRef]
  18. Peng, Y.-T.; Zhou, C.-Y.; Lin, P.; Wen, D.-Y.; Wang, X.-D.; Zhong, X.-Z.; Pan, D.-H.; Que, Q.; Li, X.; Chen, L.; et al. Preoperative Ultrasound Radiomics Signatures for Noninvasive Evaluation of Biological Characteristics of Intrahepatic Cholangiocarcinoma. Acad. Radiol. 2020, 27, 785–797. [Google Scholar] [CrossRef]
  19. Yang, C.; Huang, M.; Li, S.; Chen, J.; Yang, Y.; Qin, N.; Huang, D.; Shu, J. Radiomics Model of Magnetic Resonance Imaging for Predicting Pathological Grading and Lymph Node Metastases of Extrahepatic Cholangiocarcinoma; Elsevier: Amsterdam, The Netherlands, 2020; Available online: https://www.sciencedirect.com/science/article/pii/S0304383519305919 (accessed on 16 October 2022).
  20. Razumilava, N.; Gores, G.J. Classification, Diagnosis, and Management of Cholangiocarcinoma; Elsevier: Amsterdam, The Netherlands, 2013; Available online: https://www.sciencedirect.com/science/article/pii/S1542356512010506 (accessed on 17 October 2022).
  21. Vazhayil, A.; KP, S. DeepProteomics: Protein family classification using Shallow and Deep Networks. arXiv 2018, arXiv:1809.04461. [Google Scholar]
  22. Turecek, D.; Holy, T.; Jakubek, J.; Pospisil, S.; Vykydal, Z. PixEDLMan: A multi-platform data acquisition and processing software package for Medipix2, Timepix and Medipix3 detectors. J. Instrum. 2011, 6, C01046. [Google Scholar] [CrossRef]
  23. Bozic, I.; Antal, T.; Ohtsuki, H.; Carter, H.; Kim, D.; Chen, S.; Karchin, R.; Kinzler, K.W.; Vogelstein, B.; Nowak, M.A. Accumulation of driver and passenger mutations during tumor progression. Proc. Natl. Acad. Sci. USA 2010, 107, 18545–18550. [Google Scholar] [CrossRef]
  24. Gene: TP53 (ENSG00000141510)-Summary-Homo_Sapiens-Ensembl Genome Browser 108. Available online: http://asia.ensembl.org/Homo_sapiens/Gene/Summary?g=ENSG00000141510;r=17:7661779-7687538 (accessed on 13 November 2022).
  25. IntOGen-Cancer Driver Mutations in Breast Adenocarcinoma. Available online: https://intogen.org/search?cancer=BRCA (accessed on 13 November 2022).
  26. Shah, A.A.; Khan, Y.D. Identification of 4-carboxyglutamate residue sites based on position based statistical feature and multiple classification. Sci. Rep. 2020, 10, 16913. [Google Scholar] [CrossRef]
  27. Levine, M.D. Feature Extraction: A Survey. Proc. IEEE 1969, 57, 1391–1407. [Google Scholar] [CrossRef]
  28. Ghoraani, B.; Krishnan, S. Time—Frequency Matrix Feature Extraction and Classification of Environmental Audio Signals. IEEE Trans. Audio Speech Lang. Process. 2011, 19, 2197–2209. [Google Scholar] [CrossRef]
  29. Hall, A.R. Generalized Method of Moments. 2004. Available online: https://books.google.com/books?hl=en&lr=&id=HQVREAAAQBAJ&oi=fnd&pg=PR9&ots=_0NfFCexpL&sig=21Uxpib37-Wz4QhTV1BowcdVcJo (accessed on 13 November 2022).
  30. Zhu, H.; Shu, H.; Zhou, J.; Luo, L.; Coatrieux, J.-L. Image analysis by discrete orthogonal dual Hahn moments. Pattern Recognit. Lett. 2007, 28, 1688–1704. [Google Scholar] [CrossRef]
  31. Malebary, S.J.; Khan, Y.D. Evaluating machine learning methodologies for identification of cancer driver genes. Sci. Rep. 2021, 11, 12281. [Google Scholar] [CrossRef]
  32. Sohail, M.U.; Shabbir, J.; Sohil, F. Imputation of missing values by using raw moments. Stat. Transit. New Ser. 2019, 20, 21–40. [Google Scholar] [CrossRef]
  33. Butt, A.H.; Khan, Y.D. CanLect-Pred: A cancer therapeutics tool for prediction of target cancerlectins using experiential annotated proteomic sequences. IEEE Access 2019, 8, 9520–9531. [Google Scholar] [CrossRef]
  34. Hochreiter, S. The vanishing gradient problem during learning recurrent neural nets and problem solutions. Int. J. Uncertain. Fuzziness Knowl.-Based Syst. 1998, 6, 107–116. [Google Scholar] [CrossRef]
  35. Wang, H.; Chen, S.; Xu, F.; Jin, Y.-Q. Application of deep-learning algorithms to MSTAR data. In Proceedings of the 25 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 3743–3745. [Google Scholar]
  36. Agnes, S.A.; Anitha, J.; Solomon, A.A. Two-stage lung nodule detection framework using enhanced UNet and convolutional LSTM networks in CT images. Comput. Biol. Med. 2022, 149, 106059. [Google Scholar] [CrossRef]
  37. Sundermeyer, M.; Schlüter, R.; Ney, H. LSTM neural networks for language modeling. In Proceedings of the Thirteenth Annual Conference of The International Speech Communication Association, Portland, OR, USA, 9–13 September 2012. [Google Scholar]
  38. Rengasamy, D.; Jafari, M.; Rothwell, B.; Chen, X.; Figueredo, G.P. Deep Learning with Dynamically Weighted Loss Function for Sensor-Based Prognostics and Health Management. Sensors 2020, 20, 723. [Google Scholar] [CrossRef]
  39. Lin, G.; Shen, W. Research on convolutional neural network based on improved Relu piecewise activation function. Procedia Comput. Sci. 2018, 131, 977–984. [Google Scholar] [CrossRef]
  40. Staudemeyer, R.C.; Morris, E.R. Understanding LSTM—A tutorial into long short-term memory recurrent neural networks. arXiv 2019, arXiv:1909.09586. [Google Scholar]
  41. Gao, Y.; Glowacka, D. Deep gate recurrent neural network. In Proceedings of the Asian Conference on Machine Learning, Hamilton, New Zealand, 16–18 November 2016; pp. 350–365. [Google Scholar]
  42. Dey, R.; Salem, F.M. Gate-variants of gated recurrent unit (GRU) neural networks. In Proceedings of the 2017 IEEE 60th International Midwest Symposium on Circuits and Systems (MWSCAS), Boston, MA, USA, 6–9 August 2017; pp. 1597–1600. [Google Scholar]
  43. Guo, H.; Tang, R.; Ye, Y.; Li, Z.; He, X.; Dong, Z. Deepfm: An end-to-end wide & deep learning framework for CTR prediction. arXiv 2018, arXiv:1804.04950. [Google Scholar]
  44. Graves, A.; Schmidhuber, J. Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Netw. 2005, 18, 602–610. [Google Scholar] [CrossRef] [PubMed]
  45. Basaldella, M.; Antolli, E.; Serra, G.; Tasso, C. Bidirectional lstm recurrent neural network for keyphrase extraction. In Proceedings of the Italian Research Conference on Digital Libraries, Udine, Italy, 25–26 January 2018; pp. 180–187. [Google Scholar]
  46. Mendes-Moreira, J.; Soares, C.; Jorge, A.M.; de Sousa, J.F. Ensemble approaches for regression: A survey. Acm Comput. Surv. 2012, 45, 1–40. [Google Scholar] [CrossRef]
  47. Breiman, L. Bagging predictors. Mach Learn 1996, 24, 123–140. [Google Scholar] [CrossRef]
  48. Schapire, R.E. The strength of weak learnability. Mach Learn 1990, 5, 197–227. [Google Scholar] [CrossRef]
  49. Stefenon, S.F.; Ribeiro, M.H.D.M.; Nied, A.; Mariani, V.C.; Coelho, L.D.S.; Leithardt, V.R.Q.; Silva, L.A.; Seman, L.O. Hybrid wavelet stacking ensemble model for insulators contamination forecasting. IEEE Access 2021, 9, 66387–66397. [Google Scholar] [CrossRef]
  50. Shah, A.A.; Malik, H.A.M.; Mohammad, A.; Khan, Y.D.; Alourani, A. Machine Learning Techniques for Identification of Carcinogenic Mutations, Which Cause Breast Adenocarcinoma. Sci. Rep. 2022, 12, 11738. [Google Scholar] [CrossRef]
  51. Shah, A.A.; Alturise, F.; Alkhalifah, T.; Khan, Y.D. Deep Learning Approaches for Detection of Breast Adenocarcinoma Causing Carcinogenic Mutations. Int. J. Mol. Sci. 2022, 23, 11539. [Google Scholar] [CrossRef]
  52. Shah, A.A.; Alturise, F.; Alkhalifah, T.; Khan, Y.D. Evaluation of Deep Learning Techniques for Identification of Sarcoma-Causing Carcinogenic Mutations. Digit. Health 2022, 8, 20552076221133703. [Google Scholar] [CrossRef] [PubMed]
  53. Sohail, A.; Nawaz, N.A.; Shah, A.A.; Rasheed, S.; Ilyas, S.; Ehsan, M.K. A Systematic Literature Review on Machine Learning and Deep Learning Methods for Semantic Segmentation. IEEE Access 2022, 10, 134557–134570. [Google Scholar] [CrossRef]
  54. Shah, A.A.; Malik, H.A.M.; Muhammad, A.; Alourani, A.; Butt, Z.A. Deep Learning Ensemble 2D CNN Approach towards the Detection of Lung Cancer. Sci. Rep. 2023, 13, 2987. [Google Scholar] [CrossRef]
  55. Amanat, S.; Ashraf, A.; Hussain, W.; Rasool, N.; Khan, Y.D. Identification of Lysine Carboxylation Sites in Proteins by Integrating Statistical Moments and Position Relative Features via General PseAAC. Biomolecules 2020, 10, 396–407. [Google Scholar] [CrossRef]
  56. Hussain, W.; Rasool, N.; Khan, Y.D. Insights into Machine Learning-Based Approaches for Virtual Screening in Drug Discovery: Existing Strategies and Streamlining through FP-CADD. Molecules 2021, 26, 463–472. [Google Scholar] [CrossRef]
  57. Hussain, W.; Rasool, N.; Khan, Y.D.; Screening, H.T. A Sequence-Based Predictor of Zika Virus Proteins Developed by Integration of PseAAC and Statistical Moments. Int. J. Environ. Res. Public Health 2020, 17, 797–804. [Google Scholar] [CrossRef]
  58. Khan, Y.D.; Alzahrani, E.; Alghamdi, W.; Ullah, M.Z. Sequence-Based Identification of Allergen Proteins Developed by Integration of PseAAC and Statistical Moments via 5-Step Rule. Biomolecules 2020, 10, 1046–1055. [Google Scholar] [CrossRef]
  59. Mahmood, M.K.; Ehsan, A.; Khan, Y.D.; Chou, K.-C. iHyd-LysSite (EPSV): Identifying Hydroxylysine Sites in Protein Using Statistical Formulation by Extracting Enhanced Position and Sequence Variant Feature Technique. Cells 2020, 9, 536–545. [Google Scholar] [CrossRef]
  60. Naseer, S.; Hussain, W.; Khan, Y.D.; Rasool, N. Optimization of Serine Phosphorylation Prediction in Proteins by Comparing Human Engineered Features and Deep Representations. Int. J. Mol. Sci. 2021, 22, 114069. [Google Scholar] [CrossRef]
  61. Naseer, S.; Hussain, W.; Khan, Y.D.; Rasool, N. Sequence-Based Identification of Arginine Amidation Sites in Proteins Using Deep Representations of Proteins and PseAAC. Biomolecules 2020, 10, 937–948. [Google Scholar] [CrossRef]
  62. Naseer, S.; Hussain, W.; Khan, Y.D.; Rasool, N. NPalmitoylDeep-PseAAC: A Predictor of N-Palmitoylation Sites in Proteins Using Deep Representations of Proteins and PseAAC via Modified 5-Steps Rule. Int. J. Mol. Sci. 2021, 22, 294–305. [Google Scholar] [CrossRef]
  63. Naseer, S.; Hussain, W.; Khan, Y.D.; Rasool, N. iPhosS (Deep)-PseAAC: Identify Phosphoserine Sites in Proteins Using Deep Learning on General Pseudo Amino Acid Compositions via Modified 5-Steps Rule. Bioinformatics 2020, 36, 5709–5711. [Google Scholar] [CrossRef] [PubMed]
  64. Feng, P.; Yang, H.; Ding, H.; Lin, H.; Chen, W.; Chou, K.-C. iDNA6mA-PseKNC: Identifying DNA N6-methyladenosine sites by incorporating nucleotide physicochemical properties into PseKNC. Genomics 2019, 111, 96–102. [Google Scholar] [CrossRef] [PubMed]
  65. Hoo, Z.H.; Candlish, J.; Teare, D. What is an ROC curve? Emerg. Med. J. 2017, 34, 357–359. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The development of cholangiocarcinoma.
Figure 1. The development of cholangiocarcinoma.
Genes 14 01104 g001
Figure 2. The relative incidence of primary liver tumors.
Figure 2. The relative incidence of primary liver tumors.
Genes 14 01104 g002
Figure 3. Methodology of the proposed framework for the detection of cholangiocarcinoma.
Figure 3. Methodology of the proposed framework for the detection of cholangiocarcinoma.
Genes 14 01104 g003
Figure 4. Word cloud of cholangiocarcinoma genes.
Figure 4. Word cloud of cholangiocarcinoma genes.
Genes 14 01104 g004
Figure 5. Extensive feature extraction techniques for the cholangiocarcinoma dataset.
Figure 5. Extensive feature extraction techniques for the cholangiocarcinoma dataset.
Genes 14 01104 g005
Figure 6. LSTM cell structure for cholangiocarcinoma.
Figure 6. LSTM cell structure for cholangiocarcinoma.
Genes 14 01104 g006
Figure 7. GRU cell structure for cholangiocarcinoma.
Figure 7. GRU cell structure for cholangiocarcinoma.
Genes 14 01104 g007
Figure 8. BLSTM cell structure for cholangiocarcinoma.
Figure 8. BLSTM cell structure for cholangiocarcinoma.
Genes 14 01104 g008
Figure 9. The process of proposed EDLM.
Figure 9. The process of proposed EDLM.
Genes 14 01104 g009
Figure 10. ROC curve of EDLM in SCT.
Figure 10. ROC curve of EDLM in SCT.
Genes 14 01104 g010
Figure 11. Decision boundary result of EDLM in SCT.
Figure 11. Decision boundary result of EDLM in SCT.
Genes 14 01104 g011
Figure 12. ROC curve of EDLM in IST.
Figure 12. ROC curve of EDLM in IST.
Genes 14 01104 g012
Figure 13. Decision boundary result of EDLM in IST.
Figure 13. Decision boundary result of EDLM in IST.
Genes 14 01104 g013
Figure 14. ROC curve of 10-fold c-val.
Figure 14. ROC curve of 10-fold c-val.
Genes 14 01104 g014
Figure 15. Decision boundary of EDLM in 10-FCVT.
Figure 15. Decision boundary of EDLM in 10-FCVT.
Genes 14 01104 g015
Table 1. The following table shows a compilation of prior work.
Table 1. The following table shows a compilation of prior work.
AuthorMethodsArea Under the Curve
Matake, et al. [14]ANN0.961
Logeswaran [15]MLP0.960
Pattanapairoj, et al. [16]C4.5, ANN0.961
Shao, et al. [17]BP-ANN0.9648
Peng, et al. [18]LASSO, SVM0.930
Yang, et al. [19]Random Forest0.90
Table 2. Genes involved in cholangiocarcinoma and mutations.
Table 2. Genes involved in cholangiocarcinoma and mutations.
SymbolMutationsSamplesSymbolMutationsSamples
TP539896SF3B1136
KRAS6561LRP1B266
ARID1A2830IDH255
SMAD42428TGFBR255
IDH12826AR95
PIK3CA2119PCDH1765
ARID22118FBXW785
PBRM11918GNAS114
BAP12616KDM6A64
CDKN2A1515PIK3R154
NF11714CDH1183
KMT2C1711TRRAP73
ERBB21310BCOR53
EPHA21610FGFR283
ERBB3149CLTC83
ATM189KEAP112
ELF399NCOR122
NRAS108ARAF32
PTEN98KMT2D162
APC137EXT232
STK1197MAP2K422
RBM10127BIRC6112
RNF4322
Table 3. Comparison of results of evaluation tools of LSTM, GRU, BLSTM, and EDLM.
Table 3. Comparison of results of evaluation tools of LSTM, GRU, BLSTM, and EDLM.
Self-Consistency Test (SCT)Independent Set Test (IST)10-Fold Cross-Validation Test (10-FCVT)
AccSnSpMCCAccSnSpMCCAccSnSpMCC
LSTM94%97%93%0.8993%95%91%0.8792%91%93%0.85
GRU94%95%93%0.8894%96%92%0.8992%91%93%0.84
BLSTM99%100%98%0.9898%100%96%0.9592%91%93%0.84
EDLM94%95%93%0.8896%98%94%0.9292%94%93%0.86
Table 4. Comparison of results of proposed models to other previous results.
Table 4. Comparison of results of proposed models to other previous results.
Previous Results
AuthorModelsArea Under the Curve
Matake, et al. [14]ANN0.961
Logeswaran [15]MLP0.960
Pattanapairoj, et al. [16]C4.5, ANN0.961
Shao, et al. [17]BP-ANN0.9648
Peng, et al. [18]LASSO, SVM0.930
Yang, et al. [19]Random Forest0.90
Proposed Results
ModelsArea Under the Curve
LSTM0.98
GRU0.97
BLSTM0.99
EDLM0.98
Table 5. The results of the proposed ELDM on prostate cancer.
Table 5. The results of the proposed ELDM on prostate cancer.
Evaluation MatricesValuesEvaluation MatricesValues
Accuracy98.87%Precision98.87%
Sensitivity99.50%Recall98.87%
Specificity99.63%F1 Score98.87%
MCC0.98AUC0.99
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shah, A.A.; Alturise, F.; Alkhalifah, T.; Faisal, A.; Khan, Y.D. EDLM: Ensemble Deep Learning Model to Detect Mutation for the Early Detection of Cholangiocarcinoma. Genes 2023, 14, 1104. https://doi.org/10.3390/genes14051104

AMA Style

Shah AA, Alturise F, Alkhalifah T, Faisal A, Khan YD. EDLM: Ensemble Deep Learning Model to Detect Mutation for the Early Detection of Cholangiocarcinoma. Genes. 2023; 14(5):1104. https://doi.org/10.3390/genes14051104

Chicago/Turabian Style

Shah, Asghar Ali, Fahad Alturise, Tamim Alkhalifah, Amna Faisal, and Yaser Daanial Khan. 2023. "EDLM: Ensemble Deep Learning Model to Detect Mutation for the Early Detection of Cholangiocarcinoma" Genes 14, no. 5: 1104. https://doi.org/10.3390/genes14051104

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop