Next Article in Journal
Plazomicin against Multidrug-Resistant Bacteria: A Scoping Review
Next Article in Special Issue
A Robust Computer-Aided Automated Brain Tumor Diagnosis Approach Using PSO-ReliefF Optimized Gaussian and Non-Linear Feature Space
Previous Article in Journal
Wound Healing and Anti-Inflammatory Effects of a Newly Developed Ointment Containing Jujube Leaves Extract
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

RNN and BiLSTM Fusion for Accurate Automatic Epileptic Seizure Diagnosis Using EEG Signals

by
Nagwan Abdel Samee
1,
Noha F. Mahmoud
2,*,
Eman A. Aldhahri
3,
Ahsan Rafiq
4,*,
Mohammed Saleh Ali Muthanna
5 and
Ijaz Ahmad
6,7,8
1
Department of Information Technology, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, Riyadh 11671, Saudi Arabia
2
Rehabilitation Sciences Department, Health and Rehabilitation Sciences College, Princess Nourah bint Abdulrahman University, Riyadh 11671, Saudi Arabia
3
Department of Computer Science and Artificial Intelligence, College of Computer Sciences and Engineering, University of Jeddah, Jeddah 23218, Saudi Arabia
4
School of Automation, Chongqing University of Posts and Telecommunications, Chongqing 400065, China
5
Institute of Computer Technologies and Information Security, Southern Federal University, 347922 Taganrog, Russia
6
CAS Key Laboratory of Human-Machine Intelligence-Synergy Systems, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
7
Shenzhen College of Advanced Technology, University of Chinese Academy of Sciences, Shenzhen 518055, China
8
Guangdong-Hong Kong-Macao Joint Laboratory of Human-Machine Intelligence-Synergy Systems, Chinese Academy of Sciences, Shenzhen 518055, China
*
Authors to whom correspondence should be addressed.
Life 2022, 12(12), 1946; https://doi.org/10.3390/life12121946
Submission received: 17 October 2022 / Revised: 17 November 2022 / Accepted: 18 November 2022 / Published: 22 November 2022
(This article belongs to the Special Issue Artificial Intelligence Applications for Imaging in Life Sciences)

Abstract

:
Epilepsy is a common neurological condition. The effects of epilepsy are not restricted to seizures alone. They comprise a wide spectrum of problems that might impair and reduce quality of life. Even with medication, 30% of epilepsy patients still have recurring seizures. An epileptic seizure is caused by significant neuronal electrical activity, which affects brain activity. EEG shows these changes as high-amplitude spiky and sluggish waves. Recognizing seizures on an electroencephalogram (EEG) manually by a professional neurologist is a time-consuming and labor-intensive process, hence an efficient automated approach is necessary for the identification of epileptic seizure. One technique to increase the speed and accuracy with which a diagnosis of epileptic seizures could be made is by utilizing computer-aided diagnosis systems that are built on deep neural networks, or DNN. This study introduces a fusion of recurrent neural networks (RNNs) and bi-directional long short-term memories (BiLSTMs) for automatic epileptic seizure identification via EEG signal processing in order to tackle the aforementioned informational challenges. An electroencephalogram’s (EEG) raw data were first normalized after undergoing pre-processing. A RNN model was fed the normalized EEG sequence data and trained to accurately extract features from the data. Afterwards, the features were passed to the BiLSTM layers for processing so that further temporal information could be retrieved. In addition, the proposed RNN-BiLSTM model was tested in an experimental setting using the freely accessible UCI epileptic seizure dataset. Experimental findings of the suggested model have achieved avg values of 98.90%, 98.50%, 98. 20%, and 98.60%, respectively, for accuracy, sensitivity, precision, and specificity. To further verify the new model’s efficacy, it is compared to other models, such as the RNN-LSTM and the RNN-GRU learning models, and is shown to have improved the same metrics by 1.8%, 1.69%, 1.95%, and 2.2% on using 5-fold. Additionally, the proposed method was compared to state-of-the-art approaches and proved to be a more accurate categorization of such techniques.

1. Introduction

One of the most frequent long-lasting neurological disorders is epilepsy. The impacts of epilepsy are far-reaching; they are not limited to the physical manifestations of seizures alone. Rather, they encompass a wide range of difficulties that can lead to impairment and a drastically diminished quality of life [1]. Having epilepsy can have a negative impact on your emotional, mental, and physical well-being. People with epilepsy, and especially those with refractory epilepsy (i.e., seizures that are not well-controlled with seizure medications), face a number of challenges, including a lack of social support and family function, difficulties with cognition, medical, and psychiatric co-occurring disorders, and stigma [2,3]. Epilepsy is a neurological disorder disease that affects over 50 million individuals all over the world [4], and unfortunately, the therapies that are now available, such as anticonvulsants and surgery, can have significant negative consequences on the patients who undergo them. These sudden discharges momentarily interrupt both the patient’s behavior and their ability to function normally. Epilepsy cannot be brought under complete control with the therapies, drugs, or surgical procedures that are now available [5,6,7].
Seizures can strike without warning, and epilepsy can have a significant impact on a person’s social and psychological life; as a result, the condition is regarded as potentially fatal [8]. Therefore, epileptic seizure prediction significantly contributes to an increase in the quality of life of seizure patients through a number of different aspects. These aspects include the generation of an alarm to offer the appropriate action prior to the occurrence of a seizure, the introduction of new treatment approaches, and the developing of new strategies for understanding the nature of the disease [9,10]. In general, conventional therapy for epileptic seizures frequently results in a variety of unwanted side effects, which makes it exceptionally challenging to maintain seizure control in patients [11,12]. According to estimates provided by the World Health Organization (WHO), approximately fifty thousand people have epilepsy in the world [13]. In addition, it is said that 70% of all occurrences of epilepsy-related seizures occur in poor nations, which means that the drugs and treatment facilities necessary for accurately diagnosing epilepsy are not always available.
Due to the fact that epilepsy is a disorder associated with the electrical activity of the brain, the electroencephalogram, or EEG, signal is typically employed for its diagnosis. The interpretation of EEG signals can be used to improve more advanced forms of human interaction [11,13,14]. EEG recordings are now done digitally so that they may be seen on a computer display device. This also makes it possible for the recordings to be automatically evaluated. The use of standardized terminology for seizure categorization also makes it simpler for clinicians who care for people who have epilepsy and who conduct research on epilepsy to communicate with one another. Since the regular functioning of many brain cells is disrupted during an epileptic seizure, the electroencephalogram can be used to detect the onset of a seizure (EEG). Because effective seizure management is contingent on making an accurate diagnosis, it is critical to determine whether or not a patient suffers from epilepsy, as well as the specific type of disorder they have. The categorization of seizures into their various subtypes is helpful for guiding subsequent testing, treatment, and assessment of prognosis or outlook. During an EEG recording, the most difficult part is trying to locate the ictal spikes and seizures. In order to detect epileptic activity, it is necessary for a specialist to perform an in-depth analysis of the entire duration of the EEG recordings [15], which can be a time-consuming process. Both the sheer volume of long-term EEG recordings and their growing prevalence in clinical practice have contributed to a reduction in the likelihood that a specialist will incorrectly interpret the data or fail to arrive at the correct detection.
Different researchers have created a DL model to effectively detect epileptic from EEG signals so as to address the problem described in the previous paragraph. Various machine learning and deep learning (ML/DL) models have been used in previous publications [2,16] to help discover important and distinguishable features associated with biomedical imaging and signal for use in the binary classification task [2,16,17,18,19,20,21,22,23,24,25]. Deep learning has been responsible for a significant amount of the success that has been seen in the automated feature extraction and categorization of EEG epileptic seizures [6,10,15,17,18,19,20,21,22,23,24,25,26]. Despite this, there is still space for development in this area. According to the existing studies, the proper detection of epileptic seizures using EEG signal analysis utilizing hybrid deep learning models is always appropriate, proving the efficacy and efficiency of the technique [2,15,17,19]. This is the case irrespective of the condition of the literature. As a consequence of this, the objective of the hybrid deep learning model that is proposed in this paper is to improve the performance of previously developed EEG seizure classifiers. The following is a list of the most important findings from the study:
  • A hybrid of recurrent neural networks (RNNs) and bi-directional long short-term memories (BiLSTM) is proposed for the purpose of automatically identifying epileptic seizures through the processing of EEG signals.
  • The efficacy of the newly developed model is validated by conducting a complete comparison to the existing state-of-the-art learning models.
  • The recommended method provides a number of advantages, including shorter periods of time needed for detection, a reduced proportion of false positive results, increased sensitivity, and increased specificity.
The paper is being organized as follows: The state-of-the-art of seizure detection is outlined in Section 2. In Section 3, we discuss the data and our hybrid RNN-BiLSTM model. In Section 4, the findings of the experiments are presented, along with a discussion of the study’s conclusions and directions for further research.

2. Literature Review

In the last decade, numerous methods for feature extraction and classification approaches have been developed in order to improve the overall performance of the constructed classifier and expand its capacity for seizure detection. Both the continuous wavelet transform (CWT) and the discrete wavelet transform (DWT) are examples of traditional machine learning techniques used to extract features for the goal of building a classifier for the identification of epileptic seizures [27,28,29,30]. The adequacy of the detection of seizures has been strengthened by the help of DL-based techniques [6,10,15,17,18,19,20,21,22,23,24,25,26]. The introduced algorithms aim at boosting diagnostic standards and automating identifying essential features. To better categorize epileptic seizures, Choi et al. [21] presented a hybrid model that combines one-dimensional convolutional networks (1D CNN), gated recurrent unit networks (GRU). The introduced model has an overall accuracy of 82.86 percent, along with a sensitivity of 80% and a precision of 85 percent. A gated recurrent network, also known as GRN, was described by Affes et al. [31] as a method for predicting epileptic seizures using features collected from EEG data. These features reflect the temporal aspect and the frequency aspect of the signal, respectively. Using data that were collected in the Children’s Hospital of Boston, GRN was able to predict epileptic seizures with an accuracy of 75.6% and a mean sensitivity of 89%. A DL-based framework has been presented by Raghu et al. [32] for the categorization of seven kinds of seizures alongside non-seizure EEG. This was accomplished by the utilization of convolutional neural networks and transfer learning. The model was trained and evaluated using the EEG signals collected from Temple University Hospital. The approaches of transfer learning and feature extraction, taken separately, led to the greatest possible classification accuracy of 88.30% when using Inceptionv3 pretrained DNN. The Deep Clinical Sparse Autoencoder, also known as the DCSAE, was applied by Hilal [6] and colleagues in order to classify epileptic seizures based on EEG signals collected from the benchmarking UCI-Epileptic dataset. In order to extract the important features, the coyote optimization technique has been implemented; after that, the selected features are input into the DCSAE model. They have achieved a specificity of 99.2%, sensitivity of 99.19%, and accuracy of 98.67%. The dataset, UCI-Epileptic, has been addressed by Mursalin et al. [33]. They presented an innovative method for identifying the important elements from a benchmarking EEG data taken from the University of Bonn. The collected features have been put through a number of different traditional classifiers, such as the SVM, Random Forest (RF), and the k-nearest neighbor (KNN). By utilizing an RF classifier, they were able to attain a success rate of 98.7%. A multi-view convolutional neural network, or CNN, framework was presented by Liu et al. [10] for the purpose of predicting the occurrence of epilepsy seizures. The researchers’ objective was to acquire significant features in the time/frequency domain. They were able to get an average area under the curve of 0.82 and 0.89 on two participants of the CHB-MIT dataset. In order to improve the effectiveness of traditional machine learning approaches (such as Support Vector Machines and Neural Networks) in EEG classification, Nagabushanam et al. [23] developed a method that combines the long short-term memory (LSTM) with the neural network (NN). The level of accuracy that was reached was 71.38%. The independently recurrent neural network, a new deep learning model, was recently developed by Yao et al. [34] to build a convenient method for the seizure/non-seizure classification. The suggested method may extract temporal and spatial information from the short-term to long-term range of the complete record. Cross-subject experiments were used to evaluate the CHB-MIT data, which is notoriously noisy. They accomplished 87% accuracy, 87.3% sensitivity, and 86.7% specificity.
A hybrid RNN and LSTM was proposed by Najafi et al. for the purpose of diagnosing epilepsy in [2]. After the RNN was deployed for the feature selection process, the long short-term memory was put to use for the epilepsy classification stage. At the outset, both normal and epileptic LB channels were broken down into three levels, and 15 distinct characteristics were gleaned from each level. The LSTM was then given the selected features to use in its classification process after they had been retrieved from each signal segment individually. The investigation’s findings showed that the proposed algorithm was successful in differentiating normal participants from epileptic subjects with a rate of 96.1% accuracy, 96.8% sensitivity, and 97.4% specificity. Because of the excellence of the RNN-LSTM technique in extracting significant spatial and temporal features from EEG signals, as well as its superior accuracy and sensitivity in epilepsy classification, we were motivated to investigate other similar fusion models based on recurrent neural networks and the bidirectional LSTM. This was made possible by the fact that the RNN-LSTM technique is effective in extracting significant spatial and temporal features from EEG signals. Bidirectional long short-term memory, or BiLSTM for short, is a type of recurrent neural network that is most commonly utilized in natural language processing. In contrast to a conventional LSTM, the input travels in both directions, and the system is able to make use of information from both sides. In addition to this, it is an effective method for modeling the sequential dependencies that exist between words and phrases in both the forward and backward directions of the sequence. As a consequence of this, an RNN-BiLSTM model was presented in this paper for the automated extraction of attributes from EEG data. The purpose of this model is to improve the classification’ accuracies, and sensitivities of these signals. In addition, the combination of bidirectional LSTM and RNNs model has demonstrated superior performance in the classification/forecasting tasks of other non-medical applications [35,36]. Because of this, it has been implemented and evaluated with the intention of improving the existing performance of the state-of-the-art regarding the classification of elliptic seizures.

3. Materials and Methods

Important procedures for applying the suggested model (RNN-BiLSTM) to detect epileptic episodes are outlined below. The process of seizure detection normally consists of two phases. It begins with gathering and cleaning data, then moves on to the suggested model’s automatic feature extraction and selection. This study’s primary structure is shown in Figure 1. The proposed framework’s primary components are the following steps: EEG signal preprocessing; model training; model testing; and model evaluation.

3.1. Data Collection and Preprocessing

This study was conducted using EEG data collected at the University of Bonn as a benchmarking dataset [37] and is split into five different directories and a hundred files per folder, each of which represents a particular sample or person. Each file stores the mental activity of the subject for 23 min, and 6 s. 4097 discrete data points representing the appropriate time series have been gathered. Each data point illustrates a different sampling rate of the electroencephalogram (EEG). Therefore, there are 500 individuals, and for each individual there are 4097 data points spanning 23.5 s. Data from each of the 4097 samples were broken up into 23 sets, each of which is 1 s long and contains 178 individual data points. Therefore, we have 11,500 rows of data (rows), each of which has 178 points of data of 1, and the last column is labeled y. Since each data point reflects the value of an EEG recording at a particular point in time, the chunks were randomly shuffled after division (1,2,3,4,5). In column 179, the variable representing the outcome is indicated by the letter y, while the variables representing the predictors are shown by the letters X1, X2, etc.… X178. The value that is recorded in the variable y represents the subcategory of the 178-dimensional input signal. More specifically, y values that are within the range [1, 2, 3, 4, 5]. In column 179, the outcome variable is represented by the letter y, and the predictors are indicated by the letters X1, X2, … X178, as shown in Figure 2. The y variable stores the 178th dimension of the input signal’s granularity. In this case, y is within the interval [1, 2, 3, 4, 5]. Having the patient’s eyes open during the EEG recording time indicates a value of y = 5. This means the patient was awake. The patient’s eyes were closed during the EEG recording if y = 4, indicating the patient was feeling sleepy. In addition, if y = 3, it means that the tumor has invaded the EEG-normal region of the brain. Moreover, if y = 2, then the EEG was taken from the area where the tumor was located. Moreover, if y equals 1, then seizure-related activity will be recorded. In this research, we used a binary classification approach to the identification of epileptic seizures, where a label of 0 indicates the of an epileptic seizure and labels of 1, while 2, 3, 4, and 5 indicate the absence of an epileptic one. In this experiment we used binary classification epileptic seizure versus non-epileptic seizure detection, so the class label from 2 to 5 was assigned 0, which represents the normal condition of EEG epileptic seizure detection, while 1 represents the patients have epileptic seizures, as shown in Table 1.

3.2. The Proposed RNN-BiLSTM Learning Model

Recurrent neural networks (RNNs) are a type of artificial neural network in which the effects of some nodes’ output on other nodes’ subsequent input can be cycled through. Because of this, it is able to display temporal dynamics in its behavior. Recurrent neural networks (RNNs), which are a descendant of feedforward NNs, have the ability to memories internal states to process sequences of inputs of varying lengths [38,39,40,41]. In an RNN, all of the input vector components share nearly identical weights, as opposed to the feedforward NNs where each component has its own weight. In most cases, recurrent networks outperform traditional methods because they combine the weights of multiple input vector positions into a single vector. Sequences of varying lengths can be processed by the same model by simply reusing the weights. This reduces the number of network-learning parameters (weights), which is an additional benefit. Furthermore, the outputs passed on to the next phase were both calculated using the input vector and several data from the preceding step (usually another vector). Units are shorthand for the formulas used to derive the intermediate results (blocks). Therefore, for the most basic form of a recurrent network, the relation that follows, Equations (1) and (2), can be used to define a block:
o t = f 1 ( W o o o t 1 + W o x x t + b o )
y ^ t = f 2 ( W y o o t + b y )
where x t is an input sequence vector, and t denotes the iteration at which recurrent relations are computed. The f 1 , f 2 represent the activation functions. The weight matrices and biases are denoted by the symbols W o o , W o x , W y o , b o , and b y , respectively.
In the realm of artificial intelligence, LSTM [24,41,42,43,44] is a subset of recurrent neural networks. The disappearing of gradients problem is addressed by the introduction of this architecture. Further, this network type is superior at keeping long-range connections alive by understanding how values at the beginning and end of a sequence are related. In the LSTM model, these expressions take the form of gates. There are three distinct varieties of gates: The forget gate regulates how much data are sent from one memory cell to another. The memory cell’s update-gate (or update input) is the gate responsible for determining if the cell will be updated. Additionally, it regulates the amount of data that a possible new memory cell can send to the current memory cell. The next hidden state’s value is determined by the output gate.
In this study, we introduce the fusion of the RNN and BiLSTM for binary class epilepsy identification. The network consisted of five layers: a sequence input layer, an RNN layer with 100 hidden units, a bidirectional LSTM layer with 200 hidden units, an FC layer, a SoftMax layer, and a classification output layer. The user-specified values for the deep learning layers are shown in Table 2, along with brief descriptions of their function.
The RNN module is based on the schematic depicted in Figure 3. A specific type of RNN structure, the binary long short-term memory (LSTM) network, has been shown to be reliable and effective for simulating a sequence that may be used for various purposes and has lengthy dependencies in research that include time [45].
After the RNN module, a BiLSTM layer was used. Since the collected EEG signals are arranged in a time-sequence based fashion, the present state is strongly influenced by the past environments. To address this problem, the BiLSTM model is the most efficient tool at one’s disposal. Several self-parameterized regulating gates in the BiLSTM module’s memory cell allow for the reading, writing, and clearing of state information, as shown in Figure 4. If that door were to be opened, the cell’s entire informational capability would become available. To recap, the previous cell’s process failed because the forget gate was not turned on. Assuming nothing takes place, the preceding data may be “forgotten.” As can be seen in Figure 4, the output gate can decide for itself if it needs to relay the most recent cell output and the ultimate state.
In order to address the problem of overfitting, we made use of two dropout layers that were positioned below the ReLU activation function, in addition to the BiLSTM layer. Using these dropout layers allows us to avoid the problem of overfitting. In the meantime, the dropout contributes to the objective of minimizing the error of generalization, which is being pursued in conjunction with the growth in the number of layers contained within neural networks. After the BiLSTM layer comes the FC layer, then the SoftMax layer, and finally the classification layer.

3.3. Model Training and Testing

The use of cross-validation is a reliable strategy for avoiding overfitting. The entire dataset has been divided up into sections. In order to perform a typical K-fold cross-validation, the data need to be divided up into k folds. Then, we used the remaining holdout fold as the test set while iteratively training the algorithm on k-1 folds.
In this research we used k cross-validation on to 5-fold. We used the four given folders to train the models, then tested the models on the last folder. This process repeated five times to produce a reliable and practical detection system of epilepsies from EEG signals. Determining the learning parameters for training the DNN that would yield the optimum performance is a complex optimization issue due to the large complexity and non-convexity of the objective function. In this case, the standard practice is to employ stochastic optimizer. This research makes use of the ADAM optimizer with the following settings: learning rate = 0.0001, batch size of 32, and a number of epochs of 400. These parameters were chosen by analyzing experimental validation findings and were applied uniformly across all networks for fair evaluation of their performance and resource requirements.

4. Results and Discussion

In order to carry out experiments, we made use of both a graphics processing unit (GPU) and a central processing unit (CPU) manufactured by Intel and bearing the model number Core i5-7700. In addition, the suggested procedures were trained with the assistance of Keras and Python version 3.6. Accuracy (Acc), Precision (Prec), Sensitivity (Sen), and Specificity ( S p e c i ) are the four-performance metrics that we used to evaluate the proposed model. While accuracy is the proportion of a sample over the entire population that has been correctly classified, sensitivity is the percentage of true positives that have been accurately detected or correctly identifies patients as having epileptic seizures. Sensitivity is more accurately referred to as the true positive rate (TPR), or Recall. Accuracy is presented as a sample over the entire population that has been correctly classified. Specificity demonstrates that the patents in question do not involve any epileptic seizures. Specificity can be referred to as the true negative rate (TNR). The mathematical equations for all of the important performance metrics are listed below, in Equations (2)–(6). The T P o s t , T N e g , F P o s t , and F N e g are the true positive, true negative, false positive, and false negative, respectively. The T N e g output indicates the number of correctly classified negative examples. A similar notation, T P o s t for true positive, denotes the percentage of correctly identified positive examples. A false positive value F P o s t indicates the number of falsely positive examples that were incorrectly labeled as negative ( F N e g ), which indicates the opposite number of falsely negative examples that were incorrectly labeled as positive.
A c c ( % ) = T P o s t + T N e g T P o s t + T N e g + F P o s t + F N e g  
P r e c ( % ) = T P o s T P o s + F P o s
T P R = S e n ( % ) = T P o s T P o s + F N e g  
T N R = S p e c i ( % ) = T N e g T N e g + F P o s
In addition to these metrics, a number of other metrics, such as the Confusion matrix, Matthew’s correlation coefficient, the Roc Curve, false omission rate (FOR), false discovery rate (FDR), and Negative Predictive Value, were used to perform an analysis on the results. The confusion matrix is a specialized form of the contingency table. Contingency tables summarize binary decision-making. It has two dimensions, which are labeled actual and predicted, and both dimensions contain identical sets of classes [46]. Dealing with an imbalanced target variable is a common challenge that arises when applying machine learning techniques to medical applications. When machine learning techniques are used on data that are not evenly distributed, using confusion matrices as error measures is insufficient. Therefore, in this work, we make use of the Matthews correlation coefficient, also known as MCC, which is recognized as an important error measure for data that are not balanced [47,48]. The Matthews correlation coefficient, also known as r, is a statistical measure that compares the true classes, also known as TC, to the predicted labels, also known as PL [49]. Its definition can be found in Equation (7), as shown.
M C C = c o v ( T C , P L ) σ T C × σ P L = ( T P o s × T N e g ) ( F P o s × F N e g ) ( T P o s + F P o s ) ( T P o s + F N e g ) ( T N e g + F N e g ) ( T N e g + F N e g )
where σ T C , and σ P L are the variances of the true classes TC and the predicted labels PL, and c o v ( T C , P L ) is the covariance between them. The MCC can range anywhere from −1 to +1. When the value is −1, it indicates that the prediction is completely inaccurate, whereas when the value is +1, it indicates that the prediction is spot on. When MCC is equal to zero, it indicates that we are thinking about a random classification in which the model predictions do not have any discernible correlation to the actual results.
The area under receiver operator characteristic, ROC curve, is a different type of error metric that we are presenting. The TPR, or sensitivity, is displayed on the ROC curve in relation to the FPR, or specificity [50]. The false discovery rate and the false omission rate are two additional error metrics we have employed [51]. Biology and medicine are common application domains for the FDR and the FOR [52,53]. FDR calculates the percentage of erroneous findings among a group of significant hypothesis tests. The complement of the negative predictive value, false omission rate, is a statistical technique used in multiple hypothesis testing scenarios to account for numerous comparisons. It calculates the percentage of false negatives that are wrongfully rejected. Equations (8) and (9) show the definitions of FDR and FOR. Both measures have a range of 0 to 1. For an errorfree classification, we have F N e g = F P o s = 0 , implying F D R = F O R = 0 . In contrast, when T P o s = T N e g = 0 , we get F D R = F O R = 1 .
F D R = F P o s F P o s + T P o s
F O R = F N e g F N e g + T N e g
The negative predictive value is another error metric for evaluating classification accuracy in the medical domain (NPV) [54]. Given a negative test result, NPV is the likelihood that a person does not have a disease or condition. The percentage of people with negative test results who are correctly identified or diagnosed is what NPV stands for. In other words, when a person receives a negative, normal, test result, NPV represents the likelihood that the person is healthy (e.g., intellectually normal). The definition of NPV is given in Equation (10).
F O R = T N e g T N e g + F N e g  
We have evaluated the performance of the hybrid model that has been proposed, known as RNN-BiLSTM, as well as other models (RNN-LSTM and RNN-GRU) that are comparable to it, that are currently available in the state-of-the-art for the identification of epileptic seizures. Figure 5 presents the confusion matrices that were calculated for the three different models. The combination of RNN and BiLSTM is shown to correctly recognize the classes after being subjected to an in-depth experiment involving the confusion matrix. The proposed model for the classification of epileptic seizures achieves the T P o s t , T N e g percentages of 0.98 and 0.98, respectively. On the other hand, the values that were achieved for the same measures by using the RNN-LSTM and the RNN-GRU were respectively 0.96 and 0.95 and 0.97 and 0.97. In conclusion, the method RNN-BiLSTM was superior to the other two methods, RNN-LSTM and RNN-GRU, in terms of its accuracy in detecting the classes as well as its performance.
In the experimental study, the performance measures on different folds from 1 to 5 are shown in Figure 6. The average accuracy, sensitivity, precision, and specificity have each been improved using the RNN-BiLSTM that we have proposed, which has achieved average values of 98.4%, 98.50%, and 97.40%, and 97.30%, respectively. The results demonstrate that the new model achieves improvements in the same metrics that are average performance measure 2.1%, 2.69%, 1.95%, and 2.2%, respectively, compared to RNN-LSTM and RNN-GRU models.
The ROC is regarded as an essential metric in all EEG epileptic seizure detection. The ROC plots the results of the comparison between the true negative and true positive rates. Figure 7 depicts the ROC of the suggested approaches, which clearly depict the link between true positives and true negatives. The suggested model outperforms the RNN-LSTM and RNN-GRU in terms of average ROC values.
The sensitivity and specificity attained by the proposed models has been compared to its attained false discoveries (FDR, FOR, and NPV), as shown in Figure 8. The figure also depicts the achieved results for the same metrics for the state-of-the-art models (RNN-LSTM, RNN-GRU model.). As can be seen in the figure, the suggested model achieves high sensitivity and specificity while simultaneously producing low false discoveries (FDR and FOR).
As a measure of the correlation between the correct and incorrect labels, the Matthews correlation coefficient was computed for the proposed models and RNN-LSTM, RNN-GRU models. The value for the MCC that measures the correlation between the true classes, TC, and the predicted labels using the proposed model has been recorded as 0.99. On the other hand, the attained values for the MCC are 0.90 and 0.89 for the RNN-LSTM, RNN-GRU models. MCC measures the correlation between the correct and incorrect labels. When the classifier is perfect ( F P o s = F N e g     0 ) the value of MCC is almost 1, indicating perfect positive correlation. In this, the proposed model indicated 0.99, which is close to 1, meaning that both classes are predicted well compared to other models.
Figure 9 depicts the processing time for the suggested model. We have estimated the testing time for the suggested model to be equal to the detection time. Since much of the work during the training phase is done away from the main system, it is ignored. However, it is generally agreed that testing is crucial because it proves the model works as intended. With a processing time of only 17.1 (MS), our proposed method is clearly computationally efficient. RNN-LSTM also needs less time for testing than RNN-GRU does.

5. Comparing the Proposed Model to Traditional ML Models

The proposed model’s performance has been analyzed further by comparing it to those of the most conventional ML models like KNN, SVM, and RF [55]. Transfer learning has been used to extract significant features for use in traditional classification techniques. To begin, the high-level deep features were extracted using a pretrained convolutional neural network (CNN) named AlexNet in order to bypass the overfitting problem through the transfer learning technique [56,57,58]. Once the deep features are acquired, they are given to the standard classifier mentioned above. As a part of the feature extraction process, we used AlexNet that had already been trained. Data of the image kind were fed into the AlexNet at the input layer. Therefore, in this study, we saved each EEG as a separate image. The input images have been resized into 227 × 227, the standard size for the AlexNet network.
Model parameters are used to define or express a model in ML/DL. The training phase, on the other hand, necessitates picking the right hyperparameters for the learning algorithm to utilize in order to discover the right mapping between the features and the targets in order to reach intelligence [59]. Parameters known as hyperparameters are used to regulate and set the values of model parameters that a learning algorithm ultimately learns. The learning parameters of AlexNet were tested during its training. Here, we employed a stochastic gradient descent with momentum (SGDM) optimizer. With this optimizer, we used a learning rate of 0.001, mini batches of 32, and 400 training epochs. The validation findings from the ongoing experiments were used to choose these hyperparameters.
Different hyperparameter values were tested throughout this study; the values that yielded the best results during training for the conventional SVM, KNN, and RF classification models are detailed in Table 3. The random forest classification technique makes use of many trees, and we have the ability to determine the minimum and maximum values for the number of trees required by the algorithm. This is accomplished with the use of a hyperparameter called “n estimators”. In this research, the optimal classification performance could be obtained by setting this hyperparameter to a value of 10 trees. It is well-known that KNN is an effective supervised classification algorithm. The similarity between the input features and the target is used by KNN to estimate the relationship between them. KNN is a classification method that uses the opinion of the input instance’s most similar neighbors (a number denoted by “k”) to determine the correct class. To determine the best value of k, we tested the KNN algorithm on the training dataset with several values of k and found that k = 5 yielded the best results. SVM seeks a hyperplane in the input feature space to clearly categorize the input data points in a classification task. Selecting the plane that maximizes the gap between the distributions of two classes of data is the optimization issue posed by finding the hyperplane. In this research, we have tried out a variety of kernel functions for use in SVM training. For EEG data classification, the sigmoid kernel function has proven to be the most effective.
The findings were obtained by the application of a cross-validation procedure that was carried out five times. The cross-validation test is illustrated by a bar chart in Figure 10, which compares the retrieved accuracy of the proposed model to that of the classical classifier in each fold. As depicted in Figure 10, the proposed model has outperformed the classical classification methods in the classification of epileptic/non-epileptic seizures.

6. Comparing the Performance of the Proposed Model to State-of-the-Art

In this part of the article, we have additionally compared the model against a number of benchmark methods that currently exist. The findings of the evaluation in relation to the most recent benchmarks are presented in Table 3. In terms of the evaluation metrics known as Acc, Pres, Sens, and Spec, the proposed techniques, known as RNN-BiLSTM, offer superior results. Table 4 shows that the created hybrid model outperforms previous efforts by Yao et al. [34] and Raghu et al. [32]. The performance boost achieved by hybrid RNN-LSTM [2] has been further enhanced by the suggested hybrid RNN-BiLSTM. It has also produced high performance, comparable to that of work done with the Deep Clinical Sparse Autoencoder. The suggested hybrid RNN-BiLSTM framework system has achieved comparable performance to that of Hilal et al. [6] and Mursalin et al. [33] on the benchmarking UCI-Epileptic dataset. Hilal et al. [6] employed the Deep Clinical Sparse Autoencoder for the classification of epilepsy UCI-Epileptic dataset, and Mursalin et al. [33] have analyzed the UCI-Epileptic dataset using hybrid metaheuristic feature selection and traditional ML-based classifiers. Table 4 shows that the suggested system can achieve comparable performance to the other systems compared.

7. Conclusions

Through an analysis of the patient’s EEG signal data, the RNN-BiLSTM model is presented as a method for identifying epileptic seizures that is found in this study. The RNN is extremely effective at extracting features from EEG signals, whereas the BiLSTM network is able to classify the collected data. This research incorporated publicly accessible EEG data from UCI Machine Learning repository for epileptic seizure classification, and the results serve as evidence that the proposed method is accurate in predicting clinical outcomes. The epileptic seizure recognition process consists of a total of one exam, which also includes binary recognition tasks. The accuracy, precision, sensitivity, and specificity of the proposed system have been thoroughly evaluated using a robust collection of error measures, such as the confusion matrix, the Matthews correlation coefficient, the Roc Curve, the false omission rate, the false discovery rate, and the negative predictive value. Our proposed RNN-BiLSTM has increased accuracy, sensitivity, precision, and specificity to 99.4%, 98.99%, 99.05%, and 98.8%, respectively. The data show that the new model improves the same measures by 1.8%, 1.69%, 1.95%, and 2.2%. The proposed approach reaches high levels of sensitivity and specificity while also generating a minimal number of false discoveries (FDR and FOR). In addition, the proposed method was judged against approaches that are now considered to be state-of-the-art, and it was found to be a more accurate classification of such methods.
Although the proposed method has made significant headway in the field of epileptic seizure detection, there are still some issues that need to be handled in a more comprehensive manner in the near future. When applied to multi-class issues, the strategy that has been provided does not have a recognition accuracy that is particularly extraordinary or even good. Second, in order to carry out supervised training with the suggested approach, a sizeable amount of EEG signal data that have been labeled will be required.

Author Contributions

Conceptualization, N.A.S., A.R., M.S.A.M. and I.A.; methodology, N.A.S., A.R., M.S.A.M., E.A.A. and I.A.; software, A.R. and I.A.; validation, N.A.S., N.F.M., A.R., M.S.A.M. and I.A.; formal analysis, N.A.S., N.F.M., A.R., M.S.A.M. and I.A.; investigation, N.A.S., N.F.M., A.R., M.S.A.M., E.A.A. and I.A.; resources, N.A.S., N.F.M., A.R., M.S.A.M. and I.A.; data curation, A.R., M.S.A.M. and I.A.; writing—original draft preparation, N.A.S., N.F.M., A.R., M.S.A.M., E.A.A. and I.A.; writing—review and editing, N.A.S., N.F.M., A.R., M.S.A.M. and I.A.; visualization, A.R., M.S.A.M. and I.A.; supervision, N.A.S. and N.F.M.; project administration, N.A.S. and N.F.M. All authors have read and agreed to the published version of the manuscript.

Funding

Princess Nourah bint Abdulrahman University Researchers Supporting Project Number PNURSP2022R206, Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors express their gratitude to Princess Nourah bint Abdulrahman, University Researchers Supporting Project Number PNURSP2022R206, Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Vaurio, L.; Karantzoulis, S.; Barr, W.B. The Impact of Epilepsy on Quality of Life. In Changes in the Brain: Impact on Daily Life; Springer: New York, NY, USA, 2016; pp. 167–187. ISBN 9780387981888. [Google Scholar]
  2. Ahmad, I.; Wang, X.; Zhu, M.; Wang, C.; Pi, Y.; Khan, J.A.; Khan, S.; Samuel, O.W.; Chen, S.; Li, G. EEG-Based Epileptic Seizure Detection via Machine/Deep Learning Approaches: A Systematic Review. Comput. Intell. Neurosci. 2022, 2022, 6486570. [Google Scholar] [CrossRef] [PubMed]
  3. Hirtz, D.; Thurman, D.J.; Gwinn-Hardy, K.; Mohamed, M.; Chaudhuri, A.R.; Zalutsky, R. How Common Are the “Common” Neurologic Disorders? Neurology 2007, 68, 326–337. [Google Scholar] [CrossRef] [PubMed]
  4. Carney, P.R.; Myers, S.; Geyer, J.D. Seizure Prediction: Methods. Epilepsy Behav. 2011, 22, S94. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Mahmoodian, N.; Boese, A.; Friebe, M.; Haddadnia, J. Epileptic Seizure Detection Using Cross-Bispectrum of Electroencephalogram Signal. Seizure 2019, 66, 4–11. [Google Scholar] [CrossRef] [Green Version]
  6. Hilal, A.M.; Albraikan, A.A.; Dhahbi, S.; Nour, M.K.; Mohamed, A.; Motwakel, A.; Zamani, A.S.; Rizwanullah, M. Intelligent Epileptic Seizure Detection and Classification Model Using Optimal Deep Canonical Sparse Autoencoder. Biology 2022, 11, 1220. [Google Scholar] [CrossRef] [PubMed]
  7. Firpi, H.; Goodman, E.D.; Echauz, J. Epileptic Seizure Detection Using Genetically Programmed Artificial Features. IEEE Trans. Biomed. Eng. 2007, 54, 212–224. [Google Scholar] [CrossRef]
  8. Kerr, M.P. The Impact of Epilepsy on Patients’ Lives. Acta Neurol. Scand 2012, 126, 1–9. [Google Scholar] [CrossRef] [PubMed]
  9. Auriel, E.; Landov, H.; Blatt, I.; Theitler, J.; Gandelman-Marton, R.; Chistik, V.; Margolin, N.; Gross, B.; Parmet, Y.; Andelman, F.; et al. Quality of Life in Seizure-Free Patients with Epilepsy on Monotherapy. Epilepsy Behav. 2009, 14, 130–133. [Google Scholar] [CrossRef] [PubMed]
  10. Liu, C.L.; Xiao, B.; Hsaio, W.H.; Tseng, V.S. Epileptic Seizure Prediction with Multi-View Convolutional Neural Networks. IEEE Access 2019, 7, 170352–170361. [Google Scholar] [CrossRef]
  11. Lehnertz, K.; Mormann, F.; Kreuz, T.; Andrzejak, R.G.; Rieke, C.; David, P.; Elger, C.E. Seizure Prediction by Nonlinear EEG Analysis. IEEE Eng. Med. Biol. Mag. 2003, 22, 57–63. [Google Scholar] [CrossRef] [PubMed]
  12. Daoud, H.; Bayoumi, M.A. Efficient Epileptic Seizure Prediction Based on Deep Learning. IEEE Trans. Biomed. Circuits Syst. 2019, 13, 804–813. [Google Scholar] [CrossRef] [PubMed]
  13. Islam, M.S.; El-Hajj, A.M.; Alawieh, H.; Dawy, Z.; Abbas, N.; El-Imad, J. EEG Mobility Artifact Removal for Ambulatory Epileptic Seizure Prediction Applications. Biomed. Signal Process. Control 2020, 55, 101638. [Google Scholar] [CrossRef]
  14. Zheng, T.; Luo, W. An Improved Squirrel Search Algorithm for Optimization. Complexity 2019, 2019, 6291968. [Google Scholar] [CrossRef]
  15. Shoeibi, A.; Ghassemi, N.; Alizadehsani, R.; Rouhani, M.; Hosseini-Nejad, H.; Khosravi, A.; Panahiazar, M.; Nahavandi, S. A Comprehensive Comparison of Handcrafted Features and Convolutional Autoencoders for Epileptic Seizures Detection in EEG Signals. Expert Syst. Appl. 2021, 163, 113788. [Google Scholar] [CrossRef]
  16. Da Silva Lourenço, C.; Tjepkema-Cloostermans, M.C.; van Putten, M.J.A.M. Machine Learning for Detection of Interictal Epileptiform Discharges. Clin. Neurophysiol. 2021, 132, 1433–1443. [Google Scholar] [CrossRef] [PubMed]
  17. Sharma, G.; Parashar, A.; Joshi, A.M. DepHNN: A Novel Hybrid Neural Network for Electroencephalogram (EEG)-Based Screening of Depression. Biomed. Signal Process. Control 2021, 66, 102393. [Google Scholar] [CrossRef]
  18. Ahmad, I.; Ullah, I.; Khan, W.U.; Rehman, A.U.; Adrees, M.S.; Saleem, M.Q.; Cheikhrouhou, O.; Hamam, H.; Shafiq, M. Efficient algorithms for E-healthcare to solve multiobject fuse detection problem. J. Healthc. Eng. 2021, 2021, 9500304. [Google Scholar] [CrossRef]
  19. Ma, M.; Cheng, Y.; Wei, X.; Chen, Z.; Zhou, Y. Research on Epileptic EEG Recognition Based on Improved Residual Networks of 1-D CNN and IndRNN. BMC Med. Inform. Decis. Mak. 2021, 21, 100. [Google Scholar] [CrossRef] [PubMed]
  20. Ahmad, S.; Ullah, T.; Ahmad, I.; L-Sharabi, A.A.; Ullah, K.; Khan, R.A.; Rasheed, S.; Ullah, I.; Uddin, M.N.; Ali, M.S. A novel hybrid deep learning model for metastatic cancer detection. Comput. Intell. Neurosci. 2022, 2022, 8141530. [Google Scholar] [CrossRef] [PubMed]
  21. Choi, W.; Kim, M.-J.; Yum, M.-S.; Jeong, D.-H. Deep Convolutional Gated Recurrent Unit Combined with Attention Mechanism to Classify Pre-Ictal from Interictal EEG with Minimized Number of Channels. J. Pers. Med. 2022, 12, 763. [Google Scholar] [CrossRef]
  22. Wang, X.; Ahmad, I.; Javeed, D.; Zaidi, S.A.; Alotaibi, F.M.; Ghoneim, M.E.; Daradkeh, Y.I.; Asghar, J.; Eldin, E.T. Intelligent Hybrid Deep Learning Model for Breast Cancer Detection. Electronics 2022, 11, 2767. [Google Scholar] [CrossRef]
  23. Nagabushanam, P.; Thomas George, S.; Radha, S. EEG Signal Classification Using LSTM and Improved Neural Network Algorithms. Soft Comput. 2020, 24, 9981–10003. [Google Scholar] [CrossRef]
  24. Ahmad, I.; Liu, Y.; Javeed, D.; Ahmad, S. A decision-making technique for solving order allocation problem using a genetic algorithm. In IOP Conference Series: Materials Science and Engineering; IOP Publishing: Bristol, UK, 2020; Volume 853. [Google Scholar]
  25. Vidyaratne, L.; Glandon, A.; Alam, M.; Iftekharuddin, K.M. Deep Recurrent Neural Network for Seizure Detection. In Proceedings of the International Joint Conference on Neural Networks, Vancouver, BC, Canada, 24–29 July 2016; pp. 1202–1207. [Google Scholar]
  26. Najafi, T.; Jaafar, R.; Remli, R.; Wan Zaidi, W.A. A Classification Model of EEG Signals Based on RNN-LSTM for Diagnosing Focal and Generalized Epilepsy. Sensors 2022, 22, 7269. [Google Scholar] [CrossRef] [PubMed]
  27. Xie, S.; Krishnan, S. Wavelet-Based Sparse Functional Linear Model with Applications to EEGs Seizure Detection and Epilepsy Diagnosis. Med. Biol. Eng. Comput. 2013, 51, 49–60. [Google Scholar] [CrossRef]
  28. Chen, G. Automatic EEG Seizure Detection Using Dual-Tree Complex Wavelet-Fourier Features. Expert Syst. Appl. 2014, 41, 2391–2394. [Google Scholar] [CrossRef]
  29. Chen, L.L.; Zhang, J.; Zou, J.Z.; Zhao, C.J.; Wang, G.S. A Framework on Wavelet-Based Nonlinear Features and Extreme Learning Machine for Epileptic Seizure Detection. Biomed. Signal Process. Control 2014, 10, 1–10. [Google Scholar] [CrossRef]
  30. Faust, O.; Acharya, U.R.; Adeli, H.; Adeli, A. Wavelet-Based EEG Processing for Computer-Aided Seizure Detection and Epilepsy Diagnosis. Seizure 2015, 26, 56–64. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Affes, A.; Mdhaffar, A.; Triki, C.; Jmaiel, M.; Freisleben, B. A Convolutional Gated Recurrent Neural Network for Epileptic Seizure Prediction. In Proceedings of the Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), New York, NY, USA, 14–16 October 2019; Springer: Berlin/Heidelberg, Germany, 2019; Volume 11862 LNCS, pp. 85–96. [Google Scholar]
  32. Raghu, S.; Sriraam, N.; Temel, Y.; Rao, S.V.; Kubben, P.L. EEG Based Multi-Class Seizure Type Classification Using Convolutional Neural Network and Transfer Learning. Neural Netw. 2020, 124, 202–212. [Google Scholar] [CrossRef] [PubMed]
  33. Mursalin, M.; Islam, S.S.; Noman, M.K.; Al-Jumaily, A.A. Epileptic Seizure Classification Using Statistical Sampling and a Novel Feature Selection Algorithm. arXiv 2019, arXiv:1902.09962. [Google Scholar]
  34. Yao, X.; Cheng, Q.; Zhang, G.-Q. A Novel Independent RNN Approach to Classification of Seizures against Non-Seizures. arXiv 2019, arXiv:1903.09326. [Google Scholar]
  35. Siami-Namini, S.; Tavakoli, N.; Namin, A.S. The Performance of LSTM and BiLSTM in Forecasting Time Series. In Proceedings of the Proceedings—2019 IEEE International Conference on Big Data, Los Angeles, CA, USA, 9–12 December 2019; pp. 3285–3292. [Google Scholar]
  36. Aziz Sharfuddin, A.; Nafis Tihami, M.; Saiful Islam, M. A Deep Recurrent Neural Network with BiLSTM Model for Sentiment Classification. In Proceedings of the 2018 International Conference on Bangla Speech and Language Processing, ICBSLP 2018, Sylhet, Bangladesh, 21–22 September 2018. [Google Scholar]
  37. Andrzejak, R.G.; Lehnertz, K.; Mormann, F.; Rieke, C.; David, P.; Elger, C.E. Indications of Nonlinear Deterministic and Finite-Dimensional Structures in Time Series of Brain Electrical Activity: Dependence on Recording Region and Brain State. Phys. Rev. E 2001, 64, 061907. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Cho, K.; van Merrienboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations Using RNN Encoder-Decoder for Statistical Machine Translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  39. Gers, F.A.; Schraudolph, N.N.; Schmidhuber, J. Learning Precise Timing with LSTM Recurrent Networks. J. Mach. Learn. Res. 2002, 3, 115–143. [Google Scholar]
  40. Pearlmutter, B.A. Learning State Space Trajectories in Recurrent Neural Networks. Neural Comput. 1989, 1, 263–269. [Google Scholar] [CrossRef]
  41. Yu, Y.; Si, X.; Hu, C.; Zhang, J. A Review of Recurrent Neural Networks: Lstm Cells and Network Architectures. Neural Comput. 2019, 31, 1235–1270. [Google Scholar] [CrossRef] [PubMed]
  42. Gers, F.A.; Schmidhuber, J.; Cummins, F. Learning to Forget: Continual Prediction with LSTM. Neural Comput. 2000, 12, 2451–2471. [Google Scholar] [CrossRef] [PubMed]
  43. Graves, A.; Schmidhuber, J. Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures. Neural Netw. 2005, 18, 602–610. [Google Scholar] [CrossRef]
  44. Khan, U.; Khan, S.; Rizwan, A.; Atteia, G.; Jamjoom, M.M.; Samee, N.A. Aggression Detection in Social Media from Textual Data Using Deep Learning Models. Appl. Sci. 2022, 12, 5083. [Google Scholar] [CrossRef]
  45. Sun, T.; Yang, C.; Han, K.; Ma, W.; Zhang, F. Bidirectional Spatial–Temporal Network for Traffic Prediction with Multisource Data. Transp. Res. Rec. 2020, 2674, 78–89. [Google Scholar] [CrossRef]
  46. Emmert-Streib, F.; Moutari, S.; Dehmer, M. A Comprehensive Survey of Error Measures for Evaluating Binary Decision Making in Data Science. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2019, 9, e1303. [Google Scholar] [CrossRef] [Green Version]
  47. Baldi, P.; Brunak, S.; Chauvin, Y.; Andersen, C.A.F.; Nielsen, H. Assessing the Accuracy of Prediction Algorithms for Classification: An Overview. Bioinformatics 2000, 16, 412–424. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. Matthews, B.W. Comparison of the Predicted and Observed Secondary Structure of T4 Phage Lysozyme. Biochim. Et Biophys. Acta (BBA)-Protein Struct. 1975, 405, 442–451. [Google Scholar] [CrossRef]
  49. Chicco, D.; Tötsch, N.; Jurman, G. The Matthews Correlation Coefficient (Mcc) Is More Reliable than Balanced Accuracy, Bookmaker Informedness, and Markedness in Two-Class Confusion Matrix Evaluation. BioData Min. 2021, 14, 13. [Google Scholar] [CrossRef] [PubMed]
  50. Hanley, J.A.; McNeil, B.J. The Meaning and Use of the Area under a Receiver Operating Characteristic (ROC) Curve. Radiology 1982, 143, 29–36. [Google Scholar] [CrossRef] [PubMed]
  51. Chen, X.; Robinson, D.G.; Storey, J.D. The Functional False Discovery Rate with Applications to Genomics. Biostatistics 2021, 22, 68–81. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Farcomeni, A. A Review of Modern Multiple Hypothesis Testing, with Particular Attention to the False Discovery Proportion. Stat. Methods Med. Res. 2008, 17, 347–388. [Google Scholar] [CrossRef] [PubMed]
  53. Genovese, C.R.; Wasserman, L. Exceedance Control of the False Discovery Proportion. J. Am. Stat. Assoc. 2006, 101, 1408–1417. [Google Scholar] [CrossRef]
  54. Iverson, G.L. Negative Predictive Power. In Encyclopedia of Clinical Neuropsychology; Springer: New York, NY, USA, 2011; pp. 1720–1722. [Google Scholar]
  55. Atteia, G.E.; Mengash, H.A.; Samee, N.A. Evaluation of Using Parametric and Non-Parametric Machine Learning Algorithms for Covid-19 Forecasting. Int. J. Adv. Comput. Sci. Appl. 2021, 12, 647–657. [Google Scholar] [CrossRef]
  56. Alhussan, A.; Abdel Samee, N.; Ghoneim, V.; Kadah, Y. Evaluating Deep and Statistical Machine Learning Models in the Classification of Breast Cancer from Digital Mammograms. (IJACSA) Int. J. Adv. Comput. Sci. Appl. 2021, 12, 304–313. [Google Scholar] [CrossRef]
  57. Samee, N.A.; Atteia, G.; Meshoul, S.; Al-Antari, M.A.; Kadah, Y.M. Deep Learning Cascaded Feature Selection Framework for Breast Cancer Classification: Hybrid CNN with Univariate-Based Approach. Mathematics 2022, 10, 3631. [Google Scholar] [CrossRef]
  58. Samee, N.A.; Alhussan, A.A.; Ghoneim, V.F.; Atteia, G.; Alkanhel, R.; Al-antari, M.A.; Kadah, Y.M. A Hybrid Deep Transfer Learning of CNN-Based LR-PCA for Breast Lesion Diagnosis via Medical Breast Mammograms. Sensors 2022, 22, 4938. [Google Scholar] [CrossRef] [PubMed]
  59. Atteia, G.; Samee, N.A.; El-Kenawy, E.-S.M.; Ibrahim, A. CNN-Hyperparameter Optimization for Diabetic Maculopathy Diagnosis in Optical Coherence Tomography and Fundus Retinography. Mathematics 2022, 10, 3274. [Google Scholar] [CrossRef]
Figure 1. Proposed RNN-BiLSTM model for detecting epileptic seizures.
Figure 1. Proposed RNN-BiLSTM model for detecting epileptic seizures.
Life 12 01946 g001
Figure 2. Data classes with frequency dimensional values.
Figure 2. Data classes with frequency dimensional values.
Life 12 01946 g002
Figure 3. RNN schematic.
Figure 3. RNN schematic.
Life 12 01946 g003
Figure 4. BiLSTM schematic.
Figure 4. BiLSTM schematic.
Life 12 01946 g004
Figure 5. RNN-LSTM, RNN-GRU, and the proposed RNN-BiLSTM model’s confusion matrices.
Figure 5. RNN-LSTM, RNN-GRU, and the proposed RNN-BiLSTM model’s confusion matrices.
Life 12 01946 g005
Figure 6. The attained performance measures of proposed models and RNN-LSTM, RNN-GRU models on various folds (1–4).
Figure 6. The attained performance measures of proposed models and RNN-LSTM, RNN-GRU models on various folds (1–4).
Life 12 01946 g006
Figure 7. The ROC of the proposed models and RNN-LSTM, RNN-GRU model on 5 folders.
Figure 7. The ROC of the proposed models and RNN-LSTM, RNN-GRU model on 5 folders.
Life 12 01946 g007
Figure 8. FDR, FOR, NPV, FNR, and FPR analysis of proposed model and RNN-LSTM, RNN-GRU models.
Figure 8. FDR, FOR, NPV, FNR, and FPR analysis of proposed model and RNN-LSTM, RNN-GRU models.
Life 12 01946 g008
Figure 9. Processing time (%) of proposed model and RNN-LSTM, RNN-GRU models.
Figure 9. Processing time (%) of proposed model and RNN-LSTM, RNN-GRU models.
Life 12 01946 g009
Figure 10. Comparing the attained accuracy of the proposed model to the conventional ML techniques (KNN, SVM, and RF) on 5-folds cross validation.
Figure 10. Comparing the attained accuracy of the proposed model to the conventional ML techniques (KNN, SVM, and RF) on 5-folds cross validation.
Life 12 01946 g010
Table 1. Explanation of the UCI-epileptic seizure dataset.
Table 1. Explanation of the UCI-epileptic seizure dataset.
Class DescriptionNo. of SamplesClass LabelsBinary Classification Samples
seizure230012300
First normal (Before seizure the signal of the patient)-230009200
Second normal (Healthy brain EEG recorded data)-2300--
Third normal—Eyes closed have no seizure2300--
Fourth normal—Eyes opened have no-seizure2300--
Table 2. Layers in the proposed hybrid RNN-BiLSTM model.
Table 2. Layers in the proposed hybrid RNN-BiLSTM model.
LayersOther ParametersValue
RNN layerBatch size = 100
Epochs = 100
Learning rate = 0.01
ADAM
Hidden 100 Units
BiLSTMHidden 200 Units
FC layer2 FC layers-
SoftMaxCross entropy
Classification Layer
Table 3. Hyper-parameters for the training of the conventional classifiers, including KNN, SVM, and RF.
Table 3. Hyper-parameters for the training of the conventional classifiers, including KNN, SVM, and RF.
ClassifierLearning Parameters
SVMKernel function = Sigmoid Kernel Function
KNNK = 5
RFn-estimators = 10
Table 4. Comparing the performance of the introduced model, RNN-BiLSTM, to state-of-the-art epileptic seizures detection systems.
Table 4. Comparing the performance of the introduced model, RNN-BiLSTM, to state-of-the-art epileptic seizures detection systems.
PublicationMethodEEG ClassDatasetAcc (%)Sens (%)Spec (%)
Yao et al. [34]Independent RNNbinaryCHB-MIT87%87.3%86.7%
Raghu et al. [32]Transfer learning and CNN8 classesTemple University Hospital EEG signals88.3%--
Choi et al. [21]hybrid model (1D CNN and GRU)binaryAsan Medical Center Children’s Hospital82.86 %80%-
Najafi et al. [2].hybrid RNN and LSTMbinaryHCTM hospital’s EEG data96.1%96.8%97.4%
Hilal et al. [6]Deep Clinical Sparse AutoencoderbinaryUCI-Epileptic98.67%.99.19%99.2%
Mursalin et al. [33]Hybrid metaheuristic Feature selection, and traditional ML-based classifiersbinaryUCI-Epileptic98.7%--
Proposed modelHybrid RNN-BiLSTM modelbinaryUCI-Epileptic98.4%98.30%98.10%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Samee, N.A.; Mahmoud, N.F.; Aldhahri, E.A.; Rafiq, A.; Muthanna, M.S.A.; Ahmad, I. RNN and BiLSTM Fusion for Accurate Automatic Epileptic Seizure Diagnosis Using EEG Signals. Life 2022, 12, 1946. https://doi.org/10.3390/life12121946

AMA Style

Samee NA, Mahmoud NF, Aldhahri EA, Rafiq A, Muthanna MSA, Ahmad I. RNN and BiLSTM Fusion for Accurate Automatic Epileptic Seizure Diagnosis Using EEG Signals. Life. 2022; 12(12):1946. https://doi.org/10.3390/life12121946

Chicago/Turabian Style

Samee, Nagwan Abdel, Noha F. Mahmoud, Eman A. Aldhahri, Ahsan Rafiq, Mohammed Saleh Ali Muthanna, and Ijaz Ahmad. 2022. "RNN and BiLSTM Fusion for Accurate Automatic Epileptic Seizure Diagnosis Using EEG Signals" Life 12, no. 12: 1946. https://doi.org/10.3390/life12121946

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop