Next Article in Journal
Deterministic 3D Ray-Launching Millimeter Wave Channel Characterization for Vehicular Communications in Urban Environments
Next Article in Special Issue
A Bipolar-Channel Hybrid Brain-Computer Interface System for Home Automation Control Utilizing Steady-State Visually Evoked Potential and Eye-Blink Signals
Previous Article in Journal
Decentralized Privacy-Preserving Data Aggregation Scheme for Smart Grid Based on Blockchain
Previous Article in Special Issue
CNN and LSTM-Based Emotion Charting Using Physiological Signals
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Identification of Motor and Mental Imagery EEG in Two and Multiclass Subject-Dependent Tasks Using Successive Decomposition Index

School of Automation, Northwestern Polytechnical University, 127 West Youyi Road, Xi’an 710072, China
*
Author to whom correspondence should be addressed.
Muhammad Tariq Sadiq and Xiaojun Yu are co-first author.
Sensors 2020, 20(18), 5283; https://doi.org/10.3390/s20185283
Submission received: 29 July 2020 / Revised: 2 September 2020 / Accepted: 10 September 2020 / Published: 16 September 2020
(This article belongs to the Special Issue Signal Processing Using Non-invasive Physiological Sensors)

Abstract

:
The development of fast and robust brain–computer interface (BCI) systems requires non-complex and efficient computational tools. The modern procedures adopted for this purpose are complex which limits their use in practical applications. In this study, for the first time, and to the best of our knowledge, a successive decomposition index (SDI)-based feature extraction approach is utilized for the classification of motor and mental imagery electroencephalography (EEG) tasks. First of all, the public datasets IVa, IVb, and V from BCI competition III were denoised using multiscale principal analysis (MSPCA), and then a SDI feature was calculated corresponding to each trial of the data. Finally, six benchmark machine learning and neural network classifiers were used to evaluate the performance of the proposed method. All the experiments were performed for motor and mental imagery datasets in binary and multiclass applications using a 10-fold cross-validation method. Furthermore, computerized automatic detection of motor and mental imagery using SDI (CADMMI-SDI) is developed to describe the proposed approach practically. The experimental results suggest that the highest classification accuracy of 97.46% (Dataset IVa), 99.52% (Dataset IVb), and 99.33% (Dataset V) was obtained using feedforward neural network classifier. Moreover, a series of experiments, namely, statistical analysis, channels variation, classifier parameters variation, processed and unprocessed data, and computational complexity, were performed and it was concluded that SDI is robust for noise, and a non-complex and efficient biomarker for the development of fast and accurate motor and mental imagery BCI systems.

1. Introduction

With the rampant growth in automated systems, computer-aided physical systems, and artificial intelligence, brain–computer interface (BCI) has gained significant attention from researchers as it can bind a human mind to the computer and operate complex physical applications. The healthcare realm has been overwhelmed by the development of computer-aided brain devices, namely, prosthetic arms, brain-controlled wheelchairs, mind-controlled home automation, etc., for physically impaired people [1,2,3,4,5,6,7]. The fundamental source of BCI is the low-key signal generated on the surface of the human scalp as a result of neural activity and it acts as a watershed for the plethora of brain-controlled applications.
The common practices involved to retrieve such signals are invasive and noninvasive methods. Invasive methods, as the name implies, record signals from the inside of the human brain which results in artifact-free data. On the other hand, noninvasive techniques accumulate noise artifacts that degrade the performance of BCI systems. Electroencephalography (EEG) is the commonly employed technology for the development of practical BCI systems.
Motor and mental imagery are subdomains in BCI which deal with the simulation of motor and mental activities in the brain without performing any activity in real. The inherent nature of motor and mental imagery suggests an economical, noninvasive, portable, and high temporal resolution mode of acquiring signals and the best choice is electroencephalography (EEG) [3,8]. After acquiring the signal, the subsequent process is to correctly wring out useful information from it [9,10].
The analysis of any signal processing problem comprises at least three basic procedures: preprocessing (data preparation and artifacts removing), feature extraction (identifies the most significant characteristics in signals), and classification (segregating classes between features). As noninvasive mode of signal acquisition heavily accumulates noise artifacts and it is crucial to filter out alienated signals without disturbing the original content. Recent studies proposed independent component analysis (ICA) [11], principal component analysis, and canonical correlation analysis [12] for the noise removal of EEG signals; however, these methods are not very effective for the analysis of non-stationary signals [12]. Another hybrid algorithm namely multiscale principal component analysis (MSPCA) is recently proposed and studies [13,14] revealed its robustness in denoising non-stationary and nonlinear signals.
After preprocessing the data, the subsequent steps are features estimation and classification [15]. In EEG signal processing, some widely adopted feature extraction methods are categorized as Fourier transform (FT) [16], power spectral density (PSD) [17], common spatial patterns (CSP) [18,19], autoregressive (AR) [20,21], sparse representation, and signal decomposition (SD) [22,23,24] based methods. All of these methods have their associated demerits and complications, for example, FT-based features only preserve the spectral resolution of the signal and completely loses the temporal information, PSD-based methods are susceptible to electrodes locality, AR-based techniques are sensitive to noise content, etc.
Chattarjee et al. [25] does a comparative analysis for a different time, energy, entropy, and statistical features using a different machine and deep learning classifiers for EEG signals. The maximum classification outcome of 85% was observed for energy and entropy-based features using the support vector machine (SVM) classifier. Wang et al. [26] amalgamates empirical mode decomposition (EMD) with Hilbert spectral analysis for motor imagery EEG signals and backpropagation neural network for classification purposes. The maximum recorded accuracy was 93.8%. Gupta et al. [17] extracted PSD features for EEG signals and did a comparative analysis for different univariate and multivariate features selection methods using different classifiers. A maximum classification accuracy of 85% was obtained for the combination of the Burg and linear regression features selection method using the linear discriminant analysis classifier.
Jasmine et al. [22] presents a comparative analysis for three signal decomposition techniques, i.e., EMD, discrete wavelet transform, and wavelet packet decomposition (WPD), using motor imagery EEG datasets. The highest accuracy of 92.8% was attained for higher-order statistical features extracted from WPD using K-nearest neighbors classifier. Chaudhary et al. [27] combines the non-dyadic wavelet decomposition method and CSP features extraction method for the classification of motor imagery EEG signals. Maximum classification accuracy of 85.6% was obtained for decision tree classifier. Jiacan et al. [28] presents a deep multi-view feature learning process for the classification of motor imagery EEG tasks. First, many multidomain features (time, frequency, time-frequency, and spatial) were extracted, and then a restricted Boltzmann machine network improved by t-distributed stochastic neighbor embedding (t-SNE) is employed for features learning. An average classification accuracy of 78.5% was obtained using the SVM classifier. Chen et al. [29] develops an NAO robot walking control system based on motor imagery by utilizing CSP and local characteristic scale decomposition (LCD). The experimental results yielded a classification accuracy of 87.5%.
Our study [23] proposed an instantaneous amplitude and instantaneous frequency component-based features. First, the empirical wavelet transform (EWT) was employed to decompose an EEG signal into representative modes, then the Welch PSD method was adopted for modes selection. The last step was to calculate the instantaneous components of each selected mode and classify the features with seven machine learning classifiers. The maximum accuracy achieved was 95.2% for the proposed mechanism. Our second study [24] on motor imagery EEG proposed a multivariate empirical wavelet transform (MEWT) for signal decomposition. By selecting features with correlation-based method and classifying them with three benchmark classifiers, we obtained 98% classification outcomes for the least square version of SVM classifier. All the methods discussed above, either utilized complex signal decomposition methods in combination with features selection methods or used complex features extraction methods, which are both impractical for the realization of functional BCI system. Raghu et al. [30] proposed the successive decomposition index (SDI) method for the classification of epileptic seizures. The classification outcomes suggested that SDI is a successful feature extraction method for epileptic seizures and it can be extended to other EEG domains.
Many different studies have built graphical user interface (GUI) systems for the visual implementation of their proposed approaches. EPILAB GUI was developed by Teixeirra et al. [31] for the analysis and classification of epileptic seizures. EEGLAB developed by Delorme et al. [32] presented an ICA-based EEG signal denoising method, time-frequency analysis, and visual representation of EEG signals. Moreover, Oostenveld et al. [33] reviews a MATLAB open source toolbox named FieldTrip, which does the time-frequency analysis, non-parametrical statistical tests, and reconstruction using dipoles and distributed sources of EEG and magnetoencephalography (MEG) signals. Each of these methods analyzes multidomain EEG signals, but a specialized GUI for motor and mental imagery is lagging.
For the robust, efficient, and non-complex analysis and classification of motor and mental imagery EEG signals, this article for the first time to the best of our knowledge and understanding, makes use of successive decomposition index (SDI) for feature extraction. This research attests the performance of SDI feature using six benchmark machine learning and neural network classifiers and different case studies confirms the effectiveness of proposed method. The main contributions of this study are listed as follows:
  • Successive decomposition index is proposed for the decoding of different motor and mental imagery activities in development of the BCI system.
  • Statistical analysis and novel performance evaluation criteria named polygon area metrics (PAM) are performed to confirm the efficacy of the SDI feature as a biomarker.
  • Four different channel selection schemes are employed to validate the performance of SDI features corresponding to the number of channels.
  • Classifier parameters are varied to investigate its fallouts on the proposed method.
  • A comparison was undertaken for denoised and noisy data to confirm the robustness of SDI features against noise artifacts.
  • Validate the performance of the proposed approach for multiclass mental imagery data.
  • Developed a computerized automatic detection of motor and mental imagery-successive decomposition index (CADMMI-SDI) application for the visual and practical implementation of SDI features.
The rest of paper is organized as follows. Section 2 and Section 3 deal with the datasets and the description of methods employed during the study, Section 4 describes the performance measures, Section 5 presents experimental set-up, Section 6 provides the results and discussion of the experimental outcomes, and, finally, Section 7 summarizes the study.

2. Materials

This study makes use of three motor and mental imagery publicly available datasets: IVa, IVb, and V from BCI competition III. Dataset IVa is a motor imagery dataset with two tasks right hand (RH) (Class 1) and right foot (RF) (Class 2). Five normal subjects or participants (“aa”, “al”, “av”, “aw”, and “ay”) participated for the collection of datasets. The global 10-20 system was used for the placement of 118 electrodes on the scalp. All the participants were shown a visual sign for 3.5 s and a total number of 280 trials (140 trials for each class) were recorded for an individual participant and the data were sampled at 1000 Hz. Similarly, dataset IVb is another single participant binary class motor imagery dataset with tasks left hand (LH) (Class 1) and right foot (RF) (Class 2). The data acquisition parameters for dataset IVb are similar to dataset IVa. Dataset V is a data collection of 3 individuals with imaginative roles of LH movement, RH movement, and random word (RW) production. These tasks are named as Class 1, Class 2, and Class 3, respectively. Data was collected in three cycles from 3 individuals with 32 electrodes and sampling frequency of 512 Hz. Further information for data sets is presented online at http://www.bbci.de/competition/iii/.

3. Methods

The study proposed a SDI-based framework for automated classification of two and multi-category motor and mental imagery EEG tasks in the development of computer-aided BCI systems. Figure 1 shows a clear presentation of the proposed strategy. First, the MSPCA process is used to separate noise from the raw EEG signal. Afterward, SDI is employed, that is, an inspirational case of discrete wavelet transform where a time series is pass through n levels of low-pass and high-pass filters and the coefficient at each step is used as a feature, and at last the extracted features are used as the inputs to the several machine leaning and neural network classifiers. Moreover, this study built up a layout for the realistic implementation of proposed platform for identifying motor and mental imagery EEG signals known as computerized automated detection of motor and mental imagery successive decomposition index (CADMMI-SDI). The subsequent subsections describe the details of the proposed automated framework.

3.1. Module 1: MSPCA Denoising

EEG is a noninvasive method of signal retrieval from the subject that inherits different types of noise artifacts, i.e., systematic noise, blink signal noise, cardiac signals noise, thermal noise, etc. A mathematical model of the crude form signal can be described as follows [34],
X = X E E G + X N
where X E E G is the desired EEG signal and X N is the supplemental noise artifact added to the original signal. The objective is to model a system that can effectively remove noise from the raw signal without influencing the content of X E E G . Principal Component Analysis (PCA) is conventionally adopted for determining the linear relationship between correlated data points. Furthermore, the nonlinear and non-stationary nature of the EEG signal demands a time-frequency resolution. Therefore, wavelet transform is commonly adopted and its significance is widely tested for non-stationary and nonlinear signals. A hybrid signal denoising algorithm called multiscale principal component analysis (MSPCA) is formulated by combining the properties of PCA and wavelet transform [24]. The workflow of MSPCA is given in Figure 2. We can define the procedure as follows.
  • Take a matrix A with dimensions n × m , where n is the length of each signal and m is the number of channels. Decompose each channel into B levels using wavelet transform.
  • Formulate a detailed matrix A j A and approximation matrix X i A , and calculate PCA for all B decompositions and m channels. As the Kaiser rule suggests, select principal components with eigenvalues greater than the mean of collective eigenvalues.
  • Compute the inverse wavelet transform of the selected principal components.
  • A denoised signals matrix can be obtained by taking the PCA of the results obtained in step 3.

3.2. Module 2: Successive Decomposition Index Based Feature Extraction

In the past, a large number of studies [22,24,35,36] investigated the effectiveness of wavelet and signal decomposition-based methods for motor and mental imagery EEG signals using different mother wavelets and decomposition levels. The drawbacks of such methods are the selection of suitable mother wavelets and the number of decomposition levels which requires a thorough investigation in terms of classification outcomes and time complexity. The basic requirements of a practical BCI system are robustness, non-complexity and efficiency that are lagging in current researches. To overcome the aforementioned limitations a successive decomposition index (SDI) method is employed.
The proposed SDI method is an inspiration of discrete wavelet transform (DWT). In the first level of DWT, a time signal of length n is passed through a low and high pass filter. In the next level, the output of low pass filter is again passed through a high and low pass filter and this process is iterated for a specific number of decomposition levels. Finally, the coefficients from each decomposition level are used to extract features. The basic difference between DWT and SDI is that the former has to have a predefined number of decomposition levels whether the later has no predefined decomposition levels and the coefficient from the last level is considered for further analysis. The mathematical formulation of the SDI feature is described in following steps [30].
  • Consider an EEG signal s = { s 1 , s 2 , s 3 , . , s n } , where n is the length of the signal. The first step is to compute the average of absolute values ( S + ) of the EEG signal is as follows.
    S + = 1 n i = 1 n s i
  • The next step is to compute the average difference ( S ) of the signal and it can be calculated by the successive difference mean of non-overlapping pairs of time signal. It can mathematically represented as follows,
    s ( 1 ) = s 1 s 2 2 , s 3 s 4 2 , s n 3 s n 2 2 , s n 1 s n 2
    where the length of s ( 1 ) is n / 2 . Similarly, s ( 2 ) can be calculated as
    s ( 2 ) = s 1 ( 1 ) s 2 ( 1 ) 2 , s 3 ( 1 ) s 4 ( 1 ) 2 , s n / 2 3 ( 1 ) s n / 2 2 ( 1 ) 2 , s n / 2 1 ( 1 ) s n / 2 ( 1 ) 2
    The process of calculating s ( k ) (where k is the number of iterations) continues until we get a single coefficient and that final coefficient is the average difference term S . The number of iterations required to calculate S can be determined as k = 3.33 log 10 ( n ) and the total number of coefficients at each step are n / 2 k . The next step is to calculate two new terms S + + and S as follows.
    S + + = S + + S 2
    S = S + S 2
    The terms S + + and S gives the relation between S + and S . In addition, a square matrix Z is formed from the four coefficients as follows.
    Z = S + S S S + +
  • The final step is to calculate the determinant of matrix Z multiplied by a scalar n / k followed by log 10 .
    S D I = log 10 n k S + S + + S S
The resultant SDI is a single value bio marker for an EEG signal of length n. The significance of SDI is that it measures the variations of EEG signal successively with respect of time and packs it into a single representative value. In addition, unlike other wavelet and signal decomposition-based methods, there is no need to select a suitable mother wavelet and define the number of decomposition levels rather the process of calculating SDI is linear and non-complex, which makes it a suitable choice for the development of practical motor and mental imagery BCI systems.

3.3. Module 3: Classification

To segregate the motor and metal imagery tasks, we have utilized six widely used machine learning and neural network classifiers. Their description and parameters of classifiers utilized in this study are discussed as follows.

3.3.1. Support Vector Machine

A support vector machine (SVM) is a supervised learning classifier that formulates a hyperplane to maximize the separability between two classes. For nonlinear feature sets, different kernel functions are utilized to transform it into a linear problem at the cost of augmented dimensionality. The selection of SVM in this study is based on its robustness and reliability for motor imagery tasks discussed in [37,38]. In this study, we have utilized the radial basis function, linear function, and polynomial function as kernels and the default MATLAB toolbox hyperparameters were availed for each kernel.

3.3.2. Discriminant Analysis

Discriminative analysis (DA) is a supervised learning algorithm that formulates a predictive model during the learning phase that can be applied to test data for labeling them. DA can use lines, planes, and hyperplanes to segregate the normally distributed samples and thus it can classify multidimensional data robustly. To build a DA model, we have to compute the class probability, mean, and covariance matrix along with a suitable kernel function. In this study, we have utilized three kernels: linear, pseudo-linear, and pseudo-quadratic. The effectiveness of DA for motor imagery tasks has been accredited in [39,40].

3.3.3. Multilayer Perceptron with One Hidden Layers

A multilayer perceptron with single hidden layer (ANN) is the building block of deep learning classifiers and is robust in approximating linear, nonlinear functions and pattern recognition effectively. ANN has a three-layered structure consisting of input, hidden, and output layers. The number of input nodes is same as the number of features while the number of output nodes is equal to the number of classes. The number of hidden nodes is variable and depends primarily on classification outcomes. ANN propagates the input signal from first to last layer and the backpropagation algorithm tunes the hyperparameters of the network during training phase. The studies [41,42] attests the robustness of ANN for motor imagery EEG.

3.3.4. Multilayer Perceptron with Two Hidden Layers

A multilayer perceptron with two hidden layers (MNN) is an extension of ANN. The basic difference between both algorithms is that MNN has 2 to M hidden layers depending upon the classification results while ANN has only one hidden layer. The advantage of using MNN is that it has more parameters and hence it has an extra degree of freedom to approximate a nonlinear function or recognize a pattern. The disadvantage is that, because of the large number of hyperparameters, the training and testing time exceeds ANN and hence there is a trade-off between computational time and classification outcomes.

3.3.5. Cascade Feedforward Neural Network

The architecture of cascade feedforward neural network (CFNN) resembles ANN. The core difference between both classifiers is that CFNN has a connection from the output layer to the input layer that ANN lags in its structure. This extra connection gives CFNN the ability to memorize previous inputs and their outcomes and thus it is essential in learning sequential data. The authors of [43] utilized CFNN for the classification of motor imagery tasks.

3.3.6. Feed-Forward Neural Network

Throughout the feed-forward neural network (FFNN), a multilayered structure is used with each layer containing variable number of neurons. The signal is propagated from input to output across the network and an error is computed using a cost function. This error is then repropagated across the network and each parameter is tuned. In our research, tan sigmoid was used as an activation feature. The Levenberg–Marquardt algorithm was used for fast learning [43].
There is no structural difference between ANN and FFNN. In the present study, we utilized two different MATLAB functions named “patternenet()” for ANN and “feedforwardnet()” for FFNN. The basic difference between these two functions is that ANN uses “glorot” weights and biases initializer while FFNN uses “orthogonal” initializer. The “glorot” initializer takes random samples from a normal distribution where mean is zero and variance is 2/(size of inputs + size of outputs), while the orthogonal initializer takes a matrix from a unit uniform distribution and initializes the weights and biases with Q obtained from a QR decomposition [44,45].

4. Performance Parameters

This study utilizes a 10-fold cross-validation method to fairly evaluate the classification results. For this purpose, the feature matrix containing Class 1 and Class 2 features is divided into 10 equal parts, out of that 9 parts were used for training purposes and 1 part was used for validation. In this way, each trial of the feature set is being trained upon as well as validated. To evaluate the classification outcomes, we made use of 10-fold cross-validation method with different performance metrics, namely, classification accuracy (Acc), Sensitivity (Sen), Specificity (Spe), Kappa, and F1-Score. Their mathematical expressions are given respectively as follows,
A c c = T P + T N T P + T N + F P + F N
S e n = T P T P + F N
S p e = T N T N + F P
K a p p a = T P × T N F P × F N ( ( T P + F P ) ( T P + F N ) ( T N + F P ) ( T N + F N ) )
F 1 S c o r e = 2 × Prec × S e n Prec + S e n
where T P (True positive) is the amount of adequately identify Class 1 labels, T N (True negative) is the amount of adequately identify Class 2 labels, F P (False positive) is the number of inadequately classified Class 1 labels. and F N (False negative) is the number of inadequately identified Class 2 labels.
Apart from the above mentioned five performance parameters, we utilized a novel performance evaluation criteria named polygon area metrics (PAM) [46] for the very first time for motor and mental imagery EEG classification evaluation. The PAM constructs a hexagon with six performance parameters (F measure, Jaccard Index, Classification accuracy, Area under the curve, Sensitivity and Specificity) on each edge. The performance in this case is evaluated by the area of the polygon. The greater the area occupied by the polygon, the better the performance of the classifier and vice versa.

5. Experimental Setup

All experiments and simulations in this study were performed using MATLAB R2019b on an Intel(R) Core (TM) M-5Y10c CPU @0.80GHz cpu, Windows 10 64-bit operating system, and 8 GB RAM with WEKA 3.8.4.
Numerous studies have been performed in the past for effectively classifying motor and mental imagery tasks as detailed in Section 1. Most of them utilized complex signal processing techniques that make those unfeasible for the practical implementation and it also gets difficult for physicians to understand complex signal processing tools without having a piece of proper knowledge about the field. To cope up with such challenges, we have utilized a single non-complex feature that uses iterative signal decomposition coefficients to construct a representative feature with the least computational complexity and effective classification results.
Figure 1 shows the block diagram of the proposed methodology. At first, the raw data is passed through an MSPCA filter that suppresses the noise content from the signals. Then the data is divided into individual trials. In the case of dataset IVa and IVb, the single trial dimension is 400 × 118, where 400 is the signal length and 118 is the number of channels. For dataset V, the single-trial dimension is 512 × 32, where 512 is the signal length and 32 is the number of channels. Next, each trial is given to an SDI computational function which calculates features for that trial. In the case of dataset IVa and IVb, we get 118 features for a single trial while for dataset V, we get 32 features for a single trial. In this way, a features matrix is formed with dimensions n × m, where n is the number of trials and m is the number of features (indirectly the number of channels) per trial. Last, the feature matrices of various classes are given to six benchmark classifiers to evaluate the performance of SDI features in estimating motor and mental imagery tasks.

6. Results and Discussion

6.1. Statistical Analysis

To analyze how the SDI feature segregates motor imagery tasks, we have performed a statistical analysis in this section. Figure 3 presents the SDI feature distribution for Class 1 and Class 2 tasks by utilizing channel C 3 from all subjects of dataset IVa and IVb. Figure 3 suggests that subjects “aa”, “al”, “av”, “aw”, “ay”, and dataset IVb have a highly nonlinear relationship between both task features and it is imperative to use a nonlinear classifier to trace the pattern between both classes. It can be seen in the Figure 3 that SDI feature has significantly singled out tasks for small training samples subject “ay” and later in this study we will see that subject “ay” is the best performant among all other subjects in terms of classification outcomes.
In addition, a descriptive statistical analysis in terms of mean, standard deviation, median, and Kruskal–Wallis probability (p) values (KW test) of SDI features was performed for single trial cases of each subject. The results presented in Table 1 suggest that the mean and median values of subject “aa”, “al”, “av”, “aw”, “ay”, and dataset IVb are higher for Class 2 cases than Class 1. For the subject “ay”, the mean and median values for Class 1 are higher and this trend was consistent for all trials. Moreover, the KW p values for single-trial cases of all subjects are less than 0.05 which suggests the significance of SDI features for motor imagery tasks and the high discrimination ability of extracted features between two classes.

6.2. Results by Selecting Different Number of Channels

Siuly et al. [47] conducted a comparative analysis for 18 and 118 channels motor imagery dataset IVa and IVb using two classification algorithms. Their study concludes that 118 channel results outperform 18 channels in terms of classification outcomes. In this section, a similar type of comparison is presented for dataset IVa and IVb with 18 channels, three channels and three channels selected with automated channel selection criteria. The 18 and three channels are widely adopted motor cortex channels while three-channel selection with automated channel selection criterion was proposed in our previous study [24]. The list of automated channels for each subject is given in Table 2. As motor imagery EEG signals are highly dependent upon subject physical and mental nature so for each subject, different channels are selected by the automated channel selection criteria. Figure 4 shows a visual representation of four channels selection schemes for best and worst-performing classifiers. The worst classifier is characterized in terms of least gain in accuracy while the best classifier symbolizes maximum gain in classification accuracy. This study made use of six machine learning and neural network classifiers (NN, MNN, CFNN, FFNN, SVM, and DA) out of which FFNN was the best performing classifier and SVM was the worst performer. The rest of the analysis is given as follows.
  • It is inferred from Figure 4 that 118 channels give the highest classification accuracy for dataset IVa as compared to other channel combinations. The average classification accuracy obtained using 118 channels with FFNN classifier is 97.46%. Similarly, the average accuracy for SVM classifier using 118 channels is 93.05%. Moving on to the 18-channel combination, it is observed that the average classification accuracy for FFNN and SVM classifier is 94.28% and 77.96% respectively. Furthermore, the 3-channel scheme resulted in a mean accuracy of 77.6% and 60.1% for FFNN and SVM classifiers respectively. Finally, the 3-channel automated scheme has the least average results as compared to other channels combination. The average results obtained for 3-channel automated criteria are 73.74% and 60.46% for FFNN and SVM classifiers, respectively.
  • For dataset IVb, a similar trend of classification accuracies is observed for varying channel combinations. Figure 4 shows that FFNN and SVM resulted in 99.5% accuracy each using 118 channels for dataset IVb and this is the highest among other channel selection schemes. For the combination of 18 channels, the FFNN classifier resulted in 96.2% accuracy and SVM yielded 91% classification certainty. Moving forward to the 3-channel strategy, it is noticed that FFNN turnout 82.9% accuracy and 52.4% for SVM. Likewise, 3-channel automated scheme resulted in similar pattern of results with 83.6% for FFNN and 54.6% for SVM classifier.
  • It is observed that the 118-channel combination has a maximum gain of 23.72% and 32.59% for FFNN and SVM classifiers, respectively using dataset IVa. Dataset IVb has a maximum gain of 16.6% and 47.1% for FFNN and SVM classifiers, respectively. It accredits the significance of using 118 channels for SDI features and advocates the channel comparison study performed by Siuly et al. [47].
  • One interesting observation is made that subject “ay” of dataset IVa has above 90% classification accuracy for all channel combinations and classifiers. As mentioned in the descriptive analysis section, the SDI features for subject “ay” tasks are well separated and distinguishable. We conclude that SDI feature extraction is more significant for subject with small training samples as compared to large one and this property makes it feasible for the development of practical BCI systems as disabled patients need small training to train a device.

6.3. Analysis with Sensitivity, Specificity, Kappa, F1-Score and PAM

In this section, we explain the effect of other performance measures namely sensitivity, specificity, kappa, F1-Score and most importantly a unified novel performance measure, the polygon area metric (PAM). Figure 5 shows the sensitivity, specificity, kappa, and F1-score values for FFNN and SVM classifiers using 118 channels with 10-fold cross-validation strategy. Figure 5a,e show the sensitivity values for FFNN and SVM classifier, respectively. The average sensitivity values are 98.8% and 94.8% accordingly for individual classifiers which suggests that FFNN correctly identified Class 1 instances 98 times and SVM classified them correctly 94 times. Similarly, Figure 5b,f show the specificity values for FFNN and SVM classifiers respectively. The average 10-fold specificity values are 98.25% and 95.57%, respectively, for each classifier, which indicates that FFNN classified Class 2 instances effectively 98 times and SVM classified them positively 95 times. Figure 5c,g presents the kappa scores for the aforementioned classifiers. It is noted that the average kappa for FFNN classifier is 96.93% with slight variations for subject “aw”. The average kappa for SVM classifier is 91.5% with major variations in subject “av” and “aw”. Hence, we conclude that FFNN is more stable and unbiased in classifying Class 1 and Class 2 tasks. Finally, Figure 5d,h show the F1-Score for each classifier, respectively, and the average F1-Score for individual classifier is 98.07% and 93.83% accordingly. The high value of F1-Score for FFNN classifier illustrates the high precision and recall measures.
Figure 6 shows the PAM graphs for dataset IVa all subjects and dataset IVb using FFNN and SVM classifiers for 118 channels scheme. Figure 6a–f presents the PAM graphs for FFNN classifier and Figure 6g–l shows PAM graphs for SVM classifier. It can be seen that subject “aa” and “ay” have an area of 1 unit while subjects “al”, “av”, and “a” have areas of 0.95, 0.78, and 0.85 units for FFNN classifier, respectively. Dataset IVb has an area of 0.98 units for FFNN classifier. All of these results are consistent with the above-mentioned accuracy and other performance measures outcomes. Moreover, in the case of SVM classifier, subject “aa”, “ay” and dataset IVb has an area of 0.98 units each, subject “al”, “av”, “aw”, and “ay” has an area 0.95, 0.81, and 0.79 units, respectively. The key benefit of using PAM graph is that complete classification performance is represented in a single graph with several measures instead of looking into lengthy tables.

6.4. Results by Selecting Different Parameters of Classifiers

To investigate the fallouts of classifier parameters on the proposed approach, we compared the classification accuracies for varying classifier parameters of all classifiers. Table 3 shows the averaged 10-fold accuracies of all classifiers with varying parameters for the 118-channel scheme using dataset IVa individual subjects and dataset IVb. For neural network (NN) classifiers, the number of hidden layer neurons was varied and its effect was observed accordingly. For SVM classifier, three different kernels namely radial basis function (RBF), linear kernel and the polynomial kernel were utilized, for DA classifier, linear, pseudo quadratic and pseudo linear kernels were adopted and their performance was evaluated for both datasets individually. The findings are as following:
  • The experimental results suggest that NN classifiers have no significant effect on average accuracy by varying the number of hidden neurons. For NN classifiers, the maximum mean classification accuracy was recorded for 40 neurons with 94.6% and 99.05% results for dataset IVa and IVb, respectively. For MNN classifier, the best case mean classification outcome was obtained for 30 neurons for dataset IVa with an accuracy of 91.2% and 40 neurons for dataset IVb resulting in 94.29% mean accuracy. Moving on to CFNN classifier, it is noted that each number of neurons yields 97% accuracy for dataset IVa and 99% outcomes for dataset IVb. Lastly, FFNN classifier turned out the maximum mean accuracy of 98.27% using dataset IVa and 99.52% using dataset IVb for 30 neurons each.
  • Amid the three kernels of the SVM classifier, it is observed that polynomial kernel is the best performant among others with the mean classification accuracy of 93.08% and 99.52% for datasets IVa and IVb, respectively. The linear and RBF kernels are ranked second and third accordingly with mean accuracies of 89.64% and 70.24% for dataset IVa and 96.19% and 83.33% for dataset IVb respectively.
  • In case of the DA classifier, we observe that the linear kernel is ranked highest as compared to pseudo-quadratic and pseudo-linear kernels. The average classification accuracy for linear kernel is 95.29% and 98.1% for dataset IVa and IVb each. Moving forward, the pseudo-linear kernel is ranked second with a mean classification accuracy of 93.29% and 95.1%, respectively, for datasets IVa and IVb. Last, the pseudo quadratic-kernel results in an average classification result of 92.63% and 92.24% accordingly for each dataset.
  • The thorough investigation of varying classifiers parameters suggests that polynomial and linear kernels are the best performers for SVM and DA classifiers respectively. Similarly, for NN, MNN, CFNN, and FFNN classifiers, 40 number of hidden neurons were chosen as the best parameter setting and these parameters are employed throughout this study.
  • Figure 7 shows the average accuracies of 10 times repeated 10-fold experiments for best (FFNN) and worst (SVM) case classifiers and each subject of dataset IVa and IVb. It is noted that the average results obtained for both classes results in slight variations of ±1.5%. In case of “av” subject with the FFNN and subject “aw” with the SVM, the variations are larger than 10%, which is due to the outliers caused by classifiers in some fold results but the mean results are more or less the same as calculated previously. The extensive experimentation results obtained confirms the robustness and stability of SDI features in estimating motor imagery tasks.

6.5. Results with Raw EEG and Noise-Free EEG Signals

We discussed earlier that EEG is a noninvasive mode of signal retrieval and it inherits noise artifacts while recording the data. In this section, a comparative analysis for MSPCA denoised and unprocessed (noisy) data is performed and validated if SDI feature is being affected by noise artifacts or not.
Figure 8 shows the classification accuracy for MSPCA denoised and noisy data of dataset IVa and IVb. The classification results are calculated for best-case FFNN classifier. As observed from Figure 8, the classification accuracies for noisy data are 83.1%, 84.4%, 82.5%, 85%, 92.4%, and 81.4% for subjects “aa”, “al”, “av”, “aw”, “ay”, and dataset IVb, respectively. The average results are 85.5% and 81.4% for dataset IVa and IVb respectively. We observe a significant improvement in individual and average classification results after denoising the data. The results after denoising with MSPCA are 100%, 97.3%, 90.6%, 96.3%, 100% and 99.52% for subjects “aa”, “al”, “av”, “aw”, “ay”, and dataset IVb, respectively. The average accuracies for datasets IVa and IVb are 96.8% and 99.52%, respectively. By looking at the results obtained from two case scenarios, we observe an increase of 11.3% and 18.12% in accuracy for dataset IVa and IVb jointly. A similar trend of accuracy enhancement for denoised data was observed for other classifiers and hence it is concluded that the proposed SDI based feature extraction framework is robust against noise artifacts.
It is important to note that we have also checked numerous conventional methods including such band pass filters, temporal filtering, and spatial filtering for meticulous selection of a suitable strategy in the preprocessing module and identified that MSPCA produces the best findings for the proposed SDI feature extraction approach.

6.6. Classification Performance (%) with Dataset

This section deals with the experimental results of multiclass mental imagery dataset V. At first, the dataset was denoised with MSPCA and rearranged into individual trials with dimensions 512 × 32 (where 512 is the signal length and 32 is the number of channels) for each trial. We have rearranged the multiclass problem into 3 binary class experiments for each subject. The number of cases are given in Table 4. Here cases 1 to 3 are dedicated for participant 1 (P1), cases 4 to 6 corresponds to the participant 2 (P2). and cases 7 to 9 are formed for the participant 3 (P3). Next, the SDI feature is calculated for all trials and fed into six classifiers. The classification outcomes in terms of accuracies are given in Table 5.
It is observed from Table 5 that all classifiers achieved an average accuracy of above 90% for each subject. Moreover, the average individual classification accuracy for NN, MNN, CFNN and FFNN is above 95% which shows the effectiveness of NN classifiers in segregating mental imagery tasks. The best-case scenario was observed in for FFNN classifier with an average accuracy of 99.07%, 98.16%. and 98.38% for participants 1, 2, and 3, respectively. It should be noted that FFNN was the best performer for motor imagery tasks and now it again gives the best results for mental imagery dataset. The worst-case scenario was observed for SVM classifier with accuracies 91.84%, 90.36%, and 93.81%, respectively, for first participant, second participant, and third participant. As per the experimental results, it is concluded that NN classifiers, especially FFNN classifier is intelligent in estimating mental imagery tasks.
Figure 9 shows the classification performance of SDI feature for dataset V in terms of four performance parameters (Sensitivity, Specificity, Kappa, and F1-Score). The performance parameters are shown for the best classifier which is FFNN in our case. It can be inferred from Figure 9 that the sensitivity and specificity values for all cases in each subject are above 95% and in some cases, it is 100% which shows the greatness of FFNN classifier in predicting Class 1, Class 2, and Class 3 tasks. It can also be seen that the kappa and F1-measures are above 95% in all cases which depict the stabilization and unbiased nature of FFNN classifier. Overall it can be concluded that SDI features are not only specific for motor imagery tasks but equally essential and significant for mental imagery tasks as well.

6.7. CADMMI-SDI Application

Apart from the theoretical analysis, we have developed a computerized automatic detection of motor and mental imagery using SDI (CADMMI-SDI) graphical user interface to assist physicians and laymen to utilize SDI method for their purpose without having to implement it their self. Table 6 presents the description of individual components present in the GUI while Figure 10 shows the detailed interface of our developed CADMMI-SDI. Some interesting features of the developed application are detailed in Table 6.
The demonstration of the GUI application can be seen in link https://www.youtube.com/watch?v=ugWbq4JUtuI. A copy of the GUI application is freely available and interested readers are suggested to write an email to corresponding author.

6.8. Computational Complexity of SDI Feature

Figure 11 shows the computational time for feature extraction, training and testing for all subjects and classifiers using the system specifications given in Section 5. First of all, Figure 11a presents the all trials feature extraction time for each subject of dataset IVa and dataset IVb. It can be seen that the highest feature extraction time of 1.36 s is taken by subject “al” followed by subject “aa” and dataset IVb with 1.06 s and 0.65 s, respectively. The average single-trial feature extraction time is calculated to be 0.85 milliseconds. Next, Figure 11b shows all trials training time for individual subjects and all classifiers. It is observed that CFNN classifier takes the highest training time for all subjects followed by FFNN classifier. The highest training time of 1.8 s, 1.75 s and 1.5 s was recorded for subjects “al”, “aa”, and dataset IVb, respectively, using CFNN classifier. The highest training time recorded for FFNN classifier is 1.2 s, 1.1 s and 1.08 s for dataset IVa, subject “al” and “aa”, respectively. The average single-trial training time for FFNN classifier is calculated to be 1.27 milliseconds. Last, Figure 11c shows all trials testing time for individual subjects and all classifiers. As noted, SVM classifier takes the highest testing time of 70 milliseconds and 60 milliseconds for subjects “al”, and “aa”, respectively. The time taken by FFNN classifier is minimum in most cases and the average single-trial training time is recorded to be 0.01 milliseconds. By accumulating the single trials computational times for FFNN classifier, it comes out to be 2.13 milliseconds which is very nominal as compared to other complex signal decomposition methods and it shows that besides noise robustness and classification accuracy, SDI features are computationally less complex and efficient and hence it can be employed in the production of practical BCI systems.

6.9. Performance Comparison with Other Literature

This section presents a comparative analysis of the proposed SDI framework with other recent state of art methods. Table 7 compares the classification accuracies for dataset IVa individual subjects and the best-case results are highlighted to make a fair comparison of other methods with the proposed approach. It can be seen from the table that subjects “aa” and “ay” attained 100% classification accuracy which is the highest among other methods. The results for subjects “al”,“av”, and “aw” are above 90% and very close to the best results achieved by other methods. Comparing the results of SDI feature method with our previous studies [23,24], it is worth noting that our current method outperforms the complex signal decomposition and modes selection-based methods. It can be noted from Table 7 that our method achieved the highest average classification accuracy of 97.54% with minimal heterogeneity. Moreover, there is a 24.04% maximum gain in accuracy comparing to other state of the art methods and hence it suffices that SDI feature extraction is not only efficient and non-complex but also robust in estimating motor imagery EEG signals and this is validated by a fair comparison with other widely acclaimed studies.
Table 8 shows the comparative results for multiclass dataset V. The outcomes are presented in terms of average classification accuracies and the highest case results are highlighted to make the best combination stand out. It is worth noting that the proposed SDI method outperformed all other methods in terms of individual subject results. It can be seen that the SDI method attained an average classification outcome of 99.07%, 98.16%, and 98.37% for participant 1, participant 2, and participant 3, respectively, and these are highest as compared to other methods. In terms of overall average results, the proposed SDI framework scored the highest 98.53% accuracy with a standard deviation of 0.387 that shows the consistency of overall results. Last, it is inferred from the comparison that SDI feature extraction method gains a minimum of 15.26% average classification accuracy, which is a significant improvement and it shows that the proposed method is not only useful for binary class motor imagery datasets but equally significant for multiclass mental imagery dataset as well.
Besides classification results, it is important to compare the complexity of other methods with SDI feature extraction method. As mentioned earlier in this study, our method has no signal decomposition, complex multidomain features extraction, or features selection procedures involved, which makes it computationally simple and less time-consuming. The studies in [22,23,24] use signal decomposition techniques that involves resolution of a time signal into different modes, then extraction of complex features and lastly selection of highly uncorrelated features. Such systems might be useful for the research analysis but they are not feasible to be adopted for practical BCI systems. Similarly, the studies [18,19,53] employs common spatial pattern (CSP)-based methods, which is another complex method for the analysis of EEG signals. The crux of the matter is whether we consider robustness, efficiency and complexity, the proposed SDI method outperforms all state-of-the-art methods in every aspect and gives us a feasible solution to be considered for the development of practical BCI systems.

6.10. Future Recommendations

In the present study, we utilized data with class labels, however, semisupervised learning or transductive learning methods are attracting attention these days. In future, researchers are encouraged to implement these methods for MI classification and information for these methods can be found in [58,59]. It is also worth mentioning that here, in the present study, we focused on at most three classes and presented the results in Table 5. However, for more number of classes readers should focus on more innovative strategies such as available in [60].

7. Conclusions

This study exploits the successive decomposition index (SDI) for the feature estimation of motor and mental imagery tasks. Three publicly available datasets namely dataset IVa, dataset IVb and dataset V from BCI competition III were utilized to attest the effectiveness of proposed method. Initially, the data was denoised with MSPCA and distributed into individual trials. Then, the SDI algorithm was used to calculate the feature corresponding to each trial and build a feature matrix for individual class instances. For the analysis purpose, a statistical test was performed that comprised mean, median, standard deviation, and Kruskal–Wallis nonparametric test for individual trials and it confirmed the efficacy of SDI as a potential feature. Moreover, a single evaluation metric named polygon area metric is employed to avoid looking into long tables. To validate the performance of the said method corresponding to the number of channels, four different channel selection criteria were tested and it confirmed that the 118-channel scheme has the leading results among other combinations. Furthermore, the classifier parameters were varied and a comparison between denoised and noisy data was performed to certify its effect on the classification performance of SDI feature. We also carried out a test for multiclass dataset V, and it was concluded that the proposed method is equally significant for the binary class as well as multiclass data. In the end, a computerized automated system CADMMI-SDI was developed for the practical realization of the proposed method. A comprehensive comparison of this study is made with other state of the art methods and it confirmed that the proposed method is robust, efficient, less complex and it can be utilized for the development of practical BCI systems.

Author Contributions

M.T.S., X.Y., and M.Z.A. modeled the problem and wrote the manuscript. X.Y. thoroughly checked the mathematical modeling and English. M.T.S. and M.Z.A. solved the problem using MATLAB software. M.T.S. and M.Z.A. contributed to the results and discussion. Writing—review and editing, M.T.S., X.Y., M.Z.A., and Z.Y. All authors finalized the manuscript after its internal evaluation. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Fundamental Research Funds for the Central Universities (G2018KY0308), the Chinese Postdoctoral Science Foundation (2018M641013), and Postdoctoral Science Foundation of Shaanxi Province (Grant No. 2018BSHYDZZ05).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Birbaumer, N.; Murguialday, A.R.; Cohen, L. Brain–computer interface in paralysis. Curr. Opin. Neurol. 2008, 21, 634–638. [Google Scholar] [CrossRef]
  2. Jochumsen, M.; Knoche, H.; Kjaer, T.W.; Dinesen, B.; Kidmose, P. EEG Headset Evaluation for Detection of Single-Trial Movement Intention for Brain-Computer Interfaces. Sensors 2020, 20, 2804. [Google Scholar] [CrossRef]
  3. Pfurtscheller, G.; Neuper, C.; Muller, G.; Obermaier, B.; Krausz, G.; Schlogl, A.; Scherer, R.; Graimann, B.; Keinrath, C.; Skliris, D.; et al. Graz-BCI: State of the art and clinical applications. IEEE Trans. 2003, 11, 1–4. [Google Scholar] [CrossRef] [PubMed]
  4. Sadiq, M.T.; Shabbir, N.; Kulesza, W.J. Spectral subtraction for speech enhancement in modulation domain. Ijcsi 2013, 10, 282. [Google Scholar]
  5. Chaudhary, U.; Birbaumer, N.; Curado, M. Brain-machine interface (BMI) in paralysis. Ann. Phys. Rehabil. Med. 2015, 58, 9–13. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Chaudhary, U.; Birbaumer, N.; Ramos-Murguialday, A. Brain–computer interfaces for communication and rehabilitation. Nat. Rev. Neurosci. 2016, 12, 513. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Chaudhary, U.; Mrachacz-Kersting, N.; Birbaumer, N. Neuropsychological and neurophysiological aspects of brain-computer-interface (BCI) control in paralysis. J. Physiol. Paris 2020, JP278775. [Google Scholar] [CrossRef] [Green Version]
  8. Cincotti, F.; Mattia, D.; Aloise, F.; Bufalari, S.; Schalk, G.; Oriolo, G.; Cherubini, A.; Marciani, M.G.; Babiloni, F. Non-invasive brain–computer interface system: Towards its application as assistive technology. Brain Res. Bull. 2008, 75, 796–803. [Google Scholar] [CrossRef] [Green Version]
  9. Siuly, S.; Li, Y. Improving the separability of motor imagery EEG signals using a cross correlation-based least square support vector machine for brain–computer interface. IEEE Trans. 2012, 20, 526–538. [Google Scholar] [CrossRef]
  10. Rehman, A.U.; Naqvi, R.A.; Rehman, A.; Paul, A.; Sadiq, M.T.; Hussain, D. A Trustworthy SIoT Aware Mechanism as an Enabler for Citizen Services in Smart Cities. Electronics 2020, 9, 918. [Google Scholar] [CrossRef]
  11. Wu, X.; Zhou, B.; Lv, Z.; Zhang, C. To explore the potentials of independent component analysis in brain-computer interface of motor imagery. IEEE J. Biomed Health 2019, 24, 775–787. [Google Scholar] [CrossRef] [PubMed]
  12. Jiang, X.; Bian, G.B.; Tian, Z. Removal of artifacts from EEG signals: A review. Sensors 2019, 19, 987. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Alickovic, E.; Subasi, A. Effect of multiscale PCA de-noising in ECG beat classification for diagnosis of cardiovascular diseases. Circ. Syst. Signal Process. 2015, 34, 513–533. [Google Scholar] [CrossRef]
  14. Gokgoz, E.; Subasi, A. Effect of multiscale PCA de-noising on EMG signal classification for diagnosis of neuromuscular disorders. J. Med. Syst. 2014, 38, 31. [Google Scholar] [CrossRef]
  15. Lotte, F.; Bougrain, L.; Cichocki, A.; Clerc, M.; Congedo, M.; Rakotomamonjy, A.; Yger, F. A review of classification algorithms for EEG-based brain–computer interfaces: A 10 year update. J. Neural Eng. 2018, 15, 031005. [Google Scholar] [CrossRef] [Green Version]
  16. Polat, K.; Güneş, S. Classification of epileptiform EEG using a hybrid system based on decision tree classifier and fast Fourier transform. Appl. Math. Comput. 2007, 187, 1017–1026. [Google Scholar] [CrossRef]
  17. Gupta, A.; Agrawal, R.K.; Kirar, J.S.; Andreu-Perez, J.; Ding, W.P.; Lin, C.T.; Prasad, M. On the Utility of Power Spectral Techniques With Feature Selection Techniques for Effective Mental Task Classification in Noninvasive BCI. IEEE Trans. Syst. Man Cybern. Syst. 2019, 99, 1–13. [Google Scholar] [CrossRef] [Green Version]
  18. Lotte, F.; Guan, C. Regularizing common spatial patterns to improve BCI designs: Unified theory and new algorithms. IEEE. Trans. Biomed. Eng. 2010, 58, 355–362. [Google Scholar] [CrossRef] [Green Version]
  19. Lu, H.; Eng, H.L.; Guan, C.; Plataniotis, K.N.; Venetsanopoulos, A.N. Regularized common spatial pattern with aggregation for EEG classification in small-sample setting. IEEE. Trans. Biomed. Eng. 2010, 57, 2936–2946. [Google Scholar]
  20. Rodríguez-Bermúdez, G.; García-Laencina, P.J. Automatic and adaptive classification of electroencephalographic signals for brain computer interfaces. J. Med. Syst. 2012, 36, 51–63. [Google Scholar] [CrossRef]
  21. Pfurtscheller, G.; Neuper, C.; Schlogl, A.; Lugger, K. Separability of EEG signals recorded during right and left motor imagery using adaptive autoregressive parameters. IEEE Trans. Neural Syst. Rehabilit. Eng. 1998, 6, 316–325. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Kevric, J.; Subasi, A. Comparison of signal decomposition methods in classification of EEG signals for motor-imagery BCI system. Biomed. Signal Process. 2017, 31, 398–406. [Google Scholar] [CrossRef]
  23. Sadiq, M.T.; Yu, X.; Yuan, Z.; Fan, Z.; Rehman, A.U.; Li, G.; Xiao, G. Motor imagery EEG signals classification based on mode amplitude and frequency components using empirical wavelet transform. IEEE Access 2019, 7, 127678–127692. [Google Scholar] [CrossRef]
  24. Sadiq, M.T.; Yu, X.; Yuan, Z.; Zeming, F.; Rehman, A.U.; Ullah, I.; Li, G.; Xiao, G. Motor Imagery EEG Signals Decoding by Multivariate Empirical Wavelet Transform-Based Framework for Robust Brain–Computer Interfaces. IEEE Access 2019, 7, 171431–171451. [Google Scholar] [CrossRef]
  25. Chatterjee, R.; Bandyopadhyay, T.; Sanyal, D.K.; Guha, D. Comparative analysis of feature extraction techniques in motor imagery EEG signal classification. In Proceedings of the First International Conference on Smart System, Innovations and Computing, Jaipur, India, 15–16 April 2017; Springer: Jaipur, India, 2017. [Google Scholar]
  26. Wang, L.; Xu, G.; Wang, J.; Yang, S.; Yan, W. Application of Hilbert-Huang transform for the study of motor imagery tasks. In Proceedings of the 30th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Vancouver, BC, Canada, 20–24 August 2008. [Google Scholar]
  27. Chaudhary, P.; Agrawal, R. Non-dyadic wavelet decomposition for sensory-motor imagery EEG classification. Brain Comput. Interfaces 2020, 7, 1–11. [Google Scholar] [CrossRef]
  28. Xu, J.; Zheng, H.; Wang, J.; Li, D.; Fang, X. Recognition of EEG signal motor imagery intention based on deep multi-view feature learning. Sensors 2020, 20, 3496. [Google Scholar] [CrossRef]
  29. Chen, K.; Ai, Q.; Chen, J.; Zhou, S.; Zhao, Y. NAO Robot Walking Control System Based on Motor Imagery. J. Phys. Conf. Ser. 2020, 1453, 12123. [Google Scholar] [CrossRef]
  30. Raghu, S.; Sriraam, N.; Rao, S.V.; Hegde, A.S.; Kubben, P.L. Automated detection of epileptic seizures using successive decomposition index and support vector machine classifier in long-term EEG. Neural Comput. Appl. 2019, 32, 8965–8984. [Google Scholar] [CrossRef]
  31. Teixeira, C.; Direito, B.; Feldwisch-Drentrup, H.; Valderrama, M.; Costa, R.; Alvarado-Rojas, C.; Nikolopoulos, S.; Le Van Quyen, M.; Timmer, J.; Schelter, B.; et al. EPILAB: A software package for studies on the prediction of epileptic seizures. J. Neurosci. Methods 2011, 200, 257–271. [Google Scholar] [CrossRef]
  32. Delorme, A.; Makeig, S. EEGLAB: An open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J. Neurosci. Methods 2004, 134, 9–21. [Google Scholar] [CrossRef] [Green Version]
  33. Oostenveld, R.; Fries, P.; Maris, E.; Schoffelen, J.M. FieldTrip: Open source software for advanced analysis of MEG, EEG, and invasive electrophysiological data. Comput. Intell. Neurosci. 2011, 2011, 156869. [Google Scholar] [CrossRef] [PubMed]
  34. Alickovic, E.; Subasi, A. Ensemble SVM method for automatic sleep stage classification. IEEE Trans. Instrum. Meas. 2018, 67, 1258–1265. [Google Scholar] [CrossRef] [Green Version]
  35. Taran, S.; Bajaj, V. Motor imagery tasks-based EEG signals classification using tunable-Q wavelet transform. Neural. Comput. Appl. 2019, 31, 6925–6932. [Google Scholar] [CrossRef]
  36. Chaudhary, S.; Taran, S.; Bajaj, V.; Siuly, S. A flexible analytic wavelet transform based approach for motor-imagery tasks classification in BCI applications. Comput. Methods Programs Biomed. 2020, 187, 105325. [Google Scholar] [CrossRef]
  37. Chatterjee, R.; Bandyopadhyay, T. EEG based Motor Imagery Classification using SVM and MLP. In Proceedings of the 2nd international conference on Computational Intelligence and Networks (CINE), Bhubaneswar, India, 11 January 2016. [Google Scholar]
  38. Miao, C.; Ma, Y.; Wu, W.; Zhang, Q.; Gao, Y. Pattern Classification of Motor Imagery EEG-NIRS Based on SVM with Artificial Fish Swarm Algorithm. In Proceedings of the 5th International Conference on Information, Cybernetics, and Computational Social Systems (ICCSS), Hangzhou, China, 16–19 August 2018. [Google Scholar]
  39. Kumar, S.; Sharma, R.; Sharma, A.; Tsunoda, T. Decimation filter with common spatial pattern and fishers Discriminant analysis for motor imagery classification. In Proceedings of the 2016 international joint conference on neural networks (IJCNN), Vancouver, BC, Canada, 24–29 July 2016. [Google Scholar]
  40. Fu, R.; Tian, Y.; Bao, T. Recognition method of single trial motor imagery electroencephalogram signal based on sparse common spatial pattern and Fisher discriminant analysis. J. Biomed. Eng. 2019, 36, 911–915. [Google Scholar]
  41. Maksimenko, V.A.; Kurkin, S.A.; Pitsik, E.N.; Musatov, V.Y.; Runnova, A.E.; Efremova, T.Y.; Hramov, A.E.; Pisarchik, A.N. Artificial neural network classification of motor-related eeg: An increase in classification accuracy by reducing signal complexity. Complexity 2018, 2018, 9385947. [Google Scholar] [CrossRef] [Green Version]
  42. Gonzalez, C.D.V.; Azuela, J.H.S.; Espino, E.R.; Ponce, V.H.P. Classification of motor imagery EEG signals with CSP filtering through neural networks models. In Proceedings of the Mexican International Conference on Artificial Intelligence, Guadalajara, Mexico, 22–27 October 2018; Springer: Guadalajara, Mexico, 2018. [Google Scholar]
  43. Jana, G.C.; Swetapadma, A.; Pattnaik, P.K. Enhancing the performance of motor imagery classification to design a robust brain computer interface using feed forward back-propagation neural network. Ain Shams Eng. J. 2018, 9, 2871–2878. [Google Scholar] [CrossRef]
  44. Akhter, M.P.; Jiangbin, Z.; Naqvi, I.R.; Abdelmajeed, M.; Mehmood, A.; Sadiq, M.T. Document-level text classification using single-layer multisize filters convolutional neural network. IEEE Access 2020, 8, 42689–42707. [Google Scholar] [CrossRef]
  45. Akhter, M.P.; Jiangbin, Z.; Naqvi, I.R.; Abdelmajeed, M.; Sadiq, M.T. Automatic Detection of Offensive Language for Urdu and Roman Urdu. IEEE Access 2020, 8, 91213–91226. [Google Scholar] [CrossRef]
  46. Aydemir, O. A New Performance Evaluation Metric for Classifiers: Polygon Area Metric. J. Classif. 2020, 1–11. [Google Scholar] [CrossRef]
  47. Siuly; Li, Y.; Wen, P. Comparisons between motor area EEG and all-channels EEG for two algorithms in motor imagery task classification. Biomed. Eng. Appl. Basis C 2014, 26, 1450040. [Google Scholar] [CrossRef] [Green Version]
  48. Wu, W.; Gao, X.; Hong, B.; Gao, S. Classifying single-trial EEG during motor imagery by iterative spatio-spectral patterns learning (ISSPL). IEEE. Trans. Biomed. Eng. 2008, 55, 1733–1743. [Google Scholar] [CrossRef] [PubMed]
  49. Li, Y.; Wen, P.P. Clustering technique-based least square support vector machine for EEG signal classification. Comput. Methods Programs Biomed. 2011, 104, 358–372. [Google Scholar]
  50. Song, L.; Epps, J. Classifying EEG for brain-computer interface: Learning optimal filters for dynamical system features. Comput. Intell. Neurosci. 2007, 2007, 57180. [Google Scholar] [CrossRef] [Green Version]
  51. Zhang, R.; Xu, P.; Guo, L.; Zhang, Y.; Li, P.; Yao, D. Z-score linear discriminant analysis for EEG based brain-computer interfaces. PLoS ONE 2013, 8, e74433. [Google Scholar] [CrossRef] [Green Version]
  52. Yong, X.; Ward, R.K.; Birch, G.E. Sparse spatial filter optimization for EEG channel reduction in brain-computer interface. In Proceedings of the 2008 IEEE International Conference on Acoustics, Speech and Signal Processing, Las Vegas, NV, USA, 31 March–4 April 2008. [Google Scholar]
  53. Sun, S.; Zhang, C. Adaptive feature extraction for EEG signal classification. Med. Biol. Eng. Comput. 2006, 44, 931–935. [Google Scholar] [CrossRef]
  54. Siuly, S.; Zarei, R.; Wang, H.; Zhang, Y. A new data mining scheme for analysis of big brain signal data. In Proceedings of the Australasian Database Conference, Brisbane, Australia, 25–28 September 2017. [Google Scholar]
  55. Lin, C.J.; Hsieh, M.H. Classification of mental task from EEG data using neural networks based on particle swarm optimization. Neurocomputing 2009, 72, 1121–1130. [Google Scholar] [CrossRef]
  56. Sun, S.; Zhang, C.; Lu, Y. The random electrode selection ensemble for EEG signal classification. Pattern Recognit. 2008, 41, 1663–1675. [Google Scholar] [CrossRef]
  57. Sun, S.; Zhang, C.; Zhang, D. An experimental evaluation of ensemble methods for EEG signal classification. Pattern Recognit. Lett. 2007, 28, 2157–2163. [Google Scholar] [CrossRef]
  58. Taewijit, S.; Theeramunkong, T.; Ikeda, M. Distant supervision with transductive learning for adverse drug reaction identification from electronic medical records. J. Healthc. Eng. 2017, 2017, 7575280. [Google Scholar] [CrossRef] [Green Version]
  59. Wan, S.; Mak, M.W.; Kung, S.Y. Transductive learning for multi-label protein subchloroplast localization prediction. TCBB 2016, 14, 212–224. [Google Scholar] [CrossRef] [PubMed]
  60. Melvin, I.; Ie, E.; Kuang, R.; Weston, J.; Noble, W.S.; Leslie, C. SVM-Fold: A tool for discriminative multi-class protein fold and superfamily recognition. BMC Bioinform. 2007, 8, S2. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Block diagram of the successive decomposition index for identification of motor and mental imagery activities.
Figure 1. Block diagram of the successive decomposition index for identification of motor and mental imagery activities.
Sensors 20 05283 g001
Figure 2. Multiscale principal component analysis (MSPCA) for denoising.
Figure 2. Multiscale principal component analysis (MSPCA) for denoising.
Sensors 20 05283 g002
Figure 3. Scatter plot of SDI features for dataset IVa and IVb subjects.
Figure 3. Scatter plot of SDI features for dataset IVa and IVb subjects.
Sensors 20 05283 g003
Figure 4. Bar plots for the comparison of 3-channel automated, 3-channel, 18-channel, and 118-channel results: (a) FFNN classifier and (b) SVM classifier.
Figure 4. Bar plots for the comparison of 3-channel automated, 3-channel, 18-channel, and 118-channel results: (a) FFNN classifier and (b) SVM classifier.
Sensors 20 05283 g004
Figure 5. (ad) 10-fold Sensitivity, Specificity, Kappa, F1-Score for FFNN Classifier. (eh) Ten-fold Sensitivity, Specificity, Kappa, and F1-Score for SVM Classifier.
Figure 5. (ad) 10-fold Sensitivity, Specificity, Kappa, F1-Score for FFNN Classifier. (eh) Ten-fold Sensitivity, Specificity, Kappa, and F1-Score for SVM Classifier.
Sensors 20 05283 g005
Figure 6. (af) PAM for Subjects “aa”, “al”, “av”, “aw”, “ay” and “Dataset IVb” respectively using FFNN classifier. (gl) PAM for Subjects “aa”, “al”, “av”, “aw”, “ay” and “Dataset IVb” respectively for SVM classifier.
Figure 6. (af) PAM for Subjects “aa”, “al”, “av”, “aw”, “ay” and “Dataset IVb” respectively using FFNN classifier. (gl) PAM for Subjects “aa”, “al”, “av”, “aw”, “ay” and “Dataset IVb” respectively for SVM classifier.
Sensors 20 05283 g006
Figure 7. Results obtained with 10-fold 10 times.
Figure 7. Results obtained with 10-fold 10 times.
Sensors 20 05283 g007
Figure 8. Comparison between denoised and noisy datasets.
Figure 8. Comparison between denoised and noisy datasets.
Sensors 20 05283 g008
Figure 9. Performance parameters of FFNN classifier for Dataset V.
Figure 9. Performance parameters of FFNN classifier for Dataset V.
Sensors 20 05283 g009
Figure 10. A display of CADMMI-SDI portraying all features and functionalities.
Figure 10. A display of CADMMI-SDI portraying all features and functionalities.
Sensors 20 05283 g010
Figure 11. Bar plots representing time complexity: (a) Execution Time for SDI feature extraction method. (b) Training time. (c) Testing time.
Figure 11. Bar plots representing time complexity: (a) Execution Time for SDI feature extraction method. (b) Training time. (c) Testing time.
Sensors 20 05283 g011
Table 1. Statistical analysis.
Table 1. Statistical analysis.
ParticipantsMI TasksMeanStdMedianKW p Values
“aa“”Class 1 (RH)“4.0710.5064.0882.16 × 10−19
”Class 2 (RF)”4.6330.1754.626
“al”“Class 1 (RH)”3.9940.4983.9860.06112
“Class 2 (RF)”4.3330.9314.178
“av”“Class 1 (RH)”3.6550.5163.6983.27 × 10−40
“Class 2 (RF)”5.9610.1395.979
“aw”“Class 1 (RH)”3.8110.3433.7370.0001927
“Class 2 (RF)”3.9660.3233.925
“ay”“Class 1 (RH)”5.7660.5355.7255.81 × 10−39
“Class 2 (RF)”3.9480.6604.058
“IVb”“Class 1 (LH)”3.7160.4403.7344.09 × 10−5
“Class 2 (RF)”3.9930.5243.979
Table 2. List of 3 channels automated criteria.
Table 2. List of 3 channels automated criteria.
SubjectsSelected Channels
aaCCP5, CP5, CP6
alC3, FFC7, CCP3
avFT9, P8, PPO8
awC4, CCP6, CP6
ayCCP5, C3, CFC5
IVbC3, CCP5, C4
Table 3. Classification (%) results for different parameters of the classifier.
Table 3. Classification (%) results for different parameters of the classifier.
“Classifiers”“Variations in Parameters”“aa”“al”“av”“aw”“ay”“Dataset IVb”
“NN”“5 Neurons”97.6193.3077.7883.6795.4399.52
“10 Neurons”95.0091.0586.6790.6797.3399.05
“20 Neurons”97.6198.2087.2289.3398.2397.62
“30 Neurons”97.6197.3376.9493.0099.2498.57
“40 Neurons”99.4195.4985.9793.0099.1099.05
“MNN”“5 Neurons”87.3284.5581.2581.0097.5598.10
“10 Neurons”89.7188.3072.7863.0098.3392.38
“20 Neurons”89.7186.6874.8680.6799.1297.14
“30 Neurons”95.7494.5878.3391.0096.3493.33
“40 Neurons”96.9996.4279.7281.6799.5694.29
“CFNN”“5 Neurons”99.3898.6694.0398.3398.1299.05
“10 Neurons”99.4199.5593.0698.3397.2499.52
“20 Neurons”100.0097.7798.7591.0099.1099.23
“30 Neurons”99.4199.5595.2891.3310099.52
“40 Neurons”100.00100.0097.6494.3395.0099.05
“FFNN”“5 Neurons”99.3899.1194.0396.3397.1198.10
“10 Neurons”98.7598.2491.8198.0098.6798.57
“20 Neurons”100.0097.3190.6996.3397.9899.52
“30 Neurons”99.4198.6695.2898.3399.6599.52
“40 Neurons”10098.3392.796.710099.5
“SVM”“RBF”63.5762.9264.5869.0091.1283.33
“Linear”94.0888.3885.4284.6795.6596.19
“Polynomial”98.2095.5387.2285.3399.1299.52
“DA”“Linear”98.8299.0986.5394.6797.3398.10
“Pseudo Quadratic”94.1297.8184.5886.67100.0093.24
“Pseudo Linear”98.8299.0985.4284.0099.1295.10
Table 4. Different cases consider for SDI experimental work by employing dataset V.
Table 4. Different cases consider for SDI experimental work by employing dataset V.
Case 1:“Class 1 (LH)” vs.
“Class 2 (RH)”
Case 2:“Class 1 (LH)” vs.
“Class 3 (RW)”
Case 3:Class 2 (RH) vs.
“Class 3 (RW)”
Case 4:“Class 1 (LH)“ vs.
”Class 2 (RH)“
Case 5:”Class 1 (LH)“ vs.
”Class 3 (RW)“
Case 6:”Class 2 (RH)“ vs.
”Class 3 (RW)“
Case 7:”Class 1 (LH)“ vs.
”Class 2 (RH)“
Case 8:”Class 1 (LH)“ vs.
”Class 3 (RW)“
Case 9:”Class 2 (RH)“ vs.
”Class 3 (RW)“
Table 5. Classification accuracies (%) obtained with different cases by employing dataset V.
Table 5. Classification accuracies (%) obtained with different cases by employing dataset V.
ClassifiersCases“P1”“P2”“P3”
“NN”Case 1100.0098.2296.67
Case 293.2197.9897.44
Case 3100.0097.1298.43
Average97.7497.7797.51
“MNN”Case 1100.0099.8893.08
Case 298.4395.1698.30
Case 393.2190.3494.12
Average97.2195.1395.16
“CFNN”Case 198.4499.1297.12
Case 298.4999.1298.45
Case 396.3499.3496.34
Average97.7699.1997.30
“FFNN”Case 199.1298.2499.89
Case 2100.0099.1397.12
Case 398.0997.1298.12
Average99.0798.1698.38
“SVM”Case 195.2394.3299.12
Case 294.7495.2394.98
Case 385.5481.5287.34
Average91.8490.3693.81
“DA”Case 193.4594.5593.19
Case 294.1496.3495.83
Case 388.3291.7889.15
Average91.9794.2292.72
Table 6. CADMMI-SDI application.
Table 6. CADMMI-SDI application.
Application ComponentsDescription
Load EEG DataLoad Sample EEG data for a specified destination. The file type must be *.csv or *.xlsx
Test EEG SignalLoad test data from a specific folder. The file format should be *.csv or *.xlsx
ClassifiersChoose a classifier by drop-down selection.
StartA key to initiate/start the process
Channel #Input desired number of channels and press “Plot” to display. The channel number should be separated by a comma
SummaryText section to demonstrate the specifics of the process underway
Signals2D plot window to display EEG signals
Features Scatter Plot2D plot window to display SDI feature corresponding to each channel
Table 7. Performance comparison of motor imagery EEG signals in terms of classification accuracy (%) with other literature.
Table 7. Performance comparison of motor imagery EEG signals in terms of classification accuracy (%) with other literature.
Methods BySuggested MethodsClassification Accuracy (%)
“aa”“al”“av”“aw”“ay”“Avg.”“Std.”
Our Present Work“Successive decomposition index tested with feedforward neural network”/the proposed10098.392.796.710097.52.7
our previous work in 2019 [24]“Multivariate empirical wavelet transform tested with least-square support vector machines”959595100100972.7
our previous work in 2019 [23]“Empirical wavelet transform tested with least-square support vector machines”/our last work94.591.797.295.69795.22.3
work by “Wu” et al. in 2008 [48]“Iterative spatio-spectral patterns learning”93.610079.399.698.694.28.7
work by “Kevric” et al. in 2017 [22]“Wavelet packet decomposition tested with K nearest neighbors”9692.388.995.491.492.82.9
work by “Siuly” et al. in 2011 [49]“Clustering tested with least-square support vector machines“92.684.990.886.586.788.33.2
work by ”Song“ et al. in 2007 [50]“Common spatial pattern tested with support vector machines”87.497.469.796.888.687.911.2
work by “Lu” et al. in 2010 [19]“Regularized common spatial pattern tested with aggregation“76.898.274.592.27783.710.7
work by ”Zhang“ et al. in 2013 [51]“Z-score tested with linear discriminant analysis”77.710068.499.659.981.118.1
work by “Lotte“ et al. in 2010 [18]”Regularized common spatial pattern with selected subjects“.70.596.453.571.975.473.615.3
”Common spatial pattern with Tikhonov regularization“.71.496.463.371.986.977.913.4
”Common spatial pattern with weighted Tikhonov regularization“.69.698.254.671.985.375.916.6
”Spatially regularization common spatial pattern“.72.396.460.277.786.578.613.8
work by ”Yong“ et al. in 2008 [52]”Sparse spatial filter optimization“57.586.954.484.484.373.516
Table 8. Performance comparison of mental imagery EEG signals in terms of classification accuracy (%) with other literature.
Table 8. Performance comparison of mental imagery EEG signals in terms of classification accuracy (%) with other literature.
Methods BySuggested MethodsClassification Accuracy (%)
“P1”“P2”“P3”“Avg.”“Std.”
Our Present Work“Successive decomposition index tested with feedforward neural network”/The proposed88.994.192.491.82.6
work by “Siuly” et al. in 2017 [54]“Principal component analysis employed with random forest”91.875.282.883.38.3
research by “Lin” et al. in 2009 [55]“Modified partical swarm optimization employed with neural networks”78.375.256.569.911.81
work by “Siuly” et al. in 2011 [49]“Clustering employed with least-square support vector machines”68.264.852.161.78.5
experiments by “Sun” et al. in 2008 [56]“Selection of electrodes with the help of Ensemble method”68.756.444.856.711.9
work by “Sun” et al. in 2007 [57]“Ensemble Methods”70.648.940.953.415.4
work by “Sun” et al. in 2009 [53]“Automated common spatial method”67.768.159.665.14.82

Share and Cite

MDPI and ACS Style

Sadiq, M.T.; Yu, X.; Yuan, Z.; Aziz, M.Z. Identification of Motor and Mental Imagery EEG in Two and Multiclass Subject-Dependent Tasks Using Successive Decomposition Index. Sensors 2020, 20, 5283. https://doi.org/10.3390/s20185283

AMA Style

Sadiq MT, Yu X, Yuan Z, Aziz MZ. Identification of Motor and Mental Imagery EEG in Two and Multiclass Subject-Dependent Tasks Using Successive Decomposition Index. Sensors. 2020; 20(18):5283. https://doi.org/10.3390/s20185283

Chicago/Turabian Style

Sadiq, Muhammad Tariq, Xiaojun Yu, Zhaohui Yuan, and Muhammad Zulkifal Aziz. 2020. "Identification of Motor and Mental Imagery EEG in Two and Multiclass Subject-Dependent Tasks Using Successive Decomposition Index" Sensors 20, no. 18: 5283. https://doi.org/10.3390/s20185283

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop