Next Article in Journal
Influence of Simulated State of Disc Degeneration and Axial Stiffness of Coupler in a Hybrid Performance Stabilisation System on the Biomechanics of a Spine Segment Model
Next Article in Special Issue
How the Effect of Virtual Reality on Cognitive Functioning Is Modulated by Gender Differences
Previous Article in Journal
Artificial Intelligence Advances in Transplant Pathology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Emotion Recognition Using Hierarchical Spatiotemporal Electroencephalogram Information from Local to Global Brain Regions

1
Department of Electronic Convergence Engineering, Kwangwoon University, 20 Gwangun-ro, Nowon-gu, Seoul 01897, Republic of Korea
2
Department of ICT Convergence System Engineering, Chonnam National University, 77 Yongbong-ro, Buk-gu, Gwangju 61186, Republic of Korea
*
Author to whom correspondence should be addressed.
Bioengineering 2023, 10(9), 1040; https://doi.org/10.3390/bioengineering10091040
Submission received: 24 July 2023 / Revised: 26 August 2023 / Accepted: 28 August 2023 / Published: 4 September 2023

Abstract

:
To understand human emotional states, local activities in various regions of the cerebral cortex and the interactions among different brain regions must be considered. This paper proposes a hierarchical emotional context feature learning model that improves multichannel electroencephalography (EEG)-based emotion recognition by learning spatiotemporal EEG features from a local brain region to a global brain region. The proposed method comprises a regional brain-level encoding module, a global brain-level encoding module, and a classifier. First, multichannel EEG signals grouped into nine regions based on the functional role of the brain are input into a regional brain-level encoding module to learn local spatiotemporal information. Subsequently, the global brain-level encoding module improved emotional classification performance by integrating local spatiotemporal information from various brain regions to learn the global context features of brain regions related to emotions. Next, we applied a two-layer bidirectional gated recurrent unit (BGRU) with self-attention to the regional brain-level module and a one-layer BGRU with self-attention to the global brain-level module. Experiments were conducted using three datasets to evaluate the EEG-based emotion recognition performance of the proposed method. The results proved that the proposed method achieves superior performance by reflecting the characteristics of multichannel EEG signals better than state-of-the-art methods.

1. Introduction

Emotional recognition is considered an important topic in various fields, such as human–computer interaction, psychological research, and neuroscience [1,2,3,4]. Emotions are crucial in daily life and significantly influence behavior, communication, thinking, and mental health [5]. Thus, if machines can recognize human emotions, they offer innovative potential in various applications, including designing more effective computer systems, providing personalized services, and diagnosing and treating mental health [6].
In early emotion research studies, numerous attempts were made to recognize emotions using external signals, such as facial expressions, voices, and behavioral patterns [7,8,9,10]. However, these external signals can be judged differently depending on the subjective interpretation, and discrepancies may arise between outwardly expressed and actual inner emotions [11]. For this reason, with the development of noninvasive sensor technology, the research and application of emotion recognition using biosignals has recently attracted considerable attention [12]. Biological signals [13,14] measure a person’s physical condition, nervous system activity, and physiological responses, including electrocardiograms, skin conductivity, electromyography, and electroencephalography (EEG). Such biometric signals may reflect changes in the internal state of a person and provide information related to emotions and emotional reactions [15].
Among these biosignals, EEG, which is measured from the brain and is directly involved in the processing, generation, and control of emotions, is considered a crucial biosignal for emotion recognition and has several advantages over other modalities. EEG can directly evaluate brain areas related to emotions by directly measuring electrical activities in the brain, provides high temporal resolution to detect and analyze changes in emotions quickly, and can be applied to medical and clinical settings because it detects a wide range of emotions. In addition, emotion recognition using EEG is useful for personalized emotion recognition models and real-time emotion monitoring. Therefore, emotion recognition using EEG can understand and interpret an individual’s internal state and emotions more accurately [16].
Initially, single-channel EEG was primarily used to record brain waves using a single electrode [17]. However, the need for multichannel EEG recordings to analyze the activities occurring in various brain regions has gradually emerged, and now multichannel EEG systems that record brain waves simultaneously using dozens of electrodes are widely used. In emotion recognition studies, multichannel EEG can be used to understand the brain’s emotional processing mechanisms and develop more accurate and comprehensive emotion recognition systems by providing local activities associated with specific brain regions, network interactions among various brain regions, and individual differences and changes in emotions [18].
Deep learning technology, which has rapidly been developing and becoming commonplace in recent years, has achieved remarkable results in EEG-based emotion recognition, surpassing traditional machine learning methods [19,20,21,22]. Li et al. [23] used a hierarchical convolutional neural network (CNN) for EEG-based emotion classification to hierarchically extract features contained in the spatial topology of electrodes that are neglected in a one-dimensional deep model, such as a stacked autoencoder. Li et al. [24] improved the emotion recognition performance using multichannel EEG signals by constructing a hybrid deep learning model that combined CNNs and recurrent neural networks (RNNs). The proposed model effectively captured the interchannel correlation and contextual information. Chen et al. [25] proposed a hierarchical BGRU network with attention mechanism. This model reflects the hierarchical structure of EEG signals and learns important features by utilizing the attention mechanism at both the sample and epoch levels.
However, such existing approaches still have some limitations. First, the research results of EEG-based emotion recognition using deep learning methods still lag behind those of image and speech recognition. Therefore, to develop a more accurate and reliable emotion recognition model, an improved deep learning model that reflects EEG characteristics is required. Second, according to neurological studies, human emotions are closely related to various areas of the cerebral cortex, such as the amygdala, frontal lobe, and parietal lobe [26,27], and spatiotemporal information from different brain regions helps us understand human emotions [28,29,30]. However, these neurological research results are not sufficiently reflected. Because the contribution of EEG signals related to each brain region is different, a method that can utilize the spatiotemporal information of different brain regions is required to understand human emotional states. Third, information about spatial resolution, brain region separation, brain network analysis, individual differences in emotions, and capturing dynamic changes in emotions, which multichannel EEG includes, is not fully utilized. To this end, new studies have recently attempted to utilize the advantages of multichannel EEG fully to identify activities in specific regions of the brain and perform more accurate and multifaceted emotion recognition through the interaction and information combination among brain regions. However, the number of studies is extremely small. Zhang et al. [31] used a hierarchical spatiotemporal EEG feature learning model with attention-based antagonism neural network, whereas Wang et al. [32] proposed a transformer-based model to hierarchically learn the discriminative spatial information from the electrode level to the brain-region level.
Motivated by these research results and limitations, a more accurate and reliable emotion recognition model was developed by reflecting EEG characteristics based on neurological studies. In this paper, we propose an EEG-based emotion recognition method using a hierarchical spatiotemporal context feature learning model (HSCFLM) that extracts local and global EEG features by robustly learning the spatiotemporal dependencies within and among brain regions.
  • The main contributions of this paper are:
  • To understand the activity of certain brain regions based on human emotions and to enable the interaction among brain regions and the combination of information, we propose a hierarchical neural network model structure with self-attention. The hierarchical deep neural network model with self-attention comprises a regional brain-level encoding module and a global brain-level encoding module;
  • To obtain activity weights for each brain region according to emotional state, a regional brain-level encoding module was designed based on a dual-stream parallel double BGRU encoder with self-attention (2BGRUA) for extracting spatiotemporal EEG features. In particular, the spatial encoder learns the interchannel correlations and the temporal encoder captures the temporal dependencies of the time sequence of the EEG channels. Subsequently, the output of each encoder can be integrated to obtain the local spatiotemporal EEG features;
  • Next, the global spatiotemporal encoding module uses a single BGRU-based self-attention (BGRUA) to integrate important information within various brain regions to improve the emotion classification performance by learning discriminative spatiotemporal EEG features from local brain regions to the entire brain region. Thus, the influence of brain regions with a high contribution is strengthened by the learned weights, and the influence of the less dependent regions is reduced.
The remainder of this paper is organized as follows: Section 2 introduces the proposed method, Section 3 describes the experimental data and results, and Section 4 presents the conclusions of the study and future research directions.

2. Proposed Hierarchical Spatiotemporal Context Feature Learning Model for EEG-Based Emotion Recognition

The proposed method for EEG-based emotion classification is a hierarchical neural network architecture with self-attention called the HSCFLM, which mainly consists of three key modules: electrode division module, regional brain-level encoding module, and global brain-level encoding and classification module, as shown in Figure 1. First, EEG channels are grouped by brain region according to the spatial location of the electrodes, because each brain region has a different function. Each grouped regional EEG signal is preprocessed and inputted into the regional brain-level encoding module to extract the spatiotemporal features of each region. After learning the regional deep features, the global brain-level encoding module is used to learn the global emotional EEG context features from local to global brain regions, which are then input into the emotion classifier. The remainder of this section introduces these three key modules.

2.1. Electrode Division

The human brain consists of several regions, each performing its own function. These include the frontal, temporal, central, parietal, and occipital regions. Each of these regions shows different brain activity depending on the emotion, which is crucial in understanding complex human emotions. Applying this phenomenon to analyze the correlation among brain regions that respond to human emotional changes, we divided these regions into nine regions based on structure and function [33]: prefrontal, frontal, left temporal, right temporal, central, left parietal, parietal, right parietal, and occipital. Based on the nine brain regions classified above, each electrode can be divided into the corresponding brain region groups based on the position of the electrode attached to the brain to measure the multichannel EEG signal. An example is shown in Figure 2. Electrodes, divided by region, are shown on the left side of Figure 2, and electrodes that belong to the same region are marked with the same color. This information is summarized in the table on the right-hand side. Each region of the brain is expressed as R r   r = 1,2 , , 9 , and electrodes Fp1, Fp2, AF3, and AF4 are assigned to R 1 (prefrontal); F7, F3, Fz, F4, and F8 to R 2 (frontal); FC5, T7, and CP5 to R 3 (left temporal); FC6, T8, and CP6 to R 4 (right temporal); FC1, C3, Cz, C4, and FC2 to R 5 (central); P3, P7, and PO3 to R 6 (left parietal); CP1, Pz, and CP2 to R 7 (parietal); P8, P4, and PO4 to R 8 (right parietal); and O1, Oz, and O2 are assigned to R 9 (occipital).

2.2. Preprocessing and Spectral Analysis

Although there may be differences depending on the recording conditions and devices, in general, EEG signals are recorded at an extremely low-voltage level of approximately 1 to 100 μV and include various noise signals in addition to signals generated by the brain itself. Specifically, artifacts from eye blinking, which typically have a higher voltage level than normal EEG signals from emotional stimuli, can lead to a loss of emotion-related EEG signals; therefore, they must be removed to improve the accuracy of emotion recognition.
The preprocessing and spectral analysis are performed on the EEG signals for each brain region. The preprocessing consists of bandpass filtering and downsampling, removal of eye-blink artifacts, and segmentation.
First, a bandpass filter of 4–47 Hz is applied to the EEG signal, and downsampling was performed to 128 Hz. This removes background noise by maintaining only signals in the 4–47 Hz band, eliminating the rest, and reducing the amount of EEG data through downsampling.
Second, eye-blink artifacts are removed from the EEG signal using an optimally modified log-spectral amplitude speech estimator and a minima controlled recursive averaging noise estimation (OM-LSA)-based algorithm [34]. This process comprised the following steps—(Step 1) Artifact detection: Using the OM-LSA-based algorithm, the positions of eye-blink artifacts in the EEG signal are detected. In this process, the start and end points of the artifact are found, defining this as the artifact occurrence section (AOS). (Step 2) AOS length calculation: The number of sample data points in the AOS is calculated using information from its start and endpoints. (Step 3) Acquisition of sample data: Previous samples of the AOS length from the AOS start point and subsequent samples of the AOS length from the AOS endpoint are acquired. (Step 4) Overlap-and-add operation [35]: An operation is performed to overlap and add the EEG samples of the two sections acquired in the previous step. Thus, signals similar to EEG signals related to real emotions are generated. (Step 5) Data replacement and loss concealment: The eye-blink artifact located in the AOS is replaced with the EEG signal generated in the previous step. Thus, the eye-blink artifact is removed, and the information loss of the EEG signal is concealed.
Third, the EEG signals for each brain region from which noise and artifacts are removed are divided into 1 s segments, and each segment overlaps by 50%. Each segmented EEG signal is used in the subsequent step of the spectral analysis of each brain region.
The five main frequency bands of the brain waves are delta (1–4 Hz), theta (4–7 Hz), alpha (8–12 Hz), beta (13–30 Hz), and gamma (31–47 Hz), which are associated with different emotional states. By converting these EEG signals into a power spectral density (PSD), different emotional states can be effectively distinguished and recognized through the distribution pattern of the signal power for each band. To this end, each EEG segment is analyzed with a short-time Fourier transform using a 0.25 s sliding window with 80% overlap. Here, the PSD is calculated using the average power in four frequency bands, excluding the delta band related to deep sleep. Thus, the PSD feature expression S r = p t 1 r , p t 2 r , ,   p t N r r R d × N r × T for brain region R   r is the output. Here, N r and T denote the number of channels and segments in the EEG sequence for the brain region R r , respectively. Furthermore, d represents the dimensions of the PSD features extracted from a single channel. Feature vector S r is used as the input for regional brain-level encoding.

2.3. Regional Brain-Level Encoding Module

Certain emotions can be associated with specific brain regions. For example, fear-related brain regions are deeply related to the amygdala, and love-related emotions do not focus on just one area of the brain but involve the interaction of several areas of the brain. However, certain brain areas are related to the experience of love and romantic emotions. Simultaneous recordings of brain waves occurring in different brain regions using multichannel EEG can identify local spatiotemporal pattern changes in the brain activity associated with specific emotions.
For this reason, we designed a regional brain-level encoding module to effectively learn spatiotemporal EEG feature representations using both the spatial and temporal information of each brain region. The regional brain-level encoding module has a dual-stream parallel structure and consists of two submodules, a spatial encoder, and a temporal encoder, the structure of which is shown in Figure 3.
The PSD feature expression S r , obtained through preprocessing and spectrum analysis, is configured as in Equations (1) and (2) and input into spatial and temporal encoders in parallel.
H r = c 1 r , c 2 r , ,   c n r , ,   c N r r R d × N r
U r = s 1 r , s 2 r , , s t r , ,   s T r R d × T
where c n r represents the PSD feature vector of the n-th channel of the EEG channel sequence H r , and s t r represents the PSD feature vector of the t-th segment of the EEG time sequence U r .
Both the spatial and temporal encoders are implemented based on a two-layer BGRU with self-attention (2BGRUA). The spatial encoder utilizes 2BGRUA to learn the interchannel correlations of the input regional EEG channels. We also employed 2BGRUA, as shown in Figure 3, as a temporal encoder to learn the key temporal features of the time-sequential EEG data. The EEG sequences H r and U r are input into the first BGRU layer of the spatial and temporal encoders composed of 2BGRUA. The process up to the second layer of BGRU is performed using Equations (3)–(6).
B G R U H r = H ˇ r = h 1 C r , h 2 C r , , h n C r , , h N r C r
B G R U U r = U ˇ r = h 1 S r , h 2 S r , , h t S r , , h T S r
B G R U H ˇ r = H ^ r = g 1 C r , g 2 C r , , g n C r , , g N r C r
B G R U U ˇ r = U ^ r = g 1 S r , g 2 S r , , g t C r , , g T S r
where B G R U · denotes the BGRU operation; H ˇ r and U ˇ r denote the output vectors of the first BGRU layer for the input sequences H r and U r , respectively; and H ^ r and U ^ r denote the output vectors of the second BGRU layer. h n C r R 2 d h c and h t S r R 2 d h s represent the hidden state vectors output by the n-th and t-th bidirectional hidden units of the first BGRU layer, respectively. d h c and d h s denote the dimensions of the hidden state vector. Furthermore, g n C r R 2 d g c and g t S r R 2 d g s represent the hidden state vectors output by the n-th and t-th bidirectional hidden units of the second BGRU layer. The dimensions of each hidden unit state vector are d g c and d g s , respectively.
Specifically, as shown in Figure 4, the BGRU has a structure in which a hidden layer that processes sequences in reverse order is added to the structure of the GRU [36], which consists of only forward layers that process input sequences sequentially. This structure allows the BGRU to learn long-term dependencies better than the GRU by considering both the past and future states of the input sequence.
The forward and backward context feature representations for an input sequence extracted via BGRU are as follows in (7) and (8):
h t = G R U f x t , t 1 , T
h t = G R U b x t , t T , 1
where h t , h t , G R U f , G R U b , and x t denote the forward hidden sequence, backward hidden sequence, forward GRU, backward GRU, and the feature vector of the sample at time t , respectively. h t is obtained by processing the input sample sequence in order from time step t = 1 to T through G R U f , and h t is obtained by processing the data in reverse order from t = T to 1. The obtained forward and backward context feature vectors h t and h t are summarized in the bidirectional context feature expression h t , as shown in Equation (9), using a concatenation operation:
h t = h t , h t
The 2BGRUA model builds two BGRU networks in series and adds a self-attention mechanism. In this hybrid model, the first BGRU layer extracts future and past time series EEG information, and the second BGRU layer is applied to learn more significant feature representation, while better emotional feature information is provided by assigning weights to contextual feature information using a self-attention mechanism. In general, self-attention can effectively model global interactions; however, it has limited disadvantages in figuring out local dependencies. In contrast, the BGRU can effectively identify regional dependencies that reflect bidirectional temporal patterns and flows within the input sequence. Therefore, we applied the BGRU before self-attention. Figure 5 shows the structure of the self-attention layer.
First, each feature vector H ^ r and U ^ r output through the second BGRU layer is input into self-attention and is output as each query, key, and value by multiplying with the corresponding weight matrix, as shown in Equations (10) and (11):
Q C r = H ^ r W Q C r , K C r = H ^ r W K C r , V C r = H ^ r W V C r
Q S r = U ^ r W Q S r , K S r = U ^ r W K S r , V S r = U ^ r W V S r
where Q C r , K C r , and V C r denote the query, key, and value by the linear transformation of input H ^ r , respectively, and Q S r , K S r , and V S r refer to the query, key, and value by the linear transformation of input U ^ r , respectively. W Q C r , W K C r , W V C r , W Q S r , W K S r , and W V S r indicate the learnable weight matrices.
Next, the similarity between the query and key vectors is calculated using the dot product. The calculated similarity is divided by the square root of the dimension of the key vector, and the probability distribution is calculated by applying a softmax function to the result to obtain each attention score. A dot product operation is then performed between the obtained attention score and the value vector. Thus, each feature vector, A t t e n t i o n Q C r , K C r , V C r and A t t e n t i o n Q S r , K S r , V S r in which the attention weight is reflected, is obtained. These are the output feature representations of the spatial and temporal encoders, respectively, and are expressed by Equations (12) and (13).
A t t e n t i o n Q C r , K C r , V C r = s o f t m a x Q C r K C r T d K C r V C r
A t t e n t i o n Q S r , K S r , V S r = s o f t m a x Q S r K S r T d K S r V S r
where T represents the matrix transpose, d K C r and d K S r are the dimensions of the key vectors K C r and K S r , respectively. s o f t m a x ( . ) and A t t e n t i o n ( . ) denote the softmax function and self-attention operation, respectively.
Through the previous process, each output feature extracted independently from the spatial and temporal encoders is input into the spatiotemporal encoding module and fused into a spatiotemporal EEG feature L r for the corresponding brain region R r using concatenation operations, as shown in Equation (14).
L r = A t t e n t i o n Q C r , K C r , V C r , A t t e n t i o n Q S r , K S r , V S r
Similarly, the local spatiotemporal features extracted from the nine brain regions are all concatenated and output as a local spatiotemporal feature vector L , as shown in Equation (15), which is then used as an input for global brain-level encoding.
L = L 1 , L 2 , , L 9

2.4. Global Brain-Level Encoding and Classification Module

Emotions are distributed across various brain regions, and each region has temporal and spatial emotional characteristics. The brain is composed of networks, and emotions are associated with interactions among different brain regions. Therefore, emotional recognition performance can be improved by extracting spatiotemporal features from each brain region and learning global spatiotemporal features through interactions among brain regions and combining information. This helps us recognize emotions more accurately by simultaneously considering the activities of various brain regions. The global brain-level encoding model is a key module that can effectively learn spatiotemporal features extracted from various brain regions and improve emotion recognition performance by identifying globally distinct patterns according to various emotional states.
Figure 6 shows the structure of the global brain-level encoding and classification model, which consists of BGRU, self-attention, and softmax. Unlike the regional brain-level encoder, which uses a two-layer BGRU, the global brain-level encoder uses a single-layer BGRU. This is because it is only used to combine information among brain regions based on spatiotemporal information sufficiently extracted from the regional brain-level.
The local spatiotemporal feature vector L obtained through regional brain-level encoding is input into the BGRU layer, as shown in Equation (16), and the global context feature vector E is obtained.
B G R U L = E = h 1 G , h 2 G , , h 9 G R 2 d g × 9
where d g represents the dimension of each hidden unit state vector.
Next, the extracted global context feature vector E is input into the self-attention to extract a global spatiotemporal feature G . This is obtained using Equations (17) and (18):
A t t e n t i o n Q G , K G , V G = s o f t m a x Q G K G T d K G V G
G = A t t e n t i o n Q G , K G , V G
where Q G , K G , and V G denote the query, key, and value, respectively, by the linear transformation of the input L . In addition, d K G represents the dimensions of key vector K G .
The global spatiotemporal feature G extracted from the global brain-level encoder is input into the softmax classifier for emotional state prediction and used to calculate the probability distribution for the emotion classes.
First, a linear transformation is performed on feature expression G using the learned weight matrix W z and bias vector b z , which is expressed in Equation (19):
M = W z G + b z = v 1 ,   v 2 , , v c
where c is the number of emotion categories for classification, and v i represents the raw score for the i-th class.
The obtained raw score M is input into the softmax function to calculate the probability of each emotion class. The classifier outputs the class with the highest probability p as the recognized emotional state label. The calculation for p is as in Equation (20) below.
p = m a x e x p v j i = 1 c e x p v i j = 1,2 , , c
Cross-entropy is used as the loss function for model training and is expressed in Equation (21).
E = 1 Z z 1 Z Y z log Y ~ z X z , θ + λ 2 θ 2 2
where X z and Y z denote the z-th EEG sequence and the label of the sequence, respectively. Additionally, θ and λ denote the parameter set and normalization parameters of the model, respectively.
The Adam optimization algorithm is used to train the model to minimize the cross-entropy error between the predicted and actual labels. The weight parameters and biases are adjusted based on the calculated losses. The training process is continued until either the desired performance or the best possible performance was achieved.

3. Experiment and Results

The DEAP, MAHNOB-HCI, and SEED datasets were used to evaluate the performance of the proposed method. All three datasets include EEG data generated by audiovisual emotional stimuli.

3.1. Evaluation Datasets

The performance of the proposed model was evaluated using three public datasets containing EEG signals generated by audiovisual stimuli.
  • DEAP [37]: The DEAP dataset contains EEG and peripheral signals from 32 participants (16 males and 16 females between the ages of 19 and 37 years). EEG signals were recorded while each participant watched 40 music video clips. Each participant watched each video and rated the levels of arousal, valence, dominance, and liking on a continuous scale of 1 to 9 using a self-assessment manikin (SAM). Each trial contained a 63 s EEG signal. The first 3 s of the signal is the baseline signal. EEG signals were recorded at a sampling rate of 512 Hz using 32 electrodes. In this paper, EEG data from 24 participants (12 males and 12 females) were selected for the experiment;
  • MAHNOB-HCI [38]: The MAHNOB-HCI dataset includes the EEG and peripheral signals. EEG signals were collected at a sampling rate of 256 Hz from 32 electrodes while each of the 27 participants (11 males and 16 females) watched 20 selected videos. The video clip used as the stimulus had a length of approximately 34–117 s. Each participant watched each video and self-reported the levels of arousal, valence, dominance, and predictability through the SAM on a nine-point discrete scale. For the various experiments performed in this study, 22 of the 27 participants (11 male and 11 female) were selected;
  • SEED [39]: The SEED dataset provides EEG and eye-movement signals from 15 participants (7 males and 8 females). The EEG signals of each participant were collected while watching 15 Chinese movie clips, each approximately four minutes long, designed to elicit positive, neutral, and negative emotions. The sampling rate of the signal collected using the 62 electrodes was 1 kHz, which was later downsampled to 200 Hz. After watching each movie clip, each participant recorded the emotional label for each video as negative (−1), neutral (0), or positive (1). The experiment was performed using EEG data from 12 of the 15 participants (6 males and 6 females) were used.

3.2. Experimental Methods

To evaluate the emotion classification performance of the proposed method and compare it with other neural network models, the various methods listed below were applied in the experiment.
  • CNN: This method constituted two convolution layers with four 3 × 3 convolutional filters, two max-pooling layers with a 2 × 2 filter size, a dropout layer, two fully connected layers, and a softmax layer;
  • LSTM: LSTM was applied rather than a CNN. It comprised two LSTM layers: a dropout, fully connected and a softmax layers. The number of hidden units of the lower- and upper-layer LSTMs were 128 and 64, respectively. The fully connected layer contained 128 hidden units;
  • BGRU: BGRU was used rather than LSTM. The number of hidden units in the forward and backward GRU layers was 64;
  • Convolutional recurrent neural network (CRNN) [24]: A hybrid neural network consisting of a CNN and an RNN for extracting spatiotemporal features was applied to multichannel EEG sequences. This network consisted of two convolution layers, a subsampling layer, two fully connected recurrent layers, and an output layer;
  • Hierarchical-attention-based BGRU (HA-BGRU) [25]: The HA-BGRU consisted of two layers. The first layer encoded the local correlation between samples in each epoch of the EEG signal, and the second layer encoded the temporal correlation between epochs in an EEG sequence. The BGRU network and attention mechanism were applied at both sample and epoch levels;
  • Region-to-global HA-BGRU (R2G HA-BGRU): the HA-BGRU network was applied to extract regional features within each brain region and global features between regions;
  • R2G transformer encoder (TF-Encoder): in the R2G HA-BGRU method, transformer encoders were applied instead of BGRU with attention mechanism;
  • R2G hierarchical spatial attention-based BGRU (HSA-BGRU): only temporal encoding in a regional brain-level encoding module was used, and the same method as the HSCFLM was applied for the rest;
  • HSCFLM: This method mainly consisted of a regional brain-level encoding module, a global brain-level encoding, and classification module. This is the proposed method.
The CNN, LSTM, BGRU, CRNN, and HA-BGRU methods apply multichannel EEG signals to each neural network without dividing the brain into regions. However, the R2G HA-BGRU, R2GTF, R2G HAS-BGRU, and proposed HSCFLM methods divided the brain into nine defined regions, grouping the corresponding electrodes by region. Among them, a soft attention mechanism was applied to the R2G HA-BGRU method, while a self-attention mechanism was applied to the other methods.
Accuracy was used to evaluate the performance of each method and is defined by Equation (22):
A c c u r a c y = T P + T N T P + T N + F P + F N
where true positive (TP) refers to the number of positive data points correctly classified as positive. True negative (TN) refers to the number of negative data points correctly classified as negative. False positive (FP) refers to the number of negative data points incorrectly classified as positive. False negative (FN) refers to the number of positive data points incorrectly classified as negative. The accuracy of the model is defined as the ratio of correctly classified data points to the total number of data points.

3.3. Experimental Results

For the various experiments evaluating the performance of the proposed model, as shown in Table 1 and Table 2, the labels of the DEAP and MAHNOB-HCI datasets were redefined for several levels of emotion classes based on a rating value of 1–9 in terms of valence, arousal, and dominance. Additionally, by combining each of the two classes for valence and arousal presented in Table 1, the classes for the four-level emotion classification were defined as follows: high valence and high arousal (HVHA), low valence and high arousal (LVHA), low valence and low arousal (LVLA), and high valence and low arousal (HVLA). Because the SEED dataset is labeled as negative (−1), neutral (0), or positive (1), it was only applied to the three-level emotion classification experiment in terms of valence without relabeling.
Table 3, Table 4 and Table 5 show the results of two to four levels of emotion classification using the three datasets. All emotion classification results in Table 3, Table 4 and Table 5 were calculated using leave-one-subject-out (LOSO) cross-validation evaluation [40]. This evaluation method sets the EEG data of one subject as a test set and uses the EEG data of all the remaining subjects as a training set. This process was repeated until each participant’s data were used as a test set at least once. That is, the test and training data were always selected from different subjects. This approach can be used to evaluate the generalizability of the proposed model to new subjects by learning general patterns. The final classification performance results were obtained by calculating the average of all the cross-validation folds.
First, as shown in Table 3, Table 4 and Table 5, the accuracy of the emotion classification gradually decreased as the number of classes increased from two to four in terms of valence, arousal, and dominance in the DEAP and MAHNOB-HCI datasets. This is because as the number of emotion classes to be distinguished increases, the number and complexity of patterns that the model needs to learn increases.
For the two- and three-level emotion classifications using the DEAP dataset, as shown in Table 3, the classification accuracy was high in the order of dominance, valence, and arousal for most of the methods. Alternatively, it exhibited the lowest classification performance in terms of arousal. Similarly, two- and three-level classification using the MAHNOBHCI dataset, as presented in Table 4, showed the same tendency with the lowest classification performance in terms of arousal. This suggests that arousal may represent more complex EEG patterns than either dominance or valence. Additionally, the three-level emotion classification exhibited the highest performance in the SEED dataset among the three datasets. This may be related to the SEED dataset, which employs more spatiotemporal information from numerous local brain regions and more electrodes than the DEAP and MAHNOB-HCI datasets.
In all subject-independent experiments, the proposed method achieved the highest emotion classification accuracy. Table 6 shows the number of learnable parameters to obtain various feature vectors in the application of the proposed HSCFLM. During simulation by connecting the modules of the proposed method, the slopes of all parameters were calculated and optimized using the cross-entropy loss. The obtained optimal parameters were applied to the proposed architecture to provide the highest classification performance. In addition, Table 7 shows the network parameters, training time, test time per EEG data, network size, and average recognition accuracy performance. This represents the complexity of the model from a quantitative perspective. This information was obtained from hardware environments using Windows 10, Intel Core i5 8500 processor, NVIDIA GeForce GTX 1070 Ti GPU, DDR4 32GB RAM, and software environments using TensorFlow 2.9.1 and Python 3.10.8.
As shown in Table 3, Table 4 and Table 5, the R2G HA-BGRU, R2G TF-Encoder, R2G HSA-BGRU, and the proposed HSCFLM method exhibited superior performance in all types of experiments on the DEAP, MAHNOB-HCI, and SEED datasets. This indicates that learning spatiotemporal dependencies within and among brain regions by dividing the brain into areas is more effective for recognizing emotions than extracting temporal and spatial features from multichannel EEG signals without dividing the brain into regions.
Evidently, the R2G HSA-BGRU and HSCFLM methods, which utilize a hybrid structure combining BGRU and self-attention, present superior performance compared to the R2G TF-Encoder method. This suggests that the hybrid structure can more accurately capture regional and global dependencies from EEG signals to extract high-level global EEG features. Furthermore, upon comparing the performance of the R2G HSA-BGRU and HSCFLM methods, it can be observed that the method that considers both temporal and spatial information improves the accuracy of emotion recognition compared with the method that extracts only temporal information. This approach considers brain activity patterns and complex interactions to better understand and reflect complex patterns of emotions.
As a further experiment, we obtained a classification accuracy of 79.1% on the DEAP dataset by dividing the brain region into four primary regions (frontal, parietal, temporal, and occipital lobes) and measuring the performance of the four-level classification of the proposed model. This result was approximately 3.4% lower than the result in Table 4 obtained by dividing the brain into nine regions. Based on this result, it was found that the analysis of various brain regions can improve the accuracy of emotional classification.
To evaluate the performance of the proposed model, a within-subject experiments on emotion classification was also conducted, and the results of these experiments are presented in Table 8. Table 8 contains the results of four-level emotion classification experiments using the DEAP and MAHNOB-HCI datasets and three-level emotion classification experiments using the SEED dataset. In addition, a paired t-test was performed for the proposed method and other neural network models to evaluate the statistical difference in emotion classification performance. EEG data for 10, 8, and 5 subjects in the DEAP, MAHNOB-HCI, and SEED datasets, respectively, were applied. A p-value less than 0.05 indicates that the difference is statistically significant. From the experimental results, the average accuracy performance of the proposed model was 92.4%, 93.1%, and 97.2% for the DEAP, MAHNOB-HCI, and SEED datasets, respectively, showing average classification accuracy performance that was statistically significantly higher than those of the other methods.
Performance evaluation of the brain region, based on the emotional state, was performed using the proposed model, and the results are presented in Table 9. The results indicate that emotions can be recognized similarly in all brain regions. More specifically, in the DEAP database the prefrontal lobe (PF), frontal lobe (F), and parietal lobe (P) achieved better performance for arousal classification, which indicates the degree of emotional activation. In contrast, in the MAHNOB-HCI database, we demonstrated that PF, P, and the right parietal achieved better performance. It is estimated that the PF and F are located in the front of the brain and are connected to numerous brain regions that process emotion-related information; therefore, they are crucial to processing and regulating emotion-related information. These results were similar to the SEED database and were consistent with observations in neuroscience. Furthermore, PF, F, left temporal lobe, and right temporal lobe demonstrated superior performance in valence classification, which represents emotional tendencies or moods, across the DEAP, MAHNOB-HCI, and SEED databases. This suggests that the prefrontal, frontal, and temporal lobe contribute more to the valence classification based on emotional experiences and learning.

4. Conclusions, Limitations, and Future Work

In this paper, we propose an emotion classification method called the hierarchical emotion context feature learning model, which learns spatiotemporal emotion information from a local brain region to a global brain region. Our method extracts spatiotemporal EEG features that represent activity weights in each brain region in relation to emotions and utilizes a global EEG representation that incorporates correlations in various brain regions. Specifically, we extracted EEG spatiotemporal features using regional brain-level encoders. Using global spatiotemporal encoders, we learned discriminative spatiotemporal EEG features from local brain regions to entire brain regions. Comparative experiments demonstrated the superiority of the proposed method over EEG-based emotion-recognition neural models. Analyzing emotions by extracting regional and global EEG characteristics by learning the spatiotemporal dependence among brain regions through multichannel brainwave signals is a novel task and a successful attempt at emotional analysis research in emotional computing. Thus, we identified certain limitations of the proposed method and identified certain directions for future improvements.
First, the principles and learning processes of the proposed model were clarified, achieving higher classification accuracy compared to conventional methods; however, the model is complex, making it difficult to explain the intermediate decision making and overall interaction. To this end, we intend to increase the possibility of explanations by visualizing internal actions, such as the correlation and interaction among input data, learned features, and output classes. Additionally, we plan to explore methods to enhance the robustness and domain adaptability of the model. This will involve simplifying the model’s structure and parameters and reducing the complexity of the hierarchical deep-learning modules through visual verification.
Second, the experimental results tend to be lower than those achieved in the field of image and speech recognition. This suggests that the imbalance between the volume of EEG data and the complexity of the model does not yield optimal performance in accurately identifying each emotion category. We will solve this problem by extending the EEG data to ensure a balanced database. Along with this, we will aim to adjust and improve the loss function to identify class samples by adding penalty coefficients to the misclassified class samples.
Third, the EEG signals of DEAP and MAHNOB-HCI were recorded using 32 electrodes according to the international 10–20 system, whereas 62 electrodes were used to record the EEG of the SEED. Because the EEG equipment requires a strict laboratory environment, it is difficult to apply EEG-based emotion analysis in practice. Recently, to overcome this problem, portable and reliable EEG devices have been developed, which increases the likelihood of conducting large-scale experiments in real-world applications. Because the number of electrodes in portable EEG devices is relatively small, we intend to contribute to portable EEG devices that can be used in daily life by expanding our future work to explore the location of electrodes to efficiently recognize emotions using our model.
Fourth, only a few experiments were conducted for two, three, and four emotional classes in this study. Therefore, various emotional classes must be captured by introducing additional emotional classes. Thus far, we can add an emotion class by extending Russell’s emotional classification axis from two dimensions to multiple dimensions and extend a more granular emotion class using a classification method with a hierarchical structure. Furthermore, an emotional continuum representing a continuous emotional state can be generated, and various emotional classes can be defined and applied based on this continuum. In future studies, we will carefully select these principles and apply them to regression methods that can learn the relationship between the input variables and the emotional levels necessary for recognizing various emotional classes.
Fifth, in this study, brain region was divided into nine regions and applied to emotion recognition. Dividing the brain region into nine regions reduces the number of electrodes assigned to each region. As a result, the signal of each region may be more concise, but detailed information may be omitted. However, it has the advantage of being able to investigate the relationship between the region and emotions in more detail. On the other hand, if the brain region is divided into four main regions, more electrodes are assigned to each region. This makes it possible to analyze signals in each area in more detail, but it can complicate the analysis task. However, it is useful understand the possibility of generalization of what type of emotion each major area is associated with. For this reason, we will apply the proposed model to study which brain regional division method is most effective for emotion recognition.

Author Contributions

Conceptualization, H.-G.K. and D.-K.J.; Methodology, D.-K.J. and H.-G.K.; Software, D.-K.J. and H.-G.K.; Investigation, D.-K.J. and J.-Y.K.; Resources, J.-Y.K.; Data Curation, D.-K.J.; Writing—Original Draft Preparation, D.-K.J. and H.-G.K.; Writing—Review and Editing, J.-Y.K.; Visualization, D.-K.J.; Project Administration, H.-G.K.; Funding Acquisition, H.-G.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea(NRF) grant funded by the Korea government(MSIT) (No. NRF-2023R1A2C1006756).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The DEAP dataset can be found at https://www.eecs.qmul.ac.uk/mmv/datasets/deap/ (accessed on 26 September 2022). The MAHNOB-HCI dataset is available online at https://mahnob-db.eu/hci-tagging/ (accessed on 7 July 2023). The SEED dataset is available at https://bcmi.sjtu.edu.cn/home/seed/ (accessed on 21 December 2022).

Acknowledgments

This work was supported in part by the Research Grant of Kwangwoon University in 2023.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Powers, J.P.; LaBar, K.S. Regulating emotion through distancing: A taxonomy, neurocognitive model, and supporting meta-analysis. Neurosci. Biobehav. Rev. 2019, 96, 155–173. [Google Scholar] [CrossRef] [PubMed]
  2. Nayak, S.; Nagesh, B.; Routray, A.; Sarma, M. A human–computer interaction framework for emotion recognition through time-series thermal video sequences. Comput. Electr. Eng. 2021, 93, 107280. [Google Scholar] [CrossRef]
  3. Fei, Z.; Yang, E.; Li, D.D.U.; Butler, S.; Ijomah, W.; Li, X.; Zhou, H. Deep convolution network based emotion analysis towards mental health care. Neurocomputing 2020, 388, 212–227. [Google Scholar] [CrossRef]
  4. McDuff, D.; El Kaliouby, R.; Cohn, J.F.; Picard, R.W. Predicting ad liking and purchase intent: Large-scale analysis of facial responses to ads. IEEE Trans. Affect. Comput. 2014, 6, 223–235. [Google Scholar] [CrossRef]
  5. Picard, R.W. Affective computing: Challenges. Int. J. Hum. Comput. 2003, 59, 55–64. [Google Scholar] [CrossRef]
  6. Tian, W. Personalized emotion recognition and emotion prediction system based on cloud computing. Math. Probl. Eng. 2021, 2021, 9948733. [Google Scholar] [CrossRef]
  7. Schirmer, A.; Adolphs, R. Emotion perception from face, voice, and touch: Comparisons and convergence. Trends Cogn. Sci. 2017, 21, 216–228. [Google Scholar] [CrossRef]
  8. Marinoiu, E.; Zanfir, M.; Olaru, V.; Sminchisescu, C. 3D Human Sensing, Action and Emotion Recognition in Robot Assisted Therapy of Children with Autism. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 2158–2167. [Google Scholar] [CrossRef]
  9. Sönmez, Y.Ü.; Varol, A. A speech emotion recognition model based on multi-level local binary and local ternary patterns. IEEE Access 2020, 8, 190784–190796. [Google Scholar] [CrossRef]
  10. Karnati, M.; Seal, A.; Bhattacharjee, D.; Yazidi, A.; Krejcar, O. Understanding deep learning techniques for recognition of human emotions using facial expressions: A comprehensive survey. IEEE Trans. Instrum. Meas. 2023, 72, 5006631. [Google Scholar] [CrossRef]
  11. Li, L.; Chen, J.H. Emotion recognition using physiological signals. In Proceedings of the International Conference on Artificial Reality and Telexistence (ICAT), Hangzhou, China, 28 November–1 December 2006; pp. 437–446. [Google Scholar] [CrossRef]
  12. Leelaarporn, P.; Wachiraphan, P.; Kaewlee, T.; Udsa, T.; Chaisaen, R.; Choksatchawathi, T.; Laosirirat, R.; Lakhan, P.; Natnithikarat, P.; Thanontip, K.; et al. Sensor-driven achieving of smart living: A review. IEEE Sens. J. 2021, 21, 10369–10391. [Google Scholar] [CrossRef]
  13. Qing, C.; Qiao, R.; Xu, X.; Cheng, Y. Interpretable emotion recognition using EEG signals. IEEE Access 2019, 7, 94160–94170. [Google Scholar] [CrossRef]
  14. Goshvarpour, A.; Abbasi, A.; Goshvarpour, A. An accurate emotion recognition system using ECG and GSR signals and matching pursuit method. Biomed. J. 2017, 40, 355–368. [Google Scholar] [CrossRef] [PubMed]
  15. Zhang, B.; Zhou, W.; Cai, H.; Su, Y.; Wang, J.; Zhang, Z.; Lei, T. Ubiquitous depression detection of sleep physiological data by using combination learning and functional networks. IEEE Access 2020, 8, 94220–94235. [Google Scholar] [CrossRef]
  16. Alarcao, S.M.; Fonseca, M.J. Emotions recognition using EEG signals: A survey. IEEE Trans. Affect. 2017, 10, 374–393. [Google Scholar] [CrossRef]
  17. Gómez, A.; Quintero, L.; López, N.; Castro, J. An approach to emotion recognition in single-channel EEG signals: A mother child interaction. J. Phys. Conf. Ser. 2016, 705, 012051. [Google Scholar] [CrossRef]
  18. Li, P.; Liu, H.; Si, Y.; Li, C.; Li, F.; Zhu, X.; Xu, P. EEG based emotion recognition by combining functional connectivity network and local activations. IEEE Trans. Biomed. Eng. 2019, 66, 2869–2881. [Google Scholar] [CrossRef]
  19. Wang, X.W.; Nie, D.; Lu, B.L. Emotional state classification from EEG data using machine learning approach. Neurocomputing 2014, 129, 94–106. [Google Scholar] [CrossRef]
  20. Gupta, V.; Chopda, M.D.; Pachori, R.B. Cross-subject emotion recognition using flexible analytic wavelet transform from EEG signals. IEEE Sens. J. 2018, 19, 2266–2274. [Google Scholar] [CrossRef]
  21. Galvão, F.; Alarcão, S.M.; Fonseca, M.J. Predicting exact valence and arousal values from EEG. Sensors 2021, 21, 3414. [Google Scholar] [CrossRef]
  22. Seal, A.; Reddy, P.P.N.; Chaithanya, P.; Meghana, A.; Jahnavi, K.; Krejcar, O.; Hudak, R. An EEG database and its initial benchmark emotion classification performance. Comput. Math. Methods Med. 2020, 2020, 8303465. [Google Scholar] [CrossRef]
  23. Li, J.; Zhang, Z.; He, H. Implementation of EEG emotion recognition system based on hierarchical convolutional neural networks. In Proceedings of the Advances in Brain Inspired Cognitive Systems: 8th International Conference (BICS), Beijing, China, 28–30 November 2016; pp. 22–33. [Google Scholar] [CrossRef]
  24. Li, X.; Song, D.; Zhang, P.; Yu, G.; Hou, Y.; Hu, B. Emotion recognition from multi-channel EEG data through convolutional recurrent neural network. In Proceedings of the 2016 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Shenzhen, China, 15–18 December 2016; pp. 352–359. [Google Scholar] [CrossRef]
  25. Chen, J.X.; Jiang, D.M.; Zhang, Y.N. A hierarchical bidirectional GRU model with attention for EEG-based emotion classification. IEEE Access 2019, 7, 118530–118540. [Google Scholar] [CrossRef]
  26. Etkin, A.; Egner, T.; Kalisch, R. Emotional processing in anterior cingulate and medial prefrontal cortex. Trends Cogn. Sci. 2011, 15, 85–93. [Google Scholar] [CrossRef] [PubMed]
  27. Anders, S.; Lotze, M.; Erb, M.; Grodd, W.; Birbaumer, N. Brain activity underlying emotional valence and arousal: A response-related fMRI study. Hum. Brain Mapp. 2004, 23, 200–209. [Google Scholar] [CrossRef] [PubMed]
  28. Heller, W.; Nitscke, J.B. Regional brain activity in emotion: A framework for understanding cognition in depresion. Cogn. Emot. 1997, 11, 637–661. [Google Scholar] [CrossRef]
  29. Davidson, R.J. Affective style, psychopathology, and resilience: Brain mechanisms and plasticity. Am. Psychol. 2000, 55, 1196. [Google Scholar] [CrossRef]
  30. Lindquist, K.A.; Wager, T.D.; Kober, H.; Bliss-Moreau, E.; Barrett, L.F. The brain basis of emotion: A meta-analytic review. Behav. Brain Sci. 2012, 35, 121–143. [Google Scholar] [CrossRef]
  31. Zhang, P.; Min, C.; Zhang, K.; Xue, W.; Chen, J. Hierarchical spatiotemporal electroencephalogram feature learning and emotion recognition with attention-based antagonism neural network. Front. Neurosci. 2021, 15, 738167. [Google Scholar] [CrossRef]
  32. Wang, Z.; Wang, Y.; Hu, C.; Yin, Z.; Song, Y. Transformers for EEG-based emotion recognition: A hierarchical spatial information learning model. IEEE Sens. J. 2022, 22, 4359–4368. [Google Scholar] [CrossRef]
  33. Ribas, G.C. The cerebral sulci and gyri. Neurosurg. Focus 2010, 28, E2. [Google Scholar] [CrossRef]
  34. Cohen, I.; Berdugo, B. Speech enhancement for non-stationary noise environments. Signal Process. 2001, 81, 2403–2418. [Google Scholar] [CrossRef]
  35. Dorran, D. Audio Time-Scale Modification. Ph.D. Thesis, Dublin Institute of Technology, Dublin, Ireland, 2005. [Google Scholar] [CrossRef]
  36. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  37. Koelstra, S.; Muhl, C.; Soleymani, M.; Lee, J.S.; Yazdani, A.; Ebrahimi, T.; Patras, I. DEAP: A database for emotion analysis; using physiological signals. IEEE Trans. Affect. Comput. 2011, 3, 18–31. [Google Scholar] [CrossRef]
  38. Soleymani, M.; Lichtenauer, J.; Pun, T.; Pantic, M. A multimodal database for affect recognition and implicit tagging. IEEE Trans. Affect. Comput. 2011, 3, 42–55. [Google Scholar] [CrossRef]
  39. Zheng, W.L.; Lu, B.L. Investigating critical frequency bands and channels for EEG-based emotion recognition with deep neural networks. IEEE Trans. Auton. Ment. Develop. 2015, 7, 162–175. [Google Scholar] [CrossRef]
  40. Zheng, W.L.; Lu, B.L. Personalizing EEG-based affective models with transfer learning. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence (IJCAI), New York, NY, USA, 9–15 July 2016; pp. 2732–2738. [Google Scholar]
Figure 1. Block diagram of the proposed HSCFLM.
Figure 1. Block diagram of the proposed HSCFLM.
Bioengineering 10 01040 g001
Figure 2. Example of electrode division by brain region (for 32-channel EEG electrode placement based on the international 10–20 system).
Figure 2. Example of electrode division by brain region (for 32-channel EEG electrode placement based on the international 10–20 system).
Bioengineering 10 01040 g002
Figure 3. Overview of the regional brain-level encoder in the HSCFLM.
Figure 3. Overview of the regional brain-level encoder in the HSCFLM.
Bioengineering 10 01040 g003
Figure 4. Structure of the bidirectional gated recurrent unit.
Figure 4. Structure of the bidirectional gated recurrent unit.
Bioengineering 10 01040 g004
Figure 5. Structure of self-attention layer.
Figure 5. Structure of self-attention layer.
Bioengineering 10 01040 g005
Figure 6. Architecture of the global brain-level encoder and classifier in the HSCFLM.
Figure 6. Architecture of the global brain-level encoder and classifier in the HSCFLM.
Bioengineering 10 01040 g006
Table 1. Emotion classes for two-level emotion classification on the DEAP and MAHNOB-HCI datasets.
Table 1. Emotion classes for two-level emotion classification on the DEAP and MAHNOB-HCI datasets.
Rating Values
(RVs)
ValenceArousalDominance
1 R V s 5 LowLowLow
6 R V s 9 HighHighHigh
Table 2. Emotion classes for three-level emotion classification on the DEAP and MAHNOB-HCI datasets.
Table 2. Emotion classes for three-level emotion classification on the DEAP and MAHNOB-HCI datasets.
Rating Values
(RVs)
ValenceArousalDominance
1 R V s 3 NegativeActivatedControlled
4 R V s 6 NeutralModerateModerate
7 R V s 9 PositiveDeactivatedOverpowered
Table 3. Experimental results of subject-independent emotion classification using the DEAP dataset.
Table 3. Experimental results of subject-independent emotion classification using the DEAP dataset.
MethodsTwo-Level CLThree-Level CL
VALARODOMVALARODOM
CNN69.7
(11.82)
66.7
(9.50)
70.1
(11.84)
65.3
(10.09)
64.7
(10.43)
65.9
(8.94)
LSTM75.2
(11.56)
72.3
(10.30)
75.3
(9.11)
71.1
(8.89)
69.7
(11.13)
71.6
(9.46)
BGRU76.8
(11.62)
74.4
(11.45)
77.2
(8.96)
73.2
(9.02)
71.5
(8.60)
73.1
(8.81)
CRNN [24]81.1
(9.21)
78.9
(11.31)
81.4
(11.01)
77.4
(11.04)
76.1
(10.43)
77.4
(10.58)
HA-BGRU [25]83.4
(10.26)
81.5
(10.04)
84.1
(9.42)
80.2
(9.48)
78.8
(10.12)
80.1
(11.34)
R2G HA-BGRU87.6
(9.73)
85.3
(9.46)
88.1
(11.66)
83.9
(10.54)
82.9
(11.14)
84.1
(10.92)
R2G TF-Encoder88.1
(10.05)
86.4
(8.37)
88.4
(9.55)
84.5
(10.95)
83.5
(11.29)
84.5
(11.14)
R2G HSA-BGRU88.4
(8.95)
87.1
(10.79)
89.3
(10.59)
85.1
(9.31)
84.2
(10.43)
85.2
(11.20)
HSCFLM92.1
(9.16)
90.5
(9.81)
92.3
(8.94)
88.5
(8.52)
87.4
(8.35)
88.3
(9.76)
The standard deviation is provided in parentheses. CL, VAL, ARO, and DOM denote classification, valence, arousal, and dominance, respectively.
Table 4. Experimental results of subject-independent emotion classification using the MAHNOB-HCI dataset.
Table 4. Experimental results of subject-independent emotion classification using the MAHNOB-HCI dataset.
MethodsTwo-Level CLThree-Level CL
VALARODOMVALARODOM
HA-BGRU [25]84.6
(10.90)
82.7
(8.25)
84.3
(9.67)
80.2
(8.66)
79.8
(9.60)
80.7
(9.61)
R2G HA-BGRU88.6
(8.57)
87.1
(8.64)
88.3
(9.95)
84.5
(11.30)
84.4
(10.45)
84.8
(9.48)
R2G TF-Encoder89.2
(8.07)
87.8
(10.74)
88.9
(8.86)
85.1
(10.32)
84.9
(11.27)
85.4
(10.90)
R2G HSA-BGRU90.1
(10.22)
88.5
(10.95)
89.5
(9.65)
85.8
(9.89)
85.6
(8.42)
86.1
(11.13)
HSCFLM93.3
(9.74)
91.6
(10.71)
92.8
(8.99)
88.9
(10.62)
89.1
(8.89)
89.4
(10.38)
The standard deviation is provided in parentheses. CL, VAL, ARO, and DOM denote classification, valence, arousal, and dominance, respectively.
Table 5. Subject-independent experimental results in four-level classification using the DEAP, MAHNOB-HCI dataset and three-level classification using the SEED dataset.
Table 5. Subject-independent experimental results in four-level classification using the DEAP, MAHNOB-HCI dataset and three-level classification using the SEED dataset.
MethodsFour-Level CL
(HAHV vs. LAHV vs. HALV vs. LALV)
Three-Level CL
(VAL)
DEAPMAHNOB-HCISEED
HA-BGRU [25]74.9
(10.86)
75.2
(8.76)
81.9
(11.59)
R2G HA-BGRU79.4
(11.03)
79.0
(11.46)
85.8
(12.10)
R2G TF-Encoder78.9
(8.11)
79.8
(11.83)
86.8
(9.89)
R2G HSA-BGRU79.5
(10.16)
80.4
(10.34)
87.3
(9.58)
HSCFLM83.2
(9.04)
83.9
(9.86)
90.9
(10.15)
The standard deviation is provided in parentheses. CL and VAL denote classification and valence, respectively.
Table 6. Configuration of the proposed model.
Table 6. Configuration of the proposed model.
ModuleNumber of Network ParametersNumber of
Additions
Number of
Multiplications
Regional
Brain-Level
Encoding
Temporal Encoding350,976124,48913,568,576
Spatial Encoding527,616249,864627,200
Global Brain-Level
Encoding and Classification
639,744269,0973,805,440
Total1,518,336643,45018,001,216
Table 7. Network parameters, training time, test time per EEG data, network size, and average recognition accuracy of the proposed model.
Table 7. Network parameters, training time, test time per EEG data, network size, and average recognition accuracy of the proposed model.
ModelNetwork
Parameters
Training Time
(h:m:s)
Test Time Per
EEG Data (ms)
Network
Size (MB)
Average
Accuracy (%)
HSCFLM1,518,33602:59:072514.883.2
Table 8. Within-subject experimental results in four-level classification using the DEAP and MAHNOB-HCI datasets and three-level classification using the SEED dataset.
Table 8. Within-subject experimental results in four-level classification using the DEAP and MAHNOB-HCI datasets and three-level classification using the SEED dataset.
MethodsFour-Level CL
(HAHV vs. LAHV vs. HALV vs. LALV)
Three-Level CL
(VAL)
DEAPMAHNOB-HCISEED
ACC (STD)p-ValueACC (STD)p-ValueACC (STD)p-Value
HA-BGRU [25]83.6 (9.85)0.03883.8 (11.56)0.03188.3 (10.06)0.032
R2G HA-BGRU88.1 (11.28)0.01687.7 (10.37)0.01591.9 (12.28)0.024
R2G TF-Encoder87.8 (9.20)0.01788.5 (9.43)0.01892.9 (10.19)0.017
R2G HSA-BGRU88.4 (10.79)0.01589.3 (11.14)0.01293.7 (9.56)0.014
HSCFLM92.4 (9.06) 93.1 (8.95) 97.2 (10.11)
CL, VAL, ACC, and STD denote classification, valence, accuracy, and standard deviation, respectively.
Table 9. Emotion classification experiment results for each brain region using the proposed model.
Table 9. Emotion classification experiment results for each brain region using the proposed model.
Brain RegionsDEAP
Two-Level
CL
MAHNOB-HCI
Two-Level
CL
SEED
Three-Level CL
VALAROVALAROVAL
R1 (Prefrontal)86.886.587.786.585.4
R2 (Frontal)86.786.187.386.485.2
R3 (Left Temporal)86.185.986.585.885.3
R4 (Right Temporal)86.284.386.284.885.7
R5 (Central)85.885.486.985.784.5
R6 (Left Parietal)84.184.884.185.483.6
R7 (Parietal)85.786.285.486.284.5
R8 (Right Parietal)85.384.685.786.884.3
R9 (Occipital)83.683.184.183.582.7
All Regions92.190.593.391.690.9
CL, VAL, and ARO denote classification, valence, and arousal, respectively.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jeong, D.-K.; Kim, H.-G.; Kim, J.-Y. Emotion Recognition Using Hierarchical Spatiotemporal Electroencephalogram Information from Local to Global Brain Regions. Bioengineering 2023, 10, 1040. https://doi.org/10.3390/bioengineering10091040

AMA Style

Jeong D-K, Kim H-G, Kim J-Y. Emotion Recognition Using Hierarchical Spatiotemporal Electroencephalogram Information from Local to Global Brain Regions. Bioengineering. 2023; 10(9):1040. https://doi.org/10.3390/bioengineering10091040

Chicago/Turabian Style

Jeong, Dong-Ki, Hyoung-Gook Kim, and Jin-Young Kim. 2023. "Emotion Recognition Using Hierarchical Spatiotemporal Electroencephalogram Information from Local to Global Brain Regions" Bioengineering 10, no. 9: 1040. https://doi.org/10.3390/bioengineering10091040

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop