Next Article in Journal
Intravenous Injection of GluR2-3Y Inhibits Repeated Morphine-Primed Reinstatement of Drug Seeking in Rats
Previous Article in Journal
Letter Migrations between Words in Reading Aloud Can Result either from an Impairment in Orthographic Input or in Phonological Output
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

eSEE-d: Emotional State Estimation Based on Eye-Tracking Dataset

by
Vasileios Skaramagkas
1,2,*,†,
Emmanouil Ktistakis
1,3,†,
Dimitris Manousos
1,
Eleni Kazantzaki
1,
Nikolaos S. Tachos
4,5,
Evanthia Tripoliti
5,
Dimitrios I. Fotiadis
4,5 and
Manolis Tsiknakis
1,2
1
Institute of Computer Science, Foundation for Research and Technology Hellas (FORTH), GR-700 13 Heraklion, Greece
2
Department of Electrical and Computer Engineering, Hellenic Mediterranean University, GR-710 04 Heraklion, Greece
3
Laboratory of Optics and Vision, School of Medicine, University of Crete, GR-710 03 Heraklion, Greece
4
Biomedical Research Institute, Foundation for Research and Technology Hellas (FORTH), GR-451 10 Ioannina, Greece
5
Department of Materials Science and Engineering, Unit of Medical Technology and Intelligent Information Systems, University of Ioannina, GR-451 10 Ioannina, Greece
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Brain Sci. 2023, 13(4), 589; https://doi.org/10.3390/brainsci13040589
Submission received: 1 March 2023 / Revised: 22 March 2023 / Accepted: 29 March 2023 / Published: 30 March 2023
(This article belongs to the Section Behavioral Neuroscience)

Abstract

:
Affective state estimation is a research field that has gained increased attention from the research community in the last decade. Two of the main catalysts for this are the advancement in the data analysis using artificial intelligence and the availability of high-quality video. Unfortunately, benchmarks and public datasets are limited, thus making the development of new methodologies and the implementation of comparative studies essential. The current work presents the eSEE-d database, which is a resource to be used for emotional State Estimation based on Eye-tracking data. Eye movements of 48 participants were recorded as they watched 10 emotion-evoking videos, each of them followed by a neutral video. Participants rated four emotions (tenderness, anger, disgust, sadness) on a scale from 0 to 10, which was later translated in terms of emotional arousal and valence levels. Furthermore, each participant filled three self-assessment questionnaires. An extensive analysis of the participants’ answers to the questionnaires’ self-assessment scores as well as their ratings during the experiments is presented. Moreover, eye and gaze features were extracted from the low-level eye-recorded metrics, and their correlations with the participants’ ratings are investigated. Finally, we take on the challenge to classify arousal and valence levels based solely on eye and gaze features, leading to promising results. In particular, the Deep Multilayer Perceptron (DMLP) network we developed achieved an accuracy of 92% in distinguishing positive valence from non-positive and 81% in distinguishing low arousal from medium arousal. The dataset is made publicly available.

1. Introduction

Emotions are psychological, cognitive and behavioral states associated with feelings and thoughts. In the literature, we can find various types of models for the quantification of emotions from classification using the basic emotions such as [1] to coding models based on the facial movements such as the Facial Action Coding System (FACS) [2] and dimensional models. Among the various ways proposed to categorize emotions, researchers have mainly focused on dimensional scales of emotions [3]. Russell’s Circumplex Model of Affect [4] is a two-dimensional space with emotional arousal (EA) and emotional valence (EV) being the two dimensions. EA describes how calming or exciting an emotion is, while EV is the level of pleasantness. The idea behind this two-dimensional space is that emotions are not discrete states but there are correlations, positive or negative, between them. For example, anger is conceptualized as an emotional state with high arousal and negative valence while tenderness is mapped as neutral to low arousal and positive valence. This approach by Russel is significant because it provides a more complex method for classifying emotions than just categorizing them as positive or negative. By incorporating both EA and EV, the model enables the proper description and comprehension of a greater spectrum of emotions.
Recognizing emotions is crucial for interpreting human actions [5]; therefore, the attempt to automatically identify human emotions has led to a whole new, interdisciplinary research area: that of affective computing, which provides new types of human–computer interaction [3,6].
Several types of input information have been used to classify emotional states, such as: facial expressions, voice signals, gestures, physiological responses, heart rate and brain signals [5]. The most often utilized approach is the combination of physiological signals and eye-related measures [7]. At present, non-contact devices, which use infrared cameras, are most commonly used to track ocular movements [8,9,10]. Few researchers, however, have utilized ocular features as the only predictor of emotional arousal and valence levels. These studies attempt to solve either binary [11,12] or multi-class classification problems [13,14,15,16] with success rates for multi-class cases remaining below 80%, whereas binary classification approaches have proven to be more effective, with success rates reaching as high as 93%. Furthermore, neural network-based approaches have been used for the discrimination of various emotional levels. For example, researchers in [5] present a decision tree combined with neural networks for emotional state recognition based on the measures of pupil diameter and gaze positions when viewing emotion-evoking images. Emotion-eliciting images were used as stimulus in [17] toward the development of an emotion recognition system. The findings of prior research revealed that it is possible to recognize a certain emotional state with up to 90% accuracy.
This study has two primary objectives: specifically, (a) to produce a public database with eye-tracking data obtained from emotion-evoking stimuli under a well-defined experimental protocol, which is significantly larger both in terms of participants and raw eye and gaze data distributed than existing datasets, enabling the use of modern AI-based computational methods for the estimation of models that can predict human emotion, and (b) its use in estimating a computationally efficient small feature set which can predict human emotion using a small number of crucial eye-tracking parameters. On the basis of a meta-analysis of previous studies [18], we anticipate that pupil diameter, blink frequency, and fixation time will increase as emotional arousal rises and will therefore be the gaze-retrieved aspects that best predict emotional arousal. Similarly, according to [19,20], pupil diameter and fixation duration are expected to increase in proportion to the negative valence.
A dataset for emotional State Estimation based on Eye-tracking (eSEE-d) combined with self-assessment emotion evaluation is our contribution to the area. The collection includes the eye-tracking data of 48 individuals and their assessments of emotionally intense video clips. The recorded signals contain numerous metrics pertaining to gaze positions, blinks, and pupil features, enabling the extraction and analysis of a large array of eye attributes, including fixations and saccades. In eSEE-d, individuals’ personalities were profiled using a series of self-assessment questionnaires that assess depressive symptomatology, state and trait anxiety, and empathy (see Section 3 for details). Subsequently, each participant engaged in an experiment in which a series of emotion-eliciting video excerpts from popular films where presented to all participants.
The structure of the present manuscript is as follows. Initially, in Section 2, eye features involved in emotional processes are reported, databases containing eye-tracking data for emotion recognition are reported, and their content and limitations are examined. Section 3 presents the experimental scenarios, stimuli selection, annotation explanation and equipment utilized, and an overview of the experimental setup and the methods employed for the assessment of affect and personality traits is outlined. In Section 3.7, the low-level eye metrics analysis and algorithmic process for the extraction of eye and gaze features is explained. Afterwards, in Section 4, statistical correlations between personality and eye features with affective responses are investigated in detail. Additionally, the strategy and structure of a machine learning approach for the identification of emotional states based on neural networks is described and the results from the methods developed for arousal and valence recognition are presented. Furthermore, in Section 5, we interpret our results regarding the statistical and machine learning analysis and perform a benchmarking upon our findings. Finally, the conclusions drawn from the study are recapitulated and discussed.

2. Eye-Tracking Databases for Emotion Recognition

Interest in eye movement research dates back to the nineteenth century. Eye-tracking research is applied in areas such as cognitive psychology, neuropsychology, usability testing, or marketing [21]. One of the primary goals of machine learning in affective computing is the ability to recognize users’ emotions for the purpose of emotion engineering. Methods based on electroencephalography (EEG), face image processing, and voice analysis are among the most prevalent techniques for emotion identification. Even while eye-tracking is rapidly becoming one of the most widely used sensor modalities in affective computing, it is still a relatively new method for emotion detection, particularly when employed exclusively.
The establishment of emotional databases that can be linked to multiple modality signals, stimulus materials, and experimental paradigms is crucial to emotion recognition research. Multiple human senses can be stimulated to develop emotions through the use of audio–visual information employed in multisensory media studies. The examination of facial expressions or neuro-physiological signals has been the primary focus of databases for the research of affect recognition based on visual modalities [22,23,24,25,26]. Yet, despite the fact that eye movements have been shown to be valuable indicators of affective response [18], few researchers have concentrated on the creation of relevant databases. The Eye-Tracking Movie Database (ETMD) [27] is a video-oriented database comprised of 10 participants and annotated with continuous arousal and valence ratings. The twelve (12) movie clips about (3–3.5 min) used as stimuli were collected from the COGNIMUSE database [27] to elicit various levels of arousal and valence based on six basic emotions: namely, happiness, sadness, disgust, fear, surprise, and anger. In addition, the dataset is made available to the public and includes eye-tracking parameters pertaining to gaze and fixation positions as well as pupil size. Nevertheless, the aforementioned database lacks blink-related measurements. Moreover, the video clips comprise random dialogue scenes or sometimes a mix of different scenes from the movie, thus distracting the viewer and weakening their evoked emotions.
The EMOtional attention dataset (EMOd) [28] is a diversified collection of 1019 images that trigger a range of emotions with eye-tracking data taken from 16 people in an effort to investigate the relationship between image sentiment and human attention. In addition, the EMOd contains high-level perceptual qualities, such as elicited emotions, as well as intensive image context labels, including object shapes, object attitudes, and object semantic category. Regarding the eye and gaze tracking metrics, the dataset provides fixation sites, duration, and fixation maps, which are accessible to the public. However, no raw eye and gaze data are provided, thus introducing limitations regarding the management of data for different purposes.
In addition, the datasets 208 NUSEF [29] and CAT2000 [30] are somewhat useful. NUSEF is a collection of 751 emotive photographs, primarily depicting faces, nudity, and human movement. The CAT2000 training set contains 2000 images depicting various settings, including emotive imagery and cartoons. Yet, these two datasets lack emotional content and object labels. In addition, the eye-tracking data acquired from the participants in these two studies is not available to the general public.
Despite their substantial contribution to the scientific community, the databases described have certain limitations. The primary restriction is the limited number of participants and available eye and gaze metrics. This is especially crucial when examining relationships between eye movements and emotions, as some measures, such as blinks, are useful indications of emotional arousal [18]. A second limitation of the datasets mentioned is the fact that they are appropriate for studying the influence of emotional charge on the shifting of attention specifically to the visual stimuli based on fixation saliency maps, but they are not appropriate for studying the correlation of emotional states with gaze patterns and pupil characteristics. Finally, the datasets suffer from a lack of available eye and gaze metrics, which in turn restrains the variety of computational and algorithmic apporaches that can be performed.
The limitations of each of the aforementioned datasets demonstrate the need for the development of a new eye-tracking dataset intended for emotion recognition. To this goal, the eSEE-d database presented in this work is a one-of-a-kind resource that can support new eye-tracking analysis for emotion identification research. It is, to the best of our knowledge, the first publicly accessible eye-tracking based dataset that integrates eye and gaze movements signals with self-assessment of the users while viewing 10 emotion-eliciting video clips (duration: 1–2 min.), enabling the evaluation of the effect and relationship of eye movements with emotion and personality. eSEE-d is also the largest eye-tracking database in terms of the number of participants and the quantity of eye and gaze measurements, thus providing the opportunity for many different types of experimentation both in terms of data management and the development of a variety of machine and deep learning techniques.

3. Methods

In this section, we detail the methodology used to generate the dataset and the materials utilized in the research.

3.1. Participants

The experimental protocol (110/12-02-2021) was submitted and approved by the Ethical Committee of the Foundation for Research and Technology Hellas (FORTH).
There were a total of fifty-six (56) participants in the study. Seven were eliminated because they did not fulfill the inclusion criteria: six had CES-D scores above the threshold and one had binocular visual acuity worse than 0.10 logMAR. One participant was disqualified due to poor quality recordings. All analyses were conducted on the remaining forty-eight (48) subjects (27 female, 21 male). Their average binocular visual acuity at 80 cm was −0.11 ± 0.08 logMAR (range: 0.10 to (−0.29) logMAR), their mean age was 32 ± 8 years (range: 18–47 years), and their average education level was 17 ± 2 years (range: 12–21 years).
Exclusion criteria for all subjects included any known ocular disease, spectacle-corrected binocular visual acuity in 80 cm of less than 0.10 logMAR (0.8 decimal acuity equivalent), clinically significant aberrant phorias, any known cardiovascular disease, and a CES-D score of 19 or higher (see Section 3.4).

3.2. Video Set

In this work, we decided to use videos as emotion-evoking stimuli. Although picture-based elicitation approaches have certain advantages, such as being simple to acquire materials for and constructing the experimental paradigm, they are incapable of providing adequate continuous stimulation, resulting in suboptimal emotion induction [25]. In contrast, a video-based elicitation method can evoke emotion continuously due to prolonged stimulus. Hence, the video-based elicitation method can compensate for the loss of picture stimulation, which has gained considerable attention [31,32].
Ten (10) videos with sound were used in the study, which were obtained from the public database FilmStim [33] and modified to meet the needs of the study. They were cropped so that their duration was shorter than 2 minutes, and no important dialogues were included, since the participants were native Greek speakers and the videos were in English or French.
The dataset consisted of two videos for each of the four chosen emotions (anger, disgust, sadness and tenderness) and two more videos which served as emotionally neutral ones. Table 1 shows the videos used, their duration and their emotion annotation. According to Russell’s arousal–valence space (Figure 1), each emotion corresponds to its level of arousal and valence. Anger and disgust are High Arousal–Negative Valence (HANV), sadness is Low Arousal–Negative Valence (LANV) and tenderness is Low Arousal–Positive Valence (LAPV). Neutral is Medium Arousal–Medium Valence (MAMV). The first three emotions were chosen because they are commonly considered to be amongst the basic emotions [34], they are widely studied in emotion research [35] and their essence is easy to understand [33]. Tenderness, on the other hand, is not considered to be one of the basic emotions, but it has widely been used during the last years in emotion research [36,37,38]. Tenderness and amusement were the only positive emotions in the FilmStim database instead of the more generic “happiness”. Tenderness is also an attachment-related emotion, and thus, it belongs to another, underrepresented group of emotions. Finally, it is an emotion that can easily be evoked by films [33].
In addition, ten (10) neutral videos were used after each emotion-evoking video, with the objective of inducing a relaxing state prior to the next emotion-evoking video and in parallel to allow for the previously evoked emotion to fade away. The neutral videos had a duration of approximately 60 seconds, which enabled us to sample the induced emotion before a user’s emotional response returns to zero or to a baseline level for the first time (see Section 1 for details).

3.3. Emotion Annotations

The videos used were already annotated regarding the emotion that they evoke by [33]. The emotions were corresponded to the emotional arousal and valence levels of the valence–arousal space by the study group. This annotation will be referred to as “Objective annotation” from now on.
On the other hand, the “Subjective annotation” was based on each participant’s self-assessment after each video. The self-assessment consisted of a 4-word differential emotions scale (DES)—anger, disgust, sadness and tenderness—on an 11-point scale.
Similar to [39], only self-assessment scores equal to or higher than 4 were accepted as a significant indication of the presence of a specific emotion. A self-assessment score lower than 4 was treated as emotionally neutral. An emotion was selected as “Subjective annotation” if it received a higher rating (at least 1 point) than the other three emotions [37].

3.4. Demographics and Psychoemotional Scales

All participants were asked to fill the following set of questionnaires in this order:
  • Clinical cardiovascular record: a questionnaire that consisted of two questions: whether the participant had a cardiovascular record and whether he or she is taking any related medication.
  • CES-D scale: a short self-report scale designed to measure depressive symptomatology [40,41]. It consists of 20 items that request the participant rate how often over the past week they experienced symptoms associated with depression, such as restless sleep, poor appetite, and feeling lonely. A score greater than 19 was used as the cutoff score in order to identify individuals at risk for clinical depression.
  • Demographics: questionnaire with basic demographic information (gender, age, education level).
  • STAI-trait test: the State-Trait Anxiety Inventory (trait) is a commonly used measure of trait anxiety [42,43]. It consists of 20 items that are rated on a 4-point scale. It evaluates the anxiety of the participant during the last six months.
  • Basic Empathy Scale: questionnaire consisting of 20 items measuring emotional and cognitive empathy [44,45].
  • STAI-state test: the State-Trait Anxiety Inventory (state) is a commonly used measure of state anxiety [42,43]. It consists of 20 items that are rated on a 4-point scale. It evaluates the anxiety that the participant feels at the moment of the assessment.

3.5. Materials and Setup

The videos were presented on a computer screen (DELL, 24″, 1280 × 720) at 80 cm distance from the participant as shown in Figure 2. Wireless headphones were used, and the sound volume was set. However, participants were asked whether they were comfortable with the sound volume, and it was adjusted if necessary.
Using the Pupil Labs “Pupil Core” eye-tracker, eye-tracking measures were captured [46]. The binocular recordings had a sample rate of 240 Hz, an accuracy of 0.60 deg., and a precision of 0.02 deg. To minimize head movements, all measures were taken with the subjects seated in a chair with their heads supported by a chin and head rest.
Using European-wide standardized logMAR charts, standardized logMAR acuity was determined [47]. With the cover test, stereopsis was assessed.
With the room lights on, controlled photopic lighting conditions were created for recording purposes. The corneal illuminance was 400 lux when the screen was off and 450 lux when the screen was blank.

3.6. Experimental Procedure

All participants read and signed an Informative Consent Form before the trial. The participants were then escorted to the laboratory. Subsequently, tests of binocular visual acuity at 80 cm and stereopsis were administered.
Subsequently, participants were asked to complete the above-mentioned questionnaires on the computer screen.
In case there was a cardiovascular record or a CES-D score over 19 (risk for clinical depression), the procedure ended and the participant was excluded. Otherwise, the procedure continued.
Following the questionnaire part, the experimental part commenced. After the gaze-tracker calibration, the guidelines were presented and the participants were informed that they could stop the video if they did not feel comfortable, at any time they wanted, just by pressing the Space button. Next, the emotion-evoking videos were presented in a randomized order. After each emotion-evoking video, participants were presented with a neutral video of about 1 min so the evoked emotion faded away before the next emotion-evoking video. After the neutral video, the questionnaire for the emotion self-assessment was presented to the participants. We purposely decided to introduce a one-minute period before affective annotation from individuals who rated their feelings in response to the previous emotion-evoking clip under a self-assessment protocol targeting to simulate real world scenarios. This time period enables us to sample the emotion before a user’s emotional response returns to zero or to a baseline level for the first time. The design of the study is shown in Figure 3.
For the course of the video-viewing procedure, a member of the research team monitored the gaze-tracker’s output on a second monitor in case of any anomalies in the recordings or the participant needed additional assistance.
To safeguard the participants and the study team from the SARS-CoV-2 pandemic and to prevent the spread of the virus, every precaution was taken.

3.7. Data Analysis Methodology

In this subsection, the algorithm used to calculate and interpret eye and gaze-related features from the raw data captured by the gaze-tracking device is described.

3.7.1. Raw Eye-Tracking Data

The raw gaze points from the pupil core are processed and analyzed for each recording sequence to confirm that participants viewed the entire video scene each time and did not glance away from the screen or even closed their eyes for a period longer than the average blink time to avoid viewing. Specifically, we excluded every recording during which the participant closed his/her eyes for a duration longer than 10% of the total clip duration to avoid watching, as blinking causes around 5–10% data loss during a recording [48]. Additionally, we considered faulty recordings those in which participants’ attention was decoupled from the screen stimuli (mind-wandering phenomenon) for a length greater than 20% of the overall duration of the corresponding video clip [49]. The 20% cutoff was selected based on the evidence provided by a range of studies indicating that mind wandering occurs at least 20% and up to 50% of the time, even during tasks that are not designed to induce it [50,51]. Additionally, if a participant abruptly paused the video clip, the relevant recording was deleted. As a result, the final dataset consisted of 476 recordings that include those obtained for each of the valid emotion-evoking videos watched by the 48 study participants.
The output of the gaze tracker used in this study includes the gaze positions (x, y coordinates), the blink timings (start and end times), and the pupil diameter in millimeters. These metrics comprise various sorts of noise originating from both the eye tracker and the participants, as it is widely known that when collecting gaze-related data, there is typically some noise owing to eye blinking and an inability to capture corneal reflections [52]. Thus, filtering and denoising must be applied to the eye movement data in order to eliminate this undesired variance.
The raw gaze coordinates, which are in the form of normalized pixels, are converted into degrees of visual angle, and the instantaneous sample-to-sample gaze movement between two consecutive gaze points is calculated, leading to the calculation of the angular velocity, given the sampling frequency F s . To reduce velocity noise, we utilized a 5-tap velocity filter provided by [53] after one large velocity peak value (i.e., during a saccadic movement).

3.7.2. Fixation and Saccade Detection

Fixation detection algorithms categorize gaze data according to dispersion, velocity, and acceleration (or combinations thereof) parameters; [53,54]). Fixations and saccades are recognized based on the Velocity-Threshold Identification (I-VT) algorithm proposed by [54] due to its superiority when sample-by-sample comparisons are taken into account [55]. Furthermore, to determine the length of the fixations, we added an extra minimum duration threshold. The steps of the algorithm are as follows:
  • Compute point-to-point velocities for every protocol point.
  • Mark every point below the velocity threshold as a fixation point and every other location as a saccade point.
  • Collapse consecutive fixation points based on fixation time into fixation groups, omitting saccade points.
  • Map each cluster of fixations to the fixation at the centroid of their points.
  • Return fixations.
In the I-VT algorithm, the velocity threshold for saccade detection was set to 45 deg./s, as in [55]. In addition, the minimum fixation duration threshold was determined at 55 ms [56].

3.7.3. Pupil and Blink Detection

The pupil diameter and blink timings determined by the eye tracker contribute to the extraction of additional information linked to the pupil and blink. In the infrared illuminated eye camera frame, the pupil recognition algorithm locates the dark pupil [46]. The algorithm is not affected by corneal reflection and may be used by those who use contact lenses or spectacles. Based on input from user-submitted eye camera footage, the pupil recognition algorithm is constantly being improved. Based on a confidence threshold corresponding to the effective detection of the pupil area, blinks start and end times are determined. In total, 28 eye and gaze features are extracted based on fixation, saccade, blink and pupil characteristics and are presented in Table 2.
  • Fixation duration (median, variation coef., skewness, kurtosis);
  • Fixation frequency;
  • Saccade duration (median, variation coef., skewness, kurtosis);
  • Saccade velocity (mean, variation coef., skewness, kurtosis);
  • Peak saccade velocity (mean, variation coef., skewness, kurtosis);
  • Saccade amplitude (median, variation coef., skewness, kurtosis);
  • Saccade frequency;
  • Blink duration (mean);
  • Blink frequency;
  • Pupil diameter (mean, variation coef., skewness, kurtosis).

3.7.4. Pupil Estimation Implications

The effect of emotional arousal and valence on pupil size is complex due to the fact that pupil diameter and its variation are highly reliant on various factors, including lighting conditions [5,57], the luminance of the movie [14] and the adapting field size [58,59]. There have been attempts to remove the movie luminance effect on pupil diameter [60]. Recent research involves deriving the estimated pupil diameter from the measured pupil diameter using the V component of the HSV color space [14].
In the present work, the experimental setup was designed to minimize this effect as much as feasible. First, the room’s lighting settings were adjusted to be photopic so that any brightness variations in the film would be insignificant. Second, for the same reason, the participant’s distance from the screen was quite large (80 cm). Corneal illuminance during films was measured and varied between 390 and 411 lux, showing little variance among the films.
To determine if the influence of the film’s luminance was sufficiently low, a linear regression analysis was performed between pupil diameter and the V component of the HSV color space for each video. Among the 48 participants, 24 showed very weak correlation (r < 0.20), 20 showed weak correlation (0.20 < r < 0.40) and 4 showed moderate correlation (0.41 < r < 0.47). The levels of correlation were set based on [61].
It was considered that the results of linear regression analysis were satisfying, since no strong correlation between pupil diameter and V component was found in any of the participants. Thus, all analysis on pupil diameter was carried out with that acquired from the gaze tracker pupil diameter.

4. Results

In this section, the study’s findings are provided. First, a statistical study is presented, which is followed by a Machine Learning analysis to detect any correlation between ocular characteristics and arousal and valence levels.

4.1. Statistical Analysis

4.1.1. Questionnaire Analysis

Descriptive statistics of the five self-assessment scales are presented in Table 3. An independent sample t-test showed only an emotional empathy score was statistically significantly different between men and women (t(47) = 3.538, p = 0.001) with women showing a higher score by 0.58. No correlation was found between any of the scales and age or education level (p > 0.092).

4.1.2. Objective and Subjective Emotion Annotation Analysis

Table 4 shows the Hit Rate and the Mean Rating for each target emotion. Hit Rate is the percentage of the videos for which the participants had indicated that they had felt the target emotion (objective annotation) at least one point more intensely than any of the other three untargeted emotions [39]. Mean Rating is the mean value of the scores for each target emotion. Its scale is from 0 to 10, as decided based on methodological and theoretical criteria [62].
The relation between objective and subjective arousal is presented in Figure 4, where the subjective annotations are displayed with reference to the objective ones. Each pie chart shows the distribution of the subjective annotations corresponding to the target emotion class of the valence–arousal space. For this analysis, the valence–arousal space is divided into four quadrants, and the emotions are placed in the space according to its arousal and valence levels. The quadrants are “High Arousal–Positive Valence (HAPV)”, “High Arousal–Negative Valence (HANV)”, “Low Arousal–Negative Valence (LANV)” and “Low Arousal–Positive Valence (LAPV)”. The origin of the system is “Medium Arousal–Medium Valence”.
The videos with objective annotation HANV, i.e., anger and disgust, are subjectively annotated as: (a) HANV in a percentage of 77%, (b) MAMV, i.e., neutral, in 12% and (c) LANV in 10%. The ones with objective annotation LANV, i.e., sadness, are subjectively annotated as: (a) LANV in 71%, (b) MAMV in 21% and (c) HANV in 7%. The subjective annotation of the videos with objective annotation “tenderness” (LAPV) are annotated as: (a) LAPV in 71% and (b) MAMV in 27%. The neutral videos received a subjective annotation as neutral (MAMV) in a percentage of 85%. These results reveal that in a substantial percentage, the videos did evoke to participants the target emotion.
Based on [39], we evaluated the “Discreteness” of the videos, i.e., whether the rating of the emotion that was identified as “subjective annotation” was statistically significant greater than the ratings of the rest of emotions. We used a t-test to make a pairwise comparisons between the subjective annotation and each of the rest of the emotions. In all comparisons, the rating of the emotion that was identified as “subjective annotation” was statistically significant greater than the ratings of the rest of the emotions (p < 0.001). Table 5 shows the mean ratings of each emotion in each subjective annotation.

4.1.3. Eye Feature Analysis

As there are several samples of the same subject at the same level of arousal or valence, our dataset is comprised of dependent observations. Hence, an analysis based on the assumption of independent observations within each group or between the groups themselves, such as ANOVA, is rejected. In order to discover which ocular characteristics are affected by the arousal and valence levels, a Mixed Linear Model (MLM) evaluation was conducted on each feature individually. Mixed Linear Models are an extension of simple linear models that are used to evaluate both fixed and random effects. MLMs provide more accurate estimates of the effects, better statistical power and non-inflated Type I errors compared to traditional analyses [63]. In the MLM analysis, arousal or valence level was selected as the fixed component, while participant ID was selected as the random factor. In addition, a Bonferroni post hoc test was performed to compare class characteristics..
In Table 6 and Table 7, only the features that were affected in a statistically significant manner by the arousal and valence levels are presented. The Bonferroni post hoc test showed that five features were statistically significantly different between low and medium arousal level, six features were different between low and high, and nine features were different between medium and high. As far as the valence level is concerned, seven features were statistically significantly different between negative and medium valence level, three features were different between negative and positive, and six were different between medium and positive.
In order to evaluate the simultaneous effect of the arousal and valence levels on each eye feature, the subjective emotion class of the valence–arousal space was selected as fixed factor in the Mixed Linear Model (MLM) analysis. Again, a Bonferroni post hoc test for evaluating features among the classes was performed. Table 8 shows the features that were affected in a statistically significant manner by the emotion class. One feature was statistically significantly different between LAPV and HANV, six features were statistically significantly different between MAMV and HANV and between MAMV and LAPV, five were statistically significantly different between LANV and HANV, two were statistically significantly different between LANV and MAMV, and four were statistically significantly different between LANV and LAPV.
Estimated mean values of the Mixed Linear Model and standard errors of all eye features for three classes (low, medium, high) of emotional arousal and three classes of emotional valence (negative, medium, positive) are presented in Table 2.
When the self-assessment scores were added to the Mixed Linear Model, no statistically significant effect of them on any of the eye features was found.

4.2. Feature Selection Process

Even though deep learning-based models enable a feature extraction process, it might be a good idea to remove irrelevant features before training the model. This may reduce memory and time consumption since deep learning procedures usually require a large amount of data. Moreover, feature selection could enhance the ability of the model to learn the most and least significant features and thus exclude it from the future data collection, resulting in improved performance [64].
For the machine learning analysis presented in Section 4.3, we developed a model including specific eye-tracking features for the estimation of the levels of arousal and valence as presented in Section 4.1. To this goal, we created three different feature sets for our predictive models. Specifically, we produced a correlation matrix of the features that were statistically significantly different among the arousal and valence levels and the emotion classes (Table 6 and Table 7), and from the pairs that were highly correlated (r > 0.3), we kept the features that better distinguish among the levels of arousal and valence. Thus, for the estimation of the levels of EA, the feature set includes the metrics of Fixation and Blink frequency, Pupil diameter, Fixation duration kurtosis, Saccade duration variation and PD variation. For the identification of EV levels, our feature set comprises of eight eye-tracking metrics: Fixation and Blink frequency, Saccade amplitude and duration, Pupil diameter, Fixation duration kurtosis, Saccade duration variation and Pupil diameter kurtosis. Finally and for the synchronous estimation of EA and EV levels, the selected feature set includes the metrics of Saccade and Blink frequency, Saccade amplitude, Pupil diameter, Fixation duration kurtosis, Saccade duration variation and Pupil diameter kurtosis.

4.3. Machine Learning Analysis

This section examines the relationships between fixation, saccade, blink, and pupil-related ocular characteristics with arousal and valence levels. Consequently, the EV outcome measure can take on three values: negative, neutral, or positive, while the EA response variables are set to high, medium, and low. Specifically, the divided EA instances low, medium and high are marked as class LA, MA and HA, respectively. In the same manner, negative, neutral and positive EV examples are denoted as classes NV, MV and PV, respectively.
In our work, we used a set of DMLP neural networks written in the Python 3.8 environment for our models and generated a partition with 80% of the samples for training the networks and the remaining samples for testing, according to the Pareto principle [65]. Figure 5 illustrates the basic structure of our neural networks. The first layer contains N ϵ { 6 , 7 , 8 , 28 } neurons based on the size of the data vector per experiment followed by m hidden layers, where m ϵ Z [ 2 , 3 ] . Additionally, each hidden layer contains n neurons, where n ϵ Z [ 8 , 256 ] . All layers are followed by a dropout layer with a 0.25 rate and Rectified Linear Unit (ReLu) as their Activation Function due to its computational efficiency [66]. Moreover, for the last layer, we used the Sigmoid function for the binary classification procedure and the Softmax function for the multi-class problems [67]. The optimizer used to accelerate the learning process is “Adam” with binary cross-entropy as the loss function. Each network is characterized by a single output y.
In this study, several neural network topologies were compared. Several numbers of hidden neurons have been utilized for training and testing in order to discover the optimal network layout for each classification attempt. The network with the least number of hidden neurons and the smallest testing error has been determined to be the optimal one. The resulting neural network with reduced complexity has been trained and evaluated again. We next conducted a 10-fold cross-validation and evaluated the performance of the neural networks using the average f1-score, Area Under Curve (AUC), and accuracy rate of well-classified testing data. The achieved results for the superior neural network architectures are presented in Table 9, Table 10 and Table 11. Table 9 represents the results of identifying various levels of EA, Table 10 presents the results of the discrimination attempt between different EV levels, while Table 11 provide the synchronous identification attempt of arousal and valence levels.
From Table 9, we observe that the selected neural network is capable of discriminating the presence of a high emotional arousal level to an accuracy of 74%. However, without the presence of the medium class, it is more challenging for the neural network to discern between high and low arousal. Such behavior is not observed when HA is separated from MA or when MA is separated from LA. Additionally, especially with reference to the multi-class categorization of the three classes HA, MA, and LA, the network’s percentage of right predictions is maintained at 74% by adding a third hidden layer.
Regarding the objective of identifying and categorizing the different levels of emotional valence, the PV class is sufficiently distinct from the MV and NV classes. Simultaneously, class PV is easily differentiated from NV and MV at a rate of 82% and 90%, respectively. In comparison, NV is deemed to be weakly discriminated from MV. Despite the addition of an extra hidden layer, it appears to affect the success rate of predictions when categorizing the three classes.
An additional approach was examined in order to investigate the potential of creating a combined model to synchronously identify the levels of EA and EV. Based on the annotation given by the participants, we split the response variable into four classes depending on the combined arousal and valence levels. The respective classes, the structure of the network as well as the results are presented in Table 11.
It is worth noting that a different method was explored in relation to the models constructed and discussed in the preceding paragraphs. This strategy entailed the inclusion of the scores from the five questionnaires in the input layer of the networks. We executed the neural networks by adding in the input layer of the network each of the five questionnaires’ scores separately and in aggregate to determine their effect on the models’ performance. The results indicated that adding the scores had no discernible effect on the model’s efficiency, either increasing or decreasing.

5. Discussion

In this work, we presented an eye-tracking dataset for the assessment of emotional arousal and valence levels based simply on eye and gaze characteristics. The dataset consists of eye and gaze-recording signals from 48 participants who viewed 10 emotionally evocative videos. The participants scored each emotional clip according to four basic emotions, which were then translated into arousal and valence levels.
Our statistical research revealed that numerous characteristics can be used to distinguish between arousal and valence levels (Table 6 and Table 7). Based on the current research, we anticipated that pupil diameter, blink frequency, and fixation length would be the characteristics that best predict arousal and valence levels. This hypothesis was validated, although other variables, such as fixation frequency and saccade amplitude, contributed considerably to our prediction model. As hypothesized, our attempt to discriminate emotional states revealed that High Arousal–Negative Valence (HANV) and Low Arousal–Positive Valence (LAPV) could be distinguished with excellent accuracy. In addition to this comparison, additional characteristics such as fixation frequency, fixation duration, fixation duration kurtosis, saccade frequency, saccade amplitude, saccade duration variation and skewness, blink frequency, and pupil diameter can be used to distinguish HANV from other emotional states. Low Arousal–Negative Valence (LANV) can be discriminated from other emotional states by saccade amplitude, saccade velocity skewness and kurtosis, saccade duration, and pupil diameter (Table 8). In general, it appears that pupil diameter, blink frequency, fixation frequency, and saccade amplitude were strongly influenced by arousal, valence, and emotional states. These results demonstrate the ability of the eye-tracking data collected during the designed experimental protocol to indicate the level of EA and EV.
Regarding our neural network analysis, the highest success rate was observed during the binary classification between not positive and positive emotional valence level, achieving 92% accuracy whereas PV is effectively distinguished among the rest of the two levels. However, integrating the “neutral” class proved challenging, resulting in a significant decrease in network performance, especially when differentiating NV and MV. By adding an additional hidden layer to the shallow neural network model, the neural network was still able to reliably predict the three levels of EV despite a decline in performance. In terms of forecasting emotional arousal levels, both binary and multi-class identification tests provided positive results, with up to 81% accurate predictions; nevertheless, the LA class had a significant impact on the networks’ performance. Positive results demonstrated that a model is able to forecast synchronously the amounts of EA and EV with a 72% success rate.
Overall, we tested a wide variety of eye features related to fixations, saccades, blinks and pupil and investigated the potential of several neural network models on different classification scenarios. In agreement with prior research, our findings reveal the effect of emotional charge on eye movements and pupillary responses. Specifically and in line with the ideas of [11,12,16], high arousal and positive valence levels can be successfully identified based solely on eye-tracking features. Furthermore, when comparing our work to those of [13,14,15], our results demonstrate significant improvements in terms of accuracy which can be attributed in part to the significant larger size of the training database. In addition, we have verified that using neural networks for emotion level estimation produces encouraging results, superior to [5,17], thus indicating potential toward the development of an emotion identification system with high discretization ability.
Although the videos used in this study were obtained from a public database, they are extracted from well-known film productions. Hence, it could be argued that the fact some participants being familiar with a certain video could result in participants not being as emotionally charged as they would have been upon first viewing. One more potential limitation of our study is that the participants rated the videos based on the five eliciting emotions rather than rating directly the levels arousal and valence, which were set later in accordance with the emotion ratings. Moreover, the inclusion of tenderness as a target emotion at the expense of the basic emotion of happiness means that the set did not contain any high intensity positive emotions and thus may result in a relative imbalance. However, tenderness and amusement were the only positive valence emotions in the FilmStim database. Furthermore, it is important to take into consideration that evoking high intensity–positive valence emotions to the subjects within a lab setting, and solely based on movie clip content, is extremely difficult and would probably lead to eliciting neutral valence emotions labeled wrongfully as high valence ones (i.e., happiness) and thus diminishing the integrity of eSEE-d.
We used Hit Rate as a variable that indicates the relation between the objective and the subjective annotation, and we saw that only 53% of the videos that were objectively annotated as anger were also subjectively annotated as anger. Another 20% was annotated as sadness and another 12% was annotated as disgust. This result confirms a rather old finding that anger is difficult to discretely elicit with brief films because it tends to co-exist with other negative emotions [39]. This limitation does not directly affect our further results because all analysis was performed having the subjective annotation as a basis and not the objective one. This co-existence of emotions, however, may have an impact on the subjective annotation as well. The DES self-assessment questionnaire gives the participants the choice to rate more than one emotion, but we defined subjective annotation as the emotion that received a higher rating (at least 1 point) than the other three emotions. In order to evaluate our definition of subjective annotation, we performed pairwise comparisons between the subjective annotation and each of the rest of the emotions. The t-test shows that although there is co-existence of the negative emotions, the rating of the emotion that was set as subjective annotation was statistically significantly greater that the rest of the emotions and confirms that the subjective annotation reflects mainly the corresponding emotion.
The FilmStim database consists of both colored and black and white videos. For the present study, videos from both categories were selected. Pupil size and pupil responses are known to be greatly dependent on luminance contrast, and this is why the study setup was such that luminance effect would be minimized, and this is confirmed in Section 3.7.4. Although there are studies that have shown that pupil size depends on color, apart from luminance [68], it is accepted that pupillary responses are affected mainly by luminance contrast and not color contrast [69]. Thus, since it is shown that luminance effect was not significant in our study, color effect would be even lower. It would be interesting though to have an exact evaluation of this effect in our study.
A final issue that needs to be pointed out is the fact that a there was a 1-min time interval between the emotion-evoking video and the self-assessment questionnaire. This approach was selected to imitate as effectively as feasible a real-world setting in which a person would be unable to appraise their emotions immediately after their manifestation. Emotions are dynamic, complicated processes that develop over time. So, a complete theoretical knowledge of how they operate cannot be attained until their time-related characteristics are well understood [70]. A recent review article [71] presents research on the determinants of emotion duration, building on a previous preliminary review [72]. One central temporal characteristic of emotions is the duration of emotional experience, which has been defined as the amount of time that elapses between the beginning and end point of an emotional episode. In contrast to moods, emotions begin with the occurrence of an external or internal event [73], despite the fact that the beginning of the feeling does not always correspond with the commencement of the event. An emotional episode concludes when the intensity of the emotional response returns to zero or to a baseline level, either for the first time [74], for several consecutive times [75], or permanently [73].
Despite the fact that duration definitions and associated metrics vary across accessible studies, the first definition (i.e., initial return to zero or baseline) is most frequently employed. Regardless of the time conceptualization adopted, it has been discovered that duration is highly diverse, with emotions ranging from a few seconds to several hours or even longer. In this sense, not all feelings are created equal: some emotions have been discovered to last for a long time, while others tend to go away rapidly. However, according to modern neurology, the average duration of an emotion in the human brain is 90 s, and therefore, we assume that the 1-min time does not affect the self-assessment.

6. Conclusions

In this study, we presented a dataset consisting of eye movements recorded while each subject watched emotion-eliciting films and evaluated five emotions, which were then converted into emotional arousal and valence levels. Despite the significant contribution to advancing the relevant research to date of the available public databases, the eSEE-d dataset consists of a significantly larger number of participants, and it includes a plethora of eye and gaze metrics. Thus, it provides the necessary depth and breadth of information for both experimenting with novel machine learning methods in attempting to develop optimal models for predicting emotional state using eye-tracking features.
The results of this study indicate the potential of neural networks for differentiating between different emotional states and emphasize the urgent need for further research. In this approach, we are currently investigating new models as well as the application of advanced deep learning methods in order to develop a model for simultaneously assessing valence and arousal levels with high efficiency. In addition, we intend to experiment with alternate, more robust feature selection methods as opposed to the statistical methods provided in this paper. Deep feature extraction architectures will be utilized for this task, and a different data-handling approach will be tested in order to predict the broad variety of different emotional states. Lastly, we intend to compare our findings with those of other studies that apply multimodal techniques, i.e., employ additional biosignals, and study the necessity and potential of merging eye and gaze data with other biometrics for improved performance in terms of computing expense.
This dataset is made accessible to the public, and we strongly encourage other researchers and academics to test their methods and algorithmic approaches on this immensely challenging database.

Author Contributions

Conceptualization, E.K. (Emmanouil Ktistakis); Data curation, V.S.; Formal analysis, V.S., E.K. (Emmanouil Ktistakis), D.M. and E.T.; Funding acquisition, D.I.F. and M.T.; Investigation, V.S., E.K. (Emmanouil Ktistakis), D.M., E.K. (Eleni Kazantzaki), N.S.T. and E.T.; Methodology, V.S. and E.K. (Emmanouil Ktistakis); Project administration, D.I.F. and M.T.; Resources, V.S. and N.S.T.; Software, V.S. and D.M.; Supervision, M.T.; Validation, V.S., E.K., D.M., E.K. (Emmanouil Ktistakis), N.S.T. and E.T.; Visualization, V.S.; Writing—original draft, V.S. and E.K. (Emmanouil Ktistakis); Writing—review and editing, V.S. and E.K. (Emmanouil Ktistakis). All authors have read and agreed to the published version of the manuscript.

Funding

This project has received funding from the European Union’s Horizon 2020 research and innovation program under grant agreement No 826429 (Project: SeeFar). This paper reflects only the author’s view, and the Commission is not responsible for any use that may be made of the information it contains.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The resulting eSEE-d is available to the academic community (DOI: 10.5281/zenodo.5775674).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
FACSFacial Action Coding System
EAEmotional Arousal
EVEmotional Valence
ETMDEye-Tracking Movie Database
EMOdEMOtional attention dataset
EEGElectroencephalography
FORTHFoundation for Research and Technology Hellas
HANVHigh Arousal–Negative Valence
LANVLow Arousal–Negative Valence
LAPVLow Arousal–Positive Valence
MAMVMedium Arousal–Medium Valence
DESDifferential Emotions Scale
MLMMixed Linear Model

References

  1. Ekman, P.; Friesen, W.V. Constants across cultures in the face and emotion. J. Personal. Soc. Psychol. 1971, 17, 124–129. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Ekman, P.; Friesen, W.V. Facial Action Coding System: A Technique for the Measurement of Facial Movement; Consulting Psychologists Press: Palo Alto, CA, USA, 1978. [Google Scholar]
  3. Toisoul, A.; Kossaifi, J.; Bulat, A.; Tzimiropoulos, G.; Pantic, M. Estimation of continuous valence and arousal levels from faces in naturalistic conditions. Nat. Mach. Intell. 2021, 3, 42–50. [Google Scholar] [CrossRef]
  4. Russell, J.A. A circumplex model of affect. J. Personal. Soc. Psychol. 1980, 39, 1161–1178. [Google Scholar] [CrossRef]
  5. Aracena, C.; Basterrech, S.; Snáel, V.; Velásquez, J.D. Neural Networks for Emotion Recognition Based on Eye Tracking Data. In Proceedings of the 2015 IEEE International Conference on Systems, Man, and Cybernetics, Hong Kong, China, 9–12 October 2015; pp. 2632–2637. [Google Scholar]
  6. Di Luzio, F.; Rosato, A.; Panella, M. A randomized deep neural network for emotion recognition with landmarks detection. Biomed. Signal Process. Control. 2023, 81, 104418. [Google Scholar] [CrossRef]
  7. Zhai, J.; Barreto, A.; Chin, C.; Li, C. Realization of stress detection using psychophysiological signals for improvement of human-computer interactions. In Proceedings of the IEEE SoutheastCon 2005, Ft. Lauderdale, FL, USA, 8–10 April 2005; pp. 415–420. [Google Scholar] [CrossRef]
  8. Lin, Z.; Liu, Y.; Wang, H.; Liu, Z.; Cai, S.; Zheng, Z.; Zhou, Y.; Zhang, X. An eye tracker based on webcam and its preliminary application evaluation in Chinese reading tests. Biomed. Signal Process. Control 2022, 74, 103521. [Google Scholar] [CrossRef]
  9. JothiPrabha, A.; Bhargavi, R.; Deepa Rani, B.V. Prediction of dyslexia severity levels from fixation and saccadic eye movement using machine learning. Biomed. Signal Process. Control 2023, 79, 104094. [Google Scholar] [CrossRef]
  10. Ktistakis, E.; Skaramagkas, V.; Manousos, D.; Tachos, N.S.; Tripoliti, E.; Fotiadis, D.I.; Tsiknakis, M. COLET: A dataset for COgnitive workLoad estimation based on eye-tracking. Comput. Methods Programs Biomed. 2022, 224, 106989. [Google Scholar] [CrossRef]
  11. Alghowinem, S.; Goecke, R.; Wagner, M.; Parker, G.; Breakspear, M. Eye movement analysis for depression detection. In Proceedings of the 2013 IEEE International Conference on Image Processing, Melbourne, VIC, Australia, 15–18 September 2013; pp. 4220–4224. [Google Scholar] [CrossRef]
  12. Al-gawwam, S.; Benaissa, M. Depression Detection From Eye Blink Features. In Proceedings of the 2018 IEEE International Symposium on Signal Processing and Information Technology (ISSPIT), Louisville, KY, USA, 6–8 December 2018; pp. 388–392. [Google Scholar] [CrossRef]
  13. Zheng, L.; Mountstephens, J.; Teo, J. Multiclass Emotion Classification Using Pupil Size in VR: Tuning Support Vector Machines to Improve Performance. J. Physics Conf. Ser. 2020, 1529, 052062. [Google Scholar] [CrossRef]
  14. Tarnowski, P.; Kołodziej, M.; Majkowski, A.; Rak, R.J. Eye-Tracking Analysis for Emotion Recognition. Comput. Intell. Neurosci. 2020, 2020, 2909267. [Google Scholar] [CrossRef]
  15. Zheng, W.L.; Dong, B.N.; Lu, B.L. Multimodal emotion recognition using EEG and eye tracking data. In Proceedings of the 2014 36th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBC 2014, Chicago, IL, USA, 26–30 August 2014; pp. 5040–5043. [Google Scholar] [CrossRef]
  16. Skaramagkas, V.; Ktistakis, E.; Manousos, D.; Tachos, N.S.; Kazantzaki, E.; Tripoliti, E.E.; Fotiadis, D.I.; Tsiknakis, M. A machine learning approach to predict emotional arousal and valence from gaze extracted features. In Proceedings of the 2021 IEEE 21st International Conference on Bioinformatics and Bioengineering (BIBE), Kragujevac, Serbia, 25–27 October 2021; pp. 1–5. [Google Scholar] [CrossRef]
  17. Raudonis, V.; Dervinis, G.; Vilkauskas, A.; Paulauskaite, A.; Kersulyte, G. Evaluation of Human Emotion from Eye Motions. Int. J. Adv. Comput. Sci. Appl. 2013, 4, 79–84. [Google Scholar] [CrossRef] [Green Version]
  18. Skaramagkas, V.; Giannakakis, G.; Ktistakis, E.; Manousos, D.; Karatzanis, I.; Tachos, N.; Tripoliti, E.E.; Marias, K.; Fotiadis, D.I.; Tsiknakis, M. Review of eye tracking metrics involved in emotional and cognitive processes. IEEE Rev. Biomed. Eng. 2021, 16, 260–277. [Google Scholar] [CrossRef]
  19. Kimble, M.O.; Fleming, K.; Bandy, C.; Kim, J.; Zambetti, A. Eye tracking and visual attention to threating stimuli in veterans of the Iraq war. J. Anxiety Disord. 2010, 24, 293–299. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Quigley, L.; Nelson, A.L.; Carriere, J.; Smilek, D.; Purdon, C. The effects of trait and state anxiety on attention to emotional images: An eye-tracking study. Cogn. Emot. 2012, 26, 1390–1411. [Google Scholar] [CrossRef] [PubMed]
  21. Holmqvist, K.; Nystrom, M.; Andersson, R.; Dewhurst, R.; Jarodzka, H.; Weijer, J.V.D. Eye Tracking: A Comprehensive Guide to Methods and Measures; Reprint Edition; Oxford University Press: Oxford, UK, 2015; ISBN 9780198738596. [Google Scholar]
  22. Mollahosseini, A.; Hasani, B.; Mahoor, M.H. AffectNet: A Database for Facial Expression, Valence, and Arousal Computing in the Wild. IEEE Trans. Affect. Comput. 2019, 10, 18–31. [Google Scholar] [CrossRef] [Green Version]
  23. Giannopoulos, P.; Perikos, I.; Hatzilygeroudis, I. Deep Learning Approaches for Facial Emotion Recognition: A Case Study on FER-2013. In Advances in Hybridization of Intelligent Methods: Models, Systems and Applications; Hatzilygeroudis, I., Palade, V., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 1–16. [Google Scholar]
  24. Kosti, R.; Alvarez, J.M.; Recasens, A.; Lapedriza, A. EMOTIC: Emotions in Context Dataset. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 2309–2317. [Google Scholar]
  25. Xue, J.; Wang, J.; Hu, S.; Bi, N.; Lv, Z. Ovpd: Odor-video elicited physiological signal database for emotion recognition. IEEE Trans. Instrum. Meas. 2022, 71, 1–12. [Google Scholar] [CrossRef]
  26. Wu, M.; Teng, W.; Fan, C.; Pei, S.; Li, P.; Lv, Z. An investigation of olfactory-enhanced video on eeg-based emotion recognition. IEEE Trans. Neural Syst. Rehabil. Eng. 2023, 31, 1602–1613. [Google Scholar] [CrossRef]
  27. Zlatintsi, A.; Koutras, P.; Evangelopoulos, G.; Malandrakis, N.; Efthymiou, N.; Pastra, K.; Potamianos, A.; Maragos, P. COGNIMUSE: A multimodal video database annotated with saliency, events, semantics and emotion with application to summarization. EURASIP J. Image Video Process. 2017, 2017. [Google Scholar] [CrossRef] [Green Version]
  28. Fan, S.; Shen, Z.; Jiang, M.; Koenig, B.L.; Xu, J.; Kankanhalli, M.S.; Zhao, Q. Emotional Attention: A Study of Image Sentiment and Visual Attention. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7521–7531. [Google Scholar] [CrossRef]
  29. Ramanathan, S.; Katti, H.; Sebe, N.; Kankanhalli, M.; Chua, T.-S. An Eye Fixation Database for Saliency Detection in Images. In Proceedings of the Computer Vision—ECCV 2010; Daniilidis, K., Maragos, P., Paragios, N., Eds.; Springer: Berlin/Heidelberg, Germany, 2010; pp. 30–43. [Google Scholar]
  30. Borji, A.; Itti, L. CAT2000: A Large Scale Fixation Dataset for Boosting Saliency Research. arXiv 2015, arXiv:1505.03581. [Google Scholar]
  31. Katsigiannis, S.; Ramzan, N. Dreamer: A database for emotion recognition through eeg and ecg signals from wireless low-cost off-the-shelf devices. IEEE J. Biomed. Health Inform. 2018, 22, 98–107. [Google Scholar] [CrossRef] [Green Version]
  32. Song, T.; Zheng, W.; Lu, C.; Zong, Y.; Zhang, X.; Cui, Z. Mped: A multi-modal physiological emotion database for discrete emotion recognition. IEEE Access 2019, 7, 12177–12191. [Google Scholar] [CrossRef]
  33. Schaefer, A.; Nils, F.; Sanchez, X.; Philippot, P. Assessing the effectiveness of a large database of emotion-eliciting films: A new tool for emotion researchers. Cogn. Emot. Cogn. Emot. 2010, 24, 1153–1172. [Google Scholar] [CrossRef]
  34. Johnson-Laird, P.N.; Oatley, K. The language of emotions: An analysis of a semantic field. Cogn. Emot. 1989, 3, 81–123. [Google Scholar] [CrossRef]
  35. Ekman, P. An Argument for Basic Emotions. Cogn. Emot. 1992, 6, 169–200. [Google Scholar] [CrossRef]
  36. Quinto, L.; Thompson, W.F. Composers and performers have different capacities to manipulate arousal and valence. Psychomusicol. Music. Mind Brain 2013, 23, 137–150. [Google Scholar] [CrossRef]
  37. Zhao, G.; Zhang, Y.; Ge, Y. Frontal EEG asymmetry and middle line power difference in discrete emotions. Front. Behav. Neurosci. 2018, 12, 225. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Martínez-Rodrigo, A.; Fernandez-Aguilar, L.; Zangroniz, R.; Latorre, J.M.; Pastor, J.M.; Fernandez-Caballero, A. Film mood induction and emotion classification using physiological signals for health and wellness promotion in older adults living alone. Expert Syst. 2020, 37. [Google Scholar] [CrossRef]
  39. Gross, J.J.; Levenson, R.W. Emotion elicitation using films. Cogn. Emot. 1995, 9, 87–108. [Google Scholar] [CrossRef]
  40. Radloff, L.S. The CES-D Scale: A Self-Report Depression Scale for Research in the General Population. Appl. Psychol. Meas. 1977, 1, 385–401. [Google Scholar] [CrossRef]
  41. Fountoulakis, K.; Iacovides, A.; Kleanthous, S.; Samolis, S.; Kaprinis, S.G.; Sitzoglou, K.; Kaprinis, G.S.; Bech, P. Reliability, Validity and Psychometric Properties of the Greek Translation of the Center for Epidemiological Studies-Depression (CES-D) Scale. BMC Psychiatry 2001, 1, 3. [Google Scholar] [CrossRef] [Green Version]
  42. Spielberger, C.D.; Gorsuch, R.L.; Lushene, R.; Vagg, P.R.; Jacobs, G.A. Manual for the State-Trait Anxiety Inventory; Consulting Psychologists Press: Palo Alto, CA, USA, 1983. [Google Scholar]
  43. Fountoulakis, K.N.; Papadopoulou, M.; Kleanthous, S.; Papadopoulou, A.; Bizeli, V.; Nimatoudis, I.; Iacovides, A.; Kaprinis, G.S. Reliability and psychometric properties of the Greek translation of the State-Trait Anxiety Inventory form Y: Preliminary data. Ann. Gen. Psychiatry 2006, 5, 2. [Google Scholar] [CrossRef] [Green Version]
  44. Jolliffe, D.; Farrington, D.P. Development and validation of the Basic Empathy Scale. J. Adolesc. 2006, 29, 589–611. [Google Scholar] [CrossRef] [PubMed]
  45. Stavrinides, P.; Georgiou, S.; Theofanous, V. Bullying and empathy: A short-term longitudinal investigation. Educ. Psychol. 2010, 30, 793–802. [Google Scholar] [CrossRef]
  46. Kassner, M.; Patera, W.; Bulling, A. Pupil: An Open Source Platform for Pervasive Eye Tracking and Mobile Gaze-Based Interaction. In Proceedings of the 2014 ACM International Joint Conference on Pervasive and Ubiquitous Computing: Adjunct Publication, Seattle, WA, USA, 13–17 September 2014; pp. 1151–1160. [Google Scholar] [CrossRef]
  47. Plainis, S.; Orphanos, Y.; Tsilimbaris, M.K. A Modified ETDRS Visual Acuity Chart for European-Wide Use. Optom Vis Sci. 2007, 84, 647–653. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. How Is the Sample Percentage Calculated in Studio, Lab and Controller? Available online: https://connect.tobii.com/s/article/Sample-percentage-calculated-in-Studio-Lab-and-Controller? (accessed on 28 March 2023).
  49. Hollander, J.; Huette, S. Extracting blinks from continuous eye-tracking data in a mind wandering paradigm. Conscious. Cogn. 2022, 100, 103303. [Google Scholar] [CrossRef]
  50. Faber, M.; Bixler, R.; D’Mello, S.K. An automated behavioral measure of mind wandering during computerized reading. Behav. Res. Methods 2018, 50, 134–150. [Google Scholar] [CrossRef] [Green Version]
  51. Killingsworth, M.A.; Gilbert, D.T. A wandering mind is an unhappy mind. Science 2010, 330, 932. [Google Scholar] [CrossRef] [Green Version]
  52. Chartier, S.; Renaud, P. An Online Noise Filter for Eye-Tracker Data Recorded in a Virtual Environment. In Proceedings of the 2008 Symposium on Eye Tracking Research & Applications; Association for Computing Machinery: New York, NY, USA, 26 March 2008; pp. 153–156. [Google Scholar]
  53. Duchowski, A.T. Eye Movement Analysis. In Eye Tracking Methodology: Theory and Practice; Springer London: London, UK, 2003; pp. 111–128. [Google Scholar]
  54. Salvucci, D.D.; Goldberg, J.H. Identifying Fixations and Saccades in Eye-Tracking Protocols. In Proceedings of the 2000 Symposium on Eye Tracking Research & Applications; Association for Computing Machinery: New York, NY, USA, 8 November 2000; pp. 71–78. [Google Scholar]
  55. Andersson, R.; Larsson, L.; Holmqvist, K.; Stridh, M.; Nyström, M. One algorithm to rule them all? An evaluation and discussion of ten eye movement event-detection algorithms. Behav. Res. Methods 2016, 49, 616–637. [Google Scholar] [CrossRef] [Green Version]
  56. Zaidawi, S.M.K.A.; Prinzler, M.H.U.; Lührs, J.; Maneth, S. An Extensive Study of User Identification via Eye Movements across Multiple Datasets. arXiv 2021, arXiv:2111.05901. [Google Scholar] [CrossRef]
  57. Ellis, C.J. The pupillary light reflex in normal subjects. Br. J. Ophthalmol. 1981, 65, 754–759. [Google Scholar] [CrossRef] [Green Version]
  58. Atchison, D.A.; Girgenti, C.C.; Campbell, G.M.; Dodds, J.P.; Byrnes, T.M.; Zele, A.J. Influence of field size on pupil diameter under photopic and mesopic light levels. Clin. Exp. Optom. 2011, 94, 545–548. [Google Scholar] [CrossRef] [Green Version]
  59. Watson, A.B.; Yellott, J.I. A unified formula for light-adapted pupil size. J. Vis. 2012, 12, 12. [Google Scholar] [CrossRef] [PubMed]
  60. Sun, X.; Hong, T.; Li, C.; Ren, F. Hybrid spatiotemporal models for sentiment classification via galvanic skin response. Neurocomputing 2019, 358, 385–400. [Google Scholar] [CrossRef]
  61. Meghanathan, N. Assortativity Analysis of Real-World Network Graphs based on Centrality Metrics. Comput. Inf. Sci. 2016, 9, 7. [Google Scholar] [CrossRef] [Green Version]
  62. Philippot, P. Inducing and assessing differentiated emotion-feeling states in the laboratory. Cogn. Emot. 1993, 7, 171–193. [Google Scholar] [CrossRef] [PubMed]
  63. Singmann, H.; Kellen, D. An Introduction to Mixed Models for Experimental Psychology. In New Methods in Cognitive Psychology; Routledge: Oxfordshire, UK, 2019; p. 28. ISBN 9780429318405. [Google Scholar]
  64. Chen, Z.; Pang, M.; Zhao, Z.; Li, S.; Miao, R.; Zhang, Y.; Feng, X.; Feng, X.; Zhang, Y.; Duan, M.; et al. Feature selection may improve deep neural networks for the bioinformatics problems. Bioinformatics 2019, 36, 1542–1552. [Google Scholar] [CrossRef] [PubMed]
  65. Gholamy, A.; Kreinovich, V.; Kosheleva, O. Why 70/30 or 80/20 Relation between Training and Testing Sets: A Pedagogical Explanation; ScholarWorks: Allendale Charter Twp, MI, USA, 2018. [Google Scholar]
  66. Nair, V.; Hinton, G.E. Rectified Linear Units Improve Restricted Boltzmann Machines. In Proceedings of the 27th International Conference on International Conference on Machine Learning; Omnipress: Madison, WI, USA, 21 June 2010; pp. 807–814. [Google Scholar]
  67. Narayan, S. The generalized sigmoid activation function: Competitive supervised learning. Inf. Sci. 1997, 99, 69–82. [Google Scholar] [CrossRef]
  68. Bouma, H. Size of the Static Pupil as a Function of Wave-length and Luminosity of the Light Incident on the Human Eye. Nature 1962, 193, 690–691. [Google Scholar] [CrossRef]
  69. Carle, C.F.; James, A.C.; Maddess, T. The Pupillary Response to Color and Luminance Variant Multifocal Stimuli. Investig. Ophthalmol. Vis. Sci. 2013, 54, 467–475. [Google Scholar] [CrossRef] [Green Version]
  70. Davidson, R.J. Affective Style and Affective Disorders: Perspectives from Affective Neuroscience. Cogn. Emot. 1998, 12, 307–330. [Google Scholar] [CrossRef]
  71. Verduyn, P.; Delaveau, P.; Rotge, J.Y.; Fossati, P.; Mechelen, I.V. Determinants of Emotion Duration and Underlying Psychological and Neural Mechanisms. Emot. Rev. 2015, 7, 330–335. [Google Scholar] [CrossRef]
  72. Mechelen, I.V.; Verduyn, P.; Brans, K. The Duration of Emotional Episodes. In Changing Emotions; Psychology Press: London, UK, 2013; ISBN 9780203075630. [Google Scholar]
  73. Frijda, N.H. The laws of emotion. Am. Psychol. 1988, 43, 349–358. [Google Scholar] [CrossRef] [PubMed]
  74. Verduyn, P.; van Mechelen, I.; Tuerlinckx, F. The relation between event processing and the duration of emotional experience. Emotion 2011, 11, 20–28. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  75. Sbarra, D.A. Predicting the onset of emotional recovery following nonmarital relationship dissolution: Survival analyses of sadness and anger. Personal. Soc. Psychol. Bull. 2006, 32, 298–312. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Valence and arousal space with the corresponding location of the emotional classes that were used in the study.
Figure 1. Valence and arousal space with the corresponding location of the emotional classes that were used in the study.
Brainsci 13 00589 g001
Figure 2. Graphical illustration of the experimental apparatus.
Figure 2. Graphical illustration of the experimental apparatus.
Brainsci 13 00589 g002
Figure 3. Design of the experimental study.
Figure 3. Design of the experimental study.
Brainsci 13 00589 g003
Figure 4. Subjective annotation in reference to objective annotation, in the valence–arousal space. HANV: High Arousal–Negative Valence, LANV: Low Arousal–Negative Valence, LAPV: Low Arousal–Positive Valence, MAMV: Medium Arousal–Medium Valence.
Figure 4. Subjective annotation in reference to objective annotation, in the valence–arousal space. HANV: High Arousal–Negative Valence, LANV: Low Arousal–Negative Valence, LAPV: Low Arousal–Positive Valence, MAMV: Medium Arousal–Medium Valence.
Brainsci 13 00589 g004
Figure 5. Architecture of the implemented neural networks.
Figure 5. Architecture of the implemented neural networks.
Brainsci 13 00589 g005
Table 1. Emotion-evoking videos.
Table 1. Emotion-evoking videos.
VideoDuration (min)Emotion
American history X1.17Anger
Benny and Joon2.01Tenderness
Blue (2)0.40Neutral
Ghost1.19Tenderness
Man bites dog (1)1.32Disgust
Schindler’s list (1)0.58Sadness
Schindler’s list (2)1.55Anger
The dreamlife of angels2.00Sadness
The lover0.43Neutral
Trainspotting (2)1.00Disgust
The FilmStim database contains several video clips that belong to the same movie. The digits in parenthesis in the first column refer to the video clip number belonging to the respective movie.
Table 2. Estimated mean values ± standard errors of all features for three classes of EA and EV.
Table 2. Estimated mean values ± standard errors of all features for three classes of EA and EV.
EA EV
LAMAHANVMVPV
FeatureMeanMeanMeanMeanMeanMean
Fix. freq. (fix./s)2.10 ± 0.062.09 ± 0.062.20 ± 0.062.17 ± 0.062.09 ± 0.062.09 ± 0.07
Fix. duration (ms)314.50 ± 8.65306.94 ± 8.90295.52 ± 8.72303.62 ± 8.33307.48 ± 8.89310.83 ± 9.59
Variation0.87 ± 0.020.89 ± 0.020.87 ± 0.020.87 ± 0.020.89 ± 0.020.89 ± 0.02
Skewness2.51 ± 0.082.38 ± 0.082.64 ± 0.082.58 ± 0.072.38 ± 0.082.56 ± 0.10
Kurtosis9.48 ± 0.687.92 ± 0.7310.92 ± 0.7110.35 ± 0.597.91 ± 0.739.59 ± 0.93
Sac. freq. (sac./s)2.48 ± 0.182.70 ± 0.193.05 ± 0.182.83 ± 0.172.68 ± 0.192.54 ± 0.21
Sac. amplitude (deg.)14.01 ± 0.1413.88 ± 0.1414.00 ± 0.1414.08 ± 0.1313.89 ± 0.1413.77 ± 0.15
Variation0.18 ± 0.010.20 ± 0.010.18 ± 0.010.18 ± 0.010.20 ± 0.010.18 ± 0.01
Skewness0.23 ± 0.200.37 ± 0.210.34 ± 0.210.24 ± 0.200.36 ± 0.210.42 ± 0.23
Kurtosis7.44 ± 0.686.59 ± 0.737.85 ± 0.707.60 ± 0.626.58 ± 0.737.76 ± 0.86
Sac. velocity (°/s)223.90 ± 11.07240.77 ± 11.34234.27 ± 11.15231.71 ± 10.76240.63 ± 11.34220.16 ± 12.05
Variation0.78 ± 0.020.74 ± 0.020.74 ± 0.020.76 ± 0.020.74 ± 0.020.76 ± 0.02
Skewness1.35 ± 0.151.08 ± 0.151.20 ± 0.151.22 ± 0.141.08 ± 0.151.47 ± 0.16
Kurtosis2.43 ± 0.811.41 ± 0.852.21 ± 0.822.05 ± 0.761.40 ± 0.853.18 ± 0.95
Peak velocity (°/s)299.83 ± 13.31320.39 ± 13.60310.35 ± 13.39308.00 ± 12.97320.27 ± 13.60295.27 ± 14.38
Variation0.75 ± 0.020.72 ± 0.020.71 ± 0.020.73 ± 0.020.72 ± 0.020.74 ± 0.02
Skewness1.00 ± 0.140.78 ± 0.150.87 ± 0.140.89 ± 0.140.78 ± 0.151.08 ± 0.15
Kurtosis1.10 ± 0.560.59 ± 0.581.03 ± 0.570.92 ± 0.540.58 ± 0.581.54 ± 0.64
Sac. duration (ms)22.03 ± 0.4623.63 ± 0.4722.63 ± 0.4622.46 ± 0.4323.63 ± 0.4721.87 ± 0.52
Variation0.99 ± 0.030.96 ± 0.031.04 ± 0.031.02 ± 0.030.96 ± 0.030.97 ± 0.04
Skewness2.73 ± 0.112.60 ± 0.112.96 ± 0.112.87 ± 0.012.60 ± 0.112.74 ± 0.13
Kurtosis9.87 ± 0.919.55 ± 0.9712.21 ± 0.9311.29 ± 0.839.51 ± 0.9710.05 ± 1.16
Blink freq. (blinks/s)0.21 ± 0.020.27 ± 0.030.22 ± 0.030.23 ± 0.020.27 ± 0.030.19 ± 0.03
Blink duration (ms)215.32 ± 4.06214.88 ± 4.17220.45 ± 4.09218.83 ± 3.91214.79 ± 4.16214.61 ± 4.48
Pupil diameter (mm)3.85 ± 0.103.74 ± 0.103.95 ± 0.103.87 ± 0.103.74 ± 0.104.00 ± 0.11
Variation0.067 ± 0.0030.076 ± 0.0040.076 ± 0.0030.073 ± 0.0030.076 ± 0.0040.067 ± 0.004
Skewness−0.29 ± 0.07−0.28 ± 0.08−0.29 ± 0.07−0.23 ± 0.06−0.27 ± 0.08−0.47 ± 0.10
Kurtosis2.34 ± 0.341.71 ± 0.361.86 ± 0.351.80 ± 0.301.71 ± 0.363.09 ± 0.44
Fixation duration, saccade amplitude and saccade duration are median values. Saccade velocity, peak saccade velocity, blink duration and pupil diameter are mean values.
Table 3. Questionnaire scores.
Table 3. Questionnaire scores.
QuestionnaireMeanSt. dev.MinMax
CES-D10.44.4319
Cognitive empathy3.70.42.95.0
Emotional empathy3.30.61.64.8
State anxiety53.36.23674
Trait anxiety54.96.94270
Table 4. Results for target emotions.
Table 4. Results for target emotions.
Target EmotionHit Rate (%)Mean Rating for Target Emotion
Anger535.9
Disgust797.4
Sadness716.0
Tenderness715.1
Neutral85-
Table 5. Emotion ratings for each subjective emotion annotation.
Table 5. Emotion ratings for each subjective emotion annotation.
AnnotationAngerDisgustSadnessTenderness
Anger7.74.55.70.2
Disgust4.28.44.00.1
Sadness3.83.37.50.1
Tenderness0.40.10.16.4
Table 6. Results of Mixed Linear Model (MLM) and post hoc tests for arousal levels.
Table 6. Results of Mixed Linear Model (MLM) and post hoc tests for arousal levels.
MLMPost Hoc
FeaturepLA-MALA-HAMA-HA
Fixation frequency0.0091.0000.0300.027
Fixation duration0.0050.7440.0030.230
Fixation dur. Sk.0.0430.7000.4780.038
Fixation dur. Kurt.0.0070.2900.3280.005
Saccade frequency0.0010.6040.0000.095
Saccade amp. Var.0.0430.0481.0000.139
Saccade vel. Var0.0140.0650.0231.000
Saccade vel. Sk.0.0450.0430.3670.818
Peak sac. vel. Var.0.0190.1320.0231.000
Saccade dur.0.0000.0000.2570.029
Saccade dur. Var0.0040.5970.0890.004
Saccade dur. Sk.0.0050.8300.0760.005
Saccade dur. Kurt.0.0211.0000.0600.045
Blink frequency0.0010.0011.0000.004
Pupil diameter0.0000.080.0930.000
Pupil diam. Var.0.0080.0420.0141.000
Statistically significant differences (p < 0.05) are in bold. LA: low arousal, MA: medium arousal, HA: high arousal, Sk: skewness, Kurt: kurtosis.
Table 7. Results of Mixed Linear Model (MLM) and post hoc tests for valence levels.
Table 7. Results of Mixed Linear Model (MLM) and post hoc tests for valence levels.
FeaturepNV-MVNV-PVMV-PV
Fixation frequency0.0450.1200.1871.000
Fixation dur. Kurt.0.0200.0151.0000.419
Saccade amplitude0.0000.0470.0010.701
Saccade amp. Var0.0470.0631.0000.128
Saccade vel. Sk0.0110.5350.0650.009
Peak sac.vel. Sk.0.0230.6190.1160.019
Saccade duration0.0000.0040.4000.001
Saccade dur. Var0.0030.0070.0741.000
Saccade dur. Sk.0.0310.0290.7510.950
Blink frequency0.0000.0050.0680.000
Pupil diameter0.0000.0130.0180.000
Pupil diam. Kurt.0.0101.0000.0130.021
Statistically significant differences (p < 0.05) are in bold. NV: negative valence, MV: medium valence, PV: positive valence, Sk: skewness, Kurt: kurtosis.
Table 8. Results of Mixed Linear Model (MLM) and post hoc tests for emotion classes.
Table 8. Results of Mixed Linear Model (MLM) and post hoc tests for emotion classes.
MLMBonferroni post hoc
Featurep123456
Fixation frequency0.0191.0001.0000.4831.0000.0460.086
Fixation duration0.0121.0001.0000.0141.0000.3530.176
Fixation dur. Kurt.0.0190.8191.0001.0001.0000.0101.000
Saccade frequency0.0021.0001.0000.0031.0000.1820.034
Saccade amplitude0.0000.0050.0000.0781.0001.0000.081
Saccade velocity Var0.0170.0521.0000.0221.0001.0000.820
Saccade velocity Sk.0.0210.9180.5591.0000.0121.0000.133
Peak saccade velocity Var.0.0190.0831.0000.0171.0001.0000.907
Peak saccade velocity Sk.0.0410.8491.0001.0000.0301.0000.208
Saccade duration0.0010.0081.0001.0000.0010.0590.451
Saccade duration Var.0.0040.4401.0001.0001.0000.0070.068
Saccade duration Sk.0.0151.0001.0000.2791.0000.0120.496
Blink frequency0.0000.2990.0741.0000.0000.0070.352
Pupil diameter0.0001.0000.0000.0000.0000.0001.000
Pupil diameter Var.0.0250.2121.0000.1020.2971.0000.148
Pupil diameter Sk.0.0370.9610.0220.6800.4701.0000.540
Pupil diameter Kurt.0.0201.0000.0431.0000.0361.0000.064
Statistically significant differences (p < 0.05) are in bold. 1 = LANV-MAMV, 2 = LANV-LAPV, 3 = LANV-HANV, 4 = MAMV-LAPV, 5 = MAMV-HANV, 6 = LAPV-HANV. HANV: High Arousal–Negative Valence, LANV: Low Arousal–Negative Valence, LAPV: Low Arousal–Positive Valence, MAMV: Medium Arousal–Medium Valence. Var: variation Sk: skewness, Kurt: kurtosis.
Table 9. EA classification results.
Table 9. EA classification results.
ClassesHidden Layer SizeAUCf1-ScoreAcc.
LA + MA/HA128, 320.730.740.74
HA/LA128, 640.660.670.67
HA/MA128, 640.740.730.75
MA/LA128, 640.800.790.81
HA/MA/LA256, 64, 320.720.730.74
Table 10. EV classification results.
Table 10. EV classification results.
ClassesHidden Layer SizeAUCf1-ScoreAcc.
NV + MV/PV64, 160.910.910.92
PV/NV32, 80.820.820.82
PV/MV32, 80.870.890.90
MV/NV64, 320.580.570.59
PV/MV/NV128, 64, 320.840.840.85
Table 11. Synchronous EA and EV classification results.
Table 11. Synchronous EA and EV classification results.
ClassesHidden Layer SizeAUCf1-ScoreAcc.
MAMV/HANV/LANV/LAPV128, 64, 160.730.760.72
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Skaramagkas, V.; Ktistakis, E.; Manousos, D.; Kazantzaki, E.; Tachos, N.S.; Tripoliti, E.; Fotiadis, D.I.; Tsiknakis, M. eSEE-d: Emotional State Estimation Based on Eye-Tracking Dataset. Brain Sci. 2023, 13, 589. https://doi.org/10.3390/brainsci13040589

AMA Style

Skaramagkas V, Ktistakis E, Manousos D, Kazantzaki E, Tachos NS, Tripoliti E, Fotiadis DI, Tsiknakis M. eSEE-d: Emotional State Estimation Based on Eye-Tracking Dataset. Brain Sciences. 2023; 13(4):589. https://doi.org/10.3390/brainsci13040589

Chicago/Turabian Style

Skaramagkas, Vasileios, Emmanouil Ktistakis, Dimitris Manousos, Eleni Kazantzaki, Nikolaos S. Tachos, Evanthia Tripoliti, Dimitrios I. Fotiadis, and Manolis Tsiknakis. 2023. "eSEE-d: Emotional State Estimation Based on Eye-Tracking Dataset" Brain Sciences 13, no. 4: 589. https://doi.org/10.3390/brainsci13040589

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop