Next Article in Journal
Automatic Calibration of a Two-Axis Rotary Table for 3D Scanning Purposes
Next Article in Special Issue
Wearable Sensors for Assessing the Role of Olfactory Training on the Autonomic Response to Olfactory Stimulation
Previous Article in Journal
A New Ammonium Smart Sensor with Interference Rejection
Previous Article in Special Issue
Robust T-End Detection via T-End Signal Quality Index and Optimal Shrinkage
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

EEG-Based Estimation on the Reduction of Negative Emotions for Illustrated Surgical Images

1
Division of Software Convergence, Sangmyung University, Seoul 03016, Korea
2
Department of Computer Science, Sangmyung University, Seoul 03016, Korea
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2020, 20(24), 7103; https://doi.org/10.3390/s20247103
Submission received: 3 November 2020 / Revised: 9 December 2020 / Accepted: 9 December 2020 / Published: 11 December 2020
(This article belongs to the Special Issue Intelligent Biosignal Analysis Methods)

Abstract

:
Electroencephalogram (EEG) biosignals are widely used to measure human emotional reactions. The recent progress of deep learning-based classification models has improved the accuracy of emotion recognition in EEG signals. We apply a deep learning-based emotion recognition model from EEG biosignals to prove that illustrated surgical images reduce the negative emotional reactions that the photographic surgical images generate. The strong negative emotional reactions caused by surgical images, which show the internal structure of the human body (including blood, flesh, muscle, fatty tissue, and bone) act as an obstacle in explaining the images to patients or communicating with the images with non-professional people. We claim that the negative emotional reactions generated by illustrated surgical images are less severe than those caused by raw surgical images. To demonstrate the difference in emotional reaction, we produce several illustrated surgical images from photographs and measure the emotional reactions they engender using EEG biosignals; a deep learning-based emotion recognition model is applied to extract emotional reactions. Through this experiment, we show that the negative emotional reactions associated with photographic surgical images are much higher than those caused by illustrated versions of identical images. We further execute a self-assessed user survey to prove that the emotions recognized from EEG signals effectively represent user-annotated emotions.

1. Introduction

From the early days of brain science, many researchers have studied negative human emotions that come from surgical images showing blood, injection, and injury (BII). Usually, the negative emotion caused by surgical images is classified as fear and disgust. Even though emotional responses can be different according to the subject’s experience, knowledge, and personality, the emotional responses are located in close distances when mapped on widely used Russell’s emotional coordinate [1].
Ordinary people rarely have the chance to view surgical images. However, there are certain situations in which people should carefully inspect surgical images. For example, when we or our family members face surgical operation, surgeons explain the process of the operation by showing surgical images. Even though we feel very negative emotions in response to the images, we still have to study them very carefully to understand the procedure. In cases such as this, the negative emotional response to surgical images can prevent people from understanding necessary surgical procedures.
Some studies have tried to reduce impact of the image by simplifying the color of the images through image abstraction schemes [2,3]. They employed existing image abstraction techniques on surgical images to produce illustrated expressions in order to reduce the negative affective responses. Even though many of the existing image abstraction techniques fail to present the details of the surgical images, some surgeons participating in these studies recommend the illustrated surgical images for explaining the images to ordinary people or training students [4,5].
Even though several image abstraction algorithms such as [4,5] have demonstrated effectiveness for abstracting surgical images, it is difficult to retain the diverse and fine details of human organic structures which is especially important for educational purposes while reducing negative emotional response. It would take a lot of cost and effort to develop an algorithm that satisfies the above requirements. Before undertaking such an effort, we need objective evidence that the illustrated surgical images can reduce the severity of negative emotional reactions to photographic surgical images. Unfortunately, the works of Besancon et al.’s [2,3] lack such an evidence, as they did not include quantitative study such as EEG-based emotion measurement on the emotional reactions comparing abstracted surgical images from original images.
We argue that objective and quantitative strategies including biosignal-based methods are needed to confirm the hypothesis that illustrated surgical images can reduce negative emotional reactions. The cost of developing an automatic algorithm that produces illustrative representations of surgical images with preserving sufficient details would be surprisingly expensive; before deciding whether to develop such an algorithm, we need concrete evidence that supports our hypothesis. Among the various kinds of biosignals, the electroencephalogram (EEG) is very widely used to measure emotional responses. Recently, many deep learning-based methods employing EEG have been presented [6,7,8,9,10,11,12,13,14,15,16,17,18].
We employ an EEG-based emotion recognition framework in order to present a quantitative measure of the difference in emotional reactions to illustrated surgical images and photographic surgical images. To this end, we employ professional scientific illustrators to produce illustrations of surgical images, and execute a user study with a deep learning-based emotion recognition model with EEG biosignals. Our study seeks to provide a confidence that the negative emotional reactions caused by surgical images can be reduced through abstraction via illustration.
We employ two groups of participants who view photographic surgical images and illustrated surgical images, respectively. Their emotional responses are measured through an EEG capturing device and processed with a deep learning-based emotion recognition model. To show the effectiveness of the model, we additionally execute a nine-metric user survey for the participants, then compare those results with those of the model. From the results of this experiment, we suggest that the illustrated surgical images successfully reduce the negative emotional responses caused by their photographic counterparts.

2. Related Work

2.1. Emotional Reactions to Surgical Images

Many researchers have studied human emotional reactions on BII (blood, injection, and injury) scenes including body mutilation [19], surgical procedures [20,21], blood drawing [22], open-heart surgery [23], and surgical amputation [24]. Other studies presented human reactions to repelling scenes including homicide scenes [25], spiders [21], vomit [23], and dirty toilets [26]. To estimate human reactions to these scenes, most studies employed either estimation of physical reactions or subjective methods. For estimating physical reactions, they examined heart rate [19,23], facial expression [27], electromyography [23,28], skin conductance [28], neural activation by fMRI [26], eye tracking [29], and visuomotor processing [30]. The subjective methods include user survey [20,21], rate of refusal to watch [23], and experienced vasovagal symptoms [22]. Even though these schemes are used to estimate human reactions to repelling scenes including surgery images, Cisler et al. found that there is no universal scheme to consistently estimate human reactions [31].
Surgical images incur various human reactions including anxiety, fear, disgust, and vicarious pain [31,32]. Among them, fear and disgust are suggested to be the most prominent. Cisler et al. and Olatunji et al. reported disgust as the most representative emotion [31,33]. Their reason is that fear is an emotion of avoiding danger, which is not induced by watching images of bodily injury or surgery. Chapman and Anderson classified blood-injury disgust as a subtype of physical disgust [34]. Olatunji et al. further divided blood-injury disgust into contamination disgust and animal-reminder disgust, where animal-reminder disgust is defined as the reminder of one’s mortality and inherent animal nature [23].

2.2. CNN-Based Emotion Recognition from EEG

From the great success of AlexNet [35] and VGGNet [36] in image classification, deep convolutional neural networks are employed in various EEG analytic tasks.
Tang et al. [6] proposed an early deep CNN-based classification model for single-trial EEG. Their model, which is composed of five layers, is applied to classify motor imagery of left and right hand movement. The model recorded F1 scores of 87.76% 86.64% for classification of motor imagery. While the performance of the model does not show significant improvement over conventional hand-crafted feature-based models, this study has shown the promise of CNN for EEG recognition.
Schirrmeister et al. [7] presented a deep CNN to analyze EEG biosignals. Their model is composed of four blocks, each of which executes convolution and max pooling operations. The first block is distinguished from the others, as it executes spatial filtering between the convolution and max pooling operations. The result of fourth block is linearly classified to four soft max units including left hand, right hand, feet, and rest. They also visualized the features of their CNN to analyze the relationships between the features and EEG signals.
Li et al. [17] presented a hybrid model of convolutional neural network and recurrent neural network to recognize emotions from a multi-channel EEG dataset. As a preprocessing of their data, they employ wavelet and scalogram transform to encapsulate multi-channel neuro-signals into grid-like frames. Their model extracts task-related features and mines correlation between channels, incorporating contextual information. They demonstrate their accuracy by estimating valence and arousal.
Salama et al. [8] presented a 3-dimentional CNN approach for recognizing emotions from multi-channel EEG signals. Their model is constructed using two convolutional layers and two max pooling layers, followed by a fully connected layer. The 3D representation of EEG data is fed into a data augmentation phase, which improves the performance of their 3D CNN model. The simple structure of their model leaves many aspects to be improved. They achieved 87.44% accuracy for valence and 88.49% for arousal.
Moon et al. [9] applied CNN models for EEG-based emotion recognition. They tested three CNN models: CNN-2, CNN-5, and CNN-10, which are distinguished by the number of convolutional layers. The CNN-2 model has one convolutional layer followed by one max pooling layer; CNN-10 has five convolutional layers followed by five max pooling layers. Among the three models, CNN-5, which has three convolutional layers and two max pooling layers, shows the best accuracy for PSD, PCC, and PLV features.
Chiarelli et al. [10] presented a hybrid framework to construct a brain-computer interface using EEG and infrared spectroscopy (fNIRS). They employed DNN, which recorded unprecedented classification outcomes, for their framework. They performed a guided left and right hand motor task on 15 participants. The left-to-right classification accuracy of the DNN was estimated and compared to a stand-alone EEG and fNIRS. The results of their multi-modal recording and DNN classifier significant improvement over the state-of-the-art techniques.
Lawhern et al. [11] presented an EEGNet, a DNN-based approach for constructing brain-computer interface using EEG biosignals. Their EEGNet has three blocks of layers: Conv2D, DepthwiseConv2D, and SeparableConv2D. In this network, they introduce depthwise convolution and separable convolution to process the EEG signals effectively. They tested their method on three datasets: P300 Event-related Potential, Feedback Error-related Negativity, and Movement-related Corticial Potential. The results demonstrate EEGNet’s improved performance over reference algorithms.
Croce et al. [12] applied a CNN model to a large dataset of independent component (IC)s extracted from multi-channel EEG and magnetoencephalographic (MEG) signals. Their aim was to classify brain IC and artifactual IC from the biosignals. The EEG, MEG, and combined EEG + MEG signals were processed through a CNN model to compare its accuracy of classification with state-of-the-art models; their classification accuracies reached 92.4% for EEG, 95.4% for MEG and 95.0% for EEG + MEG.
Yang et al. [13] proposed a CNN-based approach to recognize valence and arousal from unstationary EEG signals. Their model has a multi-column structure of independent modules, each of which was designed using DenseNet [37]. The independent decisions from the modules were merged using a voting strategy to make a final decision. Their model was trained and optimized using DEAP dataset, and applied to distinguish the emotional responses between photographs and artwork images [38] and to verify the influence of contrast on valence [39].

2.3. RNN-Based Emotion Recognition from EEG

The spatio-temporal aspect of EEG biosignals invites the use of a recurrent neural network (RNN), which is known as an effective model for processing time-serial data, to analyze EEG signals.
Khosrowabadi et al. [14] presented a biologically inspired feedforward neural network (ERNN), which has six layers, to recognize human emotions from EEG biosignals. The ERNN model employs a serial-in/parallel-out shift register to simulate the short term memory of emotion. This model with a radial basis function shows very competitive accuracy compared with other feature extraction methods.
Alhagry et al. [16] presented an RNN-based emotion recognition model from EEG biosignals. Their model has two long short term memory (LSTM) layers, one dropout layer and one fully connected layer. Since the EEG biosignal captured from subjects watching a movie clip has a time-serial property, the RNN structure demonstrates competitive accuracy in emotion recognition. Their model showed 85.65% accuracy for arousal and 85.45% for valence.
Soleymani et al. [15] employed an LSTM RNN with conditional random fields to trace the emotions captured from EEG biosignals of subjects watching video. They also captured facial expressions of the subjects. The combination of EEG biosignals and facial expressions was able to provide adequate information for emotion recognition.
Xing et al. [18] presented a framework consisting of a linear EEG mixing model and an LSTM RNN model. For EEG mixing, they employed stack auto encoder (SAE), which is similar to the standard auto encoder; the difference lies in the processing of source signals, which are separated by brain region. The EEG source signals processed by SAE are fed into the LSTM RNN model, which then extracts features from them. This model achieved 81.10% accuracy for valence and 74.38% for arousal.

3. Overview of Our Framework

Our assumption is that the negative emotions caused by photographic surgical images would hinder non-professional people in understanding necessary information conveyed by the images. Since these people are not trained with surgical images, the negative emotions such as fear and disgust arise at their first glance of the images. To reduce negative emotion and to increase the understanding of the information conveyed by surgical images, anatomy textbooks tend to use illustrations, rather than photographs. From this point, we build our assumption that illustrated surgical images evoke less negative emotion from ordinary people than photographic surgical images.
To prove our assumption, we chose 10 photographic surgical images and produced their illustrated versions by hiring professional scientific illustrators. The photographic surgical images and their illustrated counterparts are presented in Figure 1, Figure 2 and Figure 3. Before producing the illustrations, the illustrators were instructed to preserve as much fine details as possible. The target images were collected from various sources and are released under ’fair use’ purpose.
From the ten pairs of photographic and illustrated surgical images, we hired 40 participants to record their emotional reactions to both photographic surgical images and illustrated images. To record their emotional reactions, we employed two different processes. One was the usage of EEG biosignals; the EEG signals captured from participants were processed through a deep multi-channel emotion recognition model [13] and quantized into valence and arousal scores. The other process was a self-assessed user survey. The participants were given a 9-metric survey to record their valence and arousal.
The 40 participants were randomly partitioned into two groups. The first group underwent the above processes for photographic surgical images, while the second worked with illustrated images. The results of these two groups are analyzed and discussed; the outline of this study is illustrated in Figure 4.

4. Deep Emotion Recognition Model

4.1. Structure of the Model

In this section, we describe our deep emotion recognition model, which was presented in our previous study [13]. Our model is based on a multi-column structure: five independent modules process the EEG signal and make estimations of valence and arousal. Results from these modules are ensembled into concerted valence and arousal scores, effectively recognizing emotional responses.

4.2. Dataset Preparation

For the training of our model, we employ the DEAP dataset [40], one of the most widely used EEG datasets. The DEAP dataset consists of preprocessed EEG signals and their corresponding labels which describe emotional states. As instructed by the original authors [40], and similar to our previous study [38], we downsample the EEG signal in DEAP to 128 Hz and process it with a 4.0–45.0 Hz band pass filter. Therefore, we extract 128 × 60 samples from a trial for 40 channels of the dataset. Among the 40 channels, we exclude 8 for normalization and employ 32 channels for the input of our model. For each input channel, we prepare 32 consecutive samples for an input for each module of our model, effectively creating 32 × 32-sized input data. Figure 5 (a) illustrates the sampling process of EEG data, (b) shows the structure of a recognition module, and (c) shows the overall multi-column structure of the model.

4.3. Model Training

The DEAP dataset is constructed using the EEG signals captured from 32 participants. We segment the dataset into three groups: training, validation and test. Out of the 32, EEG signals from 22 participants are used for training, validation and the remaining 5 for test. Each participant executed 40 experiments. Therefore, the numbers of the EEG signal data for training, validation, and test are 880, 200, and 200, respectively.

5. Implementation and Experiment

5.1. Implementation

We implemented our emotion recognition model with Pytorch library on a machine with an Intel Core i7 CPU, 64 GB main memory, and nVidia GTX 2080TI GPU. We employed LiveAmp32 with LiveCap [41], which supports 32 channels following a standard 10/20 system [42].

5.2. Preparation of Surgical Images

While preparing our dataset of surgical images, we surveyed various open emotional image datasets including IAPS [43], GAPED [44], NAPS [45], CAPD [46], SMID [47], ISEE [48] and COMPASS [49]. Some of the datasets are specialized for scary images (SFIP) [50], disgusting images (DIRTI) [51], natural disaster images (NDPS) [52], and adult images (BAPS-adult) [53]. However, there was no existing dataset dedicated to surgical images; therefore, we collected several “fair-use” images from various sources for our experiment.

5.3. Preparing of User Annotation

For user annotation of emotional responses, we presented participants with a nine-point metric separated into valence and arousal. They were asked to mark the metric for the photographic and illustrated surgical images they saw. The leftmost point means very negative reaction, which matches to 1 in EEG-based estimation, and the rightmost point means very positive reaction, which matches to 1. The mid-point signifies a neutral reaction, which matches to 0. The nine-point metric form for user annotation is presented in Figure 6.

5.4. Experiment

For the experiment, we hired 40 participants and separated them into two groups: g r o u p 1 watched photographic surgical images and g r o u p 2 watched the illustrated versions. The characteristics of the two groups are suggested in Table 1. In the case of a participant watching both photographic and illustrated surgical images in a short timeframe, the emotional response from the images they watched first can affect that from the images that they watched later. This was our motivation for employing disjoint groups. We address the issue that personal differences between participants may affect their emotional responses by increasing the number of participants. Before participants watched the images, we explained what they were about to see and allowed withdrawal from the experiment. The participants were asked to watch a 100 s movie clip, wich each image lasting 10 s. In the first round of our experiment, we extracted EEG biosignals from the participants for objective responses. We performed the second round by asking the participants to mark their valence and arousal on a 9-point metric for subjective responses. To avoid the diminishment of emotional reactions in the second round, the participants were instructed to remember their emotions during the first round and mark those. The results of both rounds of the experiment on each group are illustrated in Table 2 and Figure 7. We further visualize the comparison of personal responses for EEG-based and user-annotated emotion in Figure 8 and Figure 9.

6. Analysis

We have two research questions regarding our experiment.
  • RQ1 Are emotional responses from illustrated surgical images discernably less negative than those from photographic surgical images?
  • RQ2 Is our emotional recognition model reliable? In other words, is there sufficient evidence that the emotions recognized by our model resemble self-assessed ones?

6.1. Analysis 1: t-Test

We have set up our null hypothesis for RQ1 as follows:
  • H 0 There is no notable difference between emotional responses from photographic surgical images and illustrated images.
To answer RQ1, a t-test is executed between g r o u p 1 who watch photographic surgical images and g r o u p 2 who watch illustrated surgical images. The p values of this t-test are presented in Table 3. According to the very small p values in Table 3, we can reject H 0 in favor of the alternative hypothesis.
Another t-test regarding the emotions estimated by EEG and the emotions annotated by users, leads us to answer RQ2. RQ2’s null hypothesis is:
  • H 1 There is no notable difference between EEG-based assessment and user-annotated approach.
The p values for this second t-test are presented in Table 4. For g r o u p 1 , who watched photographic surgical images, the valence and arousal estimated from EEG biosignals and annotated by users are very closely related; the strong negative emotion recognized from photographic surgical images is consistent regardless of the recognition scheme. However, for g r o u p 2 , who watched illustrated surgical images, the valence is closely related, while the arousal is not. Therefore, we cannot reject H 1 except in case of arousal. We assume that pictures of organs, blood, and flesh effect similar negative emotions in viewers, even though they are reduced by the illustrated representation. Therefore, the different approaches for estimating emotions show consistently similar valence scores; for arousal, however, ordinary people have only rarely seen surgical images even, even in their illustrated form. Users who have watched illustrated surgical images in the first round of the emotion recognition experiment through EEG biosignals may pay less attention in the second stage experiment using user annotation. Therefore, the arousal values for illustrated surgical images in user annotation can be lower than those in EEG-based estimation.

6.2. Analysis 2: Effect Size

We estimate the effect size by calculating Cohen’s d values for the pairs of the emotions. The formula for Cohen’s d (X, Y) is suggested as follows:
d ( X , Y ) = E x p ( X ) E x p ( Y ) S D p o o l e d ,
where E x p ( X ) and E x p ( Y ) are the mean values of the distributions X and Y, respectively, and S D p o o l e d is the pooled standard deviation of X and Y.
We estimate Cohen’s d to measure the difference between the emotional reactions of photographic and illustrated surgical images in the following four combinations in Table 5:
(i)
the valence estimated by EEG between photographic and illustrated surgical images,
(ii)
the arousal estimated by EEG between photographic and illustrated surgical images,
(iii)
the valence estimated by user-annotation between photographic and illustrated surgical images,
(iv)
the arousal estimated by user-annotation between photographic and illustrated surgical images.
The Cohen’s d values for these four matches are greater than 0.8, which denotes that the effect size is very large.
We also estimate Cohen’s d to measure the difference between the emotions estimated by EEG and the emotions estimated by user-annotation in the following four combinations in Table 6:
(i)
the valence for photographic surgical images estimated by EEG biosignal and user annotation,
(ii)
the arousal for photographic surgical images estimated by EEG biosignal and user annotation,
(iii)
the valence for illustrated surgical images estimated by EEG biosignal and user annotation,
(iv)
the arousal for illustrated surgical images estimated by EEG biosignal and user annotation.
The Cohen’s d values for cases (i)∼(iii) are less than 0.23, which denotes that the effect size is small, and the d value for case (iv) implies there is relatively larger effect size. The reason case (iv) has a medium effect size can be described as similar to the reason results from the case (iv) in the prior t-test are more weakly related.

7. Discussion

7.1. Discussion 1: Comparison of Performances

In the relevant literatures, many models, including conventional machine learning techniques or deep learning techniques, have been employed to estimate valence and arousal from EEG signal. According to [13], models using machine learning-based approaches such as SVM or decision tree show 71.66% average accuracy for valence and 69.37% for arousal, while the models using deep learning schemes such as CNN or RNN show 81.4% for valence and 80.5% for arousal. Clearly, emotion recognition schemes based on deep learning techniques outperform those based on conventional machine learning techniques. The accuracy of our model is compared to that of several important existing studies that estimate emotion through valence and arousal, as shown in Table 7.

7.2. Discussion 2: Increase of Valence

Our experiment reveals that the valence estimated from illustrated surgical images is significantly higher than the valence from photographic surgical images. We assume that the unpleasant feelings from photographic blood and flesh are decreased by substituting the color of photographic blood and flesh with a similar color that has higher saturation or intensity. Classic artistic media, such as pencil or watercolor brush, produce similar effects. It is also notable to invoke the conclusion of the work of Yang et al’s work [38] that the artwork images induce higher valence than photographs. Since the illustrated surgical image can be regarded as a kind of artwork, the increased valence for the illustrated surgical image reinforces the conclusion of [38].
The change in valence in our study, however, is greater than that in [38]. We assume that emotional reactions to surgical images generally more negative: fearsome or disgusting. Therefore, the increase of valence between the surgical image and its illustrated version is greater than the increase of valence between a photograph and an artwork image when the image itself is neutral.

7.3. Discussion 3: Decrease of Arousal

Our result shows a decrease in arousal for the illustrated surgical images compared to the photographic surgical images. Since photographic surgical images are distinguished from other images, their engendered arousal is very great. The reason for the photographic surgical images showing higher arousal is reasoned to be that seeing flesh and blood usually occurs only in very frightening or alerting situations. The decrease in arousal for the illustrated images can be explained by the fact that the realistic colors of flesh and blood are converted to less threatening colors frequently seen in animations or cartoons. The simple and friendly color of the illustrated flesh and blood reduces the sense of actual alert or frightening, which results in the decrease of arousal.

7.4. Discussion 4: Evaluation from a Surgeon

We have asked a surgeon to evaluate the illustrated surgical images. The surgeon marked some regions of the photographic and illustrated surgical images and suggested the following opinions:
(1)
Color transform of the illustrated images is reasonable. Replacing vivid colors such as red and violet by less vivid colors can help reducing negative reactions from the people who do not have an experience in the surgical images
(2)
In the illustrated versions, the reflections on the surface of organs are illustrated as a narrow spot with higher brightness (the yellow circles in Figure 10). Since the spots on organs can be from some disease or from the reflection, the reflections should be illustrated in different style.
(3)
The blood vessels, which play important role in many diagnosis cases, are not illustrated in a consistent way. In some figures, they are preserved in a very salient way (the blue circles in Figure 10), and in others, they are omitted (the green circles in Figure 10). Presenting details such as blood vessels should be expressed in a consistent way.
As a conclusion, the surgeon suggested a positive answer for using the illustrated images to reduce negative emotional reactions from ordinary people. However, he suggested several points to improve for educational or professional purposes.

7.5. Discussion 5: Limitations

The limitation of this study is that we have not taken into account the opinions of experts, such as surgeons or pathologists in producing the illustrative surgical images. Surgeons may provide productive insights in the illustration of surgical images, for example, which fine details are important and must be kept. They may also be able to suggest proper colors to replace the original colors of flesh and blood.
This limitation of this study can be addressed in two points. Similar studies [2,3] hired a small group of surgeons to confirm that the illustrated surgical images could be used for communication and education. In another point, we collected a series of opinions from a surgeon for the illustrative surgical images in Section 7.4. The opinions can be employed to give a guidance for developing an automatic algorithm that produces illustrative surgical images.

8. Conclusions and Future Work

In this paper, we produced illustrated surgical images to prove their ability to reduce the negative emotional responses engendered by photographic surgical images. We executed emotion recognition processes on 40 participants to compare their emotional responses to photographic and illustrated surgical images. The emotional responses were estimated in a bi-modal approach: a deep learning-based emotion recognition model from EEG biosignals was combined with a 9-point metric user annotation. From the results, we conclude that illustrated surgical images indeed capable of reducing the negative emotions of participants.
In our future research, we will study relevant methods to create appropriate illustrated images. We will consult with experts including surgeons and pathologists to enrich the illustration schemes on surgical images. We will also examine the illustrated surgical images from experts to improve the quality of the illustrations. These approaches will help developing an automatic algorithm for generating illustrated surgical images that satisfy both ordinary people and experts.

Author Contributions

Conceptualization, H.Y., J.H. and K.M.; methodology, H.Y.; software, J.H.; validation, J.H.; formal analysis, J.H.; investigation, K.M., resources, J.H.; data curation, J.H.; writing–original draft preparation, H.Y.; writing–review and editing, K.M.; visualization, K.M.; supervision, K.M.; project administration, K.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Russell, J. Evidence for a three-factor theory of emotions. J. Res. Personal. 1977, 11, 273–294. [Google Scholar] [CrossRef]
  2. Besancon, L.; Semmo, A.; Biau, D.; Frachet, B.; Pineau, V.; Sariali, E.H.; Taouachi, R.; Isenberg, T.; Dragicevic, P. Reducing affective responses to surgical images and videos through color manipulation and stylization. In Proceedings of the Expressive, Vancouver, BC, Canada, 17–19 August 2018; pp. 11:1–11:13. [Google Scholar]
  3. Besancon, L.; Semmo, A.; Biau, D.; Frachet, B.; Pineau, V.; Sariali, E.H.; Soubeyrand, M.; Taouachi, R.; Isenberg, T.; Dragicevic, P. Reducing affective responses to surgical images and videos through stylization. Comput. Graph. Forum 2020, 39, 462–483. [Google Scholar] [CrossRef]
  4. Kyprianidis, J.; Dollner, J. Image abstraction by structure adaptive filtering. In Proceedings of the EG UK—Theory and Practice of Computer Graphics, Manchester, UK, 9–11 June 2008; pp. 51–58. [Google Scholar]
  5. Kyprianidis, J.; Kang, H. Image and video abstraction by coherence-enhancing filtering. Comput. Graph. Forum 2011, 30, 593–602. [Google Scholar] [CrossRef]
  6. Tang, Z.; Li, C.; Sun, S. Single-trial EEG classification of motor imagery using deep convolutional neural networks. Optik 2017, 130, 11–18. [Google Scholar] [CrossRef]
  7. Schirrmeister, R.T.; Springenberg, J.T.; Fiederer, L.D.J.; Glasstetter, M.; Eggensperger, K.; Tangermann, M.; Hutter, F.; Burgard, W.; Ball, T. Deep learning with convolutional neural networks for EEG decoding and visualization. Hum. Brain Map. 2017, 38, 5391–5420. [Google Scholar] [CrossRef] [Green Version]
  8. Salama, E.S.; El-Khoribi, R.A.; Shoman, M.E.; Shalaby, M.A.E. EEG-based emotion recognition using 3D convolutional neural networks. Int. J. Adv. Comput. Sci. Appl. 2018, 9, 329–337. [Google Scholar] [CrossRef]
  9. Moon, S.-E.; Jang, S.; Lee, J.-S. Convolutional neural network approach for EEG-based emotion recognition using brain connectivity and its spatial information. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Calgary, AB, Canada, 15–20 April 2018; pp. 2556–2560. [Google Scholar]
  10. Chiarelli, A.M.; Croce, P.; Merla, A.; Zappasodi, F. Deep learning for hybrid EEG-fNIRS brain-computer interface: Application to motor imagery classification. J. Neural. Eng. 2018, 15, 036028. [Google Scholar] [CrossRef]
  11. Lawhern, V.J.; Solon, A.J.; Waytowich, N.R.; Gordon, S.M.; Hung, C.P.; Lance, B.J. EEGNet: A compact convolutional neural network for EEG-based brain–computer interfaces. J. Neural. Eng. 2018, 15, 056013. [Google Scholar] [CrossRef] [Green Version]
  12. Croce, P.; Zappasodi, F.; Marzetti, L.; Merla, A.; Pizzella, V.; Chiarelli, A.M. Deep Convolutional Neural Networks for Feature-Less Automatic Classification of Independent Components in Multi-Channel Electrophysiological Brain Recordings. IEEE Trans. Biom. Eng. 2019, 66, 2372–2380. [Google Scholar] [CrossRef]
  13. Yang, H.; Han, J.; Min, K. A Multi-Column CNN Model for Emotion Recognition from EEG Signals. Sensors 2019, 19, 4736. [Google Scholar] [CrossRef] [Green Version]
  14. Khosrowabadi, R.; Chai, Q.; Kai, K.A.; Wahab, A. ERNN: A biologically inspired feedforward neural network to discriminate emotion from EEG signal. IEEE Trans. Neural Netw. Learn. Syst. 2014, 25, 609–620. [Google Scholar] [CrossRef] [PubMed]
  15. Soleymani, M.; Asghari-Esfeden, S.; Fu, Y.; Pantic, M. Analysis of EEG signals and facial expressions for continuous emotion detection. IEEE Trans. Affect. Comput. 2016, 7, 17–28. [Google Scholar] [CrossRef]
  16. Alhagry, S.; Fahmy, A.A.; El-Khoribi, R.A. Emotion recognition based on EEG using LSTM recurrent neural network. Int. J. Adv. Comput. Sci. Appl. 2017, 8, 355–358. [Google Scholar] [CrossRef] [Green Version]
  17. Li, X.; Song, D.; Zhang, P.; Yu, G.; Hou, Y.; Hu, B. Emotion recognition from multi-channel EEG data through convolutional recurrent neural network. In Proceedings of the IEEE International Conference on Bioinformatics and Biomedicine, Kansas City, MO, USA, 13–16 November 2017; pp. 352–359. [Google Scholar]
  18. Xing, X.; Li, Z.; Xu, T.; Shu, L.; Hu, B.; Xu, X. SAE + LSTM: A New framework for emotion recognition from multi-channel EEG. Front. Nuerorobot. 2019, 13, 37. [Google Scholar] [CrossRef]
  19. Klorman, R.; Weissberg, R.; Wiesenfeld, A. Individual differences in fear and autonomic reactions to affective stimulation. Psychophysiology 1977, 14, 45–51. [Google Scholar] [CrossRef]
  20. Sawchuk, C.; Lohr, J.; Westendorf, D.; Meunier, S.; Tolin, D. Emotional responding to fearful and disgusting stimuli in specific phobics. Behav. Res. Ther. 2002, 40, 1031–1046. [Google Scholar] [CrossRef]
  21. Tolin, D.; Lohr, J.; Sawchuk, C.; Lee, T. Disgust and disgust sensitivity in blood-injection-injury and spider phobia. Behav. Res. Ther. 1997, 35, 949–953. [Google Scholar] [CrossRef]
  22. Gilchrist, P.; Ditto, B. The effects of blood-draw and injection stimuli on the vasovagal response. Psychophysiology 2012, 49, 815–820. [Google Scholar] [CrossRef]
  23. Olatunji, B.; Haidt, J.; McKay, D.; David, B. Core, animal reminder, and contamination disgust: Three kinds of disgust with distinct personality, behavioral, physiological, and clinical correlates. J. Res. Personal. 2008, 42, 1243–1259. [Google Scholar] [CrossRef]
  24. Rohrmann, S.; Hopp, H. Cardiovascular indicators of disgust. Int. J. Psychophysiol. 2008, 68, 201–208. [Google Scholar] [CrossRef]
  25. Hare, R.; Wood, K.; Britain, S.; Shadman, J. Autonomic responses to affective visual stimulation. Psychophysiology 1970, 7, 408–417. [Google Scholar] [CrossRef] [PubMed]
  26. Schienle, A.; Stark, R.; Walter, B.; Blecker, C.; Ott, U.; Kirsch, P.; Sammer, G.; Vaitl, D. The insula is not specifically involved in disgust processing: An fMRI study. Neuroreport 2002, 13, 2023–2026. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Lumley, M.; Melamed, B. Blood phobics and nonphobics: Psychological differences and affect during exposure. Behav. Res. Ther. 1992, 30, 425–434. [Google Scholar] [CrossRef]
  28. Lang, P.; Greenwald, M.; Bradley, M.; Hamm, A. Looking at pictures: Affective, facial, visceral, and behavioral reactions. Psychophysiology 1993, 30, 261–273. [Google Scholar] [CrossRef]
  29. Armstrong, T.; Hemminger, A.; Olatunji, B. Attentional bias in injection phobia: Overt components, time course, and relation to behavior. Behav. Res. Ther. 2013, 51, 266–273. [Google Scholar] [CrossRef]
  30. Haberkamp, A.; Schmidt, T. Enhanced visuomotor processing of phobic images in blood-injury-injection fear. J. Anxiety Disord. 2014, 28, 291–300. [Google Scholar] [CrossRef]
  31. Cisler, J.; Olatunji, B.; Lohr, J. Disgust, fear, and the anxiety disorders: A critical review. Clin. Psychol. Rev. 2009, 29, 34–46. [Google Scholar] [CrossRef] [Green Version]
  32. Benuzzi, F.; Lui, F.; Duzzi, D.; Nichelli, P.; Porro, C. Does it look painful or disgusting? Ask your parietal and cingulate cortex. J. Neurosci. 2008, 28, 923–931. [Google Scholar] [CrossRef]
  33. Olatunji, B.; Cisler, J.; McKay, D.; Phillips, M. Is disgust associated with psychopathology? Emerging research in the anxiety disorders. Psychiatry Res. 2010, 175, 1–10. [Google Scholar] [CrossRef]
  34. Chapman, H.; Anderson, A. Understanding disgust. Ann. N. Y. Acad. Sci. 2012, 1251, 62–76. [Google Scholar] [CrossRef]
  35. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  36. Simonyan, K.; Andrew, Z. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  37. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  38. Yang, H.; Han, J.; Min, K. Distinguishing emotional responses to photographs and artwork using a deep learning-based approach. Sensors 2019, 19, 5533. [Google Scholar] [CrossRef] [Green Version]
  39. Yang, H.; Han, J.; Min, K. Emotion variation from controlling contrast of visual contents through EEG-Based deep emotion recognition. Sensors 2020, 20, 4543. [Google Scholar] [CrossRef]
  40. Koelstra, S.; Muhl, C.; Soleymani, M.; Lee, J.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. DEAP: A Database for Emotion Analysis; Using Physiological Signals. IEEE Trans. Affect. Comput. 2012, 3, 18–31. [Google Scholar] [CrossRef] [Green Version]
  41. BCI+: LiveAmp. Compact Wireless Amplifier for Mobile EEG Applications. BCI+ Solutions by Brain Products. Available online: bci.plus/liveamp/ (accessed on 12 December 2019).
  42. Klem, G.H.; Lüders, H.O.; Jasper, H.H.; Elger, C. The ten-twenty electrode system of the International Federation. The International Federation of Clinical Neurophysiology. Electroencephalogr. Clin. Neurophysiol. Suppl. 1999, 52, 3–6. [Google Scholar]
  43. Lang, P.; Bradley, M.; Cuthbert, B. International Affective Picture System (IAPS): Technical Manual and Affective Ratings; Technical Report A-8; University of Florida: Belle Glade, FL, USA, 2008. [Google Scholar]
  44. Dan-Glauser, E.; Scherer, K. The Geneva affective picture database (GAPED): A new 730-picture database focusing on valence and normative significance. Behav. Res. Methods 2011, 43, 468. [Google Scholar] [CrossRef] [PubMed]
  45. Marchewka, A.; Zurawski, L.; Jenorog, K.; Grabowska, A. The Nencki Affective Picture System (NAPS): Introduction to a novel, standardized, wide-range, high-quality, realistic picture database. Behav. Res. Methods 2014, 46, 596–610. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Moyal, N.; Henik, A.; Anholt, G. Categorized Affective Pictures Database (CAP-D). J. Cogn. 2018, 1, 41. [Google Scholar] [CrossRef] [Green Version]
  47. Crone, D.; Bode, S.; Murawski, C.; Laham, S. The Socio-Moral Image Database (SMID): A novel stimulus set for the study of social, moral and affective processes. PLoS ONE 2018, 13, e0190954. [Google Scholar] [CrossRef] [Green Version]
  48. Kim, H.; Lu, X.; Costa, M.; Kandemir, B.; Adams Jr., R.; Li, J.; Wang, J.; Newman, M. Development and validation of Image Stimuli for Emotion Elicitation (ISEE): A novel affective pictorial system with test-retest repeatability. Psychiatry Res. 2018, 261, 414–420. [Google Scholar] [CrossRef]
  49. Weierich, M.; Kleshchova, O.; Reider, J.; Reilly, D. The Complex Affective Scene Set (COMPASS): Solving the Social Content Problem in Affective Visual Stimulus Sets. Collabra Psychol. 2019, 5, 53. [Google Scholar] [CrossRef]
  50. Michalowski, J.; Drozdziel, D.; Matuszewski, J.; Koziejowski, W.; Jednorog, K.; Marchewka, A. The Set of Fear Inducing Pictures (SFIP): Development and validation in fearful and nonfearful individuals. Behav. Res. Methods 2017, 49, 1407–1419. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  51. Haberkamp, A.; Glombiewski, J.; Schmidt, F.; Barke, A. The DIsgust-RelaTed-Images (DIRTI) database: Validation of a novel standardized set of disgust pictures. Behav. Res. Ther. 2017, 89, 86–94. [Google Scholar] [CrossRef] [PubMed]
  52. Merlhiot, G.; Mermillod, M.; Le Pennec, J.; Mondillon, L. Introduction and validation of the Natural Disasters Picture System (NDPS). PLoS ONE 2018, 13, e0201942. [Google Scholar] [CrossRef]
  53. Szymanska, M.; Comte, A.; Tio, G.; Vidal, C.; Monnin, J.; Smith, C.; Nezelof, S.; Vulliez-Coady, L. The Besançon affective picture set-adult (BAPS-Adult): Development and validation. Psychiatry Res. 2019, 271, 31–38. [Google Scholar] [CrossRef]
Figure 1. The first comparison of photographic surgical images and illustrated surgical images.
Figure 1. The first comparison of photographic surgical images and illustrated surgical images.
Sensors 20 07103 g001
Figure 2. The second comparison of photographic surgical images and illustrated surgical images.
Figure 2. The second comparison of photographic surgical images and illustrated surgical images.
Sensors 20 07103 g002
Figure 3. The third comparison of photographic surgical images and illustrated surgical images.
Figure 3. The third comparison of photographic surgical images and illustrated surgical images.
Sensors 20 07103 g003
Figure 4. Overview of our framework.
Figure 4. Overview of our framework.
Sensors 20 07103 g004
Figure 5. Structure and dataflow of our model, which is constructed according to [13].
Figure 5. Structure and dataflow of our model, which is constructed according to [13].
Sensors 20 07103 g005
Figure 6. Nine-point metric for user annotation.
Figure 6. Nine-point metric for user annotation.
Sensors 20 07103 g006
Figure 7. Results plotted in Russell’s emotion circumplex model.
Figure 7. Results plotted in Russell’s emotion circumplex model.
Sensors 20 07103 g007
Figure 8. Results from photographic surgical images. Subjects 01∼20 participated in the experiment for photographic surgical images. The left box, which corresponds to EEG-based recognition has a range of ( 1 , 1 ) × ( 1 , 1 ) , while the right box corresponding to user-annotated emotion has a range of ( 1 , 9 ) × ( 1 , 9 ) . The x-axis of each box represents valence and the y-axis represents arousal.
Figure 8. Results from photographic surgical images. Subjects 01∼20 participated in the experiment for photographic surgical images. The left box, which corresponds to EEG-based recognition has a range of ( 1 , 1 ) × ( 1 , 1 ) , while the right box corresponding to user-annotated emotion has a range of ( 1 , 9 ) × ( 1 , 9 ) . The x-axis of each box represents valence and the y-axis represents arousal.
Sensors 20 07103 g008
Figure 9. Results from illustrated surgical images. Subjects 21∼40 participated in the experiment for illustrated surgical images. The left box, which corresponds to EEG-based recognition has a range of ( 1 , 1 ) × ( 1 , 1 ) , while the right box corresponding to user-annotated emotion has a range of ( 1 , 9 ) × ( 1 , 9 ) . The x-axis of each box represents valence and the y-axis represents arousal.
Figure 9. Results from illustrated surgical images. Subjects 21∼40 participated in the experiment for illustrated surgical images. The left box, which corresponds to EEG-based recognition has a range of ( 1 , 1 ) × ( 1 , 1 ) , while the right box corresponding to user-annotated emotion has a range of ( 1 , 9 ) × ( 1 , 9 ) . The x-axis of each box represents valence and the y-axis represents arousal.
Sensors 20 07103 g009
Figure 10. The marks drawn by a surgeon for the evaluation of the illustrated images.
Figure 10. The marks drawn by a surgeon for the evaluation of the illustrated images.
Sensors 20 07103 g010
Table 1. Gender and age distribution of the participants.
Table 1. Gender and age distribution of the participants.
TotalGenderAge
FemaleMale20 s30 s>40
g r o u p 1 2010101361
g r o u p 2 2010101280
Table 2. Results of our experiment valence and arousal values in EEG-estimated matrix are rearranged to (−1∼1) scale.
Table 2. Results of our experiment valence and arousal values in EEG-estimated matrix are rearranged to (−1∼1) scale.
EEG-EstimatedUser-Annotated
Photographic SurgicalIllustrated SurgicalPhotographic SurgicalIllustrated Surgical
Part. No.Val.Arou.Part. No.Val.Arou.Part. No.Val.Arou.Part. No.Val.Arou.
01−0.860.3421−0.680.2101262123
02−0.910.6422−0.390.2402192245
03−0.760.4723−0.540.0903272314
04−0.850.5524−0.480.1904372445
05−0.730.4925−0.450.1705372544
06−0.930.5126−0.530.1806172635
07−0.810.5827−0.590.1307272725
08−0.840.6328−0.520.2508182836
09−0.850.6529−0.630.2109182925
10−0.860.6130−0.610.31110193036
11−0.840.5931−0.430.2611383136
12−0.880.5232−0.470.2412173256
13−0.920.5133−0.510.2313173335
14−0.780.4834−0.560.2814253435
15−0.830.5335−0.640.2715173536
16−0.890.5536−0.650.2616173617
17−0.790.5937−0.590.2317173735
18−0.800.5338−0.580.2618273836
19−0.760.5139−0.520.3119463926
20−0.740.5540−0.610.3320274036
Table 3. p values for t-test between photographic and illustrated surgical images.
Table 3. p values for t-test between photographic and illustrated surgical images.
EEG-EstimatedUser-Annotated
Photographic SurgicalIllustrated SurgicalPhotographic SurgicalIllustrated Surgical
( group 1 )( group 2 )( group 1 )( group 2 )
valence 2.27 × 10 15 7.6 × 10 4
arousal 2.07 × 10 17 2.2 × 10 7
Table 4. p values for t-test between EEG-estimated and user-annotated valence and arousal.
Table 4. p values for t-test between EEG-estimated and user-annotated valence and arousal.
Photographic Surgical ( group 1 )Illustrated Surgical ( group 2 )
EEG-EstimatedUser-AnnotatedEEG-EstimatedUser-Annotated
valence 0.72134 0.844513
arousal 0.942176 0.007345
Table 5. Cohen’s d values to measure the difference in emotional reaction between photographic and illustrated surgical images.
Table 5. Cohen’s d values to measure the difference in emotional reaction between photographic and illustrated surgical images.
EEG-EstimatedUser-Annotated
Photographic SurgicalIllustrated SurgicalPhotographic SurgicalIllustrated Surgical
( group 1 )( group 2 )( group 1 )( group 2 )
valence(i) 1.78(ii) 1.01
arousal(iii) 1.82(iv) 1.37
Table 6. Cohen’s d values to measure the difference of the emotional reaction estimation methods: EEG-biosignal and user-annotation.
Table 6. Cohen’s d values to measure the difference of the emotional reaction estimation methods: EEG-biosignal and user-annotation.
Photographic Surgical ( group 1 )Illustrated Surgical ( group 2 )
EEG-EstimatedUser-AnnotatedEEG-EstimatedUser-Annotated
valence(i) 0.12(ii) 0.06
arousal(iii) 0.23(iv) 0.64
Table 7. Comparison to existing models that recognizes valence and arousal using DEAP dataset.
Table 7. Comparison to existing models that recognizes valence and arousal using DEAP dataset.
Existing ModelsClassifierAccuracy (%)
ValenceArousal
Khosrowabadi et al. 2014 [14]RNN71.4370.83
Alhagry et al. 2017 [16]LSTM RNN85.0085.00
Li et al. 2017 [17]CRNN72.0674.12
Salama et al. 2018 [8]3D CNN87.4488.49
Xing et al. 2019 [18]LSTM81.1074.38
Oursmulti-column90.0190.65
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yang, H.; Han, J.; Min, K. EEG-Based Estimation on the Reduction of Negative Emotions for Illustrated Surgical Images. Sensors 2020, 20, 7103. https://doi.org/10.3390/s20247103

AMA Style

Yang H, Han J, Min K. EEG-Based Estimation on the Reduction of Negative Emotions for Illustrated Surgical Images. Sensors. 2020; 20(24):7103. https://doi.org/10.3390/s20247103

Chicago/Turabian Style

Yang, Heekyung, Jongdae Han, and Kyungha Min. 2020. "EEG-Based Estimation on the Reduction of Negative Emotions for Illustrated Surgical Images" Sensors 20, no. 24: 7103. https://doi.org/10.3390/s20247103

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop