Next Article in Journal
Development and Application of a Stability Index Estimation Algorithm Based on Machine Learning for Elderly Balance Ability Diagnosis in Daily Life
Next Article in Special Issue
Deep Transfer Learning Using Real-World Image Features for Medical Image Classification, with a Case Study on Pneumonia X-ray Images
Previous Article in Journal
Effects of a Subanesthetic Ketamine Infusion on Inflammatory and Behavioral Outcomes after Closed Head Injury in Rats
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of Impulsive Aggression Based on Video Images

1
School of Optics and Photonics, Beijing Institute of Technology, Beijing 100081, China
2
Beijing Key Laboratory for Precision Optoelectronic Measurement Instrument and Technology, Beijing 100081, China
3
Yangtze Delta Region Academy of Beijing Institute of Technology, Jiaxing 314019, China
*
Author to whom correspondence should be addressed.
Bioengineering 2023, 10(8), 942; https://doi.org/10.3390/bioengineering10080942
Submission received: 22 June 2023 / Revised: 27 July 2023 / Accepted: 1 August 2023 / Published: 8 August 2023
(This article belongs to the Special Issue Artificial Intelligence in Advanced Medical Imaging - 2nd Edition)

Abstract

:
In response to the subjectivity, low accuracy, and high concealment of existing attack behavior prediction methods, a video-based impulsive aggression prediction method that integrates physiological parameters and facial expression information is proposed. This method uses imaging equipment to capture video and facial expression information containing the subject’s face and uses imaging photoplethysmography (IPPG) technology to obtain the subject’s heart rate variability parameters. Meanwhile, the ResNet-34 expression recognition model was constructed to obtain the subject’s facial expression information. Based on the random forest classification model, the physiological parameters and facial expression information obtained are used to predict individual impulsive aggression. Finally, an impulsive aggression induction experiment was designed to verify the method. The experimental results show that the accuracy of this method for predicting the presence or absence of impulsive aggression was 89.39%. This method proves the feasibility of applying physiological parameters and facial expression information to predict impulsive aggression. This article has important theoretical and practical value for exploring new impulsive aggression prediction methods. It also has significance in safety monitoring in special and public places such as prisons and rehabilitation centers.

1. Introduction

Impulsive aggression is a term used to express the tendency of individuals to act aggressively without careful consideration [1,2]. Compared with normal individuals, special individuals (such as those who are compelled to undergo treatment and those who are imprisoned) have higher impulsiveness [3]. They are prone to quick and unplanned reactions, ignoring the negative consequences of these reactions in response to internal or external stimuli. This behavioral tendency poses a threat to site management and personnel safety. Therefore, it is very important and practically significant to screen and predict special individuals with impulsive aggression in advance, reducing the possibility of impulsive aggressive behavior through intervention and treatment.
However, the existing research methods for assessing impulsive aggression primarily rely on questionnaires. Questionnaires are dependent on the subjective perceptions and judgements of the subjects concerning their own feelings. As each individual may have varying perceptions and experiences regarding the same issues, the measurement outcomes tend to be subjective. While behavioral and neurological measures offer a more objective approach, they cannot fully replace questionnaire measures due to their demanding requirements for measurement environments and equipment.
At the same time, multiple studies have shown that there is a significant correlation between impulsive aggression and negative emotions, particularly anger, anxiety, and other negative emotions [4,5,6]. Moreover, studies have also found that individuals often exhibit intensified negative emotions when engaging in aggressive behavior [7]. Thus, it is possible to predict impulsive aggression by detecting an individual’s negative emotions [8].
Heart rate variability (HRV) refers to fluctuations in the time interval between consecutive heartbeats and is usually used to reflect the functional activity of the autonomic nervous system [9]. Previous studies have shown that when individuals experience negative emotions, their sympathetic nervous system activity increases, their respiratory rate accelerates, their heart rate (HR) significantly increases, the high frequency (HF) of the power spectrum of heart rate variability decreases, and the ratio of low frequency to high frequency (LF/HF) increases, especially when experiencing emotions such as anger and anxiety [9]. Conversely, when individuals experience positive emotions, such as happiness, their RR intervals become longer, their heart rate slows down, and the low-frequency power (LF) of their heart rate variability power spectrum also decreases [9]. The regulation of heart rate variability is associated with the functional activities of the sympathetic and parasympathetic nervous systems, where the high-frequency component is mainly controlled by the parasympathetic nervous system, and the low-frequency component is regulated by the sympathetic nervous system [10]. Thus, by analyzing the time-domain and frequency-domain parameters of heart rate variability, it is possible to analyze an individual’s current autonomic nervous system state and further analyze their emotional state [11]. Facial expressions can also reflect an individual’s current emotional and psychological states [12]. Ekman and Friesen defined the six culturally universal facial expressions that represent six basic emotions (happiness, disgust, surprise, sadness, anger, and fear) in the 1970s [12]. When experiencing negative emotions, individuals typically express emotions such as disgust, sadness, anger, and fear, while when experiencing positive emotions, they express emotions such as happiness and surprise [13]. As impulsive aggression is closely related to an individual’s emotional state, analyzing an individual’s facial expressions to predict and prevent impulsive aggression is possible.
Therefore, this study breaks through the traditional questionnaire survey method and proposes an impulse attack behavior prediction method that integrates physiological parameters and facial expression information. The method uses video to extract the subject’s heart rate variability and facial expression information through IPPG technology and facial recognition technology, then predicts impulsive aggression by fusing heart rate variability characteristic parameters with facial expression information. This method is non-contact and non-interfering and can realize the early screening and prediction of impulsive aggression tendencies in special individuals. It is more objective than questionnaires and has important significance for exploring new impulsive aggression prediction methods.
The remainder of this paper is organized as follows: In Section 2, we provide an overview of related works in the field. Section 3 outlines our proposed video-based impulse attack behavior prediction method, providing comprehensive and detailed descriptions. The experimental setup and introduction are presented in Section 4, followed by a display of the experimental results in Section 5. Finally, in Section 6, we present our conclusion, summarizing the key findings and contributions of this study.

2. Related Works

In recent years, numerous scholars have conducted research on the psychological status of special populations. For instance, Maria et al. conducted predictive research on the trait aggression of recidivists [14], while Ricarte et al. studied suicidal behavior of incarcerated males [15]. These studies utilized the Buss–Perry aggression questionnaire and impulsive premeditated aggression scale to assess whether the subjects exhibited impulsive aggression.
Additionally, Hausam et al. scored and classified prisoners’ aggression based on information at the time of imprisonment and observations of their behavior through prison guards during the initial weeks of incarceration [16]. Seid et al. measured antisocial personality disorder (ASPD) through face-to-face interviews and The Diagnostic and Statistical Manual of Mental Disorders (DSM-5) [17].
Furthermore, Wang et al. proposed an “implicit-based and explicit secondary ‘implicit + explicit’” screening evaluation system [18] to assess impulsiveness among special individuals. Shi et al. developed the “Rain Man” painting test to evaluate the impulsive aggression of individuals compelled to undergo treatment [19]. This test evaluates psychological pressure and aggressiveness by scoring different details of the drawn pictures.
Overall, the evaluation of the psychological status of special individuals currently relies predominantly on questionnaires, inquiries, and observations. The assessment criteria are subject to the tester’s subjective understanding and judgment, and the measurement results can be significantly influenced by subjective factors, as the subjects may conceal information in the questionnaire survey. The summary of related works is showed in Table 1.

3. Materials and Methods

3.1. Overall Process

The overall framework of this study model is shown in Figure 1. Firstly, the subjects’ facial videos were captured via a camera and saved in a .AVI format. Then, the face detection algorithm was used to obtain the facial image in each frame of the video. The physiological parameters and expression information were extracted from the selected facial images. Finally, the physiological parameters and expression information were input into the random forest classification model to predict the level of impulsive aggression of each subject.

3.2. Physiological Parameters

Imaging photoplethysmography (IPPG) is a non-invasive biomedical detection method which is widely used for non-contact collection of an individual’s heart rate and heart rate variability parameters [20,21]. This technology uses imaging equipment to collect video information containing the measured area, record the pulse signal that changes the light intensity caused by blood volume changes in the form of video images, and extract physiological parameters such as heart rate and blood oxygen saturation through video image processing [22,23]. This study used this method to collect the subjects’ physiological parameters, as shown in Figure 2.
Using a CCD camera with a frame rate of 100 fps, a video containing the face was obtained. The face detection algorithm was used to obtain the facial image in each frame of the video, and the tracking was performed. Next, the key point detection model was used to obtain the facial key point information in the image and select the region of interest. To avoid interference from areas such as the eyes and lips, this study selected the cheek area with relatively rich facial capillaries as the region of interest (ROI) [24], as shown in Figure 3.
The selected ROI area is indicated by the green box in the figure, and its specific position is determined by two facial key points: A (X37, Y29) and B (X46, Y34). After obtaining the sensitive area, the IPPG signal of the G channel was extracted for data preprocessing [24], including smoothing prior filtering (SPA), moving average detrending, Butterworth bandpass filtering, wavelet filtering, and other methods. Firstly, the low-frequency interference of the signal was removed using the smoothing prior method. The smoothing prior filter uses a relatively universal parameter estimation algorithm known as regularized least squares, as shown in Equation (1):
θ λ = arg min θ H θ z 2 + λ 2 D d H θ 2
where θ is the regression parameter, λ is the regularization parameter, H is the observation matrix, Dd is the discrete approximation of the d-order derivative operator, and z is the RR interval time series, which can be expressed as:
z = R 2 R 1 , R 3 R 2 , , R N R N 1 T
where N is the number of R peaks detected.
There was still high frequency noise in the signal after filtering, which exceeded the frequency of the pulse wave information. This was filtered out using moving average filtering. The formula is:
y ( n ) = x ( n ) + x ( n + 1 ) + + x ( n + M 1 ) M
where x(n) is the input signal, y(n) is the output signal, and M is the step length.
After the moving average filtering, the signal was input into a Butterworth bandpass filter (0.83–3.33 Hz) corresponding to the individual heartbeat frequency range. The signal is then subjected to multi-level Daubechies wavelet decomposition, and the high-frequency components of the 5th, 6th, and 7th layers were reconstructed to obtain the final signal. Finally, the preprocessed signal was normalized, and the pulse wave information was extracted to calculate heart rate variability (HRV) parameters.
This paper analyzes HRV information based on three aspects: time domain, frequency domain, and nonlinear domain [25,26]. The extracted HRV parameters are shown in Table 2.

3.3. Expression Parameters

In order to achieve accuracy and stability in facial expression recognition, this study adopted the ResNet-34 residual network model for expression recognition [27]. This is currently a commonly used facial expression recognition network. The problem of gradient disappearance and gradient explosion is solved through residual block avoidance, and at the same time, deeper features can be learned to capture complex facial emotion and expression information. The detailed structure of ResNet-34 is shown in Figure 4.
The selected dataset was the JAFFE facial expression dataset [28], which includes 7 facial expressions: 0—anger, 1—disgust, 2—fear, 3—happy, 4—sad, 5—surprise, and 6—neutral. The negative emotions are anger, disgust, fear, and sadness, while the positive emotions include happy and surprise. An increase in the proportion of overall negative emotions to the total number of emotions indicates a heightened likelihood of the individual being in a negative emotional state. Consequently, during such times, they are more susceptible to experiencing higher levels of impulsive aggression and are more prone to engaging in impulsive aggressive behavior [7]. Therefore, this study used the impulsive emotion value EMO to quantify an individual’s facial expression information, which was calculated as follows:
E MO = S emo N emo × 100 %
where Nemo represents the total number of facial expressions in the video and Semo represents the number of facial expressions corresponding to negative emotions in the video.

3.4. Random Forest Model

Based on the extracted feature parameters of heart rate variability and facial expression information, the individual’s impulsive aggression were predicted using a random forest classification model [29]. Random forest is an ensemble learning method that combines multiple decision trees to achieve accurate prediction. Its decision function can be expressed as:
H ( x ) = c j , i = 1 T h i j ( x ) > 0.5 k = 1 N i = 1 T h i k ( x ) reject , other
where x is the sample, H is the final outcome, T is the total number of learners, hi is the learner, N is the total number of categories, cj is the category, h i j denotes the output of hi on category cj, and i, j, and k are summation variables.
The structure of the random forest model is shown in Figure 5. The model consists of multiple decision trees, each of which is trained using Bagging ensemble learning technology. During classification, the physiological parameters and facial expression information extracted above are input as features into the random forest model. After passing through each decision tree, a classification result is obtained, and the final result is selected by voting based on the most frequently occurring result. Compared with a single decision tree, the random forest model can effectively overcome the overfitting problem of the decision tree and has better tolerance for noise and outliers.

4. Experimental Setup and Study Description

The experimental setup is shown in Figure 6. The light source was directed towards the participant’s face and reflected onto a color CCD industrial camera. The collected data were transmitted to a computer for processing via a data line. At the same time, a contact-based physiological parameter detection (CBPPD) device was used for synchronous detection to ensure the accuracy of the non-contact measurement results. The frame rate of the CCD industrial camera (GS3-U3-23S6C-C) was set to 100 fps with a resolution of 1024 × 1024, and the lens was a Kanda Mark M1214-MP2 industrial lens with a light intensity of 1000 lx and a color temperature of 4000 K. In order to improve the accuracy of the data collection, the participants were required to maintain a sitting position before the experiment and keep their head relatively fixed during the experiment. Figure 7 shows the actual experimental setup.
This experiment collected samples from 23 volunteers between the ages of 18 and 28, including 11 men and 12 women. All the volunteers were informed of the purpose, methods, and content of the experiment and participated with informed consent. All the volunteers were healthy and had no heart-related diseases. Two hours before the experiment, the subjects were asked to maintain a calm state, not to consume food or beverages that could excite or relax the nerves, and not to engage in intense exercise. During the experiment, 8 items related to impulsive aggression were selected from the State-Trait Anger Expression Inventory to assess the participants’ current emotional experience and impulsiveness [30]. Each item contained four options, with scores of 0, 1, 2, and 3, for a total score of 24. If the questionnaire score exceeded 12 points, the subject was considered to be in an impulsive and aggressive state.
As research suggests, behaviors such as violence and bullying have a higher likelihood of triggering negative emotions and impulsive aggression in individuals [31]. Thus, for our experiment, we deliberately selected a 5 min segment from the film and television work titled “The Glory”, which revolves around school violence and revenge, as our emotional induction material. To ensure the efficacy of the emotional induction process, we enlisted the assistance of several volunteers to watch the selected video segment and complete questionnaires. This was done to confirm the video’s ability to genuinely induce negative emotions and impulsive aggression in the subjects.
In order to ensure the effectiveness of the experiment, all the subjects watched this video content for the first time, then filled out the questionnaire honestly and without deliberate concealment to verify the accuracy and authenticity of the data. The specific information of the volunteers is shown in Table 3.
The specific experimental procedure is shown in Figure 8. First, the subjects rested for 30 min in a quiet room to keep their body in a relaxed state. After 30 min, the subjects were asked to sit still in a chair while their facial video was collected for a period of 5 min. These data served as the baseline data for the subjects’ calm state, and the subjects were also asked to complete the State-Trait Anger Expression Inventory questionnaire, which served as the score for the subjects’ emotional experience and impulsiveness in their calm state.
After a 5 min rest, the subjects began the emotional induction experiment. While watching the film, the subjects’ emotions were induced using guided language, and their facial videos were collected using a camera for a duration of 5 min. These data served as the data for the subjects’ negative emotional state. After the emotional induction experiment ended, the subjects were immediately asked to complete the State-Trait Anger Expression Inventory questionnaire again, and the score from this round was used as the score for the subjects’ emotional experience and impulsiveness in their induced emotional state. If there was a significant difference between the questionnaire results obtained after the task and those obtained in the calm state survey, it was considered successful in inducing negative emotions and impulsive aggression in that volunteer. Only the samples that generated negative emotions and impulsive aggression were used to evaluate the establishment of the model.
In order to ensure the effectiveness of the experimental method and the model we proposed, before the formal experiment, we used the same experiment and model to perform binary classification detection on the basic emotion “anger”, and the result accuracy was 93.67%. Therefore, we considered the experiment and the proposed model to be reliable in the subsequent detection and classification of impulsive aggression.

5. Results and Discussion

In the experiment, a total of 20 valid samples were collected after removing three invalid samples. Next, the samples were processed using a sliding window, with a window length of 3 min and a sliding step of 30 s, resulting in 100 groups of experimental data. The average values of all the data are shown in Table 4.
After obtaining the experimental characteristic parameters, we screened the characteristic parameters using an ANOVA (with a significance level of α < 0.05). The selected characteristic parameters should meet two requirements. First, there should be a clear difference between the parameter in the calm state and the impulsive state so the model can distinguish between whether the individual is in a calm state or an impulsive state. Second, the differences in parameters between different individuals in the same state should be small, so as to avoid misjudgment due to individual differences. The sum of squares between groups (SSB) is used to measure the difference or variability between different groups. When αSSB < 0.05, we can consider that there is a significant difference between the calm state and the impulsive state. The sum of squares within groups (SSW) is used to measure the degree of dispersion of individual observations within a group. When αSSW > 0.05, we can consider that there is not a significant difference between different individuals in the same state. The SSB and SSW of each feature are shown in Table 5.
After screening, the selected feature parameters that met the requirements included the mean HR, LF/HF, SD2/SD1, and EMO parameters. Figure 9 shows the distribution of these feature parameters in the calm and impulsive states as boxplots. Compared to the calm state, when subjects were in an impulsive aggressive state, their mean HR, LF/HF, SD2/SD1, and EMO increased.
The selected four HRV and facial expression parameters were used to train the impulsivity classification model. The training was performed using a random forest classification model, and a five-fold cross-validation was applied due to the small sample size. The model parameters were also controlled to prevent overfitting, and the optimal parameters were determined using a grid search method. The final parameters and the accuracy of the model classification results are shown in the Table 6 below:
After searching for model parameters, the final determined parameters are shown in the table. The model selected is the random forest classification model, with a maximum depth of 30, a minimum leaf node size of 4, a minimum sample size of 5 on each leaf node, and 10 decision trees. At this setting, the accuracy of the classification model was 89.39%, effectively categorizing whether an individual was in an impulsive aggression state and predicting their impulsive aggressiveness.
In order to verify the performance of the model, we also established impulsive aggression prediction models based on physiological parameters (containing only HRV feature parameters) and impulsive aggression prediction models based on expressions (containing only expression parameters) as comparative analyses. The accuracy of the prediction results for the three models are shown in Table 7 below.
Table 7 shows an accuracy comparison between our proposed model and the single models. It can be seen that our prediction model is more accurate than the others.

6. Conclusions

This study proposes a model for predicting impulsive aggression based on the fusion of physiological and facial expression information from video images. Compared with existing methods for predicting aggressive behavior, this model overcomes the shortcomings of strong subjectivity, low accuracy, and high concealment. The method utilizes imaging photoplethysmography (IPPG) and the ResNet-34 facial recognition model to extract physiological and facial expression information. It also uses a random forest classification model to predict an individual’s impulsive aggression. The experimental results show that the accuracy of the model reached 89.39%, which was higher than single models. It could effectively classify individuals in impulsive aggression states, and it predicted an individual’s impulsive aggression. This method is non-contact and non-invasive and can be used to screen and predict the tendency of impulsive aggression in special individuals in advance, making it more objective. Moreover, this method also provides important reference value for exploring new methods for predicting impulsive aggression. However, the current experiment also has the problem of a small experimental sample dataset. In the future, we hope to obtain more samples and further train the model to obtain more accurate results.

Author Contributions

Conceptualization, L.D.; methodology, L.D. and L.K.; software, B.Z.; experiment, B.Z.; validation, B.Z.; data curation, M.L.; writing—original draft preparation, B.Z.; writing—review and editing, M.H. and X.C.; project administration, L.D. and Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by NSFC, grant numbers 61935001, 61705010, and 11774031. This research was also funded by the National Key Research and Development Program of China, grant number 2020YFB2009304.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Medical and Laboratory Animal Ethics Committee of Beijing Institute of Technology (protocol code is 049 and date of approval is 9 March 2023).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data is stored in the Baidu network disk sharing link, the data link is: https://pan.baidu.com/s/1Ejz89G9x3zN7lpWWE2mvXw, and the extraction code is: cdgj.

Acknowledgments

The authors are grateful to the volunteers who volunteered to participate in the experiment.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hook, R.W.; Grant, J.E.; Ioannidis, K.; Tiego, J.; Yücel, M.; Wilkinson, P.; Chamberlain, S.R. Transdiagnostic measurement of impulsivity and compulsivity a review of self-report tools. Neurosci. Biobehav. Rev. 2021, 120, 455–469. [Google Scholar] [CrossRef]
  2. Sobolev, M.; Vitale, R.; Wen, H.; Kizer, J.; Leeman, R.; Pollak, J.P.; Baumel, A.; Vadhan, N.P.; Estrin, D.; Muench, F. The Digital Marshmallow Test (DMT) diagnostic and monitoring mobile health app for impulsive behavior development and validation study. JMIR Mhealth Uhealth 2021, 9, e25018. [Google Scholar] [CrossRef] [PubMed]
  3. Yang, X.; Kovarik, C.; Wang, Y.; Yu, S. A multisite cross-sectional study of anxiety symptoms and the associated factors among Chinese drug users undergoing compulsory detoxification treatment. Front. Public Health 2021, 9, 524068. [Google Scholar] [CrossRef]
  4. Strüber, D.; Lück, M.; Roth, G. Sex, aggression and impulse control: An integrative account. Neurosci. Crime 2020, 14, 93–121. [Google Scholar]
  5. Davidson, R.J.; Putnam, K.M.; Larson, C.L. Dysfunction in the neural circuitry of emotion regulation—A possible prelude to violence. Science 2000, 289, 591–594. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Coccaro, E.F.; Fanning, J.R.; Keedy, S.K.; Lee, R.J. Social cognition in intermittent explosive disorder and aggression. J. Psychiatr. Res. 2016, 83, 140–150. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Siep, N.; Tonnaer, F.; van de Ven, V.; Arntz, A.; Raine, A.; Cima, M. Anger provocation increases limbic and decreases medial prefrontal cortex connectivity with the left amygdala in reactive aggressive violent offenders. Brain Imaging Behav. 2019, 13, 1311–1323. [Google Scholar] [CrossRef] [Green Version]
  8. Ferreira-Garcia, R.; Faria, C.G.F.; Nardi, A.E.; da Rocha Freire, R.C. Negative affect mediates impulsivity in generalized anxiety disorder. Psychopathology 2019, 52, 327–333. [Google Scholar] [CrossRef]
  9. Chen, Y.-C.; Hsiao, C.-C.; Zheng, W.-D.; Lee, R.-G.; Lin, R. Artificial neural networks-based classification of emotions using wristband heart rate monitor data. Medicine 2019, 98, e16863. [Google Scholar] [CrossRef]
  10. Ritsert, F.; Elgendi, M.; Galli, V.; Menon, C. Heart and Breathing Rate Variations as Biomarkers for Anxiety Detection. Bioengineering 2022, 9, 711. [Google Scholar] [CrossRef]
  11. Tagnithammou, T.; Monacelli, É.; Ferszterowski, A.; Trénoras, L. Emotional state detection on mobility vehicle using camera: Feasibility and evaluation study. Biomed. Signal Process. Control 2021, 66, 102419. [Google Scholar] [CrossRef]
  12. Ekman, P.; Friesen, W.V. Unmasking the Face: A Guide to Recognizing Emotions from Facial Clues; ISHK: Los Altos, CA, USA, 2003. [Google Scholar]
  13. Bulagang, A.F.; Weng, N.G.; Mountstephens, J.; Teo, J. A review of recent approaches for emotion classification using electrocardiography and electrodermography signals. Inform. Med. Unlocked 2020, 20, 100363. [Google Scholar] [CrossRef]
  14. Baños, M.C.Z.; Álvarez, J.J.C.; Martín, O.L.; Lorca, M.M.; Jiménez, M.V.J.; Trives, J.J.R. Functioning of psychopathy and trait aggression as predictive variables of criminal recidivism. An. Psicol. Ann. Psychol. 2019, 35, 364–370. [Google Scholar] [CrossRef]
  15. Ricarte, J.J.; Caravaca-Sánchez, F.; Barry, T.J.; Aizpurua, E. Suicide behaviours in incarcerated males: Links to psychopathic traits, forms of aggression to others, personal characteristics, and current penitentiary variables. J. Forensic Leg. Med. 2022, 89, 102357. [Google Scholar]
  16. Hausam, J.; Lehmann, R.J.B.; Dahle, K.P. A person-centered approach to prison behavior based on officers’ observations: Relations to risk, prison misconduct, and recidivism. Front. Psychiatry 2020, 11, 241. [Google Scholar] [CrossRef]
  17. Seid, M.; Anbesaw, T.; Melke, S.; Beteshe, D.; Mussa, H.; Asmamaw, A.; Shegaw, M. Antisocial personality disorder and associated factors among incarcerated in prison in Dessie city correctional center, Dessie, Ethiopia: A cross-sectional study. BMC Psychiatry 2022, 22, 53. [Google Scholar] [CrossRef]
  18. Wang, C.G.; Luo, G.L.; Xu, W.F. Construction of the “implicit + explicit” screening and evaluation system for impulsive aggression among drug addicts. Chin. J. Prison. 2020, 35, 142–145. (In Chinese) [Google Scholar]
  19. Shi, L.Y.; Guo, D.Q.; Li, C.S.; Tian, X.J.; Luo, Y.; Liu, L.; Huo, R.; Zhang, Y. Analysis of the stress and coping effects of compulsory drug addicts based on the “Man in the Rain” drawing test. Chin. J. Drug Depend. 2021, 30, 215–220. (In Chinese) [Google Scholar]
  20. Kong, L.Q.; Chen, F.; Zhao, Y.J.; Dong, L.Q.; Liu, M.; Hui, M. Non-contact Psychological Stress Detection Combining Heart Rate Variability and Expression. Acta Opt. Sin. 2021, 41, 0310003. (In Chinese) [Google Scholar] [CrossRef]
  21. Luo, J.; Zhen, J.; Zhou, P.; Chen, W.; Guo, Y. An iPPG-based device for pervasive monitoring of multi-dimensional cardiovascular hemodynamics. Sensors 2021, 21, 872. [Google Scholar] [CrossRef]
  22. Favilla, R.; Zuccala, V.C.; Coppini, G. Heart rate and heart rate variability from single-channel video and ICA integration of multiple signals. IEEE J. Biomed. Health Inform. 2018, 23, 2398–2408. [Google Scholar] [PubMed]
  23. Yu, X.; Laurentius, T.; Bollheimer, C.; Leonhardt, S.; Antink, C.H. Noncontact monitoring of heart rate and heart rate variability in geriatric patients using photoplethysmography imaging. IEEE J. Biomed. Health Inform. 2020, 25, 1781–1792. [Google Scholar] [CrossRef] [PubMed]
  24. Tohma, A.; Nishikawa, M.; Hashimoto, T.; Yamazaki, Y.; Sun, G. Evaluation of remote photoplethysmography measurement conditions toward telemedicine applications. Sensors 2021, 21, 8357. [Google Scholar] [CrossRef]
  25. Mejía-Mejía, E.; May, J.M.; Torres, R.; Kyriacou, P.A. Pulse rate variability in cardiovascular health. A review on its applications and relationship with heart rate variability. Physiol. Meas. 2020, 41, 07TR01. [Google Scholar] [CrossRef] [PubMed]
  26. Henriques, T.; Ribeiro, M.; Teixeira, A.; Castro, L.; Antunes, L.; Costa-Santos, C. Nonlinear methods most applied to heart-rate time series: A review. Entropy 2020, 22, 309. [Google Scholar] [CrossRef] [Green Version]
  27. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  28. Lyons, M.; Kamachi, M.; Gyoba, J. The Japanese Female Facial Expression (JAFFE) Dataset. In Proceedings of the Third International Conference on Automatic Face and Gesture Recognition, Nara, Japan, 14–16 April 1998; pp. 14–16. [Google Scholar] [CrossRef]
  29. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  30. Giegling, I.; Olgiati, P.; Hartmann, A.M.; Calati, R.; Möller, H.-J.; Rujescu, D.; Serretti, A. Personality and attempted suicide. Analysis of anger, aggression and impulsivity. J. Psychiatr. Res. 2009, 43, 1262–1271. [Google Scholar] [CrossRef]
  31. Schaefer, A.; Nils, F.; Sanchez, X.; Philippot, P. Assessing the effectiveness of a large database of emotion-eliciting films: A new tool for emotion researchers. Cogn. Emot. 2010, 24, 1153–1172. [Google Scholar] [CrossRef]
Figure 1. Overall flowchart. Firstly, the face video of the special individuals is captured by the camera and saved in video format. Then, the face detection algorithm is used to obtain the face image of each frame of the video. The physiological and facial parameters are extracted from the selected face images. Finally, the physiological and facial parameters are input into a random forest classification model simultaneously to obtain the impulsive aggression prediction results of the special individuals.
Figure 1. Overall flowchart. Firstly, the face video of the special individuals is captured by the camera and saved in video format. Then, the face detection algorithm is used to obtain the face image of each frame of the video. The physiological and facial parameters are extracted from the selected face images. Finally, the physiological and facial parameters are input into a random forest classification model simultaneously to obtain the impulsive aggression prediction results of the special individuals.
Bioengineering 10 00942 g001
Figure 2. Process of non-contact extraction of physiological parameters. In order to extract the parameters, we obtained the video signal first. Then, using facial detection and tracking, along with key point detection, we successively selected the region of interest area. After that, we extracted the channel and preprocessed the signal. Finally, by extracting the pulse waves, we calculated the heart rate variability characteristic parameters.
Figure 2. Process of non-contact extraction of physiological parameters. In order to extract the parameters, we obtained the video signal first. Then, using facial detection and tracking, along with key point detection, we successively selected the region of interest area. After that, we extracted the channel and preprocessed the signal. Finally, by extracting the pulse waves, we calculated the heart rate variability characteristic parameters.
Bioengineering 10 00942 g002
Figure 3. Key point detection and ROI area selection. (a) Schematic diagram of 68 key points on the face. (b) The image detected by the 68 facial features. The green box is the selected region of interest. A and B are the two vertices of the box, which are A (X37, Y29) and B (X46, Y34).
Figure 3. Key point detection and ROI area selection. (a) Schematic diagram of 68 key points on the face. (b) The image detected by the 68 facial features. The green box is the selected region of interest. A and B are the two vertices of the box, which are A (X37, Y29) and B (X46, Y34).
Bioengineering 10 00942 g003
Figure 4. ResNet-34 residual neural network framework. The ResNet-34 residual neural network includes a convolutional layer, a pooling layer, a series of residual structures, average pooling downsampling, and a fully connected layer to obtain the final output.
Figure 4. ResNet-34 residual neural network framework. The ResNet-34 residual neural network includes a convolutional layer, a pooling layer, a series of residual structures, average pooling downsampling, and a fully connected layer to obtain the final output.
Bioengineering 10 00942 g004
Figure 5. Random forest model structure. The x is the input of the model, and Tree 1 to Tree T are the different decision trees in the model. Inputting x into each decision tree will result in the classification result h1 to hT. Finally, by voting on the results of all trees, we obtain the classification result H of the model.
Figure 5. Random forest model structure. The x is the input of the model, and Tree 1 to Tree T are the different decision trees in the model. Inputting x into each decision tree will result in the classification result h1 to hT. Finally, by voting on the results of all trees, we obtain the classification result H of the model.
Bioengineering 10 00942 g005
Figure 6. Schematic of the experimental setup. The contact-based physiological parameter detection (CBPPD) device was attached to the index finger of the participant’s left hand for PPG signal detection. The LED provided uniform illumination of 1000 lx. The camera captured the subject’s facial video image and sent the data to the desktop to calculate the result.
Figure 6. Schematic of the experimental setup. The contact-based physiological parameter detection (CBPPD) device was attached to the index finger of the participant’s left hand for PPG signal detection. The LED provided uniform illumination of 1000 lx. The camera captured the subject’s facial video image and sent the data to the desktop to calculate the result.
Bioengineering 10 00942 g006
Figure 7. Experimental setup (dark room for actual experiments). The actual experiment was carried out in a dark room. The contact-based physiological parameter detection (CBPPD) device was attached to the index finger of the subject’s left hand for PPG signal detection. The LED provided uniform illumination of 1000 lx. The camera captured the subject’s facial video image and sent it to the desktop to calculate the result.
Figure 7. Experimental setup (dark room for actual experiments). The actual experiment was carried out in a dark room. The contact-based physiological parameter detection (CBPPD) device was attached to the index finger of the subject’s left hand for PPG signal detection. The LED provided uniform illumination of 1000 lx. The camera captured the subject’s facial video image and sent it to the desktop to calculate the result.
Bioengineering 10 00942 g007
Figure 8. Experimental flow chart showing the experiment process. First, the subjects rested for 30 min to ensure that they were in a calm state. Then, video footage was captured when they were calm to serve as the control data, and they were asked to complete the questionnaire. Next, negative emotions were induced in the subjects, and video footage was captured simultaneously as the experimental data in the impulsive aggression state. Finally, the subjects completed the questionnaire again.
Figure 8. Experimental flow chart showing the experiment process. First, the subjects rested for 30 min to ensure that they were in a calm state. Then, video footage was captured when they were calm to serve as the control data, and they were asked to complete the questionnaire. Next, negative emotions were induced in the subjects, and video footage was captured simultaneously as the experimental data in the impulsive aggression state. Finally, the subjects completed the questionnaire again.
Bioengineering 10 00942 g008
Figure 9. Box plot of feature parameters. The figure shows boxplots of four selected feature parameters: (a) mean HR, (b) LF/HF, (c) SD2/SD1, and (d) emotion. In each boxplot, the red color represents the distribution of data in the calm state, while the green color represents the distribution of data in the impulsive state.
Figure 9. Box plot of feature parameters. The figure shows boxplots of four selected feature parameters: (a) mean HR, (b) LF/HF, (c) SD2/SD1, and (d) emotion. In each boxplot, the red color represents the distribution of data in the calm state, while the green color represents the distribution of data in the impulsive state.
Bioengineering 10 00942 g009
Table 1. Summary of related works.
Table 1. Summary of related works.
YearPurposeMeasures
Maria et al. [14]2019Trait aggression of recidivistsQuestionnaire
Ricarte et al. [15]2021Self-injurious behaviours and suicide attempts among incarcerated peopleQuestionnaire
Hausam et al. [16]2020Scored and classified prisoners’ aggressionBehavior rating scale
Seid et al. [17]2022Measured antisocial personality disorder (ASPD) of incarcerated peopleFace-to-face interviews and scale
Wang et al. [18]2019Screening for highly impulsive aggressive behavior in drug addictsBehavioral tasks and questionnaire
Shi et al. [19]2021Measuring stress in drug addictsDraw-a-Person-in-the-Rain (DAPR) test
Table 2. Heart rate variability characteristic parameters.
Table 2. Heart rate variability characteristic parameters.
IndexDefinitionUnit
Mean HRAverage number of heartbeats per minutems
PNN50Heart rate of adjacent R–R intervals greater than 50 ms as a percentage of all NN intervals%
LFLow frequency power (0.04–0.15 Hz)ms2
HFHigh frequency power (0.15–0.4 Hz)ms2
LF/HFPower ratio of low frequency band to high frequency band%
SD1Standard deviation of the vertical line in the Poincare chartms
SD2Standard deviation along the marked line in the Poincare chartms
SD2/SD1Correlation dimension
Table 2 shows the name, definition, and unit of the Heart rate variability characteristic parameters. Here, the mean HR and PNN50 are the time domain parameters. LF, HF, and LF/HF are the frequency domain parameters, and SD1, SD2, and SD2/SD1 are the nonlinear domain parameters.
Table 3. Volunteer information.
Table 3. Volunteer information.
Test InformationStatistical Results (Mean ± Standard Deviation)
Age (years)22.5 ± 3.5
Gender (Male:Female)11:12
Height (cm)173.6 ± 14.8
Weight (kg)60.7 ± 15.7
Table 4. Comparison of characteristic parameters.
Table 4. Comparison of characteristic parameters.
IndexCalm StateImpulsive State
Mean HR (BPM)76.1882.06
PNN50 (%)40.0951.62
SDNN (ms)92.3691.75
RMSSD (ms)58.9399.58
LF/HF0.500.86
SD1 (ms)94.57147.60
SD2 (ms)95.71176.66
SD2/SD11.011.19
EMO (%)20.9547.60
The data in the table are the comparison results of one of the subject’s parameters before and after emotion induction (corresponding to a calm state and impulsive state).
Table 5. ANOVA table of feature parameters.
Table 5. ANOVA table of feature parameters.
IndexSSBSSW
Mean HR6.7 × 10−250.42
PNN501.7 × 10−220.03
SDNN0.180.48
RMSSD1 × 10−310.02
LF/HF5.9 × 10−340.16
SD11 × 10−790.04
SD21 × 10−790.04
SD2/SD14.2 × 10−170.37
EMO1.2 × 10−330.90
The data in the table are the results of the ANOVA for each parameter. SSB represents the between-group analysis result, and SSW represents the within-group analysis result.
Table 6. Results of model training.
Table 6. Results of model training.
ModelMax DepthMin Samples LeafMin Samples SplitEstimatorsAccuracy/%
Random Forest30451089.39
The table shows the accuracy of the random forest classification model used in the study. The selected model is the random forest classification model, with a maximum depth of 30, a minimum leaf node size of 4, a minimum sample size of 5 on each leaf node, and 10 decision trees. At this setting, the accuracy of the classification model was 89.39%.
Table 7. Comparison of the experimental results.
Table 7. Comparison of the experimental results.
ModelAccuracy/%
Physiological parameter80.65
Expression parameter75.32
Physiological and expression parameters89.39
The table shows that the accuracy of the model based on physiological parameters was 80.65%. The accuracy of the model based on the expression parameters was only 75.32%, while the accuracy of the model integrating physiological and expression parameters was the highest: 89.39%.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, B.; Dong, L.; Kong, L.; Liu, M.; Zhao, Y.; Hui, M.; Chu, X. Prediction of Impulsive Aggression Based on Video Images. Bioengineering 2023, 10, 942. https://doi.org/10.3390/bioengineering10080942

AMA Style

Zhang B, Dong L, Kong L, Liu M, Zhao Y, Hui M, Chu X. Prediction of Impulsive Aggression Based on Video Images. Bioengineering. 2023; 10(8):942. https://doi.org/10.3390/bioengineering10080942

Chicago/Turabian Style

Zhang, Borui, Liquan Dong, Lingqin Kong, Ming Liu, Yuejin Zhao, Mei Hui, and Xuhong Chu. 2023. "Prediction of Impulsive Aggression Based on Video Images" Bioengineering 10, no. 8: 942. https://doi.org/10.3390/bioengineering10080942

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop