Next Article in Journal
GLH: From Global to Local Gradient Attacks with High-Frequency Momentum Guidance for Object Detection
Next Article in Special Issue
A Two-Stage Voting-Boosting Technique for Ensemble Learning in Social Network Sentiment Classification
Previous Article in Journal
Piecewise Business Bubble System under Classical and Nonsingular Kernel of Mittag–Leffler Law
Previous Article in Special Issue
Strong Generalized Speech Emotion Recognition Based on Effective Data Augmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dual-ATME: Dual-Branch Attention Network for Micro-Expression Recognition

1
School of Computer, Jiangsu University of Science and Technology, Zhenjiang 212100, China
2
Key Laboratory of Behavior Sciences, Institute of Psychology, Chinese Academy of Sciences, Beijing 100101, China
3
Department of Psychology, University of the Chinese Academy of Sciences, Beijing 100049, China
*
Authors to whom correspondence should be addressed.
Entropy 2023, 25(3), 460; https://doi.org/10.3390/e25030460
Submission received: 30 January 2023 / Revised: 26 February 2023 / Accepted: 28 February 2023 / Published: 6 March 2023
(This article belongs to the Special Issue Machine and Deep Learning for Affective Computing)

Abstract

:
Micro-expression recognition (MER) is challenging due to the difficulty of capturing the instantaneous and subtle motion changes of micro-expressions (MEs). Early works based on hand-crafted features extracted from prior knowledge showed some promising results, but have recently been replaced by deep learning methods based on the attention mechanism. However, with limited ME sample sizes, features extracted by these methods lack discriminative ME representations, in yet-to-be improved MER performance. This paper proposes the Dual-branch Attention Network (Dual-ATME) for MER to address the problem of ineffective single-scale features representing MEs. Specifically, Dual-ATME consists of two components: Hand-crafted Attention Region Selection (HARS) and Automated Attention Region Selection (AARS). HARS uses prior knowledge to manually extract features from regions of interest (ROIs). Meanwhile, AARS is based on attention mechanisms and extracts hidden information from data automatically. Finally, through similarity comparison and feature fusion, the dual-scale features could be used to learn ME representations effectively. Experiments on spontaneous ME datasets (including CASME II, SAMM, SMIC) and their composite dataset, MEGC2019-CD, showed that Dual-ATME achieves better, or more competitive, performance than the state-of-the-art MER methods.

1. Introduction

As an important nonverbal cue for emotional understanding, facial micro-expressions (MEs) are very brief and involuntary facial expressions that usually last from 0.04 to 0.2 s [1]. In contrast to ordinary expressions, MEs are considered the unconscious leakage of genuine feelings when people try to hide them. Therefore, MEs can help reveal hidden emotions and have promising applications in many fields, such as clinical diagnosis [2] and national security [3].
Micro-expression recognition (MER) has attracted more and more attention from researchers in recent decades as a critical way of understanding human emotions [4]. In addition to traditional human-based psychological studies, many researchers have tried to investigate automatic MER methods based on computer vision techniques. In recent years, research on MER has gradually progressed with the development of published ME datasets [5,6,7,8,9,10,11,12]. There are two main approaches: methods based on Hand-crafted features [13,14,15,16,17,18] and methods based on Deep learning [19,20,21,22,23,24,25]. However, a crucial research issue for these two methods is extracting salient and discriminative features from MEs. Firstly, the subtle and fleeting properties of MEs require improvement in the representativeness of both manual features and deep learning-based features. Secondly, while deep learning has demonstrated strong capabilities of feature learning, it relies on extensive data for training. The number of spontaneous annotated ME samples is limited to only just over 2000, which limits the ability of deep learning networks to extract high-level features related to MEs. Thus, extracting discriminative features of MEs remains an area of ongoing exploration.
Enhancing MER performance by expanding the data scale is highly challenging. The data collection and annotation for ME is especially complicated. With limited data, single-scale features may not be sufficient for discriminating different categories of MEs. Combining multiple channels and effectively integrating various scale features, it is possible to improve the model’s ability to learn ME features.
In order to learn ME characteristics efficiently, we propose a simple and effective method called Dual -branch ATtention network for Micro-Expression recognition (Dual-ATME), which uses feature selection, based on both Hand-crafted Attention Region Selection (HARS) and Automated Attention Region Selection (AARS), for MER. Our proposed method includes three sub-modules: Data preprocessing, Dual-ATME module and Feature fusion for ME classification.
In data preprocessing, to better capture the ME motion information, we computed dual-scale optical flow features between the onset and apex frames. Precisely, one scale feature is from the full face and the other is from the HARS-based puzzled counterpart. In the Dual-ATME module phase, in order to simultaneously learn global and local ME features, we utilize the bi-Inception network for HARS-based and AARS-based (full-face) feature extraction, respectively. Furthermore, in order to be able to better focus on ME-related features in the full-face learning network, we employ a combined channel and spatial attention mechanism module. After obtaining the features at both scales, we narrow the distance between them by means of similarity comparison, which empowers the network to refine the ME features common to both. Finally, emotional category prediction is performed by fusing the HARS-based and AARS-based features.
The main contributions of our work are as follows:
  • We propose the Dual-ATME framework, which extracts HARS- and AARS-based features to perform MER. By adding a parallel artificially-selected ROI ME feature learning module to a standalone deep attention mechanism, we enable the proposed Dual-ATME to effectively learn more discriminative ME features. In particular, based on experimental results, we find that manual feature extraction, based on prior knowledge, is essential for MER with limited data size.
  • We design a simple and effective joint loss to optimize feature discrimination in our proposed framework. In particular, in addition to the traditional loss for ME classification, we use a similarity comparison loss to close the distance of the dual-scale ME features in the embedding space.
  • Our Dual-ATME method is extensively evaluated on multiple ME datasets. The experimental results showed that our method demonstrates superior, or comparable, MER performance to state-of-the-art (SOTA) methods on the composite dataset benchmark and single dataset evaluation.
The rest of this paper is organized as follows. Section 2 introduces related work on MER. Section 3 presents the details of our proposed algorithm. Section 4 reports the experimental results on the composite dataset benchmark and single dataset evaluation, as well as ablation studies and visual analysis of our proposed modules. Finally, Section 5 discusses the conclusions and future research directions.

2. Related Work

2.1. Micro-Expression Recognition

MER is a challenging task that involves recognizing and interpreting subtle facial movements that convey emotions. Generally, these methods could be roughly categorized into hand-crafted methods and deep learning methods.

2.1.1. Hand-Crafted Methods

Hand-crafted methods capture the distinctive features of MEs by manually designing visual descriptors, which are then fed into a classifier for emotion recognition. Local binary pattern from three orthogonal planes (LBP-TOP) [13] combines the temporal and spatial properties of the three orthogonal planes and is broadly deployed in ME analysis. Due to its low computational complexity, many improved methods based on LBP-TOP have been proposed. Liong et al. [14] extracted LBP-TOP based on facial regions of interest (ROIs), which further improved the performance of LBP-TOP. Huang et al. successively proposed Spatiotemporal local quantized pattern (STCLQP) [15] and discriminative spatiotemporal LBP, with revisited integral projection (DiSTLBP-RIP) [16], revealing discriminative information about MEs. Furthermore, using optical flow algorithms to extract motion features for MER has also made adequate progress. Liu et al. [17] proposed using the Main direction mean optical flow (MDMO) to reduce the dimensionality of features. Liong et al. [18] proposed the Bi-weighted oriented optical flow (Bi-WOOF), which applies two schemes to locally and globally weight the Histogram of oriented optical flow (HOOF) [26] descriptors. Bi-WOOF improves the efficiency of MER by representing the motion changes of the whole ME clip using only the disparities of the onset and apex frames. However, hand-crafted feature extraction methods rely on specialized knowledge and complex parameter-tuning processes. In addition, these methods have inferior robustness and find it difficult to adapt to the impact of data change.

2.1.2. Deep Learning Methods

Deep learning-based methods use neural networks to learn and extract features from the data automatically. In recent decades, deep neural networks have achieved prominent results in computer vision and have greatly inspired the exploration of deep learning for MER. The milestone work based on deep learning for MER was [27], which employed CNN networks and transfer learning methods for MER. However, due to the small sample size, end-to-end methods have difficulty extracting salient features directly and their performance did not surpass some hand-crafted methods. However, by using deep learning techniques to further mine hand-crafted features, such as using optical flow as the input of deep network for feature learning [21,22,23,28,29,30,31], the performance of MER further improved. Gan et al. [29] proposed OFF-apexNet, which extracts optical flow features from each video’s onset and apex frames and then feeds the horizontal and vertical components of optical flows into a dual-stream CNN. Based on OFF-apexNet, Liong et al. [19] proposed the Shallow triple stream three-dimensional CNN (STSTNet). Besides the horizontal and vertical optical flows, STSTNet also extracted a hand-crafted feature, i.e., optical strain, achieving the SOTA performance for the composite dataset evaluation benchmark. Zhou et al. [20] designed the Dual-inception Network using horizontal and vertical optical flow features extracted from the onset and apex frames. In FR [32], Zhou et al. fused expression-shared and expression-specific features for MER, further improving performance. In addition, Xia et al. [22,31] modeled the spatio-temporal variation of ME using recurrent convolutional networks (RCNs).
These methods have achieved SOTA performance on MER tasks. In particular, methods based on ROIs improve the performance by selecting the local regions related to ME [33]. However, the small sample size and subtle characteristics of ME samples limit the combination of deep learning and MER. Hence, learning ME features effectively is critical to further improving the performance of MER.

2.2. Attention Mechanism in Computer Vision

The attention mechanism plays an essential role in human perception analysis [34,35]. It is based on the principle of assigning higher weights to salient regions in an image and suppressing the useless features. In computer vision, attention mechanisms have gained significant attention because of the ability to effectively focus on the representations of ROIs in images or videos. Recently, these methods have been applied to various visual tasks, such as image classification, object detection and facial expression analysis [36,37,38]. Rodriguez et al. [39] fused spatial attention maps with multi-scale information from diverse convolutional layers to represent the importance of each region in the image. Wang et al. [40] proposed a Residual attention network (RAN) with stacked attention modules to produce attention-aware features. By refining the feature map, RAN achieves prominent performance, while being robust to noisy inputs. Hu et al. [41] proposed the Squeeze-and-Excitation network, using global average pooling features to compute inter-channel attention. Based on this, CBAM [42] exploited both spatial and channel attention modules, further improving the performance of image classification and object detection. CBAM demonstrates that exploiting both spatial and channel attention is superior to using only the channel counterpart.
Furthermore, for MER, attention mechanisms direct the model to focus on critical facial regions and suppress irrelevant facial regions and backgrounds. ME-PLAN [38] combines a 3D residual prototype network and a local attention module to learn a precise prototype of ME features while focusing on local facial movements. Su et al. [43] accentuated the key facial components most relevant to MEs to enhance spatial encoding and capture motive information and non-rigid deformation more effectively. In MMNet [44], Li et al. introduced a Continuous Attention module and a Position Calibration module based on the vision transformer [45], allowing for efficient capture of ME features by focusing on local, subtle muscle movements. Recent methods have also focused on incorporating spatio-temporal and channel information to represent MEs better [46,47,48]. Wang et al. [48] proposed a dual-stream spatio-temporal attention network (DSTAN) that captures MEs’ appearance features and discriminative motion areas. Moreover, DSTAN utilizes a temporal attention mechanism to model the importance of ME clips in a temporal sequence, further boosting the performance of MER. The attention mechanism-based approach described above showed promising results on visual tasks by adaptively enforcing the model to weight different features for the input.
Our attention module draws on the design structure of CBAM, focusing on salient features of facial regions from both channel and spatial perspectives. With this simple and effective design, we improve the network’s discriminative ability for different emotional categories of ME features while avoiding overfitting.

3. Proposed Method

3.1. Framework Overview

Figure 1 illustrates the overview of our proposed Dual-ATME framework, which consists of three modules: Data preprocessing, Dual-ATME module and Feature fusion for classification module. The Dual-ATME utilizes a dual-branch Inception network as the backbone. It employs Hand-crafted Attention Region Selection (HARS) and Automated Attention Region Selection (AARS) to focus on facial discriminative features in manual and automated manners, respectively.
In particular, during data preprocessing, for each original ME clip, we obtain the optical flow features x e n t i r e i for the full face and the puzzled counterpart x p u z z l e d i after HARS. Next, in the Dual-ATME module, we feed x e n t i r e i and x p u z z l e d i into the dual-branch backbone network for independent feature learning. Then, we execute the contrastive loss L c o n to pull the distance between HARS-based and AARS-based features from the same ME samples closer, while pushing away the distance between different ME samples. Finally, feature fusion for the ME classification module is performed for expression-refined feature fusion and label prediction.

3.2. Data Preprocessing

In this subsection, we introduce the processes of face cropping, HARS and optical flow-based feature extraction.

3.2.1. Face Cropping

In the ME clip, the ME movement starts at the onset frame and reaches its highest intensity at the apex frame. Studies have shown that by comparing these two frames, ME features can be effectively extracted while avoiding redundant information [20,22,49]. Therefore, we only used the onset and apex frames to capture ME movement changes and to simplify the process. In other words, a pair of onset and apex represented one ME sample in our study.
For each pair of onset and apex frames, we first used the Dlib tool [50] to detect 68 facial landmarks for the onset frame, denoted as ϕ = { ϕ i j | i = 1 , . . . , N ; j = 1 , . . . , 68 } , where N represents the total number of ME samples and the range of j represents the 68 facial landmarks, as shown in Figure 2.
Then, based on Equation (1), we performed face alignment:
θ = a t a n 2 ( d y , d x ) × 180 / π = a t a n 2 ( ϕ i 45 ( y ) ϕ i 36 ( y ) , ϕ i 45 ( x ) ϕ i 36 ( x ) ) × 180 / π
where a t a n 2 ( ) is the azimuth calculation function, d y and d x denote the vertical and horizontal offsets of the line between the two eyes, respectively. We then flipped the face image in the opposite direction of θ and, then, re-performed the face key point detection.
Next, we determined the cropping face region based on the updated coordinates of the facial landmarks. Specifically, for the left, right and bottom boundaries of the face region in ith ME samples, we used the landmarks on the sides of the cheeks and the lower jaw, that is, ϕ i 1 , ϕ i 15 and ϕ i 8 . In addition, for the top boundary, we first calculated the average distance between the left and right sides below the eyebrows and above the eyes, i.e.,
d = [ ( ϕ i 38 ( y ) ϕ i 20 ( y ) ) + ( ϕ i 43 ( y ) ϕ i 23 ( y ) ) ] / 2
Then, the top boundary was defined as ϕ i 19 + d , conserving an expansion area above the eyebrows. Finally, we cropped the face region based on these landmarks.

3.2.2. Hand-Crafted Attention Region Selection (HARS)

The facial regions with the most muscle activity when MEs occur are the eyebrow and mouth regions [31,51,52]. In other words, these regions contribute most of the discriminative ME information. Furthermore, we conducted a statistical analysis of facial action units (AUs) on ME samples from the CASME II and SAMM datasets of MEGC2019. Based on the AU annotations provided for each ME sample by CASME II and SAMM, we counted and ranked the occurrences of AUs. As shown in Figure 3, it was observed that AUs with higher frequencies were concentrated in the areas around the eyebrows and mouth. Therefore, to reduce the influence of interference from irrelevant regions on MER, our proposed HARS treats the eyebrow and mouth regions as ROIs for local feature extraction. In this way, we can extract more semantic ME-related information.
First, based on the landmarks of the mouth and the eyebrows, we performed region detection and cropping to obtain the boundary borders of each ROI (shown as red dotted boxes in Figure 2). Specifically, for the left eyebrow of the i-th ME sample, we calculated the distance between ϕ i 19 and ϕ i 37 as le_h and the distance between ϕ i 17 and ϕ i 21 as le_l. Then, in order to capture a richer set of information around the left eyebrow, we expanded 1 / 4 · l e _ l to the left and right sides of the left eyebrow, while expanding 1 / 4 · l e _ h upwards to the top of the left eyebrow. Additionally, to reduce interference from the eye area, we set the lower boundary of the left eyebrow area to ϕ i 37 ( y ) 1 / 4 · l e _ h . We obtained the boundary of the left eyebrow region through these steps, i.e.,
[ ϕ i 19 ( y ) 1 4 · l e _ h : ϕ i 27 ( y ) 1 4 · l e _ h , ϕ i 17 ( x ) 1 4 · l e _ l : ϕ i 21 ( x ) + 1 4 · l e _ l ]
Symmetrically, the boundary of the right eyebrow region was
[ ϕ i 24 ( y ) 1 4 · r e _ h : ϕ i 44 ( y ) 1 4 · r e _ h , ϕ i 22 ( x ) 1 4 · r e _ l : ϕ i 26 ( x ) + 1 4 · r e _ l ]
For the mouth area, we calculated the distance between the bottom of the nose ( ϕ i 33 ) and the top of the lips ( ϕ i 51 ) as m_h and the distance between the two corners of the mouth ( ϕ i 48 and ϕ i 54 ) as m_l. Given the four basic landmarks of the lip ( ϕ i 48 , ϕ i 51 , ϕ i 54 and ϕ i 57 ), we expanded 1 / 4 · m _ l horizontally and 1 / 2 · m _ h vertically. Finally, the boundary of the mouth region was
[ ϕ i 51 ( y ) 1 2 · m _ h : ϕ i 57 ( y ) + 1 2 · m _ h , ϕ i 48 ( x ) 1 4 · m _ l : ϕ i 54 ( x ) + 1 4 · m _ l ]
Second, after obtaining the boundary coordinates of the eyebrow and mouth regions, we cropped three ROIs for each ME sample: the left eyebrow, right eyebrow and mouth region. Finally, we concatenated the resized left and right eyebrows horizontally and then vertically concatenated them with the mouth. The final puzzled counterpart of the ME sample was formed, as shown in the middle part of Figure 1b.
It is worth noting that, as shown in Figure 1, HARS-based and AARS-based feature extraction are parallel in our proposed network. That is, in addition to the ROI-based puzzled counterpart obtained by HARS, we introduce full face-based AARS, described in Section 3.3.2, thus, achieving MER under the similarity comparison of the two kinds of features.

3.2.3. Optical Flow Extraction

As previously mentioned, the action information in ME is very subtle and it is difficult to extract features of RGB ME samples for MER directly. Based on the theory of brightness constancy [53], optical flow is an appropriate feature to represent action information. We used the TV-L1 [54] algorithm to extract the optical flow features.
Specifically, we estimated and extracted the optical flow information for each ME clip between their onset and apex frames. Inspired by Liong et al. [19], we extracted horizontal and vertical optical flows ( u , v ) and further added optical strain to form the final optical flow image, which enriches the motion variations of MEs. Specifically, optical strain, as a derivative of optical flow, is capable of approximating the intensity of facial deformation and is less affected by factors such as lighting conditions and skin color. It can be defined as:
ϵ = 1 2 [ O f + ( O f ) T ]
where O f = [ u , v ] T denotes the optical flow vector, including horizontal and vertical components and ∇ denotes the derivative of O f .
In sum, the final constructed optical flow feature in our study was ( u , v , ϵ ) R 3 × w × h . For the whole face, we estimated the optical flow feature x e n t i r e between the onset and apex frames ( I onset e , I apex e ) using the TV-L1 algorithm as follows:
x e n t i r e = ( u e , v e , ϵ e ) = T V L 1 ( I onset e , I apex e )
Similarly, for HARS-based puzzled counterparts of onset and apex frames ( I onset p , I apex p ), the optical flow feature x p u z z l e d was obtained as:
x p u z z l e d = ( u p , v p , ϵ p ) = T V L 1 ( I onset p , I apex p )

3.3. Dual-ATME Module

The Dual-ATME consists of two components: the Dual-branch Inception feature extraction module and the ME feature similarity estimation module. The Dual-branch Inception feature extraction module extracts multi-scale features from the ME sample (full face) and its puzzled counterpart. In particular, we apply AARS on full face-based sub-network to learn ME features effectively. Subsequently, the ME feature similarity estimation module estimates the similarity between extracted features and constrains the similarity of the same group of features to be as high as possible.

3.3.1. Backbone: Dual-Branch Inception Feature Extraction Module

Differences in emotion types and individual expressions for ME can lead to diverse distribution of ME movements on the face. Hence, it is challenging to adapt fixed-size convolution kernels in traditional CNNs to each ME sample. Larger kernels are more suitable for extracting global information, while smaller kernels perform better in extracting local information. To learn both local and global ME features, we used the Dual-branch Inception module as the backbone to extract multi-scale features of MEs based on optical flow input.
As shown in Figure 4, each branch of our Dual-branch Inception module consists of a bi-Inception module, composed of two Inception blocks [55]. An Inception block combines feature maps of different sizes and receptive fields, enabling the network to capture richer information. Specifically, the single Inception block uses three different sizes of convolution kernels, namely 1 × 1 , 3 × 3 and 5 × 5 , to obtain feature maps of various sizes. Additionally, 1 × 1 convolution layers are inserted between the 3 × 3 and 5 × 5 convolution layers to reduce the number of channels in the feature map, thus reducing the computational cost. In our Dual-branch Inception module, the filters for the first and second layers were set to 6 and 12, respectively. Furthermore, we used max-pooling to aggregate the information from the feature maps after each Inception block in each layer. Therefore, in our constructed bi-Inception module, the dimension of the output from the first and the second Inception blocks were 24 × 14 × 14 and 48 × 7 × 7 , respectively.
Following the suggestion in [19], the inputs of the two bi-Inception blocks are optical flow features, including x e n t i r e i R 3 × 28 × 28 and its corresponding puzzled counterpart x p u z z l e d i R 3 × 28 × 28 , respectively. In particular, as presented in the Dual-ATME module in Figure 1,
  • x e n t i r e i enters the upper branch, where the bi-Inception network automatically extracts discriminative facial features with the help of the AARS module (See Section 3.3.2).
  • x p u z z l e d i enters the lower branch. As mentioned before, the puzzled counterparts are manually obtained with good discriminability through HARS. Moreover, experiments also demonstrated that adding an attention block to this branch did not significantly improve performance (See Table 1). Thus, we did not implement an Attention block in the lower branch to reduce the model parameters.
The core of the Dual-branch Inception module extracts different feature maps by using two deep convolutional layers simultaneously, improving the robustness and classification performance of the model. For the two feature maps extracted by the Dual-branch Inception module, on the one hand, we conducted ME feature comparison learning (described in Section 3.3.3) to pull the ME sample and its puzzled counterpart closer. On the other hand, the two feature maps were cascaded after the Dual-branch Inception module by channel dimension and then sent to the classification module (described in Section 3.4).

3.3.2. Automated Attention Region Selection (AARS)

As introduced in Section 3.2.2, we propose two Attention Region Selection modules, i.e., HARS and AARS. Contrary to manual ROI selection in HARS, AARS exploits the attention mechanism to automatically focus on the crucial features of the original facial region from the channel and spatial dimensions, respectively and assigns higher weights to these features.
Precisely, we deployed the Convolutional Block Attention Module (CBAM) [42] attention mechanism module. CBAM was chosen because it includes not only channel attention but also spatial attention to further enhance valuable features in the feature map and suppresses unnecessary features.
The architecture of CBAM in our study is illustrated in Figure 5. The output of the bi-Inception Module V e _ t = I n c ( x e n t i r e ) R 48 × 7 × 7 is given as the input to CBAM. We fed V e _ t through the Channel Attention Module (CAM) and the Spatial Attention Module (SAM) in sequence, to obtain the refined features V e = M ( V e _ t ) .
In detail, CAM outputs the channel attention map using the max-pooling and average-pooling in a weight-shared network, as well as the output of a Multi-layer Perceptron (MLP). SAM pools the output of CAM along the channel axis and passes it through a convolutional layer to obtain the spatial attention map. The CAM and SAM are as follows:
  • CAM. As shown in the left of Figure 5, we first extracted the spatial context information from V e _ t by passing through the max-pooling and the average-pooling layers. Then, the features were fed into a weight-shared MLP with two hidden layers. Finally, the two features output from the MLP were summed element-wise and the result activated by sigmoid to obtain the channel attention features M c . The channel attention weight function M c ( · ) can be represented as:
    M c ( V e _ t ) = σ M L P M a x ( V e _ t ) + M L P A v g ( V e _ t )
    where σ denotes the s i g m o i d function and M a x ( · ) and A v g ( · ) denote max-pooling and average-pooling, respectively.
  • SAM. Subsequently, as shown in the right of Figure 5, the input feature for SAM was V e _ t c = M c ( V e _ t ) · V e _ t . Then, we applied max-pooling and average-pooling to V e _ t c along the channel dimension. Next, we concatenated the generated feature maps along the channel dimension and applied convolutional and sigmoid operations to obtain the final spatial attention feature map M s . The spatial attention weight function M s ( · ) can be represented as:
    M s ( V e _ t c ) = σ [ M a x ( V e _ t c ) ; A v g ( V e _ t c ) )
    where [ · ; · ] denotes stacking two feature maps along channel dimension.
Finally, the final output feature ( V e ) of the CBAM module ( M ( · ) ) can be expressed as:
V e = M ( V e _ t ) = M s ( V e _ t c ) · ( M c ( V e _ t ) · V e _ t ) = M s ( M c ( V e _ t ) · V e _ t ) · ( M c ( V e _ t ) · V e _ t )

3.3.3. ME Feature Similarity Estimation

Cosine similarity [56] can be used to measure the similarity or dissimilarity between two embedding vectors and is commonly used in fields such as machine learning, natural language processing and computer vision. In our study, we wanted to measure the similarity between the ME feature maps of the full face and its puzzled counterpart. Hence, we used L 2 normalized cosine similarity for the comparison. First, we normalized the two feature maps from 1 to 1. Then, the L 2 normalized cosine similarity s i m ( · , · ) was calculated as follows:
s i m ( V e , V p ) = V e · V p V e · V p
where V e = M ( V e _ t ) = M ( I n c ( x e n t i r e ) ) and V p = I n c ( x p u z z l e d ) . The value of s i m ( V e , V p ) varies from −1 to 1, where 1 indicates a perfect match and 1 indicates a complete mismatch. If the value is 0, it suggests that the two vectors are orthogonal (perpendicular) to each other and have no correlation. Therefore, to evaluate the difference between two features, we set the contrastive loss as:
L c o n = 1 s i m ( V e , V p )
This way, different versions of the same ME sample (full-face and puzzled) were matched in the high-level representation space to achieve instance-level approximation.

3.4. Feature Fusion for ME Classification

In the feature fusion stage, two 48 × 7 × 7 dimensional features V e and V p from the dual-ATME module are concatenated along the channel dimension to get a 96 × 7 × 7 feature. Next, we flatten the feature into a one-dimensional vector and feed it into the final classification module, which consists of two fully-connected layers. To prevent overfitting, we include a Dropout layer with a probability of 0.5 after the first fully-connected layer. The output of the last fully-connected layer is passed through a softmax activation to obtain the ME category predictions.

3.5. Joint Loss Function

In the proposed Dual-ATME, the ME feature similarity estimation (contrastive module) and the classification module are jointly trained. Thus, the joint loss function of the entire network is represented as follows:
L = L c l s + λ L c o n
where L c l s and L c o n denote the classification loss and contrastive loss, respectively. λ represents the regularization parameter, which determines the weight of L c o n in the overall loss. Following the setting in [22], we perform the Focal loss [57] as the classification loss. Focal loss could effectively reduce the loss weight for well-classified examples and focus on complex examples with higher losses, improving the recognition performance on unbalanced datasets. By optimizing the joint loss, Dual-ATME can extract discriminative features for MER.

4. Experiments

In this section, we first describe our experimental configuration, which includes datasets, validation protocols and experimental settings. Next, we compare Dual-ATME with SOTA MER methods. We also conducted adequate ablation studies to demonstrate the effectiveness of each module in our framework. Finally, we provide an attention visualization analysis of Our Dual-ATME.

4.1. Datasets and Validation Protocols

4.1.1. Datasets

MEGC2019-CD. A 3DB-combined dataset called MEGC2019-CD was proposed by Micro-Expression Grand Challenge (MEGC2019) for Composite Dataset Evaluation (CDE). It is a composite of three spontaneous datasets: SMIC, CASME II and SAMM, with three emotion categories: Negative (containing Repression, Anger, Contempt, Disgust, Fear and Sadness), Positive (i.e., Happiness) and Surprise). The detailed information of these three datasets is described as follows and shown in Table 2.
SMIC. The Spontaneous Micro-Expression Corpus (SMIC) consists of three different portions captured by different types of cameras: a conventional visual camera (VIS), a near-infrared camera (NIR) and a high-speed camera (HS). We only used the HS subset of SMIC, captured by a high-speed camera and consistent with CASME II and SAMM. The SMIC-HS subset contains 164 video clips from 16 subjects, recorded using a 100 fps high-speed camera with a resolution of 640 × 480. All MEs in SMIC are divided into three categories: Negative (70), Positive (51) and Surprise (43).
CASME II. The Chinese Academy of Sciences Micro-Expression II (CASME II) dataset contains 255 ME samples from 26 participants, captured using a 200 fps high-speed camera. The raw resolution of each frame is 640 × 480 and the facial region is 280 × 340 pixels. The CASME II dataset is divided into seven categories: Disgust (63), Fear (2), Happiness (32), Repression (27), Sadness (7), Surprise (25) and Others (99). In MEGC2019-CD, it is re-classified into three classes: Negative (88, including Disgust and Repression), Positive (32) and Surprise (25).
SAMM. The Spontaneous Actions and Micro-Movement (SAMM) dataset consists of 159 ME clips from 29 participants, captured using a high-speed camera at 200 fps. The original resolution for each ME frame in SAMM is 2040 × 1088 and the facial area is approximately 400 × 400 pixels. This dataset includes eight raw emotion classes: Anger (57), Contempt (12), Disgust (9), Fear (8), Happiness (26), Sadness (6), Surprise (15) and Others (26). In MEGC2019-CD, it is re-grouped into three categories: Negative (92, including Anger, Contempt, Disgust, Fear and Sadness), Positive (26) and Surprise (15).

4.1.2. Validation Protocols

Validation Protocol: Leave-one-subject-out (LOSO) cross-validation is a type of cross-validation where each subject in a dataset is used as the test set once, while the remaining subjects are used as the training set. This method is useful when the subjects in the datasets may have inherent differences. Considering the small sample size in MEGC2019-CD and the significant variation in subjects, we used the LOSO cross-validation method to evaluate the model’s performance.
Evaluation Metrics: To evaluate the performance of different methods on the CDE benchmark, we used three metrics: accuracy (Acc), unweighted average recall (UAR) and unweighted F1-score (UF1). These metrics were used to measure the performance on both the composite and individual datasets. The Acc, UAR and UF1 are defined as follows:
Acc = T P c N t o t a l
UAR = 1 C i = 1 C T P c N c
UF 1 = 1 C i = 1 C 2 × T P c 2 × T P c + F P c + F N c
where C is the number of classes, N c is the number of samples with the c-th class, N t o t a l is the total samples and T P c , F P c and F N c are the number of true positive, false positive and false negative samples in the c-th class, respectively.

4.2. Experimental Setting

For each dataset, we used the Dlib [50] tool to detect 68 facial landmarks and used these landmarks to crop the facial regions. During training, the facial images were resized to 28 × 28 to serve as the input to the network, with random horizontal flipping applied for data augmentation. During testing, the input images were only resized to 28 × 28 and then fed into the trained model. Our Dual-ATME method was implemented using the Pytorch toolbox, with the backbone network being bi-Inception blocks [55].
The number of training epochs in our Dual-ATME framework was set as 60. The Adam optimizer was employed, with β 1 and β 2 set to 0.5 and 0.999, respectively. The initial learning rate was 0.001 and a cosine learning rate schedule was applied. For each training iteration, 128 ME samples were used in a mini-batch. All experiments were conducted on a single NVIDIA-RTX-4090 GPU. Our code is available at https://github.com/HaoliangZhou/Dual-ATME (accessible since 27 February 2023).

4.3. Experimental Results

We compared our proposed method with hand-crafted feature extraction methods and classical deep learning methods on the widely used ME datasets SMIC-HS, CASME II and SAMM. We also performed CDE on their combined dataset MEGC2019-CD.
In terms of the choice of comparison methods, among the hand-crafted feature-based methods, we compared our proposed method with LBP-TOP [13] and Bi-WOOF [18]. Among the deep learning methods, we chose the SOTA methods from the MEGC2019 and MER projects with open source code, including STSTNet [19], Dual-Inception [20], RCN(_a,_w,_c and _f) [22], KFC-MER [43] and MMNet [44]. Given the transient and subtle nature of ME samples, preprocessing operations on the images can significantly affect the results. To ensure comparability and fairness, we reproduced all these methods using the same inputs and data configuration, including the same number of samples, classes and cross-validation protocol. In addition, we used full-face images as the model input when implementing these methods.
Table 3 provides a comprehensive experimental results overview of all methods on the MEGC2019-CDE dataset. Among the hand-crafted MER methods, the best algorithm was LBP-TOP, which had a UAR and UF1 of 0.5753 and 0.5857, respectively. The RCN_a method achieved the highest UAR and UF1 performance of 0.6351 and 0.6339 among the deep learning methods. Besides these methods, our proposed Dual-ATME, combining HARS and AARS, achieved the best performance on the MEGC2019-CD dataset. In addition, it performed the best on all datasets, except CASME II (n.b. shown in bold) and was second best on CASME II, slightly behind RCN_a. In most cases, our method outperformed its competitors significantly, with a UAR and UF1 that were 4.42% and 4.49% higher than the second-best method (RCN_a), respectively.
In addition, Figure 6 shows the confusion matrix of our Dual-ATME on CASME II, SAMM, SMIC and the MEGC2019-CD datasets. On the MEGC2019-CD datasets, Dual-ATME obtained accuracies of 0.80, 0.57 and 0.65 for “Negative”, “Positive” and “Surprise”, respectively. In addition, by comparing the three single datasets, we found that Dual-ATME achieved the highest performance in CASME II, i.e., obtaining 0.93, 0.62 and 0.72 for “Negative”, “Positive” and “Surprise”, respectively. However, in SAMM, Dual-ATME showed the lowest accuracy values in recognizing positive and surprise emotion categories. The reason was that the SAMM dataset had fewer samples than the other two datasets. Moreover, SAMM contains 13 ethnic groups. The rich diversity of subjects, i.e., differences in subjects, can affect the experimental results of LOSO-cross-validation.
From the perspective of different emotional categories, the ME recognition performance on specific emotion type would be empirically improved with more sample data. This is because the deep learning-based model is data-driven, meaning that representational features need to be learned from as much data as possible. As shown in Figure 6, the number of “Negative” samples outnumbered the “Positive” and “Surprise” in each dataset. Thus, the model consistently achieved the highest recognition performance in the “Negative” category compared to the other two categories.

4.4. Ablation Study

To demonstrate the effectiveness of our method, we conducted ablation studies to assess the contribution of the proposed modules and components to the final performance. For all experiments, we evaluated the performance using the MEGC2019-CD dataset.

4.4.1. Effectiveness of the Proposed Modules

To evaluate the performance of the crucial modules in Dual-ATME, we conducted ablation studies for HARS and AARS on the MEGC2019-CD dataset. As reported in Table 1, standalone HARS and AARS only achieved sub-optimal performance. Moreover, HARS obtained higher performance than AARS on MER, demonstrating that manual feature extractionm based on prior experiencem plays an essential role in cases of limited data size. The reason is that with limited data, neural networks may not have enough samples to learn complex representations from raw data. In contrast, hand-crafted features, based on prior experience, are more likely to capture the salient information for emotion category identification.
As shown in Figure 7, the Grad-CAM-based [59] feature visualization also proved the above conclusion. Specifically, AU12 represents the upturned corners of the mouth and it was found that the network focused better on the mouth region in the puzzled counterpart. Additionally, the combination of HARS and AARS could achieve collaborative enhancement. As shown in the first row of Figure 7, for this ME sample, AARS only focused on the left corner of the mouth in the full-face image (column (b)). However, the Puzzled image of HARS, based on prior knowledge, was less disturbed, so it not only focused on the left corner of the mouth, but also on the right corner (column (d)). Thus, better MER performance was achieved by fusing the features of HARS and AARS in parallel.
Overall, Dual-ATME combines HARS and AARS and achieved the best results among all variations. HARS-based features are designed to capture concrete, human-understandable concepts, while AARS-based features automatically capture more abstract and data-driven concepts. Combining these two types of features can achieve collaborative enhancement and improve overall performance.

4.4.2. Different Combinations of Face Regions

To validate the effectiveness of the HARS selection strategy based on the eyebrows and mouth regions, we conducted experiments on different combinations of facially discrete areas, including eyebrows, cheeks, nose and mouth. Specifically, we designed the following four combinations: (a) eyebrows + nose, (b) eyebrows + mouth, (c) cheeks + nose, (d) cheeks + mouth and (e) full-face. Thus, the inputs of our Dual-ATME framework were full-face.
As shown in Table 4, the combination of eyebrows and mouth achieved the best performance on MEGC2019CD, SMIC and SAMM datasets, indicating that variations in the eyebrows and mouth regions are more prominent when ME occurs. This result was further supported by the statistical results based on psychological AU annotation shown in Figure 3, which demonstrated the reasonableness and effectiveness of our HARS selection strategy focusing on the eyebrows and mouth regions.

4.4.3. Different Values of Weight Coefficient λ

To evaluate the recognition performance of the proposed method, we varied the value of λ in Equation (11), as listed in Table 5. A larger value of λ meant that the contrastive loss ( L c o n ) played a more significant proportion of the overall loss and vice versa. Since ME classification is the main task of our model, the classification loss should have the highest weight factor. At the same time, the contrastive loss has an auxiliary role, serving to close the distance between two features in the same group, so we chose a smaller value of λ . In this way, we could balance the effects of both losses and obtain the optimal contrastive performance without sacrificing the classification performance. Specifically, we fixed the weight coefficient of the classification loss to 1 and ranged λ from 0 to 0.1.
The results in Table 5 show that the highest performance was achieved when λ was set to 0.01. It performed the best in all datasets, except for CASME II and SAMM, which still achieved suboptimal competitive performance. This was because if the value of λ is too large, the contrastive loss affects the joint loss function too much, degrading the classification performance. Conversely, a miserly value of λ causes the effect of the contrastive loss to be insignificant, resulting in ineffectiveness in closing the distance between two features.

4.4.4. Different Loss Functions and Optimizers

We also evaluated the performance of different classification loss functions and optimizers. Concretely, we compared the performance of standard CE loss, weighted CE loss and Focal loss, respectively. When deciding the weights of weighted CE loss, we calculated the weights of Negative, Positive and Surprise emotions as 0.2857, 0.1329 and 0.1695, respectively, based on the sample sizes of the three types of MEs in the MEGC2019-CD dataset. We also investigated the improvement of model accuracy by different optimizers, i.e., SGD and Adam.
Table 6 shows the results for Acc, UAR and UF1 using various classification loss functions and optimizers. Focal loss achieved the best performance in MEGC2019-CD, SMIC and SAMM datasets and the competitive sub-optimal performance in the CASME II dataset. It automatically focuses on the categories with the fewest samples and gives these features a higher weight. Meantime, the standard cross-entropy loss achieved the best performance on the CASME II dataset. In our experiments, the Adam optimizer was used by default because it performs better than SDG.

5. Conclusions and Perspective

In this paper, we proposed a Dual-branch Attention Network (Dual-ATME) for MER, which consists of three stages: Data preprocessing, Dual-ATME module and Feature fusion for ME classification. In particular, in our Dual-ATME framework, HARS and AARS were combined to extract identity features. Priori experience-based HARS captures noteworthy information about human-understandable concepts. Meantime, Attention mechanism-based AARS automatically captures the complex abstract hidden information within the data. Finally, the above two components were combined to achieve collaborative enhancement and extract discriminative features for MER effectively. The recognition performance of the model was further improved by adding contrast loss in our joint loss to close the two types of features of the same ME sample. Experimental results showed the superiority of our proposed method for performing MER.
In the future, we will further explore automated ROI selection based on large-scale datasets that may help improve MER performance. Specifically, the selection of ROIs is optimized by drawing on the hotspot regions of the attention mechanism. Moreover, we will focus more on spatio-temporal dynamic information to further learn the dynamic representational features of MEs by using the temporal attention mechanism.

Author Contributions

Conceptualization, methodology, H.Z., S.H. and J.L.; software, validation and data creation, H.Z.; writing—original draft preparation, H.Z.; writing—review and editing, H.Z., J.L., S.H. and S.-J.W.; supervision, S.H., J.L. and S.-J.W.; funding acquisition, H.Z., J.L., S.H. and S.-J.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported, in part, by grants from the National Natural Science Foundation of China (62276118, 62106256, U19B2032, 62276252), in part, by grants from Graduate student scientific research innovation projects in Jiangsu province of China (KYCX22_3853) and, in part, by a grant from the Youth Innovation Promotion Association CAS.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ekman, P. Telling Lies: Clues to Deceit in the Marketplace, Politics and Marriage, revised ed.; WW Norton & Company: New York, NY, USA, 2009. [Google Scholar]
  2. Frank, M.; Herbasz, M.; Sinuk, K.; Keller, A.; Nolan, C. I see how you feel: Training laypeople and professionals to recognize fleeting emotions. In Proceedings of the Annual Meeting of the International Communication Association, New York, NY, USA; 2009; pp. 1–35. Available online: http://www.allacademic.com/meta/p15018index.html (accessed on 27 February 2023).
  3. O’Sullivan, M.; Frank, M.G.; Hurley, C.M.; Tiwana, J. Police lie detection accuracy: The effect of lie scenario. Law Hum. Behav. 2009, 33, 530. [Google Scholar] [CrossRef] [PubMed]
  4. Zhao, G.; Li, X. Automatic micro-expression analysis: Open challenges. Front. Psychol. 2019, 10, 1833. [Google Scholar] [CrossRef] [PubMed]
  5. Li, X.; Pfister, T.; Huang, X.; Zhao, G.; Pietikäinen, M. A spontaneous micro-expression database: Inducement, collection and baseline. In Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition, Shanghai, China, 22–26 April 2013; pp. 1–6. [Google Scholar]
  6. Yan, W.J.; Li, X.; Wang, S.J.; Zhao, G.; Liu, Y.J.; Chen, Y.H.; Fu, X. CASME II: An improved spontaneous micro-expression database and the baseline evaluation. PLoS ONE 2014, 9, e86041. [Google Scholar] [CrossRef] [PubMed]
  7. Davison, A.K.; Lansley, C.; Costen, N.; Tan, K.; Yap, M.H. SAMM: A Spontaneous Micro-Facial Movement Dataset. IEEE Trans. Affect. Comput. 2018, 9, 116–129. [Google Scholar] [CrossRef]
  8. Qu, F.; Wang, S.J.; Yan, W.J.; Li, H.; Wu, S.; Fu, X. CAS(ME)2: A database for spontaneous macro-expression and micro-expression spotting and recognition. IEEE Trans. Affect. Comput. 2017, 9, 424–436. [Google Scholar] [CrossRef]
  9. Li, J.; Dong, Z.; Lu, S.; Wang, S.J.; Yan, W.J.; Ma, Y.; Liu, Y.; Huang, C.; Fu, X. CAS(ME)3: A Third Generation Facial Spontaneous Micro-Expression Database with Depth Information and High Ecological Validity. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 2782–2800. [Google Scholar] [CrossRef] [PubMed]
  10. See, J.; Yap, M.H.; Li, J.; Hong, X.; Wang, S.J. Megc 2019–the second facial micro-expressions grand challenge. In Proceedings of the 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), Lille, France, 14–18 May 2019. [Google Scholar]
  11. Ben, X.; Ren, Y.; Zhang, J.; Wang, S.J.; Kpalma, K.; Meng, W.; Liu, Y.J. Video-based facial micro-expression analysis: A survey of datasets, features and algorithms. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 5826–5846. [Google Scholar] [CrossRef]
  12. Li, X.; Cheng, S.; Li, Y.; Behzad, M.; Shen, J.; Zafeiriou, S.; Pantic, M.; Zhao, G. 4DME: A Spontaneous 4D Micro-Expression Dataset With Multimodalities. IEEE Trans. Affect. Comput. 2022, 1–18. [Google Scholar] [CrossRef]
  13. Zhao, G.; Pietikainen, M. Dynamic texture recognition using local binary patterns with an application to facial expressions. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 915–928. [Google Scholar] [CrossRef]
  14. Liong, S.T.; See, J.; Phan, R.C.W.; Wong, K.; Tan, S.W. Hybrid facial regions extraction for micro-expression recognition system. J. Signal Process. Syst. 2018, 90, 601–617. [Google Scholar] [CrossRef]
  15. Huang, X.; Zhao, G.; Hong, X.; Zheng, W.; Pietikäinen, M. Spontaneous facial micro-expression analysis using spatiotemporal completed local quantized patterns. Neurocomputing 2016, 175, 564–578. [Google Scholar] [CrossRef]
  16. Huang, X.; Wang, S.J.; Liu, X.; Zhao, G.; Feng, X.; Pietikäinen, M. Discriminative spatiotemporal local binary pattern with revisited integral projection for spontaneous facial micro-expression recognition. IEEE Trans. Affect. Comput. 2017, 10, 32–47. [Google Scholar] [CrossRef]
  17. Liu, Y.J.; Zhang, J.K.; Yan, W.J.; Wang, S.J.; Zhao, G.; Fu, X. A main directional mean optical flow feature for spontaneous micro-expression recognition. IEEE Trans. Affect. Comput. 2015, 7, 299–310. [Google Scholar] [CrossRef]
  18. Liong, S.T.; See, J.; Wong, K.; Phan, R.C.W. Less is more: Micro-expression recognition from video using apex frame. Signal Process. Image Commun. 2018, 62, 82–92. [Google Scholar] [CrossRef]
  19. Liong, S.T.; Gan, Y.S.; See, J.; Khor, H.Q.; Huang, Y.C. Shallow triple stream three-dimensional cnn (STSTNet) for micro-expression recognition. In Proceedings of the 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), Lille, France, 14–18 May 2019. [Google Scholar]
  20. Zhou, L.; Mao, Q.; Xue, L. Dual-inception network for cross-database micro-expression recognition. In Proceedings of the 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), Lille, France, 14–18 May 2019. [Google Scholar]
  21. Li, J.; Wang, Y.; See, J.; Liu, W. Micro-expression recognition based on 3D flow convolutional neural network. Pattern Anal. Appl. 2019, 22, 1331–1339. [Google Scholar] [CrossRef]
  22. Xia, Z.; Peng, W.; Khor, H.Q.; Feng, X.; Zhao, G. Revealing the invisible with model and data shrinking for composite-database micro-expression recognition. IEEE Trans. Image Process. 2020, 29, 8590–8605. [Google Scholar] [CrossRef] [PubMed]
  23. Zhao, S.; Tao, H.; Zhang, Y.; Xu, T.; Zhang, K.; Hao, Z.; Chen, E. A two-stage 3D CNN based learning method for spontaneous micro-expression recognition. Neurocomputing 2021, 448, 276–289. [Google Scholar] [CrossRef]
  24. Song, B.; Zong, Y.; Li, K.; Zhu, J.; Shi, J.; Zhao, L. Cross-Database Micro-Expression Recognition Based on a Dual-Stream Convolutional Neural Network. IEEE Access 2022, 10, 66227–66237. [Google Scholar] [CrossRef]
  25. Zhu, J.; Zong, Y.; Chang, H.; Xiao, Y.; Zhao, L. A Sparse-Based Transformer Network With Associated Spatiotemporal Feature for Micro-Expression Recognition. IEEE Signal Process. Lett. 2022, 29, 2073–2077. [Google Scholar] [CrossRef]
  26. Chaudhry, R.; Ravichandran, A.; Hager, G.; Vidal, R. Histograms of oriented optical flow and binet-cauchy kernels on nonlinear dynamical systems for the recognition of human actions. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 1932–1939. [Google Scholar]
  27. Patel, D.; Hong, X.; Zhao, G. Selective deep features for micro-expression recognition. In Proceedings of the 2016 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 2258–2263. [Google Scholar]
  28. Khor, H.Q.; See, J.; Phan, R.C.W.; Lin, W. Enriched long-term recurrent convolutional network for facial micro-expression recognition. In Proceedings of the 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), Xi’an, China, 15–19 May 2018; pp. 667–674. [Google Scholar]
  29. Gan, Y.S.; Liong, S.T.; Yau, W.C.; Huang, Y.C.; Tan, L.K. OFF-ApexNet on micro-expression recognition system. Signal Process. Image Commun. 2019, 74, 129–139. [Google Scholar] [CrossRef]
  30. Khor, H.Q.; See, J.; Liong, S.T.; Phan, R.C.; Lin, W. Dual-stream shallow networks for facial micro-expression recognition. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 36–40. [Google Scholar]
  31. Xia, Z.; Hong, X.; Gao, X.; Feng, X.; Zhao, G. Spatiotemporal recurrent convolutional networks for recognizing spontaneous micro-expressions. IEEE Trans. Multimed. 2019, 22, 626–640. [Google Scholar] [CrossRef]
  32. Zhou, L.; Mao, Q.; Huang, X.; Zhang, F.; Zhang, Z. Feature refinement: An expression-specific feature learning and fusion method for micro-expression recognition. Pattern Recognit. 2022, 122, 108275. [Google Scholar] [CrossRef]
  33. Li, J.; Wang, T.; Wang, S.J. Facial Micro-Expression Recognition Based on Deep Local-Holistic Network. Appl. Sci. 2022, 12, 4643. [Google Scholar] [CrossRef]
  34. Itti, L.; Koch, C.; Niebur, E. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 1254–1259. [Google Scholar] [CrossRef]
  35. Corbetta, M.; Shulman, G.L. Control of goal-directed and stimulus-driven attention in the brain. Nat. Rev. Neurosci. 2002, 3, 201–215. [Google Scholar] [CrossRef] [PubMed]
  36. Cao, C.; Liu, X.; Yang, Y.; Yu, Y.; Wang, J.; Wang, Z.; Huang, Y.; Wang, L.; Huang, C.; Xu, W.; et al. Look and think twice: Capturing top-down visual attention with feedback convolutional neural networks. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 2956–2964. [Google Scholar]
  37. Sønderby, S.K.; Sønderby, C.K.; Maaløe, L.; Winther, O. Recurrent spatial transformer networks. arXiv 2015, arXiv:1509.05329. [Google Scholar]
  38. Zhao, S.; Tang, H.; Liu, S.; Zhang, Y.; Wang, H.; Xu, T.; Chen, E.; Guan, C. ME-PLAN: A deep prototypical learning with local attention network for dynamic micro-expression recognition. Neural Netw. 2022, 153, 427–443. [Google Scholar] [CrossRef] [PubMed]
  39. Rodriguez, P.; Velazquez, D.; Cucurull, G.; Gonfaus, J.M.; Roca, F.X.; Gonzalez, J. Pay attention to the activations: A modular attention mechanism for fine-grained image recognition. IEEE Trans. Multimed. 2019, 22, 502–514. [Google Scholar] [CrossRef]
  40. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual attention network for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3156–3164. [Google Scholar]
  41. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  42. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  43. Su, Y.; Zhang, J.; Liu, J.; Zhai, G. Key facial components guided micro-expression recognition based on first & second-order motion. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Shenzhen, China, 5–9 July 2021; pp. 1–6. [Google Scholar]
  44. Li, H.; Sui, M.; Zhu, Z.; Zhao, F. MMNet: Muscle motion-guided network for micro-expression recognition. arXiv 2022, arXiv:2201.05297. [Google Scholar]
  45. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  46. Chen, B.; Zhang, Z.; Liu, N.; Tan, Y.; Liu, X.; Chen, T. Spatiotemporal convolutional neural network with convolutional block attention module for micro-expression recognition. Information 2020, 11, 380. [Google Scholar] [CrossRef]
  47. Gajjala, V.R.; Reddy, S.P.T.; Mukherjee, S.; Dubey, S.R. MERANet: Facial micro-expression recognition using 3D residual attention network. In Proceedings of the Twelfth Indian Conference on Computer Vision, Graphics and Image Processing, Jodhpur, India, 19–22 December 2021; pp. 1–10. [Google Scholar]
  48. Wang, Y.; Huang, Y.; Liu, C.; Gu, X.; Yang, D.; Wang, S.; Zhang, B. Micro expression recognition via dual-stream spatiotemporal attention network. J. Healthc. Eng. 2021, 2021, 7799100. [Google Scholar] [CrossRef]
  49. Song, B.; Li, K.; Zong, Y.; Zhu, J.; Zheng, W.; Shi, J.; Zhao, L. Recognizing spontaneous micro-expression using a three-stream convolutional neural network. IEEE Access 2019, 7, 184537–184551. [Google Scholar] [CrossRef]
  50. King, D.E. Dlib-ml: A machine learning toolkit. J. Mach. Learn. Res. 2009, 10, 1755–1758. [Google Scholar]
  51. Yan, W.J.; Wang, S.J.; Chen, Y.H.; Zhao, G.; Fu, X. Quantifying micro-expressions with constraint local model and local binary pattern. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 296–305. [Google Scholar]
  52. Li, J.; Soladie, C.; Seguier, R. Local Temporal Pattern and Data Augmentation for Micro-Expression Spotting. IEEE Trans. Affect. Comput. 2020, 14, 811–822. [Google Scholar] [CrossRef]
  53. Horn, B.K.; Schunck, B.G. Determining optical flow. Artif. Intell. 1981, 17, 185–203. [Google Scholar] [CrossRef]
  54. Zach, C.; Pock, T.; Bischof, H. A duality based approach for realtime tv-l 1 optical flow. In Proceedings of the Joint Pattern Recognition Symposium, Heidelberg, Germany, 12–14 September 2007; pp. 214–223. [Google Scholar]
  55. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  56. Melacci, S.; Sarti, L.; Maggini, M.; Bianchini, M. A neural network approach to similarity learning. In Proceedings of the IAPR Workshop on Artificial Neural Networks in Pattern Recognition, Paris, France, 2–4 July 2008; pp. 133–136. [Google Scholar]
  57. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  58. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  59. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
Figure 1. The overview of our proposed Dual-ATME.
Figure 1. The overview of our proposed Dual-ATME.
Entropy 25 00460 g001
Figure 2. Face cropping and HARS based on 68 detected facial landmarks. Specifically, the eyebrows and the mouth regions were selected for further ME local feature learning.
Figure 2. Face cropping and HARS based on 68 detected facial landmarks. Specifically, the eyebrows and the mouth regions were selected for further ME local feature learning.
Entropy 25 00460 g002
Figure 3. AU statistics for the ME sample of the MECG2019 dataset.
Figure 3. AU statistics for the ME sample of the MECG2019 dataset.
Entropy 25 00460 g003
Figure 4. Bi-Inception block in our Dual-ATME.
Figure 4. Bi-Inception block in our Dual-ATME.
Entropy 25 00460 g004
Figure 5. The overview of CBAM. The module has two sub-modules: Channel Attention Module (CAM) and Spatial Attention Module (SAM).
Figure 5. The overview of CBAM. The module has two sub-modules: Channel Attention Module (CAM) and Spatial Attention Module (SAM).
Entropy 25 00460 g005
Figure 6. Confusion matrices for MER with Dual-ATME on CASME II, SAMM, SMIC and the MEGC2019-CD datasets.
Figure 6. Confusion matrices for MER with Dual-ATME on CASME II, SAMM, SMIC and the MEGC2019-CD datasets.
Entropy 25 00460 g006
Figure 7. Optical flow (x) and Attention map (V) visualization in our Dual-ATME of the whole face and puzzled conterpart in CASME II, SAMM, SMIC datasets. (Emotion type: Positive, AU12).
Figure 7. Optical flow (x) and Attention map (V) visualization in our Dual-ATME of the whole face and puzzled conterpart in CASME II, SAMM, SMIC datasets. (Emotion type: Positive, AU12).
Entropy 25 00460 g007
Table 1. Ablation studies for the proposed modules of our method on the MEGC2019-CD dataset, including HARS and AARS and their combined version, i.e., our Dual-ATME. Note that w/ is an abbreviation for with and A. is an abbreviation for Attention. The best results are highlighted in bold.
Table 1. Ablation studies for the proposed modules of our method on the MEGC2019-CD dataset, including HARS and AARS and their combined version, i.e., our Dual-ATME. Note that w/ is an abbreviation for with and A. is an abbreviation for Attention. The best results are highlighted in bold.
ModuleMEGC2019-CDCASME IISMICSAMM
HARSHARS (w/ A.)AARSUARUF1UARUF1UARF1UARUF1
0.6450.6480.7760.7850.5650.5630.5720.576
0.6330.6320.7680.7960.5330.5350.5030.516
0.6620.6620.7770.7840.6190.6140.5150.527
0.6800.6790.7510.7650.6580.6460.5380.562
Table 2. The emotion categories of MEGC2019-CD.
Table 2. The emotion categories of MEGC2019-CD.
DatasetsNegativePositiveSurpriseTotal
SMIC [5]705143164
CASME II [6]883225145
SAMM [7]922615133
MEGC19-CD (In total) [10]25010983442
Table 3. Performance comparisons among different methods on the MEGC2019-CD dataset. The best results are highlighted in bold. All the results of the SOTA methods were obtained by reproducing the experiments.
Table 3. Performance comparisons among different methods on the MEGC2019-CD dataset. The best results are highlighted in bold. All the results of the SOTA methods were obtained by reproducing the experiments.
MethodsMEGC2019-CDCASME IISMICSAMM
AccUARUF1AccUARUF1AccUARUF1AccUARUF1
LBP-TOP [13]0.6430.5750.5860.7860.7160.7230.5550.5350.5440.5940.4340.436
Bi-WOOF [18]0.6610.5930.6040.7730.6980.7130.5920.5740.5800.6240.4390.443
ResNet18 [58]0.6430.5750.5860.7860.7160.7230.5550.5350.5440.5940.4340.436
STSTNet [19]0.6880.6100.6240.8210.7450.7690.5430.5290.5320.7120.5050.531
Dual-Incep [20]0.6800.6310.6290.8140.7540.7740.5750.5710.5710.6490.4930.496
RCN_a [22]0.6810.6350.6340.8340.8040.8060.5670.5580.5560.6540.5000.502
RCN_w [22]0.6610.5900.6000.7580.6810.7060.5670.5520.5540.6690.4790.489
RCN_c [22]0.6810.5980.6160.7790.7080.7370.5730.5530.5580.7060.4790.503
RCN_f [22]0.6670.5950.6070.7720.6980.7270.5610.5450.5470.6840.4870.499
KFC-MER [43]0.3130.2350.2550.2760.2200.2290.3450.2510.2830.3160.2460.240
MMNet [44]0.6010.5140.5280.7660.6990.7190.4570.4380.4410.5940.3420.326
Dual-ATME0.7200.6800.6790.8170.7510.7650.6460.6580.6460.7140.5380.562
Table 4. Ablation studies for the different combinations of face areas on the MEGC2019-CD dataset. Here, E., F., M., N. are the abbreviations for Eyebrows, Cheeks, Mouth and Nose, respectively. E.&N. is an abbreviation for the combination of Eyebrows and Nose. The best results are highlighted in bold.
Table 4. Ablation studies for the different combinations of face areas on the MEGC2019-CD dataset. Here, E., F., M., N. are the abbreviations for Eyebrows, Cheeks, Mouth and Nose, respectively. E.&N. is an abbreviation for the combination of Eyebrows and Nose. The best results are highlighted in bold.
CombinationsMEGC2019-CDCASME IISMICSAMM
UARUF1UARUF1UARUF1UARUF1
E.&N.0.6300.6310.7740.7810.5570.5490.4870.503
E.&M.0.6800.6790.7510.7650.6580.6460.5380.562
F.&N.0.6240.6310.7670.7720.5270.5300.5570.566
F.&M.0.6220.6290.7600.7660.5420.5450.5260.541
Full-face0.6630.6700.8100.8280.5910.5910.5430.553
Table 5. Ablation studies for the different values of λ (represent the balance factor for contrastive loss L c o n ) on the MEGC2019-CD dataset. The best results are highlighted in bold.
Table 5. Ablation studies for the different values of λ (represent the balance factor for contrastive loss L c o n ) on the MEGC2019-CD dataset. The best results are highlighted in bold.
λ MEGC2019-CDCASME IISMICSAMM
UARUF1UARUF1UARUF1UARUF1
00.6470.6460.7490.7580.6020.6010.5220.518
0.00010.6680.6660.7410.7550.6370.6280.5460.559
0.0010.6680.6680.7450.7620.6350.6230.5310.547
0.010.6800.6790.7510.7650.6580.6460.5380.562
0.10.6490.6510.7880.7940.5960.5940.4980.503
Table 6. Ablation studies for the different loss functions and optimizers on the MEGC2019-CD dataset. Note that w-CE denotes the weighted Cross-Entropy loss. The best results are highlighted in bold.
Table 6. Ablation studies for the different loss functions and optimizers on the MEGC2019-CD dataset. Note that w-CE denotes the weighted Cross-Entropy loss. The best results are highlighted in bold.
LossOptimizerMEGC2019-CDCASME IISMICSAMM
UARUF1UARUF1UARUF1UARUF1
CESGD0.6440.6420.7650.7830.5910.5810.5040.516
CEAdam0.6730.6710.7480.7680.6430.6340.5140.527
w-CEAdam0.6680.6660.7510.7690.6430.6340.5200.533
FocalAdam0.6800.6790.7510.7650.6580.6460.5380.562
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, H.; Huang, S.; Li, J.; Wang, S.-J. Dual-ATME: Dual-Branch Attention Network for Micro-Expression Recognition. Entropy 2023, 25, 460. https://doi.org/10.3390/e25030460

AMA Style

Zhou H, Huang S, Li J, Wang S-J. Dual-ATME: Dual-Branch Attention Network for Micro-Expression Recognition. Entropy. 2023; 25(3):460. https://doi.org/10.3390/e25030460

Chicago/Turabian Style

Zhou, Haoliang, Shucheng Huang, Jingting Li, and Su-Jing Wang. 2023. "Dual-ATME: Dual-Branch Attention Network for Micro-Expression Recognition" Entropy 25, no. 3: 460. https://doi.org/10.3390/e25030460

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop