Next Article in Journal
A Fuzzy-Based Fast Feature Selection Using Divide and Conquer Technique in Huge Dimension Dataset
Previous Article in Journal
Unified Theory of Zeta-Functions Allied to Epstein Zeta-Functions and Associated with Maass Forms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

TKRM: Learning a Transfer Kernel Regression Model for Cross-Database Micro-Expression Recognition

1
School of Electronics and Information, Northwestern Polytechnical University, Xi’an 710072, China
2
School of Biological Science and Medicial Engineering, Southeast University, Nanjing 210096, China
3
College of Information Engineering, Yancheng Institute of Technology, Yancheng 224051, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(4), 918; https://doi.org/10.3390/math11040918
Submission received: 21 January 2023 / Revised: 9 February 2023 / Accepted: 9 February 2023 / Published: 11 February 2023
(This article belongs to the Topic Machine and Deep Learning)

Abstract

:
Cross-database micro-expression recognition (MER) is a more challenging task than the conventional one because its labeled training (source) and unlabeled testing (target) micro-expression (ME) samples are from different databases. In this circumstance, a large feature-distribution gap may exist between the source and target ME samples due to the different sample sources, which decreases the recognition performance of existing MER methods. In this paper, we focus on this challenging task by proposing a simple yet effective method called the transfer kernel regression model (TKRM). The basic idea of TKRM is to find an ME-discriminative, database-invariant and common reproduced kernel Hilbert space (RKHS) to bridge MEs belonging to different databases. For this purpose, TKRM has the ME discriminative ability of learning a kernel mapping operator to generate an RKHS and build the relationship between the kernelized ME features and labels in such RKHS. Meanwhile, an additional novel regularization term called target sample reconstruction (TSR) is also designed to benefit kernel mapping operator learning by improving the database-invariant ability of TKRM while preserving the ME-discriminative one. To evaluate the proposed TKRM method, we carried out extensive cross-database MER experiments on widely used micro-expression databases, including CASME II and SMIC. Experimental results obtained proved that the proposed TKRM method is indeed superior to recent state-of-the-art domain adaptation methods for cross-database MER.

1. Introduction

Micro-expression (ME) is defined as a set of brief, subtle and unconscious facial movements, whose duration boundary is often within half a second [1,2,3]. It has been demonstrated to be an important cue to reveal people’s hidden intentions and, hence, it can be widely applied in police interrogation [4], national and public security [5], psychology and healthcare [6], and deceit analysis [7]. However, unlike conventional facial expressions, it is very difficult for individuals to recognize MEs even once they have gained professional training due to the brief and subtle characteristics of MEs. Therefore, it is of value to investigate the problem of automatic ME recognition (MER) to enable the computer to understand the emotion states from micro-expression video clips, helping people better understand MEs.
Recently, researchers have made great progress in the research of MER. For example, lots of publicly available ME databases, e.g., CASME series [8,9,10,11], SMIC series [12,13], and SAMM [14], have been released, which support the advancement of MER research. In order to effectively describe MEs, researchers have also designed a number of handcrafted spatiotemporal descriptors, e.g., local binary pattern from three orthogonal planes (LBP-TOP) [12,15], spatial temporal LBP with integral projection (STLBP-IP) [16], and a histogram of image gradient-TOP (HIGO-TOP) [17], as well as feature-enhancement methods [18]. Recently, inspired by the success of deep neural networks in computer vision tasks, deep learning methods [19,20,21] have also been used to learn discriminative features for recognizing MEs.
Although MER research has experienced remarkable progress, it is also worth mentioning that most existing MER methods that perform well were designed and evaluated under an ideal assumption that the training and test ME samples both belong to the same database. In other words, these ME samples were often recorded by the same camera and under the same environment; hence, their features can be viewed as abiding by the same probability distributions. In practical applications, however, this assumption may easily be broken, which leads to the feature distribution mismatch between the training and testing ME samples. Hence, the performance of existing MER methods would be significantly degraded. To address this issue, several researchers have shifted their focus to a more challenging but interesting MER task, i.e., cross-database MER [22], in which the training and testing ME samples derive from different databases. Recently, numerous methods have been proposed to manage this MER task and achieve promising performance, e.g., data augmentation [23], transfer subspace learning [22], and deep transfer learning methods [24].
In this paper, we also focus on this challenging but interesting topic and try to approach it from the angle of kernel learning. By utilizing the nonlinear mapping ability of kernel learning, we propose a novel transfer subspace learning method called the transfer kernel regression model (TKRM) to cope with the cross-database MER tasks, an overview of which is shown in Figure 1. In TKRM, we first build a simple and straightforward kernelized regression model to seek a reproduced kernel Hilbert space (RKHS) to build the relationship between the source ME features and label vectors. Meanwhile, a novel regularization term called target sample reconstruction (TSR) is designed to further enhance the TKRM model so that it is robust to the database variance existing between the source and target ME databases and also preserves the ME’s discriminative ability. Hence, the proposed TKRM model is also applicable in the task of accurately predicting the ME labels of target samples.
In summary, the major contributions of this paper are threefold:
  • We propose a novel transfer subspace learning model called TKRM to deal with the cross-database MER problem. The major advantage of the proposed TKRM is seeking a common reproduced kernel Hilbert space for the distribution alignment between the ME samples from different databases.
  • So that TKRM possesses the database-invariant ability, a well-designed regularization term called TSR is designed, which allows any target ME sample to be reconstructed by a few source ones.
  • A large number of comprehensive experiments are conducted to compare the proposed TKRM method and recent state-of-the-art transfer subspace learning methods. The results verify the effectiveness and superiority of TKRM in coping with the cross-database MER problem.
The rest of the manuscript is organized as follows: Section 2 reviews the recent progress in cross-database MER methods, including both transfer subspace learning and deep transfer learning methods. Section 3 provides the details of the proposed TKRM model, including how to build, optimize, and use the TKRM for cross-database MER tasks. In Section 4, we desribe the extensive cross-database MER experiments that were conducted to evaluate the proposed TKRM model and we also provide in-depth discussion. Finally, the conclusions and directions for future work are provided in Section 5.

2. Related Work

The earliest work of cross-database MER may be traced to [22], in which Zong et al. designed extensive cross-database MER tasks based on two widely used spontaneous ME databases, including CASME II [9] and SMIC [13]. They also proposed a straightforward transfer subspace learning method called target sample re-generator (TSRG) to remove the feature distribution mismatch between the training (source) and testing (target) ME samples. Subsequently, Zong et al. [25] further extended TSRG to a more generalized framework called domain regeneration (DR). Recently, in order to advance the research of cross-database MER and provide a standard platform to evaluate emerging methods, a cross-database MER benchmark was presented in the works of [23,26]. Subsequently, researchers have shifted their focus to achieving better performance in the tasks of cross-database MER. To achieve this goal, some researchers continue to develop the transfer subspace learning methods by considering the ME-aware discriminative clues. For example, in the works of [26,27,28], the authors attempted to learn more discriminative features from the facial local region cues to better describe MEs, while considering the database-invariant performance.
On the other hand, several researchers have tried to manage the problem of cross-database MER using deep transfer learning methods. For example, Xia et al. [24] designed a style aggregated and attention transfer framework (SA-AT) consisting of two stages enabling the ResNet model learned from the source ME samples to also be applicable to target ME. Li et al. [29] proposed a novel deep domain adaption method called the discriminative region transfer network (DRTN), which uses an adversarial-based adaptation structure and attention mechanism to learn the database-invariant ME features from discriminative facial regions. More recently, Song et al. [30] designed a dual-stream convolutional neural network (DSCNN) to cope with cross-database MER tasks. The major novelty of DSCNN is that the temporal dynamic and salient cues in ME samples are simultaneously considered to ensure that the network can learn more discriminative and robust features.

3. TKRM for Cross-Database MER

In this section, we first formulate the TKRM model in detail. Then, we design an efficient algorithm to learn the optimal parameters of the proposed TKRM model. Finally, we present how to use the TKRM model to predict the target ME samples that are different from the source ones.

3.1. Formulation of TKRM

Assume that we are given a labeled source ME database and their corresponding feature matrix is denoted by X s R d × n s , where d is the dimension of the feature vector used for describing ME, and n s is the source ME sample number. Let L s R c × n s be the ME label matrix, where c is the number of micro-expression categories. The ith column of L s , denoted by l s i , is a c-dimensional vector whose jth entry is set as 1, while the others are fixed at 0 if this ME sample’s label is jth ME. Accordingly, we are also given an unlabeled target ME database, whose feature matrix is denoted by X t R d × n t , where n t is the the target sample number.
As described previously, the basic idea of the proposed TKRM is seeking a common RKHS for both source and target ME samples by resorting to a kernel mapping operator to bridge them such that their feature distribution mismatch can be removed, which is shown in Figure 1. Following this idea, we first build the relationship between the source ME features and their corresponding label vectors via a simple kernel regression model, whose optimization problem can be formulated as follows:
min ϕ ( U ) L s ϕ ( U ) T ϕ ( X s ) F 2 ,
where ϕ ( · ) is a kernel mapping operator, and ϕ ( U ) and ϕ ( X s ) denote the projection matrix and source ME features in the reproduced kernel Hilbert space (RKHS) generated by ϕ ( · ) , respectively. From Equation (1), it is clear to see that the proposed TKRM can own the ME discriminative ability and we can easily predict the ME category of the source ME feature vector based on the projection matrix ϕ ( U ) in the RKHS.
Meanwhile, according to Figure 1, the ϕ ( U ) learned by our TKRM model should also be applicable to predicting the target ME samples’ labels. In other words, the RKHS should be a common space for target ME samples as well as the source ones. Unfortunately, in the case of cross-database MER, a large feature distribution gap exists between the source and target ME databases. Therefore, we hope to find a common RKHS, in which the feature distributions of the source and target domains are close, i.e., samples of two domains can be bridged by a linear combination. Since the source data have emotion labels, reconstructing each target sample with source data can not only accurately supervise the emotion prediction of the target samples, but also effectively align the feature distributions between two domains. To this end, we also map the target ME features onto the same RKHS, which can be denoted by ϕ ( X t ) . Then, we design a novel regularization term called TSR for the TKRM model to help modify the RKHS so that it is also suitable for the target ME samples. The objective function of TSR can be formulated as follows:
f ( ϕ ( U ) ) = ϕ ( U ) T ϕ ( X t ) ϕ ( U ) T ϕ ( X s ) W F 2 + λ i = 1 n t w i 1 ,
where W = [ w 1 , , w n t ] n s × n t is a coefficient matrix, · 1 denotes the L 1 norm of a vector and λ is a trade-off parameter to control the balance between the L 1 norm and reconstruction terms. Note that, by minimizing the objective function of TSR in Equation (2), the learned projection matrix ϕ ( U ) in the common RKHS forces the target ME samples to be a linear combination of several source ME ones, which indicates the projected target ME samples will be assimilated into the source ones. Furthermore, the sparse property of the L 1 regularization term also effectively reduces the dimension of the parameter space to avoid over-fitting in the optimization process. Therefore, it is a good choice to jointly minimize the objective function in Equation (1) and the one shown in Equation (2); in this way, we will arrive at the optimization problem of the proposed TKRM, which is expressed as follows:
min ϕ ( U ) , W L s ϕ ( U ) T ϕ ( X s ) F 2 + λ 1 ϕ ( U ) T ϕ ( X t ) ϕ ( U ) T ϕ ( X s ) W F 2 + λ 2 i = 1 n t w i 1 ,
where λ 1 and λ 2 = λ 1 × λ are the trade-off parameters controlling the balance among three terms in the objective function of TKRM.
However, it should be pointed out that it is difficult to solve the current version of the optimization problem for TKRM because the row dimension of the kernel mapping operator ϕ ( U ) is infinite. Using the kernel trick [31], the kernel function can be used to convert the optimization problem to be a solvable one. Specifically, let ϕ ( U ) = [ ϕ ( X s ) , ϕ ( X t ) ] P , where P R ( n s + n t ) × c is a reconstruction of the coefficient matrix used to enforce ϕ ( U ) to be linearly combined by the source ME features ϕ ( X s ) and target ME features ϕ ( X t ) in the RKHS. By substituting this equation into Equation (3), we are able to arrive at the new formulation of the optimization problem for TKRM, as follows:
min P , W L s P T K s F 2 + λ 1 P T K t P T K s W F 2 + λ 2 i = 1 n t w i 1 ,
where K s = [ K s s T , K t s T ] T and K t = [ K s t T , K t t T ] T . Among them, K s s = ϕ ( X s ) T ϕ ( X s ) , K t s = ϕ ( X t ) T ϕ ( X s ) , K s t = ϕ ( X s ) T ϕ ( X t ) , and K t t = ϕ ( X t ) T ϕ ( X t ) , and can be directly calculated by the kernel function, e.g., Gaussian, polynomial, and linear ones.
In addition, we would like to further regularize the model parameter matrix P of TKRM by adding an L 1 norm term with respect to P to the objective function, which results in a sparse column vector for P . This was inspired by the works of [18,23], in which a sparse coefficient matrix led to a more effective reconstruction in the RKHS and helped to avoid overfitting for optimization. Consequently, the eventual optimization problem of the proposed TKRM model can be written as follows:
min P , W L s P T K s F 2 + λ 1 P T K t P T K s W F 2 + λ 2 i = 1 n t w i 1 + λ 3 P 1 .
Note that similar to λ 1 and λ 2 , λ 3 is a trade-off parameter as well. P 1 = i = 1 c p i 1 and p i is the ith column of P .

3.2. Optimization of TKRM

We would like to use the alternated direction method (ADM) [32] to learn the optimal solution of the TKRM model, i.e., alternatively updating P and W until convergence. Specifically, we initialize W and repeat the following three major steps:
  • Fix W and update P . In this step, the original optimization problem will be reduced to:
    min P L s P T K s F 2 + λ 1 P T Δ K s t F 2 + λ 3 P 1 ,
    where Δ K s t = K t K s W . The inexact augmented Lagrangian multiplier (IALM) [33] method can be adopted to efficiently learn the optimal solution. Concretely, we first introduce an additional variable Q satisfying P = Q . Then, the original optimization problem of TKRM, which is an unconstrained one, can be converted to a constrained one as follows:
    min P , Q L s Q T K s F 2 + λ 1 Q T Δ K s t F 2 + λ 3 P 1 , s . t . P = Q .
    Subsequently, the Lagrangian function for the above optimization problem can be written as follows:
    L ( P , Q , W , T , μ ) = L s Q T K s F 2 + λ 1 Q T Δ K s t F 2 + λ 3 P 1 + T T ( P Q ) + μ 2 P Q F 2 ,
    where T is the Lagrangian multiplier matrix and μ is the trade-off parameter.
    By alternatively minimizing the Lagrangian function with respect to the variables until convergence, the optimal P will be obtained. The detailed updating procedures are as follows:
    (a)
    Fix P , T , and μ , and update Q ; the optimization problem with respect to Q can be reformulated as:
    min Q L s Q T K s F 2 + λ 1 Q T Δ K s t F 2 + T T ( P Q ) + μ 2 P Q F 2 ,
    which has a closed-form solution, as follows:
    Q = ( 2 K s K s T + λ 1 Δ K s t Δ K s t T + μ I ) 1 ( 2 K s L s T + T + μ P ) ,
    where I is an ( n s + n t ) × ( n s + n t ) identity matrix.
    (b)
    Fix Q , T , and μ , and update P . In this step, the optimization problem with respect to P is:
    min P λ 3 P 1 + T T ( P Q ) + μ 2 P Q F 2 ,
    which can be rewritten as the following formulation:
    min P λ 3 μ P 1 + 1 2 P ( Q T μ ) F 2 .
    Referring to Step 2 of optimization procedures in [22], the solution of Equation (12) can be determined according to the following criterion, i.e.,
    (1)
    P i j = ( P i j Q i j μ ) λ 3 μ , if P i j Q i j μ > λ 3 μ ,
    (2)
    P i j = ( P i j Q i j μ ) + λ 3 μ , if P i j Q i j μ < λ 3 μ ,
    (3)
    P i j = 0 , otherwise,
    where P i j , Q i j , and T i j are the entry in the ith row and jth column of matrices P , Q , and T , respectively.
    (c)
    Update T and μ : T = T + μ ( P Q ) and μ = min { ρ μ , μ m a x } , where ρ is a scaled parameter greater than 1 and μ m a x is a preset maximal value for μ .
    (d)
    Check convergence: P Q F < ϵ or the interaction reaches maximal value I t e r m a x .
  • Fix P and update W : In this step, the original optimization problem is:
    min w i P T k t i P T K s w i F 2 + λ 2 λ 1 w i 1 ,
    where k t i is the ith column of K t corresponding to the kernel function of the ith target ME sample. This is a standard Lasso problem and can be efficiently solved by a number of typical algorithms, e.g., coordinate descent.
  • Check convergence or reach the preset maximal iterations.

3.3. Convergence Analysis for the Optimization of TKRM

As Section 3.2 shows, we proposed an efficient iterative algorithm to learn the optimal solutions of the proposed TKRM. In this section, we would like to analyze the convergence of this algorithm from a mathematical angle. Specifically, it is clear to see that the optimization problem for TKRM in Equation (5), whose objective function is denoted by f ( P , W ) , is divided into two minimization subproblems, i.e., Equations (6) and (13), whose objective functions are denoted by f ( P , W ^ ) and f ( P ^ , W ) , in the proposed optimization algorithm, where W ^ and P ^ are constants learned by their corresponding minimization subproblems in a previous round of optimization. Then, since all the terms in the objective function of TKRM are L 1 and Frobenius norms, which are continuous and lower-bounded, the objective function value of TKRM in Equation (5), f ( P , W ) , would sequentially decrease if these two minimization subproblems, Equations (6) and (13), were iteratively optimized. Consequently, the convergence of the proposed iterative algorithm for learning the optimal solutions of the proposed TKRM can be guaranteed by the following inequation:
f ( P ^ t , W ^ t ) < f ( P ^ t , W ^ t 1 ) < f ( P ^ t 1 , W ^ t 1 ) ,
where t denotes the updating step index in the optimization algorithm.
This inequation holds obviously in the proposed iterative algorithm, which can be supported by the following two conclusions drawn from the divided minimization subproblem: (1) given a constant W ^ , the minimization subproblem in Equation (6) would undoubtedly result in f ( P ^ t , W ^ ) < f ( P ^ t 1 , W ^ ) at the tth step of the proposed iterative algorithm; (2) similarly, based on the minimization subproblem in Equation (13), we can also obtain that f ( P ^ , W ^ t ) < f ( P ^ , W ^ t 1 ) for a constant P ^ at the tth step.

3.4. Target Sample’s ME Label Prediction

Suppose the optimal solution of the TKRM model is denoted by P ^ . Then, we can conveniently predict the ME labels of the target samples, even though the source of these samples is quite different from that of the training samples. Let x t i be a target ME sample’s feature vector. Thus, we can first calculate its ME label vector, which is obtained with the following optimization problem:
min l t i l t i P ^ T k t i 2 ,
where k t i = [ ϕ ( x t i ) T ϕ ( X s ) , ϕ ( x t i ) T ϕ ( X t ) ] T . It is noted that the inner product in the RKHS associated with k t i must be calculated using the same kernel function as the one chosen in the training stage. Equation (15) is a quadratic programming (QP) problem and can be easily solved with a number of of QP algorithms, including the interior point method.
Finally, according to the optimal label vector l t i , the eventual ME label would be determined as the jth ME label if the jth entry of l t i has a maximal value among all the entries.

4. Experiments

4.1. Micro-Expression Databases and Experiment Setting

In this section, we adopt the CASME II [9] and SMIC [13] micro-expression databases to design cross-database MER tasks to evaluate the proposed TKRM model. The SMIC database includes recordings with high speed (HS), normal visual (VIS), and near-infrared (NIR) cameras, resulting in three subsets, i.e., HS, VIS, and NIR. It is the first publicly available ME database containing diverse ME samples. In all the subsets, the ME samples are divided into three different ME types including P o s i t i v e , N e g a t i v e , and S u r p r i s e . Similarly to the HS subset in SMIC, CASME II is also recorded using a high-speed camera. Each sample of CASME II is assigned one of five ME labels, i.e., H a p p y , D i s g u s t , S u r p r i s e , R e p r e s s i o n , and O t h e r s . Figure 2 provides an example of the above four ME sample sets, which demonstrates the sample difference existing among these databases.
By using any two of the above four types of ME samples to alternatively serve as the source and target, we are able to obtain six different cross-database MER tasks. We denote these tasks as C H , H C , C V , V C , C N , and N C , respectively, where C, H, V, and N are abbreviations for CASME II, HS, VIS, and NIR. Since the source and target samples in cross-database MER require consistent ME labels, we select and relabel parts of samples from CASME II. Specifically, the H a p p y samples are given the P o s i t i v e label, the D i s g u s t and R e g r e s s i o n samples are relabeled as N e g a t i v e , and the label for S u r p r i s e samples remains unchanged. Table 1 shows the new statistics of the relabeled CASME II and SMIC databases used in our experiments.

4.2. Comparison Methods and Performance Metrics

To see whether the proposed TKRM method performs well in coping with cross-database MER tasks, we choose seven state-of-the-art transfer subspace learning methods, including transfer component analysis (TCA) [34], geodesic flow kernel (GFK) [35], subspace alignment (SA) [36], transfer kernel learning (TKL) [37], target sample re-generator (TSRG) [22], domain regeneration in the original feature space with unchanged target domain (DRFS-T) [25], domain regeneration in the label space (DRLS) [25], and region selective transfer regression (RSTR) [26]. To analyze the experimental results, m e a n F 1 - s c o r e and A c c u r a c y are used as the performance metrics. Among them, m e a n F 1 - s c o r e is the primary metric and calculated by 1 c i = 1 c F 1 ( i ) , where F 1 ( i ) is the F1 score associated with ith predicted ME samples. As for the other metric, A c c u r a c y , it is defined as the correct predictions divided by the total testing sample number.
Note that unlike end-to-end deep learning methods, subspace learning methods require the features to describe MEs; hence, we choose uniform LBP-TOP with the parameters R = 3 and P = 8 to serve as the spatiotemporal descriptors to extract the features of MEs in the experiments. In addition, a multi-scale spatial division scheme, including four types of grids, i.e., 1 × 1 , 2 × 2 , 4 × 4 , and 8 × 8 , is used to divide the facial image into 85 facial local blocks before extracting the LBP-TOP. Hence, a supervector consisting of 85 LBP-TOP feature vectors is eventually served to describe each ME sample. As for the parameter setting of transfer subspace learning methods, we follow the benchmark evaluation work in [23,26] and directly take their results for comparison because our experiment setting is identical to theirs. Regarding our TKRM, the kernel function is set as linear. Three trade-off parameters need also to be set for our TKRM. Still following the experiment setting of [23,26], we search for them from a preset parameter interval, i.e., [0.1:0.1:1,2:1:10], and then report the best results when it reaches the best m e a n F 1 - s c o r e for each experiment.

4.3. Experimental Results and Discussion

Table 2 and Table 3 summarize the m e a n F 1 - s c o r e and A c c u r a c y for our KDAL and seven comparison transfer subspace learning methods. Apart from these, it can be seen that we also directly use the SVM classifier to perform the same cross-database MER experiments to serve as the baselines. As Table 2 and Table 3 show, it is easily found that nearly all the transfer subspace learning methods outperformed the baseline method, i.e., SVM without any knowledge transfer operation, which indicates that transfer learning is indeed an effective method to improve the model robustness to the database variance and provides a possible direction to advance the research of MER.
Second, it is noted that the proposed TKRM achieves the best performance in terms of both m e a n F 1 - s c o r e and A c c u r a c y averaged among all six experiments. By further observing the detailed results of six experiments, it can also be obtained that our TKRM also achieved the best m e a n F 1 - s c o r e and A c c u r a c y in three tasks including H C , C V , and N C . These observations indicates that our TKRM can achieve a more promising and satisfactory performance than recent state-of-the art transfer subspace learning methods for cross-database MER tasks.
Finally, it can also be noted that nearly all the transfer subspace learning methods performed better in the experiments between CASME II and VIS than the others. For example, we can find that seven methods, including our TKRM, achieved a m e a n F 1 - s c o r e of over 60.00% in at least one task ( C V , V C , or both). However, they do not all have improved performance in the four remaining tasks. Therefore, it can be concluded that the experiments between CASME II and HS, or between CASME II and NIR, are more challenging cross-database MER tasks. Nevertheless, our TKRM can still achieve a m e a n F 1 - s c o r e of over 50.00% for all four tasks, which demonstrates the superior performance of TKRM over all comparison methods.

4.4. Performance of TKRM with Different Kernel Functions

In the above experiments, our TKRM only uses the linear kernel function and achieved promising performance in cross-database MER tasks. It should be pointed out that the nonlinear kernel function is also suitable for TKRM. To see how the kernel function affects the performance of TKRM, we further carry out several additional experiments by choosing different nonlinear kernel functions for TKRM. Specifically, we choose three tasks as the representatives and set the kernel function as the polynomial, i.e., f ( x , y ) = ( x T y ) d , PolyPlus, i.e., f ( x , y ) = ( x T y + 1 ) d , and Gaussian, i.e., f ( x , y ) = e x y 2 2 t , where f ( · ) is the kernel function, x and y are the feature vectors, and d is the dimension of feature vectors. For polynomial and PolyPlus functions, their parameter d is set to 1.01, 1.05 and 1.1, respectively. For the Gaussian, its parameter t is fixed at 1, 5 and 10, respectively. Table 4 depicts the results of the experiments, which enables us to obtain several interesting observations and conclusions. It is clearly observed that, for most tasks, the linear-kernel-based TKRM obtained the best results in terms of both m e a n F 1 - s c o r e and A c c u r a c y . However, it is still noted that the polynomial kernel with the parameter d = 1.01 achieved the highest m e a n F 1 - s c o r e , reaching 60.02. This indicates that the performance of the proposed TKRM with a suitable nonlinear kernel would achieve a further increase. Nevertheless, it should be pointed out that the performance of the nonlinear version of TKRM varies sensitively with respect to the change in the kernel function’s type and parameter, which can be demonstrated by the comparison between the results achieved by the polynomial and Gaussian functions. Consequently, how to choose the kernel function and search for its optimal parameter in the KDAL model to improve its performance when coping with cross-database MER tasks is still an open question. This point should be investigated in the future.

5. Conclusions and Future Work

In this paper, we propose a new transfer kernel regression model (TKRM), which makes full use of the kernel learning’s powerful nonlinear mapping ability, to deal with the cross-database MER problem. The major advantage of TKRM is minimizing the probability distributions divergence between source and target ME samples in the kernelized feature space with an implicit method of feature reconstruction. Moreover, extensive experiments were carried out to evaluate the effectiveness and superiority of the proposed TKRM. The experimental results proved that our TKRM method can achieved improved performance compared with most recent state-of-the-art transfer subspace learning methods for the cross-database MER problem. In addition, it should be noted that the proposed TKRM method is still not satisfied because its performance in challenging tasks, e.g., tasks between SMIC (NIS) and CASME II, only reaches a low-level mean F1-score of about 50%. Consequently, the ME discriminative ability of TKRM needs to be improved so that the cross-database MER problem can be better addressed. In the future, we will further enhance the proposed TKRM model by absorbing discriminative information learned from the ME-aware cues.

Author Contributions

Conceptualization, methodology, and writing—original draft preparation, Z.C. and C.L.; funding acquisition, software and validation, formal analysis, investigation, resources, data curation, and writing—review and editing, F.Z. and Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by the Major Basic Research Project of the Natural Science Foundation of the Jiangsu Higher Education Institutions under the Grant 19KJA110002, in part by the Natural Science Foundation of China under the Grants 61902064 and 61673108, in part by the Zhishan Young Scholarship of Southeast University, and in part by the Yancheng Institute of Technology High-level Talent Research Initiation Project under the Grant XJR2022001.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: [CASME II] [http://casme.psych.ac.cn/casme/e2] (accessed on 12 September 2022) and [SMIC] [http://www.cse.oulu.fi/SMICDatabase] (accessed on 12 September 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Oh, Y.H.; See, J.; Le Ngo, A.C.; Phan, R.C.W.; Baskaran, V.M. A survey of automatic facial micro-expression analysis: Databases, methods, and challenges. Front. Psychol. 2018, 9, 1128. [Google Scholar] [CrossRef] [PubMed]
  2. Ben, X.; Ren, Y.; Zhang, J.; Wang, S.J.; Kpalma, K.; Meng, W.; Liu, Y.J. Video-based facial micro-expression analysis: A survey of datasets, features and algorithms. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 5826–5846. [Google Scholar] [CrossRef]
  3. Ekman, P.; Friesen, W.V. Facial Action Coding System; Consulting Psychologist Press: Palo Alto, CA, USA, 1978. [Google Scholar]
  4. O’Sullivan, M.; Frank, M.G.; Hurley, C.M.; Tiwana, J. Police lie detection accuracy: The effect of lie scenario. Law Hum. Behav. 2009, 33, 530. [Google Scholar] [CrossRef] [PubMed]
  5. Frank, M.G.; Maccario, C.J.; Govindaraju, V. Behavior and security. In Protecting Airline Passengers in the Age of Terrorism; ABC-CLIO: Santa Barbara, CA, USA, 2009; pp. 86–106. [Google Scholar]
  6. Frank, M.; Herbasz, M.; Sinuk, K.; Keller, A.; Nolan, C. I see how you feel: Training laypeople and professionals to recognize fleeting emotions. In Annual Meeting of the International Communication Association; Sheraton New York: New York, NY, USA, 2009; pp. 1–35. [Google Scholar]
  7. Ekman, P. Lie catching and microexpressions. Philos. Decept. 2009, 1, 5. [Google Scholar]
  8. Yan, W.J.; Wu, Q.; Liu, Y.J.; Wang, S.J.; Fu, X. CASME database: A dataset of spontaneous micro-expressions collected from neutralized faces. In Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Shanghai, China, 22–26 April 2013. [Google Scholar]
  9. Yan, W.J.; Li, X.; Wang, S.J.; Zhao, G.; Liu, Y.J.; Chen, Y.H.; Fu, X. CASME II: An improved spontaneous micro-expression database and the baseline evaluation. PLoS ONE 2014, 9, e86041. [Google Scholar] [CrossRef] [PubMed]
  10. Qu, F.; Wang, S.J.; Yan, W.J.; Li, H.; Wu, S.; Fu, X. CAS (ME) 2: A database for spontaneous macro-expression and micro-expression spotting and recognition. IEEE Trans. Affect. Comput. 2017, 9, 424–436. [Google Scholar] [CrossRef]
  11. Li, J.; Dong, Z.; Lu, S.; Wang, S.J.; Yan, W.J.; Ma, Y.; Liu, Y.; Huang, C.; Fu, X. CAS (ME) 3: A Third Generation Facial Spontaneous Micro-Expression Database with Depth Information and High Ecological Validity. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 2782–2800. [Google Scholar] [CrossRef] [PubMed]
  12. Pfister, T.; Li, X.; Zhao, G.; Pietikäinen, M. Recognising spontaneous facial micro-expressions. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 1449–1456. [Google Scholar]
  13. Li, X.; Pfister, T.; Huang, X.; Zhao, G.; Pietikäinen, M. A spontaneous micro-expression database: Inducement, collection and baseline. In Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Shanghai, China, 22–26 April 2013. [Google Scholar]
  14. Davison, A.K.; Lansley, C.; Costen, N.; Tan, K.; Yap, M.H. Samm: A spontaneous micro-facial movement dataset. IEEE Trans. Affect. Comput. 2016, 9, 116–129. [Google Scholar] [CrossRef]
  15. Zhao, G.; Pietikainen, M. Dynamic texture recognition using local binary patterns with an application to facial expressions. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 915–928. [Google Scholar] [CrossRef]
  16. Huang, X.; Wang, S.J.; Zhao, G.; Piteikainen, M. Facial micro-expression recognition using spatiotemporal local binary pattern with integral projection. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Santiago, Chile, 7–13 December 2015; pp. 1–9. [Google Scholar]
  17. Li, X.; Hong, X.; Moilanen, A.; Huang, X.; Pfister, T.; Zhao, G.; Pietikäinen, M. Towards reading hidden emotions: A comparative study of spontaneous micro-expression spotting and recognition methods. IEEE Trans. Affect. Comput. 2017, 9, 563–577. [Google Scholar] [CrossRef]
  18. Zong, Y.; Huang, X.; Zheng, W.; Cui, Z.; Zhao, G. Learning from hierarchical spatiotemporal descriptors for micro-expression recognition. IEEE Trans. Multimed. 2018, 20, 3160–3172. [Google Scholar] [CrossRef]
  19. Xia, Z.; Hong, X.; Gao, X.; Feng, X.; Zhao, G. Spatiotemporal recurrent convolutional networks for recognizing spontaneous micro-expressions. IEEE Trans. Multimed. 2019, 22, 626–640. [Google Scholar] [CrossRef]
  20. Wei, M.; Zheng, W.; Jiang, X.; Zong, Y.; Lu, C.; Liu, J. A Novel Magnification-Robust Network with Sparse Self-Attention for Micro-expression Recognition. In Proceedings of the 2022 26th International Conference on Pattern Recognition (ICPR), Montreal, QC, Canada, 21–25 August 2022; pp. 1120–1126. [Google Scholar]
  21. Zhu, J.; Zong, Y.; Chang, H.; Xiao, Y.; Zhao, L. A Sparse-Based Transformer Network with Associated Spatiotemporal Feature for Micro-Expression Recognition. IEEE Signal Process. Lett. 2022, 29, 2073–2077. [Google Scholar] [CrossRef]
  22. Zong, Y.; Huang, X.; Zheng, W.; Cui, Z.; Zhao, G. Learning a target sample re-generator for cross-database micro-expression recognition. In Proceedings of the 25th ACM International Conference on Multimedia, Mountain View, CA, USA, 23–27 October 2017; pp. 872–880. [Google Scholar]
  23. Zong, Y.; Zheng, W.; Hong, X.; Tang, C.; Cui, Z.; Zhao, G. Cross-database micro-expression recognition: A benchmark. In Proceedings of the 2019 on International Conference on Multimedia Retrieval, Ottawa, ON, Canada, 10–13 June 2019; pp. 354–363. [Google Scholar]
  24. Xia, Z.; Liang, H.; Hong, X.; Feng, X. Cross-database micro-expression recognition with deep convolutional networks. In Proceedings of the 2019 3rd International Conference on Biometric Engineering and Applications, Stockholm, Sweden, 29–31 May 2019; pp. 56–60. [Google Scholar]
  25. Zong, Y.; Zheng, W.; Huang, X.; Shi, J.; Cui, Z.; Zhao, G. Domain regeneration for cross-database micro-expression recognition. IEEE Trans. Image Process. 2018, 27, 2484–2498. [Google Scholar] [CrossRef]
  26. Zhang, T.; Zong, Y.; Zheng, W.; Chen, C.P.; Hong, X.; Tang, C.; Cui, Z.; Zhao, G. Cross-Database Micro-Expression Recognition: A Benchmark. IEEE Trans. Knowl. Data Eng. 2022, 34, 544–559. [Google Scholar] [CrossRef]
  27. Zong, Y.; Zheng, W.; Cui, Z.; Zhao, G.; Hu, B. Toward bridging microexpressions from different domains. IEEE Trans. Cybern. 2020, 50, 5047–5060. [Google Scholar] [CrossRef]
  28. Jiang, X.; Zong, Y.; Zheng, W.; Liu, J.; Wei, M. Seeking Salient Facial Regions for Cross-Database Micro-Expression Recognition. In Proceedings of the 2022 26th International Conference on Pattern Recognition (ICPR), Montreal, QC, Canada, 21–25 August 2022; pp. 1019–1025. [Google Scholar]
  29. Li, J.; Hu, R.; Mukherjee, M. Discriminative Region Transfer Network for Cross-Database Micro-Expression Recognition. In Proceedings of the ICC 2022-IEEE International Conference on Communications, Seoul, Repulic of Korea, 16–20 May 2022; pp. 5082–5087. [Google Scholar]
  30. Song, B.; Zong, Y.; Li, K.; Zhu, J.; Shi, J.; Zhao, L. Cross-Database Micro-Expression Recognition Based on a Dual-Stream Convolutional Neural Network. IEEE Access 2022, 10, 66227–66237. [Google Scholar] [CrossRef]
  31. Vapnik, V. The Nature of Statistical Learning Theory; Springer Science & Business Media: Berlin, Germany, 1999. [Google Scholar]
  32. Zheng, W. Multi-view facial expression recognition based on group sparse reduced-rank regression. IEEE Trans. Affect. Comput. 2014, 5, 71–85. [Google Scholar] [CrossRef]
  33. Lin, Z.; Liu, R.; Su, Z. Linearized alternating direction method with adaptive penalty for low-rank representation. Adv. Neural Inf. Process. Syst. 2011, 24, 1–9. [Google Scholar]
  34. Pan, S.J.; Tsang, I.W.; Kwok, J.T.; Yang, Q. Domain adaptation via transfer component analysis. IEEE Trans. Neural Netw. 2010, 22, 199–210. [Google Scholar] [CrossRef]
  35. Gong, B.; Shi, Y.; Sha, F.; Grauman, K. Geodesic flow kernel for unsupervised domain adaptation. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2066–2073. [Google Scholar]
  36. Fernando, B.; Habrard, A.; Sebban, M.; Tuytelaars, T. Unsupervised visual domain adaptation using subspace alignment. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2960–2967. [Google Scholar]
  37. Long, M.; Wang, J.; Sun, J.; Philip, S.Y. Domain invariant transfer kernel learning. IEEE Trans. Knowl. Data Eng. 2014, 27, 1519–1532. [Google Scholar] [CrossRef]
Figure 1. Basic Idea of the Proposed TKRM Method for Cross-Database MER Problem.
Figure 1. Basic Idea of the Proposed TKRM Method for Cross-Database MER Problem.
Mathematics 11 00918 g001
Figure 2. Sample Examples of Four ME Sample Sets Used in the Evaluation Experiments.
Figure 2. Sample Examples of Four ME Sample Sets Used in the Evaluation Experiments.
Mathematics 11 00918 g002
Table 1. Sample Statistics of the CASME II and SMIC databases used in the designed cross-database MER experiments.
Table 1. Sample Statistics of the CASME II and SMIC databases used in the designed cross-database MER experiments.
ME DatabaseNegativePositiveSurpriseSummation
CASME II (C)733225130
HS (H)705143164
VIS (V)28232071
NIR (N)28232071
Table 2. The results of cross-database MER experiments in terms of the mean F1-score (%) among CASME II, HS, VIS, and NIR, respectively. The best result in each experiment is underlined.
Table 2. The results of cross-database MER experiments in terms of the mean F1-score (%) among CASME II, HS, VIS, and NIR, respectively. The best result in each experiment is underlined.
MethodC→ HH→CC→VV→CC→NN→CAverage
SVM36.9732.4547.0153.6752.9523.6841.12
TCA46.3748.7068.3457.8949.9239.3751.77
GFK41.2647.7663.6160.5651.8044.6951.61
SA43.0254.4759.3952.4357.3835.9248.77
TKL38.2946.6160.4253.7853.9242.4849.25
TSRG50.4251.7159.3562.0856.2441.0553.48
DRLS-T45.2454.6062.1767.6253.6946.5354.98
DRLS49.2452.6757.5759.4248.8538.3851.02
RSTR52.9756.2258.8270.2150.0946.9355.87
TKRM52.5359.4170.2065.3150.2951.3558.18
Table 3. The results of cross-database MER experiments in terms of accuracy (%) among CASME II, HS, VIS, and NIR, respectively. The best result in each experiment is underlined.
Table 3. The results of cross-database MER experiments in terms of accuracy (%) among CASME II, HS, VIS, and NIR, respectively. The best result in each experiment is underlined.
MethodC→ HH→CC→VV→CC→NN→CAverage
SVM45.1248.4650.7053.0842.1123.8545.55
TCA46.3453.0869.0159.2350.7042.3150.73
GFK46.9550.7766.2061.5053.5246.9254.31
SA47.5662.3159.1551.5447.8936.9250.90
TKL44.5154.6260.5653.0854.9343.8551.93
TSRG51.8360.7759.1563.0856.3446.1556.22
DRLS-T46.9560.0063.3868.4656.3450.7757.65
DRLS53.0559.2357.7560.0049.8342.3753.71
RSTR54.2760.7759.1570.7750.7050.7757.74
TKRM52.4466.1570.4265.3850.7052.3159.57
Table 4. The results of cross-database MER experiments by using TKRM with different kernel functions. The best result in each experiment is underlined.
Table 4. The results of cross-database MER experiments by using TKRM with different kernel functions. The best result in each experiment is underlined.
Kernel FunctionC→ HH→CC→V
Linear52.53/52.4459.41/66.1570.20/70.42
Polynomial ( d = 1.01 )52.53/52.4460.02/65.3858.89/69.01
Polynomial ( d = 1.05 )51.99/51.8358.46/60.7767.31/67.61
Polynomial ( d = 1.1 )52.53/52.4459.39/62.3163.89/63.38
PolyPlus ( d = 1.01 )52.53/52.4460.02/65.3868.89/69.01
PolyPlus ( d = 1.05 )51.99/51.8358.46/60.7767.35/67.61
PolyPlus ( d = 1.1 )51.85/51.8359.39/62.3170.20/70.42
Gaussian ( t = 1 )37.93/40.2447.14/57.6951.25/50.70
Gaussian ( t = 5 )50.61/50.6157.22/57.6067.65/67.61
Gaussian ( t = 10 )49.45/49.3551.74/52.3166.05/66.20
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, Z.; Lu, C.; Zhou, F.; Zong, Y. TKRM: Learning a Transfer Kernel Regression Model for Cross-Database Micro-Expression Recognition. Mathematics 2023, 11, 918. https://doi.org/10.3390/math11040918

AMA Style

Chen Z, Lu C, Zhou F, Zong Y. TKRM: Learning a Transfer Kernel Regression Model for Cross-Database Micro-Expression Recognition. Mathematics. 2023; 11(4):918. https://doi.org/10.3390/math11040918

Chicago/Turabian Style

Chen, Zixuan, Cheng Lu, Feng Zhou, and Yuan Zong. 2023. "TKRM: Learning a Transfer Kernel Regression Model for Cross-Database Micro-Expression Recognition" Mathematics 11, no. 4: 918. https://doi.org/10.3390/math11040918

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop