Next Article in Journal
Developing Low-Cost Mobile Device and Apps for Accurate Skin Spectrum Measurement via Low-Cost Spectrum Sensors and Deep Neural Network Technology
Next Article in Special Issue
EPSDNet: Efficient Campus Parking Space Detection via Convolutional Neural Networks and Vehicle Image Recognition for Intelligent Human–Computer Interactions
Previous Article in Journal
Electrochemical Sensor for the Determination of Methylthiouracil in Meat Samples
Previous Article in Special Issue
Energy Saving Planner Model via Differential Evolutionary Algorithm for Bionic Palletizing Robot
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Domain Adaptive Hand Pose Estimation Based on Self-Looping Adversarial Training Strategy

School of Rail Transportation, Soochow University, 8 Jixue Road, Xiangcheng District, Suzhou 215100, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(22), 8843; https://doi.org/10.3390/s22228843
Submission received: 21 October 2022 / Revised: 7 November 2022 / Accepted: 12 November 2022 / Published: 15 November 2022

Abstract

:
In recent years, with the development of deep learning methods, hand pose estimation based on monocular RGB images has made great progress. However, insufficient labeled training datasets remain an important bottleneck for hand pose estimation. Because synthetic datasets can acquire a large number of images with precise annotations, existing methods address this problem by using data from easily accessible synthetic datasets. Domain adaptation is a method for transferring knowledge from a labeled source domain to an unlabeled target domain. However, many domain adaptation methods fail to achieve good results in realistic datasets due to the domain gap. In this paper, we design a self-looping adversarial training strategy to reduce the domain gap between synthetic and realistic domains. Specifically, we use a multi-branch structure. Then, a new adversarial training strategy we designed for the regression task is introduced to reduce the size of the output space. As such, our model can reduce the domain gap and thus improve the prediction performance of the model. The experiments using H3D and STB datasets show that our method significantly outperforms state-of-the-art domain adaptive methods.

1. Introduction

Human hand poses are one of the long-standing research topics in computer vision. In recent decades, hand pose has had a wide range of applications in VR/AR, robot control [1], and human–machine interaction. Moreover, similar to human pose estimation for action recognition, 3D hand pose estimation can be further applied to gesture recognition and sign language recognition [2,3,4,5,6,7]. Despite the great success of applying deep neural networks to pose estimation tasks [8,9,10,11,12,13,14], the lack of well-labeled datasets has limited the development of powerful deep learning methods for hand pose estimation tasks. It is not an easy task to annotate real-world images because it is time-consuming and labor-intensive. To solve the problem of a lack of data, some researchers started to study synthetic datasets [15] after finding them easily accessible [16,17,18,19,20,21]. Therefore, making use of synthetic data, which is of high sufficiency, is the mainstream method to make up for the deficiency of training data in hand pose estimation.
However, there is a domain gap in skin texture and background between synthetic and real images, and this gap can affect the performance of the trained model [17,18,19,20,21]. To solve the domain gap problem, many methods have been proposed in existing works, for example, Mueller et al. [20] proposed a Cyclegan network, which aims to make synthetic images closer to realistic images in terms of background and texture. In the unlabeled target domain, in order to make use of the output of the teacher network, a mean-teacher model was proposed by Tarvainen et al. [22] for the guidance of the training of student network. During the domain shift, the noise of the pseudo labels would affect the training process of the model, which may lower the performance. As illustrated in the literature, the regression space of key point estimation is generally continuous. Therefore, the sparsity of the regression space was discovered by Jiang et al. [19] in a probabilistic sense. Then, a domain adaptive method was proposed in an unsupervised way, which is called RegDA. The RegDA method transforms the mini–max game between regressors to achieve the minimization of two opposite goals. Based on this, RegDA reduces the domain gap by adversarial training.
To solve the issue caused by the large output space, both RegDA [19] and MarsDA [23] consider the sparsity of the regression space. This means that, if there is wrong prediction on the target domain, the distribution of the error predictions is not uniform in the pixel space. For an error predicted position of the finger joint, for example, the wrong position is always located at an adjacent fingertip or other key points. It is rarely located in the background. However, this is not absolute (see Figure 1), and when we only consider spatial locations with high probability and ignore locations with low probability, it will affect the accuracy and robustness of model prediction. The output space of the estimation model usually has a size of 64 × 64, and we can consider reducing the output space size such as 16 × 16. Inspired by this, we designed a new adversarial training strategy to reduce the output space of the regression task by adding a refinement module. With the reduced output space, we can effectively perform adversarial training to improve the prediction performance of the network and reduce the output fluctuations.
In this paper, we propose a new domain adaptive method for hand pose estimation. It can effectively reduce the domain gap and extract domain invariant features. We use the mean-teacher network to compose a multi-branch output model. We train the mean-teacher network on the synthetic data with labels. Furthermore, a set of pseudo labels are then generated for the unlabeled real-world data. However, the noise of the pseudo labels limits the accuracy of the model. To this end, we solve this problem by performing adversarial training with three branching networks. We add a refinement module to the student network to change the size of the output space to 16 × 16, thus assisting the model in adversarial training. The redesigned adversarial training strategy can effectively reduce the noise of pseudo labels. Our model is trained using accurate pseudo-labels, which ultimately results in better prediction accuracy.
We perform experiments on the H3D [24] and STB [25] datasets to evaluate the proposed method, and the results show that RegDA yields the best prediction results. The main contributions are as follows.
  • We designed a new unsupervised domain adaptive model for hand pose estimation, which designed self-looping adversarial training strategy to bridge the gap between synthetic and real-world images.
  • A new self-looping adversarial training strategy was designed to learn domain-invariant features more efficiently, which can lead to more accurate pseudo labels generated by the teacher network.
  • Achieving state-of-the-art performance on H3D and STB real-world datasets demonstrates that self-looping adversarial training strategies can effectively reduce domain differences.
We organize the rest of this paper as follows. The related work is discussed in Section 2. The proposed method is then presented in Section 3. After that, experimental evaluation and analysis are performed in Section 4. Finally, Section 5 concludes the paper.

2. Related Work

2.1. Hand Pose Estimation

In recent decades, hand posture estimation has attracted the attention of many researchers, so there are very many innovations and applications [3,5,20,26,27,28,29,30,31]. Thanks to the rapid development of deep learning algorithms, hand pose estimation based on RGB images has become a popular research topic [3,32,33,34,35]. However, deep learning algorithms require many labeled data to train the model, so that a good model can be obtained. High-quality RGB hand pose datasets are very scarce, which restricted the development of pose estimation. Some researchers have started to find solutions. Spurr et al. [36] proposed a VAE-based “cross-modal variational model” that learns the shared latent space between different modalities. Wan et al. [37] proposed a network structure based on two generative networks with the goal of 3D hand pose estimation. The network consists of a variational autoencoder for hand pose and a generative adversarial network for deep image distribution modeling.
As the rendering technology has undergone significant development, synthetic datasets are widely used by researchers in order to assist in real-world hand pose estimation tasks. To enable the adaptation from synthetic datasets to the real-world unlabeled datasets, an end-to-end network was proposed by Dibra et al. [38]. A separated potential space was proposed by Yang et al. [39] to separate the image variations, e.g., the image background content and hand pose, which can be utilized to estimate hand pose and for image synthesis. There are several widely used benchmark datasets for testing hand pose estimation methods, including the Stereo Hand Pose Tracking Benchmark (STB) [25], the Rendered Hand Pose Dataset (RHD) [17], and the Hand-3D-Studio dataset (H3D) [24]. The proposed method is tested on these benchmarks and achieves state-of-the-art performance on H3D and STB with an accuracy of 81.3% and 82.4%.

2.2. Unsupervised Domain Adaptation

The unsupervised domain adaptation methods are designed to train a model in the unlabeled target domain from a labeled source domain. Then, the trained model on the unlabeled target domain performs well. Using the adversarial learning is the mainstream approach. The domain invariant features are learned by a feature extractor to fool a domain discriminator. Domain adaptation has many applications in areas such as semantic segmentation, classification tasks, and image recognition. Domain adaptation has many applications in areas such as semantic segmentation, classification tasks, and image recognition [9,40,41,42,43]. Li et al. [44] proposed an MMD-AAE framework that aligns the features extracted from multiple domains. Sankaranarayanan et al. [45] proposed a domain adaptive method which uses an adversarial training framework for weak segmentation. However, the relationship between images and 3D poses is nonlinear, so stronger constraints are required to effectively eliminate the domain gap. These above algorithms do not obtain the expected results when applied to the hand estimation task, and for this reason, a new domain adaptive algorithm was designed.

3. The Proposed Method

In this section, we describe our unsupervised domain adaptation model in detail. Figure 2 shows the overview architecture. The acquired images of the unlabeled target domain are fed into the network, and finally, the network model outputs an accurate 2D hand pose. To this end, we design a student–teacher network. Both the teacher network and the student network are composed of a feature extractor and three regressors. The student network utilizes source and target domain data to train, while the teacher network generates pseudo-labels for the unlabeled target domain data. These pseudo labels are noisy due to the domain gap. Training the model directly with pseudo-labels will have an impact on the model prediction. To reduce this effect, we introduced an adversarial training strategy in the multi-branch structure.

3.1. Multi-Branch Domain Adaptation Module

The domain adaptation task contains data from two different domains, where the first domain is the source domain X s = x s i i = 1 N with label Y s = y s i i = 1 N and the second domain is the target domain X t = x t i i = 1 N . It worth noting that, the samples of the target domain are not the ground truth. The proposed student network includes a feature extractor, different regressors, and a refinement network. Both of the data in the source domain and the target domain are input into the model. The output of the teacher network is used as the pseudo label in the domain of the target. The teacher network and the output are denoted by symbols “*” and “∼”, respectively. The final prediction result is the output of the regressor when performing the testing phase.
Similarly to MarsDA [23], the student network is trained following the way of the standard supervised network in the training phase. The learning of the parameters θ of the student network is different from that of the parameters θ of the teacher network. The student network is learned using stochastic gradient descent (SGD), while the teacher network is not involved in gradient direction propagation. The exponential moving average normalization (EMAN) is employed to update the parameters θ of the teacher, i.e.,
θ t = m θ t 1 + ( 1 m ) θ t ,
μ t = m μ t 1 + ( 1 m ) μ t ,
σ t 2 = m σ t 1 2 + ( 1 m ) σ t 2 ,
where μ is the mean of batch normalization (BN), and σ 2 is the variance of BN. t denotes the epoch number of training. The value of the momentum m is a number close to 1, i.e., 0.999.
For a better comparison with RegDA [19] and MarsDA [23], the loss between the ground truth and the estimated heatmaps is calculated by the Kullback–Leibler (KL) divergence. In the task of hand pose estimation, the same performance can be obtained for a model trained with KL divergence as for a model trained with MSE. First, a spatial probability distribution P T H k , k { 1 , 2 , , K } is defined, which aims to normalize the heatmap of each key point H k R H × W . K is the number of key points in the hand, K = 21 .
P T H k = H k h = 1 H w = 1 W H k h , w .
Denoting Q ( · ) as the spatial softmax function:
Q ( z ) h , w = exp ( z h , w ) h = 1 H w = 1 W exp ( z h , w ) ,
We can use KL divergence to calculate the loss.
L T H s , y s = 1 K k K K L P T H y s k Q H s k ,
where H s = f 0 ψ x s R K × H × W , ψ is the feature extractor, f 0 is the main regressor, and H y s k is the heatmap of each keypoint k in the label y s .
Equation (6) represents the loss in the source domain between the heatmap predicted by the main regressor and the ground truth.
L s = L T ( H s , y s ) .
We use the pseudo-label y ˜ t of the target domain instead of the ground truth, and thus calculate the loss function in the target domain.
L T H t , y ˜ t = 1 K k K K L P T H y ˜ t k Q H t k ,
where y ˜ t = f 0 * ( ψ * ( x t ) ) is the pseudo-label. The output of regressor f 0 * is the pseudo-labels.
To bridge the domain gap between the source and target domains, we apply a feature alignment loss to align the feature distributions between synthetic and real-world domains. Global feature alignment loss is introduced.
L g l o b a l = L T ( F s , F t ) .
This loss can mitigate the effect of noise on the pseudo-label at a certain level.

3.2. Self-Looping Adversarial Training

Since the data distributions of the source and target domains are different, it is a core problem of the domain adaptation task to measure the difference between the data distributions of these two domains. Currently, adversarial training is commonly used in domain adaptation tasks to align the data distribution of these domains.
Inspired by the latest theory [19,46], we designed an multi-branch adversarial training strategy. As shown in Figure 2, we use three regressors ( f 0 , f 1 and f 2 ) and a refinement module r to implement adversarial training. f 1 is the auxiliary regressor, and f 2 is the adversarial regressor.
The size of the output space of the hand pose estimation model is 64 × 64, while the classification model is much smaller than that. Therefore, we cannot directly apply the model for the classification task to the hand pose estimation task. For this problem, RegDA and MarsDA consider the sparsity of the regression space in the sense of probability, thus constraining the output space from a whole image space into a smaller one with only K key points. However, this method only considers the locations with higher probability in the output space, which is incomplete and affects the accuracy and robustness of the model prediction. To circumvent this problem, we propose a new adversarial training strategy that reduces the size of the output space to 16 × 16 using the refinement module.
The input of the refinement module is the output of the regressor f 2 , which is denoted as R t . As shown in Figure 2, we designed a self-feedback loop. We first made the output of the refinement module the K × 16 × 16 heatmaps. Then, we upsampled the output of the refinement module and used it to supervise the regressor f 2 . As such, the gap between classification and regression may be bridged. We also proposed an error probability distribution to make the optimization of adversarial training easier. The distribution of error probability is generated using the pseudo labels in the target domain. That is, the ground error prediction, which is intended to make the distance from f 2 to the correct key points as far as possible. Hence, the optimization of f 2 is guided. The error probability distribution can be redesigned as follows:
H F H k = I R t ,
where I is the matrix whose elements are all 1. Then, we can obtain P F ( H k ) :
P F H k = H F ( H k ) h = 1 H w = 1 W H F ( H k ) h , w .
P F ( H k ) represents the probability of the distribution of errors made by the model at different locations. Therefore, we convert the mini–max game of the two regressors to the minimization problem of two opposite objectives.
From the above, we need to keep the output of the refinement module away from the correct keypoint location. Then, the output of f 2 is supervised by the output of the refinement module via the self-feedback loop:
L a d v = L T ( f 2 ( ψ ( x t ) ) , R t ) .
We concurrently supervised the output of the refinement module with the pseudo labels, i.e.,:
L r = L F ( R t , y t ) ,
where R t = r ( H t ) , H t = f 0 ( ψ ( x t ) ) , y t is also the pseudo labels, and y t R K × 16 × 16 .
To reduce the difficulty of model training, we converted the max–min game in adversarial training into two minimization strategies. These two minimization strategies are defined as follows.
L F M 2 , y ˜ t = 1 K k K K L P F H y ˜ t k Q M 2 k ,
L T M 2 , M 1 = 1 K k K K L P T H M 1 k Q M 2 k ,
where M 1 = f 1 ( ψ ( x t ) ) , M 2 = f 2 ( ψ ( x t ) ) . They are the predicted results of two regressors f 1 and f 2 .

3.3. Training Process

The purpose of the adversarial training strategy is to train the feature extractor ψ to deceive the adversarial regressor f 2 , so that the feature extractor ψ can effectively learn domain invariant features. As with MarsDA, the final training steps are divided into three phases, namely A, B, and C. It is important to know that the loss functions in these three steps are optimized simultaneously in one framework.
Firstly, the feature extractor and the three regressors are trained using the source domain data, while the main regressor f 0 is trained using the pseudo-labels of the target domain. It should be noted that we minimize the loss functions of the adversarial regressors f 2 and f 1 on the source domain.
min ψ , f 0 , f 1 , f 2 E x s , y s P L T f 0 ψ x s , y s + L T f 1 ψ x s , y s + λ 1 L T f 2 ψ x s , f 1 ψ x s + λ 2 E x t , y ˜ t Q L T f 0 ψ x t , y ˜ t + λ 3 L T F s , F t ,
where λ 1 , λ 2 , λ 3 are the weights to balance all losses.
Secondly, we minimize the losses of the adversarial regressor f 2 and the refinement module (Fix ψ , f 0 and f 1 ).
min f 2 μ E x t , y ˜ t Q λ 4 L T f 2 ψ x t , R t + λ 5 L F R t , y ˜ t .
where λ 4 , λ 5 are the weights to balance all losses.
Thirdly, we train the feature extractor ψ to minimize the loss function between the regressors f 1 and f 2 over the target domain.
min ψ μ E x t , y ˜ t Q L T f 2 ψ x t , f 1 ψ x t .
We keep repeating the above process to train the model and finally achieved the consistency of the data distribution in both domains.

4. Experiments

In the experimental section, we performed experiments using a synthetic dataset and two real-world datasets to validate our proposed method.

4.1. Datasets and Metrics

RHD: Rendered Hand Pose Dataset (RHD) [17] is a synthetic dataset with an image resolution of 320 × 320. The dataset is collected from 20 characters, where each character performs 39 different actions. This includes 4k training images as well as 3k testing images. All images are labeled with the locations of 2D and 3D keypoints. During the training process, we also cropped and resized the images to 256 × 256.
H3D: Hand-3D-Studio (H3D) [24] is a real-world dataset containing 22k images. It builds on hand poses performed by 10 people. Both sexes were represented and all the skin tones of the 10 people were different. According to RegDA, we used 3.2k images for testing, and used the rest for training. For the training process, cropped images were used with 512 × 512 resolution, which are provided by RegDA.
STB: Stereo Hand Pose Tracking Benchmark (STB) [25] is a dataset of the real-world including images with 640 × 480 resolution. This dataset has 18 k images, and 21 hand joint locations are collected as ground truths. The 18k images were divided into 15k images for training, and another 3k images for testing. In the training phase, the images are cropped and resized into 256 × 256.
We used the percentage of correct keypoints (PCK) as an evaluation metric. Specifically, we used PCK@0.05. That is to say, if the distance between the prediction and the ground truth is no more than a fraction α = 0.05 of the size of an image, the prediction is regarded as a correct result. The average PCK of the 21 keypoints are calculated. At the same time, the PCK of different hand parts are also shown, e.g., the metacarpophalangeal joint (MCP), proximal interphalangeal joint (PIP), and distal interphalangeal joint (DIP), and the fingertip. RHD→H3D stands for the domain adaptation between the source dataset RHD and the target dataset H3D. RHD→STB stands for domain adaptation from the source dataset RHD to the target dataset STB. The image processing is implemented via Python.

4.2. Implementation Details

Resnet101 [8] is employed as the extractor ψ of features. We used two convolutional layers for regressors. There is a bottleneck block in the refinement module, followed by a convolutional layer. As with MarsDA [23], we cropped and resized the training images to 256 × 256. The whole model was trained for 100 epochs. The mini-batch SGD with a momentum was 0.9. The batch size was 32. We adjusted the learning rate by l p = l 0 ( 1 + α p ) β . The p here denotes the step of the training. l 0 = 0.1 , α = 0.0001 and β = 0.75 . According to [47], we set the feature extractor learning rate to one tenth of the regressor. In the optimization phase, the weights of losses are separately set to λ 1 = 6 , λ 3 = 0.5 , λ 4 = 0.8 , λ 5 = 0.2 and λ 2 = m i n ( 0.01 * e p o c h , 0.3 ) .

4.3. Main Results

We compared the proposed method with some of the latest domain adaptive methods, and the experimental results are shown in Table 1 and Table 2. The experimental results show that the model directly trained with synthetic data does not achieve excellent performance. Although the teacher–student network improves the performance of the model, the performance improvement is limited due to the inaccurate pseudo-labeling. Methods such as MCD and DANN also struggled to obtain excellent performance in the hand pose estimation task due to the domain gap between the source and target domains. RegDA and MarsDA achieved better performance because they exploited the probabilistic sparsity of the model on the output space, allowing the model to efficiently learn domain invariant features. Compared with them, our method reduced the output space by self-looping adversarial training. Then, the gap between the source and target domains effectively were bridged, and the teacher network is allowed to generate accurate pseudo labels and help the model to be trained. The average accuracy is increased by more than 2% compared with MarsDA on the STB dataset. From Figure 3, this demonstrates that our model effectively improves the accuracy and robustness of the prediction.
To more visually demonstrate the superiority of the proposed method, we show some visualization results in Figure 4 and Figure 5. The proposed method is compared with other methods in the figures. From the figures, we can see that our method can obtain more accurate prediction results, while correcting the keypoints that other methods incorrectly predict.

4.4. Ablation Study

We conducted ablation experiments in the H3D dataset, which were used to investigate the contribution of different modules in the proposed model. The specific experimental results are shown in Table 3. The “source only” in the table refers to the model trained directly using the source domain data. “RD” refers to the RegDA network. “MT” refers to the mean-teacher network. “SAT” stands for the self-looping adversarial training strategy. “FA” refers to feature alignment. As can be seen, the new adversarial training strategy can effectively improve performance and bring 2.2% PCK improvement over the RegDA. The mean-teacher network gives a 3.8% performance improvement to the model, and feature alignment gives another 2.6% performance improvement. The self-looping adversarial strategy brings 1.6% PCK improvement. The final experimental results verify the superiority of the proposed method, which can obtain a PCK gain of 19.3% over the baseline. From the results, we can see that the method with SAT+MT+FA obtains the best performance of 81.3% in accuracy. This result indicates that the proposed self-looping adversarial training strategy significantly improves the result of hand pose estimation, which validates the effectiveness of this strategy. Furthermore, the feature alignment can also further improve the performance of the network.
We also conducted an ablation study to illustrate how different methods of adversarial training influences adaptation. Table 4 shows the results. The first row is RegDA, which minimizes two opposite goals separately. The second row is MarsDA, which optimizes the RegDA method. These are both trained by considering the sparsity of the output space in terms of probability. The last row is our method, where we reduce the size of the output space and thus perform the adversarial training. Our proposed method outperforms the first two methods to a large extent.
The training process is visualized in Figure 6. For RegDA, we can see that RegDA, like several other domain adaptation methods, suffers from the problem that there is a significant drop in accuracy as the training epoch increases. Marsda is the same as RegDA in that it considers the sparsity of the regression space in the sense of probability. Its accuracy, although improved, still suffers from the large fluctuation of the network output. However, this does not occur in our method. This demonstrates that our model effectively converts the mini–max game between two regressors into the minimization of two opposite goals. From Figure 3, we can see the training processes of different methods. It is shown that our method outperforms it in terms of accuracy and stability. In order to more effectively demonstrate the robustness of the model, we used 10 epochs as a group to calculate the variance, and the results are shown in Table 4. As can be seen from the table, the variance calculated by Marsda and Regda in 50–100 epochs has a very large fluctuation, while the proposed method differs from them in that it has a very small fluctuation. More specifically, the variance values calculated by the proposed method in 50–100 epochs are within 0.01, while the other methods are above 0.4. Compared to the student network, the teacher network is much more stable than the student network because the teacher network is a temporal aggregation of the student network. To better demonstrate the distribution of features learned by the proposed model, we used t-sne [50] to visualize the features in the source and target domains extracted by the feature extractor. The results are shown in Figure 7, from which we can see that the proposed domain adaptive model can effectively reduce the domain differences and thus obtain better prediction performance.

5. Discussions

From the experimental results on different datasets, we can find that the colors will not affect the effectiveness of the proposed method. Meanwhile, it should be noted that, although the colors in the datasets are a little different, their difference is not significant. Since the data were captured in the lab with stable lighting, the results are stable under different illuminations. If there is an extreme illumination condition, the result depends on the effectiveness of hand detection results. Of course, the distance between the camera and the hand will influence the performance due to the change in the resolution of the hand. We conduct the experiments on the NVIDIA 3090 GPU to run the deep learning algorithms. The proposed method can be used for a robot, since the resources needed for inference are much less than that of training and the inference time is fast. The proposed method is robust with different backgrounds. From the figures of experimental results, we can find that the backgrounds of the images are significantly different, and the proposed method performs well with its promised results.

6. Conclusions

In this paper, a new unsupervised domain adaptation method is proposed for hand pose estimation. A self-looping adversarial training strategy is designed for knowledge transfer between the synthetic source domain and the real-world target domain. In the adversarial module, we reduce the size of the regression space, thus effectively converting the minimax game of the two regressors to a minimization problem between the two opposite goals. Thereby, the issue of the noise in pseudo labels at the training phase can be mitigated. Meanwhile, the data distribution between the synthetic and real-world domains are aligned. Extensive experiments on two benchmark datasets show the effectiveness of our approach. The main contributions including three main aspects. (1) We designed a new unsupervised domain adaptive model for hand pose estimation, which designs a self-looping adversarial training strategy to bridge the gap between synthetic and real-world images. (2) A new self-looping adversarial training strategy was designed to more efficiently learn domain-invariant features, which can lead to more accurate pseudo labels generated by the teacher network. (3) Achieving state-of-the-art performance on H3D and STB real-world datasets demonstrates that self-looping adversarial training strategies can effectively reduce domain differences.

Author Contributions

Methodology, R.J.; Supervision, J.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (NSFC No. 61773272), and the Six Talent Peaks Project of Jiangsu Province, China (No. XYDXX-053).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, J.; Tao, D. Empowering Things with Intelligence: A Survey of the Progress, Challenges, and Opportunities in Artificial Intelligence of Things. IEEE Internet Things J. 2020, 8, 7789–7817. [Google Scholar] [CrossRef]
  2. Yang, J.; Yuan, J.; Li, Y. Parsing 3D motion trajectory for gesture recognition. J. Vis. Commun. Image Represent. 2016, 38, 627–640. [Google Scholar] [CrossRef]
  3. Ge, L.; Ren, Z.; Li, Y.; Xue, Z.; Wang, Y.; Cai, J.; Yuan, J. 3D hand shape and pose estimation from a single RGB image. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  4. Zhu, C.; Yang, J.; Shao, Z.; Liu, C. Vision based hand gesture recognition using 3D shape context. IEEE/CAA J. Autom. Sin. 2021, 8, 1600–1613. [Google Scholar] [CrossRef]
  5. Baek, S.; Kim, K.I.; Kim, T.K. Weakly-supervised Domain Adaptation via GAN and Mesh Model for Estimating 3D Hand Poses Interacting Objects. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  6. Yang, J.; Zhu, C.; Yuan, J. Real time hand gesture recognition via finger-emphasized multi-scale description. In Proceedings of the 2017 IEEE International Conference on Multimedia and Expo (ICME’17), Hong Kong, China, 10–14 July 2017; pp. 631–636. [Google Scholar]
  7. Pan, T.; Wang, Z.; Fan, Y. Optimized convolutional pose machine for 2D hand pose estimation. J. Vis. Commun. Image Represent. 2022, 83, 103461. [Google Scholar] [CrossRef]
  8. Xiao, B.; Wu, H.; Wei, Y. Simple baselines for human pose estimation and tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 466–481. [Google Scholar]
  9. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep high-resolution representation learning for human pose estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 5693–5703. [Google Scholar]
  10. Newell, A.; Yang, K.; Deng, J. Stacked hourglass networks for human pose estimation. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 483–499. [Google Scholar]
  11. Zhang, J.; Chen, Z.; Tao, D. Towards High Performance Human Keypoint Detection. Int. J. Comput. Vis. (IJCV) 2021, 129, 2639–2662. [Google Scholar] [CrossRef]
  12. Xu, Y.; Zhang, Q.; Zhang, J.; Tao, D. ViTAE: Vision Transformer Advanced by Exploring Intrinsic Inductive Bias. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Virtual, 6–14 December 2021. [Google Scholar]
  13. Vishwakarma, D. A Two-fold Transformation Model for Human Action Recognition using Decisive Pose. Cogn. Syst. Res. 2019, 61, 1–13. [Google Scholar] [CrossRef]
  14. Lu, D.; Yong, W.; Robert, L.; Dan, H.; Shan, F. A CNN model for real time hand pose estimation. J. Vis. Commun. Image Represent. 2021, 79, 103200. [Google Scholar]
  15. Vazquez, D.; Lopez, A.M.; Marin, J.; Ponsa, D.; Geronimo, D. Virtual and real world adaptation for pedestrian detection. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI) 2014, 36, 797–809. [Google Scholar] [CrossRef] [PubMed]
  16. Cai, Y.; Ge, L.; Cai, J.; Yuan, J. Weakly-supervised 3d hand pose estimation from monocular rgb images. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 666–682. [Google Scholar]
  17. Zimmermann, C.; Brox, T. Learning to estimate 3d hand pose from single rgb images. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  18. Li, C.; Lee, G.H. From Synthetic to Real: Unsupervised Domain Adaptation for Animal Pose Estimation. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021. [Google Scholar]
  19. Jiang, J.; Ji, Y.; Ximei Wang, Y.L.; Wang, J.; Long, M. Regressive Domain Adaptation for Unsupervised Keypoint Detection. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021. [Google Scholar]
  20. Mueller, F.; Bernard, F.; Sotnychenko, O.; Mehta1, D. GANerated Hands for Real-Time 3D Hand Tracking from Monocular RGB. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  21. Zhang, Y.; Chen, L.; Liu, Y.; Yong, J.; Zheng, W. Adaptive Wasserstein Hourglass for Weakly Supervised Hand Pose Estimation from Monocular RGB. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 2076–2084. [Google Scholar]
  22. Tarvainen, A.; Valpola, H. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In Proceedings of the Conference and Workshop on Neural Information Processing Systems (NIPS), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  23. Jin, R.; Zhang, J.; Yang, J.; Tao, D. Multi-Branch Adversarial Regression for Domain Adaptative Hand Pose Estimation. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 6125–6136. [Google Scholar] [CrossRef]
  24. Zhao, Z.; Wang, T.; Xia, S.; Wang, Y. Hand-3d-studio: A new multi-view system for 3d hand reconstruction. In Proceedings of the ICASSP 2020–2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 2478–2482. [Google Scholar]
  25. Zhang, J.; Jiao, J.; Chen, M.; Qu, L.; Xu, X.; Yang, Q. A hand pose tracking benchmark from stereo matching. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017. [Google Scholar]
  26. Vishwakarma, D.; Maheshwari, R.; Kapoor, R. An Efficient Approach for the Recognition of Hand Gestures from Very Low Resolution Images. In Proceedings of the 2015 Fifth International Conference on Communication Systems and Network Technologies, Gwalior, India, 4–6 April 2015; pp. 467–471. [Google Scholar] [CrossRef]
  27. Huang, Y.; Yang, J. A multi-scale descriptor for real time RGB-D hand gesture recognition. Pattern Recognit. Lett. 2021, 144, 97–104. [Google Scholar] [CrossRef]
  28. Vishwakarma, D.; Kapoor, R. An Efficient Interpretation of Hand Gestures to Control Smart Interactive Television. Int. J. Comput. Vis. Robot. 2017, 7, 454–471. [Google Scholar] [CrossRef]
  29. He, Y.; Yang, J.; Shao, Z.; Li, Y. Salient feature point selection for real time RGB-D hand gesture recognition. In Proceedings of the 2017 IEEE International Conference on Real-time Computing and Robotics (RCAR), Okinawa, Japan, 14–18 July 2017; pp. 103–108. [Google Scholar]
  30. Vishwakarma, D.K. Hand gesture recognition using shape and texture evidences in complex background. In Proceedings of the 2017 International Conference on Inventive Computing and Informatics (ICICI), Coimbatore, India, 23–24 November 2017; pp. 278–283. [Google Scholar] [CrossRef]
  31. Wu, M.Y.; Ting, P.W.; Tang, Y.H.; Chou, E.T.; Fu, L.C. Hand pose estimation in object-interaction based on deep learning for virtual reality applications. J. Vis. Commun. Image Represent. 2022, 70, 102802. [Google Scholar] [CrossRef]
  32. Li, M.; Wang, J.; Sang, N. Latent Distribution-based 3D Hand Pose Estimation from Monocular RGB Images. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 4883–4894. [Google Scholar] [CrossRef]
  33. Guo, S.; Rigall, E.; Qi, L.; Dong, X.; Li, H.; Dong, J. Graph-Based CNNs With Self-Supervised Module for 3D Hand Pose Estimation From Monocular RGB. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 1514–1525. [Google Scholar] [CrossRef]
  34. Vishwakarma, D.; Grover, V. Hand gesture recognition in low-intensity environment using depth images. In Proceedings of the 2017 International Conference on Intelligent Sustainable Systems (ICISS), Palladam, India, 7–8 December 2017; pp. 429–433. [Google Scholar] [CrossRef]
  35. Wang, Y.; Peng, C.; Liu, Y. Mask-Pose Cascaded CNN for 2D Hand Pose Estimation From Single Color Image. IEEE Trans. Circuits Syst. Video Technol. 2019, 29, 3258–3268. [Google Scholar] [CrossRef]
  36. Spurr, A.; Song, J.; Park, S.; Hilliges, O. Cross-Modal Deep Variational Hand Pose Estimation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 89–98. [Google Scholar] [CrossRef] [Green Version]
  37. Wan, C.; Probst, T.; Van Gool, L.; Yao, A. Crossing Nets: Combining GANs and VAEs with a Shared Latent Space for Hand Pose Estimation. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1196–1205. [Google Scholar] [CrossRef] [Green Version]
  38. Dibra, E.; Wolf, T.; Oztireli, C.; Gross, M. How to Refine 3D Hand Pose Estimation from Unlabelled Depth Data? In Proceedings of the 2017 International Conference on 3D Vision (3DV), Qingdao, China, 10–12 October 2017; pp. 135–144. [Google Scholar] [CrossRef]
  39. Yang, L.; Yao, A. Disentangling Latent Hands for Image Synthesis and Pose Estimation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 9869–9878. [Google Scholar] [CrossRef] [Green Version]
  40. Huang, J.; Lu, S.; Guan, D.; Zhang, X. Contextual-Relation Consistent Domain Adaptation for Semantic Segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 705–722. [Google Scholar]
  41. Gong, R.; Chen, Y.; Paudel, D.P.; Li, Y.; Chhatkuli, A.; Li, W.; Dai, D.; Gool, L.V. Cluster, Split, Fuse, and Update: Meta-Learning for Open Compound Domain Adaptive Semantic Segmentation. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021. [Google Scholar]
  42. Kim, Y.; Hong, S. Adaptive Graph Adversarial Networks for Partial Domain Adaptation. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 172–182. [Google Scholar] [CrossRef]
  43. Liu, G.; Wu, J. Unsupervised person re-identification by Intra–Inter Camera Affinity Domain Adaptation. J. Vis. Commun. Image Represent. 2021, 80, 103310. [Google Scholar] [CrossRef]
  44. Li, H.; Pan, S.J.; Wang, S.; Kot, A.C. Domain Generalization with Adversarial Feature Learning. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5400–5409. [Google Scholar] [CrossRef]
  45. Sankaranarayanan, S.; Balaji, Y.; Jain, A.; Lim, S.N.; Chellappa, R. Learning from Synthetic Data: Addressing Domain Shift for Semantic Segmentation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3752–3761. [Google Scholar] [CrossRef]
  46. Saito, K.; Watanabe, K.; Ushiku, Y.; Harada, T. Maximum Classifier Discrepancy for Unsupervised Domain Adaptation. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  47. Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; Marchand, M.; Lempitsky, V. Domain-adversarial training of neural networks. J. Mach. Learn. Res. 2016, 17, 2096-2030. [Google Scholar]
  48. Zhang, Y.; Liu, T.; Long, M.; Jordan, M. Bridging theory and algorithm for domain adaptation. In Proceedings of the International Conference on Machine Learning (ICML), Long Beach, CA, USA, 9–15 June 2019; pp. 7404–7413. [Google Scholar]
  49. Hoffman, J.; Tzeng, E.; Park, T.; Zhu, J.Y.; Isola, P.; Saenko, K.; Efros, A.A.; Darrell, T. CyCADA: Cycle Consistent Adversarial Domain Adaptation. In Proceedings of the International Conference on Machine Learning (ICML), Stockholm, Sweden, 10–15 July 2018; pp. 1989–1998. [Google Scholar]
  50. van der Maaten, L.; Hinton, G. Visualizing data using t-sne. J. Mach. Learn. Res. (JMLR) 2008, 9, 2579–2605. [Google Scholar]
Figure 1. Some visualizations on the unlabeled target domains. The first row is the prediction results of the source-only model. The second row shows that the results of our model are more accurate. The colors indicate the estimated skeletons of different fingers.
Figure 1. Some visualizations on the unlabeled target domains. The first row is the prediction results of the source-only model. The second row shows that the results of our model are more accurate. The colors indicate the estimated skeletons of different fingers.
Sensors 22 08843 g001
Figure 2. Our network is a student–teacher architecture, where the student network consists of a feature extractor, three regressors, and a refine module. The source and target images are fed into the student network, while the target images are fed into the teacher network. In addition, The teacher network also has three regressors, but the regressors f 2 * and f 1 * , which correspond to the adversarial regressors and f 1 in the student network, are not used during training and testing, so we do not draw them.
Figure 2. Our network is a student–teacher architecture, where the student network consists of a feature extractor, three regressors, and a refine module. The source and target images are fed into the student network, while the target images are fed into the teacher network. In addition, The teacher network also has three regressors, but the regressors f 2 * and f 1 * , which correspond to the adversarial regressors and f 1 in the student network, are not used during training and testing, so we do not draw them.
Sensors 22 08843 g002
Figure 3. Accuracy of different models during training.
Figure 3. Accuracy of different models during training.
Sensors 22 08843 g003
Figure 4. Qualitative results of different methods on the H3D dataset. The colors indicate the estimated skeletons of different fingers.
Figure 4. Qualitative results of different methods on the H3D dataset. The colors indicate the estimated skeletons of different fingers.
Sensors 22 08843 g004
Figure 5. Qualitative results of different methods on the STB dataset.
Figure 5. Qualitative results of different methods on the STB dataset.
Sensors 22 08843 g005
Figure 6. Accuracy of different adversarial training strategies during training.
Figure 6. Accuracy of different adversarial training strategies during training.
Sensors 22 08843 g006
Figure 7. We randomly selected 1920 samples of synthetic and real-world images and used t-SNE [50] to visualize the features learned by the source-only model and our MarsDA model, respectively. The red points are samples in the source domain while the blue points are samples in the target domain.
Figure 7. We randomly selected 1920 samples of synthetic and real-world images and used t-SNE [50] to visualize the features learned by the source-only model and our MarsDA model, respectively. The red points are samples in the source domain while the blue points are samples in the target domain.
Sensors 22 08843 g007
Table 1. Results on the task RHD→H3D. The last row (oracle) denotes the results of training with target domain labels.
Table 1. Results on the task RHD→H3D. The last row (oracle) denotes the results of training with target domain labels.
MethodMCPPIPDIPFingertipAvg
Res101 [8]67.464.263.354.861.8
MCD [46]59.156.154.746.954.6
DD [48]72.769.666.254.465.2
DANN [47]67.362.660.951.260.6
Cyclegan [49]63.863.661.353.560.1
Mean-teacher [22]72.671.267.159.466.8
RegDA [19]79.654.471.262.972.5
MarsDA [23]87.785.880.770.180.6
Our87.286.280.872.581.3
Oracle97.797.295.792.595.8
Table 2. Results on the task RHD→STB. The last row (oracle) denotes the results of training with target domain labels. The best performance is achieved by our MarsDA.
Table 2. Results on the task RHD→STB. The last row (oracle) denotes the results of training with target domain labels. The best performance is achieved by our MarsDA.
MethodMCPPIPDIPFingertipAvg
Res101 [8]67.665.465.959.963.1
MCD [46]58.156.355.446.754.9
DD [48]62.570.168.571.968.4
DANN [47]68.164.165.159.263.1
Cyclegan [49]58.258.557.958.958.1
Mean-teacher [22]68.971.269.767.268.9
RegDA [19]67.479.675.473.873.6
MarsDA [23]75.784.781.283.580.2
Our79.889.184.976.882.4
Oracle93.993.093.894.493.4
Table 3. Ablation study results of the proposed model.
Table 3. Ablation study results of the proposed model.
MethodMCPPIPDIPFingertipAvg
Source only67.464.263.354.861.8
+RD79.674.471.262.972.5
+SAT82.577.273.563.474.7
+RD+MT82.682.279.167.476.3
+RD+MT+FA85.484.679.967.678.9
+SAT+MT+FA87.286.280.872.581.3
Table 4. Model Robustness Analysis.
Table 4. Model Robustness Analysis.
Method0–1010–2020–3030–4040–5050–6060–7070–8080–9090–100
RegDA1.4710.3670.2470.4050.3760.1960.7261.0871.4320.450
MarsDA1.5780.3780.2160.1960.6590.7140.5650.8481.1880.594
Our1.3080.2510.1940.3460.1620.0640.0840.0890.0650.070
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jin, R.; Yang, J. Domain Adaptive Hand Pose Estimation Based on Self-Looping Adversarial Training Strategy. Sensors 2022, 22, 8843. https://doi.org/10.3390/s22228843

AMA Style

Jin R, Yang J. Domain Adaptive Hand Pose Estimation Based on Self-Looping Adversarial Training Strategy. Sensors. 2022; 22(22):8843. https://doi.org/10.3390/s22228843

Chicago/Turabian Style

Jin, Rui, and Jianyu Yang. 2022. "Domain Adaptive Hand Pose Estimation Based on Self-Looping Adversarial Training Strategy" Sensors 22, no. 22: 8843. https://doi.org/10.3390/s22228843

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop