Next Article in Journal
Applications Enabled by FPGA-Based Technology
Next Article in Special Issue
Research on Log Anomaly Detection Based on Sentence-BERT
Previous Article in Journal
Research Progress of Neural Synapses Based on Memristors
Previous Article in Special Issue
Preference-Aware Light Graph Convolution Network for Social Recommendation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

GGTr: An Innovative Framework for Accurate and Realistic Human Motion Prediction

1
Key Laboratory Measurement and Control of CSE, Ministry of Education, School of Automation, Southeast University, Nanjing 210096, China
2
Nanjing Center for Applied Mathematics, Nanjing 211135, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(15), 3305; https://doi.org/10.3390/electronics12153305
Submission received: 6 July 2023 / Revised: 29 July 2023 / Accepted: 30 July 2023 / Published: 1 August 2023

Abstract

:
Human motion prediction involves forecasting future movements based on past observations, which is a complex task due to the inherent spatial-temporal dynamics of human motion. In this paper, we introduced a novel framework, GGTr, which adeptly encapsulates these patterns by integrating positional graph convolutional network (GCN) layers, gated recurrent unit (GRU) network layers, and transformer layers. The proposed model utilizes an enhanced GCN layer equipped with a positional representation to aggregate information from body joints more effectively. To address temporal dependencies, we strategically combined GRU and transformer layers, enabling the model to capture both local and global temporal dependencies across body joints. Through extensive experiments conducted on Human3.6M and CMU-MoCap datasets, we demonstrated the superior performance of our proposed model. Notably, our framework shows significant improvements in predicting long-term movements, outperforming state-of-the-art methods substantially.

1. Introduction

Human motion prediction, the task of forecasting future human movements based on past observations, plays a critical role in various domains such as robotics, computer vision, healthcare, and sports analysis. Accurately predicting human motion is instrumental for facilitating effective human-robot collaboration [1,2], ensuring system security [3,4], analyzing human behavior and emotions [5,6], and supporting sports performance analysis [7]. However, predicting human motion presents significant challenges due to the complexity and diversity of human behaviors. First, human motion exhibits high variability and uncertainty. This is evident at the 3D skeletal level, due to the diversity in human body sizes, and at the movement level, due to individual idiosyncrasies. In scenarios with rapid changes, such as sudden reactions or motions, it is exceedingly difficult for predictive models to adapt quickly. Second, the interplay between different body parts and their coordinated movements further complicates the task. For instance, a motion initiated by one body part can propagate to other body parts in complex and often non-intuitive ways.
So, it becomes essential to capture both spatial and temporal features, as illustrated in Figure 1. Numerous research efforts have been made to address the challenges of modeling human motion. Traditional methods and machine learning methods such as hidden Markov models (HMM) [8], Gaussian processes (GP) [9], and restricted Boltzmann machine [10]. However, these methods may not fully capture the complex interdependencies and non-linear dynamics present in human motion. More recently, deep learning approaches, such as convolutional neural networks (CNNs) [11], graph convolutional networks (GCN) [12,13,14,15], temporal modules such as recurrent neural networks (RNNs) [16,17,18,19,20,21], and transformers [22,23,24] have been used. While existing RNN and deep learning-based models have significantly improved the prediction performance, they still have limitations. These methods often struggle to capture the dynamic and complex interactions between different body parts. The relations among body joints cannot be simply modeled by static spatial proximity; instead, they are influenced by various factors such as the individual’s physical attributes, the current body state, and the motion context. Moreover, these models frequently encounter challenges in effectively capturing both local and global temporal dependencies.
To address these challenges, we proposed a novel approach for human motion prediction that combines positional graph convolutional network (GCN) layers, gated recurrent unit (GRU) network layers, and transformer layers, termed as GGTr. This combination allows us to better capture the complex spatial-temporal patterns in human motion data. In summary, the primary contributions of this paper include:
(1)
The introduction of a novel GCN layer with positional representation, enabling better aggregation of information from adjacent body joints;
(2)
The strategic combination of GRU and transformer layers to capture both local and global temporal dependencies across body joints;
(3)
The conduction of extensive experiments on the Human3.6M and CMU-MoCap datasets, demonstrating the effectiveness and advantages of our proposed framework, our model shows significant improvements in predicting short-term movements. Experiments reveal that our model significantly outperforms state-of-the-art methods.
The remainder of this paper is organized as follows: Section 2 reviews related work. In Section 3, we first introduce some preliminary concepts and formalize the human motion prediction problem and the overall framework of the spatial temporal network model. Then, we discuss in detail the three model components to deal with the spatial and temporal dependencies, separately. In Section 4, experiments were conducted on two large-scale datasets, comparing the performance of the proposed method with baselines. Section 5 provides a summary and conclusion, as well as a discussion of future work.

2. Related Work

Human motion prediction is a challenging task due to the complexity and variability of human movements. This complexity arises from the intricate interplay of various factors such as the individual’s physical characteristics, the environment, and the task at hand. Over the years, several approaches have been proposed to tackle this problem, each with its own strengths and limitations.
Traditional methods primarily rely on data statistical approaches or prior knowledge, such as Markov prediction models [8], Gaussian process dynamical models [9] and restricted Boltzmann machine [10], which can only tackle simple human motion patterns. Although these methods have achieved some success in certain scenarios, they still face challenges in capturing complex spatial and temporal dependencies.
Chiu et al. [25] used LSTM units to model the underlying structure of human motion hierarchically, but did not adequately utilize spatial information of human motion data. Martinez et al. [18] introduced a residual structure using GRU to model the velocity of human motion sequences, Their model focused on short-term temporal modeling, ignoring long-term dependencies and spatial structure. Jain et al. [17] combined LSTM and fully connected (FC) layers in a structural RNN model to encode high-level spatio-temporal structures in human motion sequences. Guo et al. [26] employed FC layers and GRU to model local structures and capture long-term temporal dependencies, but they did not take into account the interactions between different human body limbs. The transformer model with a self-attention mechanism has been proved to be more effective than recurrent networks in various domains [23,27,28]. It applies a multi-head attention mechanism to directly learn dependencies between each pair of input and output positions without any latency.
Graph convolutional networks (GCN) have become widely used for modeling the underlying relationships of non-Euclidean data. Kipf et al. [29] proposed a layer-wise propagation rule for nodes inspired by first-order approximation of spectral convolutions on graphs. However, this approach is limited by the characteristics of the graph. Yuan et al. [30] proposed a more flexible approach by learning node connectivity based on node neighborhood. Velickovic et al. [31] introduced self-attention to determine the neighborhood structure to be considered, providing more flexibility to the network. This approach has been applied to action recognition [32] by using a GCN to capture the temporal and spatial dependencies of human body joints via a graph defined on temporally connected kinematic trees. These techniques have been applied to human motion prediction by building the human pose as a graph and using GCN to encode the spatial connectivity of human joints. Ma et al. [33] proposed two variants of GCN to extract spatial and temporal features. They built a multi-stage structure where each stage contains an encoder and a decoder and, during training, the model is trained with intermediate supervision to learn to progressively refine the prediction. References [12,13,34] extended the graph of the human pose to multi-scale the version across the abstraction levels of the human pose. However, as these models aggregate body joint features based on an input adjacency matrix, the relation between body parts is fixed and may limit the model’s ability to adapt to complex human motions.
The transformer [35] has become the dominant approach in natural language processing (NLP). The key component of the transformer is a multi-head self-attention mechanism that captures long-range dependencies. Building upon the success of the transformer in various tasks [36,37,38], researchers have increasingly focused on exploring its potential applications in 3D human motion prediction [22,23,24]. Cai et al. [22] employed a transformer-based architecture with discrete cosine transform (DCT) to capture the long-range spatial correlations and temporal dependencies in human motion dynamics. Another notable advancement is the spatio-temporal transformer (ST-Trans) mechanism proposed by Aksan et al. [23], which effectively captures the spatio-temporal dependencies of decomposed 3D human motions. However, the ST-Trans method overlooks the importance of ensuring consistency between spatial and temporal information, which is a crucial factor when dealing with time-varying data. To address this limitation, a cross-transformer approach [24] has been developed to explore effective interaction between spatial and temporal branches. This approach is designed to learn the coherence of spatial and temporal information and simultaneously enhance the model’s predictive capacity. Despite these advancements, transformer-based methods may overlook local information when dealing with human motion data, warranting further investigation into this aspect.
In summary, with the development of human motion prediction in recent years, GCN/GRU/transformer-based architectures have been well explored and results have significantly improved. In this paper, we proposed a new spatial-temporal graph convolutional network framework to address the human motion prediction problem. We used graph convolutional networks with a position-wise attention mechanism to capture the spatial dependencies of the human body joints. A gated recurrent neural network with transformer layers was used to capture both local and global information of human motion sequences in the temporal dimension.

3. The Proposed Framework

Problem Fromulation. In the realm of human motion prediction, the objective is to forecast future human movements utilizing historical motion data. Specifically, these historical motion data can be expressed as a time series within a 3D skeletal structure. This structure can be denoted as G = ( V , E ) , where V = { v 1 , v 2 , , v M } represents a set of M joints in the human body, and E is a set of edges that represent the physical connections between these joints. The relationships between joints can be demonstrated by a symmetric adjacency matrix A R M × M , where the i , j -th element, A i j , represents the biomechanical correlation between joints v i and v j . This correlation is typically gauged by the physical linkage or common motion patterns shared between two joints. Note that A i j = 0 if there is no substantial correlation between the two respective joints. The historical motion data can be represented in a sequence Z = ( z 1 , z 2 , , z τ ) on the 3D skeletal structure, where each z t R M × 1 signifies the motion status of M joints at time t. With the above notations, we can formally state the problem as follows.
Given the 3D skeletal structure G = ( V , E ) and the historical motion data Z = ( z 1 , z 2 , , z τ ) , the aim was to formulate a model f that can accept a new sequence X = ( x 1 , x 2 , , x T ) of length T as an input and predict the motion status for the subsequent T time steps, denoted as:
X p r e d = f ( G ; X | Z ) = ( x T + 1 , , x T + T ) .
Moreover, during the training phase, we employed a sliding window of length T + T over the historical motion sequence Z to generate the training samples. These samples serve to train the model f.
Overview. The conceptual framework of the proposed GGTr network, illustrated in Figure 2, is primarily composed of three crucial components. The spatial graph convolutional network (GCN) layers were designed to model the spatial correlation between the human body joints graph. These layers were applied to the input representations of the gated recurrent unit (GRU), which modeled the sequential temporal relations, also referred to as local temporal dependencies. The framework also includes a transformer layer that is specifically designed to directly apprehend the long-range or global temporal dependencies within the motion sequence. The GRU and transformer layers work in tandem to capture the temporal dependencies for each joint independently, albeit from distinct perspectives. Subsequently, we delve into the spatial dependency modeling with the GCN, followed by an explanation of how the GRU layer and the transformer layer function to capture temporal dependencies, leading to a comprehensive summary of the complete framework.

3.1. Spatial Dependence Modeling

Acquiring the complex spatial dependence is a crucial problem in human motion prediction. The traditional convolutional network (CNN) can grasp local spatial features, but its application is primarily confined to Euclidean space, such as images or regular grids. However, the human body essentially forms a graph, not a two-dimensional grid. A graph is a structure composed of nodes and edges. In the context of human motion prediction, nodes can represent joints in the human body, and edges can represent connections between joints. Graphical models can effectively represent the complex relationships between human joints, which is useful for human motion prediction. This means the CNN model cannot reflect the complex topological structure of the human body and thus cannot accurately capture spatial dependence. Recently, generalizing the CNN to the graph convolutional network (GCN), which can handle arbitrary graph-structured data, has received widespread attention.
In the human motion prediction problem, if two body joints are connected or in close proximity, their movements are likely to mutually influence each other. Thus, to capture these spatial relationships, we employed the graph convolutional networks model proposed in [29,31]. This model is used to transform and propagate motion information through the graph structure. Specifically, given input motion information X l R M × d l on the structure, where d l represents the input dimension, the output X l + 1 R M × d l + 1 can be computed, with d l + 1 denoting the output dimension.
X l + 1 = σ ( D ˜ 1 / 2 A ˜ D ˜ 1 / 2 X l W l ) ,
where σ ( . ) is a non-linear activation function. In our work, we adopted ReLU (·), standing for REctified Linear Unit, which is advantageous for its ability to speed up the convergence of stochastic gradient descent compared to sigmoid and tanh functions. A ˜ = A + I M is the adjusted adjacency matrix with I M as the M-dimensional identity matrix, and D ˜ is the modified degree matrix, with D ˜ i i = j A ˜ i j . W l are the parameters to be learned. For convenience, we represent the operation in Equation (2) as follows:
X l + 1 = G C N 0 ( X l , A ) .
In the aforementioned formulation, the operation relies solely on the human body joints structure information, which is premised on the physical proximity between body parts. However, the interplay between body joints can be significantly more intricate. Various factors such as muscle constraints, physical capabilities, motion style, and ongoing actions can impact the motion of body parts. Therefore, the information from neighboring joints should not be aggregated equally to a given central joint when performing the aggregation in Equation (2). Recently, GaAN [39] attempted to employ the attention mechanism to model the complex relationships between graph nodes. Ideally, the aforementioned factors could be used to calculate the attention score, but these factors are not always available. Furthermore, there may be other factors influencing the relations between joints that we are not aware of. Therefore, in this paper, we proposed learning a positional representation to capture these factors for each joint. Specifically, for joints v i , v j , we aimed to learn two latent positional representations h i , h j R M . We then modeled the pairwise relations between any body joints as:
e i j = a ( [ W h i | | W h j ] ) α i j = exp ( L e a k y R e L U ( a ( h i , h j ) ) ) k = 1 M exp ( L e a k y R e L U ( a ( h i , h j ) ) ) ,
where a ( . ) is a relation score function modeled with dot product as follows:
a ( h i , h j ) = ( W h i ) T W h j ,
where W is a transformation matrix to be learned. We further employed a mask to sparsify the relation matrix α in order to reduce computational complexity:
m a s k ( α ) = α i j , if A ˜ i j > 0 0 , otherwise .
Subsequently, the GCN operation can be applied to the newly learned relation matrix m a s k ( α ) :
X l + 1 = σ ( D ˜ α 1 / 2 α ˜ D ˜ α 1 / 2 X l W l ) ,
where α ˜ = m a s k ( α ) + I M and D ˜ α is the degree matrix for α ˜ . For simplicity, we represent the operation in Equation (7) as:
X l + 1 = G C N N ( A , X l ) .
Instead of the operation in (2), the operation in (8) was used to capture the spatial relations. This approach allowed us to learn a positional representation for each joint, capturing various factors that influence their relationships. By using these learned representations in combination with an attention mechanism and a sparsified relation matrix, we can effectively model complex spatial relationships between different body joints and improve the accuracy of human motion prediction.

3.2. Temporal Dependence Modeling

The gated recurrent unit (GRU) is a variant of a traditional RNN and can effectively capture the semantic association between long sequences and alleviate the problem of gradient vanishing or explosion. Its core structure can be divided into two parts for analysis: update gate and reset gate, illustrated in Figure 3. To capture the temporal dependency in human motion sequences, we adopted the gated recurrent unit (GRU) [40] to process the sequence information.
For each time step, we maintained a hidden representation that served as the output for the current step and influenced the information flow to the subsequent step. It is important to note that the GRU operation was applied individually to each joint in the body, and the parameters were shared for all joints. To incorporate spatial relations in human motion while processing the sequence, we applied a modified graph convolutional network (GCN) operation, as described in Equation (8), to the input representations of the GRU. Specifically, at each time step t, with input x t and previous step’s hidden representations h t 1 , we applied the modified GCN operation:
x ˜ t = G C N N ( A , x t ) h ˜ t 1 = G C N N ( A , h t 1 ) .
For a body joint v i at time step t, the history observation x t [ i ] contains the dynamic features for joint v i at t- t h time step. The update gate and reset gate are denoted as follows:
z t = σ ( W z x ˜ t [ i ] + U z h ˜ t 1 [ i ] + b z ) r t = σ ( W r x ˜ t [ i ] + U r h ˜ t 1 [ i ] + b r ) ,
where W z , W r , W h , U z , U r , b z , b r are the trainable parameters.
We then calculated the candidate hidden state and the updated hidden state h t [ i ] as follows:
h ˜ t [ i ] = tan h ( W h x ˜ t [ i ] + r t U h h ˜ t 1 [ i ] + b h ) h t [ i ] = ( 1 z t ) h ˜ t 1 [ i ] + z t h ˜ t [ i ] ,
where ⊙ denotes the element-wise multiplication, U h , b h are the parameters to be learned, and h t [ i ] serves as extracted features of joint v i at the next time step.
The above are the fundamental equations and steps of GRU. By controlling the update and reset gates, GRU can dynamically update and adjust the hidden state based on different patterns in the sequence, enabling it to better capture the local temporal information. However, in the human body motion prediction problem, the temporal information may not only be sequentially dependent. Hence, it is important to capture the global temporal information for accurate human motion prediction. Thus, after processing with the GRU layer, we adopted a transformer layer [35] to capture the global dependencies following the GRU.
The transformer layer, similar to the GRU layer, was applied to each joint individually. For joint v i , the output sequence ( h 1 [ i ] , , h T [ i ] ) from GRU was taken as the input for the transformer. Figure 4 illustrates that a transformer layer consists of a multi-head attention layer, a shared feed-forward neural work layer, and normalization layers between them.
Inputs of model. The input of the model is the spatial dependency embedding sequence of each joint. Let us consider the i- t h joint’s spatial dependency embedding sequence ( h 1 [ i ] , h 2 [ i ] , , h T [ i ] ) . Although the self-attention mechanism is effective in capturing hidden dependency relationships in the sequence, it fails to maintain location information. Hence, we incorporated position encoding e t as proposed by [35], between GRU and the transformer layer. This choice was motivated by the ability of this approach to facilitate the model’s learning to attend to relative positions in a sequence. The new representations h t [ i ] were computed as follows:
h t [ i ] = h t [ i ] + e t ,
where e t is defined as
e t = s i n ( t / 10000 2 i / d m o d e l ) , if i = 0 , 2 , 4 c o s ( t / 10000 2 i / d m o d e l ) , otherwise .
Temporal multi-head attention. In our model, we defined a matrix h v i R T × d by arranging ( h 1 [ i ] , h 2 [ i ] , , h T [ i ] ) in a row-wise manner across the temporal axis. The multi-head attention mechanism is composed of several attention heads, which process the input in parallel and whose outputs are then combined. This concept can be represented mathematically as follows:
M u l t i H e a d ( h v i ) = C o n c a t ( h e a d 1 , , h e a d s ) W O .
This signifies that the output of multi-head attention is the concatenation of s single head attention blocks, each one being projected with the matrix W O . Each single head attention block, h e a d s , can be given by:
h e a d s = A t t e n t i o n ( h v i ) = s o f t m a x ( Q s K s T d k ) V s ,
where W s Q , W s K , and W s V denote the queries, keys, and values of the s- t h single head attention for joint v i , respectively. We obtained the Q s , K s , and V s by a linear projection with W Q , W K , and W V :
Q s = h v i W s Q , K s = h v i W s K , V s = h v i W s V ,
where W s Q R d × d k , W s K R d × d k , and W s V R d × d v are the trainable projection matrices for the s- t h attention head and are shared by all the joints.
Feed-forward Layer. After the multi-head attention layer, the output states are then processed by a feedforward neural network layer. This layer is identical for all joints, which includes two linear transformations and a ReLU activation function in between:
F F N ( X ) = m a x ( 0 , X W 1 + b 1 ) W 2 + b 2 ,
where W 1 , W 2 are the trainable parameters of the feed-forward neural network and b 1 and b 2 are bias terms.
Residual connection and normalization. As shown in Figure 4, the residual connection and normalization operators appear in each layer. The residual connection was introduced to tackle the difficulty in training deep network optimization algorithms, and normalization was applied to prevent overfitting. Given the embedded feature Z i n , the process of the residual connection and normalization layer is denoted as follows:
Z i n = L a y e r N o r m ( Z i n + M u l t i H e a d ( Z i n ) ) h o u t = L a y e r N o r m ( Z i n + F F N ( Z i n ) ) ,
where L a y e r N o r m ( . ) denotes the normalization function.
Lastly, after each sub-layer in the transformer, such as the temporal multi-head attention layer and the feed-forward network layer, a layer normalization operation and a residual connection were applied. The final output of the transformer layer for joint v i can be denoted as h o u t v i R T × d .
In summary, this approach combines both the local and global temporal information through the GRU and transformer layers, which can be highly beneficial for accurate human motion prediction. The individual sequence of movements and the overall pattern of movements are both considered, making this method a versatile and robust choice for this task.

3.3. Loss Function

To train our model, we employed an end-to-end training technique. The mean position per joint error (MPJPE) loss [41] function between the anticipated motion sequence and the ground truth motion sequence was used to analyse the difference between the predicted outcomes and the true pose, which is defined as follows:
L = 1 M × T i = 1 M t = 1 T | | x ^ t [ i ] x t [ i ] | | 2 2 ,
where M is the number of joints in human pose, T is the number of time steps in the future series, x ^ t [ i ] is the predicted joint position at the t- t h time step of the i- t h joint, and x t [ i ] is the corresponding ground truth.
We optimised the loss function using the improved Adam method (AdamW [42] ), which mitigates the overfitting problem by adding a weight decay term and can significantly improve the robustness of the model.

4. Experiments

In this section, we present experiments on two large-scale human motion capture benchmark datasets (Human3.6M and CMU-MoCap) to demonstrate the effectiveness of the GGTr network for human motion prediction. We first introduced the experimental settings, including datasets, baselines, and parameter settings. Then, we conducted experiments to compare the performance of the GGTr with other baselines. Finally, we designed comprehensive ablation studies to evaluate the impact of the essential architectural components.

4.1. Datasets

Human3.6M [41] is the largest existing human motion analysis database, consisting of seven actors (S1, S5, S6, S7, S8, S9, and S11) performing 15 actions such as walking, eating, smoking, discussing, and directions. Each pose includes 32 joints, represented in the form of an exponential map. Following the data processing of [18], by converting these into 3D coordinates, eliminating redundant joints, global rotation, and translation, the resulting skeleton retains 17 joints that provide sufficient human motion details. These joints include key ones that locate major body parts (e.g., shoulders, knees, and elbows). We downsampled the frame rate to 25 fps and used S5 and S11 for testing and validation, while the remaining five actors were used for training.
CMU-MoCap (Available at http://mocap.cs.cmu.edu/ accessed on 5 July 2023) is a 3D human motion dataset, released by Carnegie Mellon University, that used 12 Vicon infrared MX-40 cameras to record the positions of 41 sensors attached to the human body, describing human motion. The dataset can be divided into six motion themes, including human interaction, interaction with environment, locomotion, physical activities and sports, situations and scenarios, and test motions. We adopted the same data preprocessing method as described in the literature [43], simplifying each human body and reducing the motion rate to 25 frames per second. Furthermore, seven actions were selected from the dataset to evaluate the model’s performance. No hyperparameters were adjusted on this dataset. We used only the training and testing sets, with the splitting method consistent with common practice in the literature [43].

4.2. Implementation Details

All experiments in this paper were implemented using the PyTorch deep learning framework. The experimental environment was Ubuntu 20.04, utilizing an NVIDIA A100 GPU. During the training process, a batch size of 16 was used and the number of attention heads was set to 4. The number of GRU and transformer layers were both set to 1. The model was optimized using the AdamW optimizer. The initial learning rate was set to 0.003, with a 5% decay every 5 epochs. Training was conducted for 800 epochs, and each experiment was repeated three times to obtain an average result, ensuring a more robust evaluation of the model’s performance. For the input motion prediction, a length of 25 frames (1000 ms) was considered, and the prediction generated 25 frames (1000 ms). The choice and configuration of related hyperparameters are summarized in Table 1.

4.3. Evaluation Metrics and Baselines

The evaluation metrics employed in our study were consistent with those utilized in existing algorithms [33,43]. The mean per joint position error (MPJPE) [41] was adopted as the standard measurement, which computed the average Euclidean distance (in millimeters, mm) between the predicted joint 3D coordinates and the ground truth. In addition, to further illustrate the advantages of the proposed method, a comparative analysis was conducted with several state-of-the-art approaches [15,18,22,23,33,34,43].
Res. sup. [18] is an early RNN-based method. LTD [43], MSR [34], ST-DGCN [33], and LCDC [15] utilize GCN-based methodologies. Meanwhile, LPJP [22] and STCT [23] employ transformer-based approaches. We compared the proposed method with all the above approaches on the Human3.6M dataset, and with approaches [15,18,22,34,43] on CMU-MoCap. By comparing the proposed method with these existing approaches, this study aimed to demonstrate its effectiveness and highlight its advantages in terms of accuracy and performance.

4.4. Experimental Results and Analysis

Human3.6M. Table 2 presents quantitative comparisons of our method and other approaches in predicting short-term (80 ms, 160 ms, 320 ms, 400 ms) and long-term movements (560 ms, 1000 ms) on the Human3.6M dataset. The final column provides the average performance across all tested time intervals for 15 actions.
Among the comparison methods, RNN-based techniques exhibited the poorest performance, while transformer-based approaches outperformed GCN-based methods. Above all, our method emerged as the top performer. The most significant improvements in the MPJPE metric were observed with transformer-based techniques, highlighting their aptitude for modeling 3D human motion dynamics and capturing global dependencies.
Specifically, existing methods generally performed well when predicting periodic and simpler movements, such as “walking” and “eating”. However, their performance dropped significantly when tasked with predicting more unpredictable and irregular movements like “directions”, “posing”, and “purchases”. This indicates that these methods have difficulty managing the dynamic changes and local–global dependencies inherent in human motion.
On the other hand, the algorithm proposed in our study demonstrated a high prediction accuracy, even with highly complex, non-periodic, and irregular movements. Our experimental results revealed that our proposed method surpasses most baseline methods in short-term motion prediction, with even greater improvements noted in long-term prediction. Our method delivered a superior performance in the 560 ms and 1000 ms MPJPE metrics, an accomplishment attributable to GGTr’s ability to fully capture spatial correlation and local–global temporal features, thereby bolstering the model’s prediction accuracy.
While our predictions for movements such as “walking”, “smoking”, “greeting” and “waiting together” fell short compared to those of MSR [34], this nonetheless underscores the sophisticated nature of transformer-based approaches. Looking ahead, we are committed to further refining our model to enhance its performance.
Overall, our proposed method outperformed all the baseline models on average, proving its superior performance in motion prediction. The outstanding performance across both short-term and long-term human motion prediction highlights our model’s effective capacity to capture both local and global temporal dependencies.
CMU-MoCap. To further validate the generalization of the proposed method, we compared its performance with five existing algorithms [15,18,22,34,43] on the CMU-MoCap dataset. The mean per joint position error was calculated for short-term and long-term predictions. The experimental results are shown in Table 3, which includes the actions “basketball”, “basketball signal”, “directing traffic”, “jumping”, “soccer”, “walking”, and “wash window”, as well as the average prediction error across all actions.
Through quantitative evaluation, we clearly observe that our method effectively handles various types of actions and consistently achieves a superior performance across all of them. These empirical findings reinforce the superiority of our approach in human motion prediction, both for short-term and long-term predictions. The consistent and significant performance improvement over state-of-the-art methods on the dataset demonstrates the robustness of our method.
Notably, as can be seen from Table 2 and Table 3, our model performed well on the Human3.6m and CMU-MoCap datasets for prediction tasks up to 320 ms, and even better for tasks above 320 ms. This intriguing performance pivot at 320ms may represent a transition point where prognostic difficulty shifts from short to long-term. We believe this is due to our model’s capability to capture local and global time dependencies across body joints through a strategic combination of GRU and transformer layers, effectively handling this transition. Furthermore, this phenomenon could be tied to the inherent complexity of human movement, especially when the prediction time exceeds 320 ms. This complexity poses a challenge for models that rely on short-term prognosis. However, our model, thanks to its specific structure and training procedure, seems to cope with this challenge more effectively. In future work, we plan to delve deeper into this “prognostic barrier” at 320 ms, aiming to better understand its underlying causes and how we can further improve our model’s performance at this critical transition point. This understanding will potentially enable us to optimize the dynamics of 3D joint position prediction.

4.5. Qualitative Comparison

To enhance our intuitive understanding and facilitate the evaluation of our model, we provided a visualization of the motion prediction results. Figure 5 illustrates three examples of predicted poses using our proposed method, alongside three other baseline methods. The visualization results for the “phoning”, “discussion”, and “purchases” actions of the Human3.6M dataset are presented in Figure 5. The first row in each subplot displays the ground truth pose sequences (in black), followed by the predicted poses (in blue). In other words, each row presents the prediction results from one model.
The visualization results indicate that our model is capable of adequately capturing spatial dependencies and local–global temporal dependencies. It is noticeable that the predictions generated by our method show higher similarity to the actual sequences and better continuity between frames. For instance, in the case of subtler movements such as “phoning”, our model successfully captures the long-distance temporal dependencies concealed within the movement sequence, yielding superior long-term predictions. Moreover, in the “purchases” motion visualization, the movements between hands are more coordinated. This illustrates our model’s proficiency in forecasting highly complicated irregular movements and complex periodic motions.

4.6. Ablation Study

To critically assess the contribution of each component in our model, we performed a set of ablation studies on the Human3.6M dataset. All architecture parameters were kept constant to ensure a fair comparison of each module’s impact. These experiments focused on assessing the impact of the graph convolutional network ( G C N N ), gated recurrent units (GRU), and transformer layers (Tr) on the model’s performance. The experimental results are presented in Table 4. The optimal performance was achieved by integrating these three components.
G C N N : We used the original graph convolution module in place of the proposed module. As shown in the first and third rows of Table 4, when this module is replaced, prediction performance decreases. This result clearly indicates that there is critical spatially relevant information hidden in the adjacent pose. When this information is ignored, it is difficult for the model to capture time evolution trends, resulting in degraded performance.
GRU and transformer layers(Tr): The GRU unit, designed to capture local temporal dependence, shows a remarkable performance, enhancing the accuracy of short-term predictions. The transformer layer exhibits exceptional ability in handling global temporal dependence, which improves the accuracy of long-term predictions. We removed the transformer layer directly from the proposed method. It can be seen that the long-term prediction performance of the model was significantly reduced, and the short-term prediction performance was also slightly reduced.

5. Conclusions

In this paper, we have proposed a novel framework for human motion prediction that leverages the power of position-wise enhanced graph convolutional networks, gated recurrent unit networks, and transformer layers. By strategically combining these networks, our model effectively captures spatial information across body joints and temporally aligns these dependencies, both locally and globally. The proposed framework has shown significant improvements in predicting long-term movements, surpassing existing state-of-the-art methods by a substantial margin. Experiments on the Human3.6M and CMU-MoCap datasets provide evidence supporting the effectiveness of our proposed model. The efficacy of our approach accentuates the potential of integrating advanced neural network architectures for improved understanding and prediction of complex human motion dynamics. Future work will explore the integration of more sophisticated attention mechanisms and deep learning architectures to further enhance prediction accuracy and efficiency.

Author Contributions

Conceptualization, B.H.; methodology, B.H.; software, B.H.; validation, B.H.; formal analysis, B.H.; investigation, B.H.; resources, B.H.; data curation, B.H.; writing—original draft preparation, B.H.; writing—review and editing, B.H.; visualization, B.H.; supervision, X.L.; project administration, X.L.; funding acquisition, X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under grants 62233003 and 62073072, the Key Projects of the Key R&D Program of Jiangsu Province under grants BE2020006 and BE2020006-1, and Shenzhen Natural Science Foundation under grants JCYJ20210324132202005 and JCYJ20220818101206014.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets generated and/or analyzed during the current study are publicly available. The Human3.6M dataset can be accessed through the reference in [41]. The CMU-MoCap dataset is publicly available and can be accessed online at http://mocap.cs.cmu.edu/ accessed on 5 July 2023. The use of these datasets is governed by their respective usage policies.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
GGTrGraph Convolutional, Gated Recurrent Unit, and Transformer layers
GCNGraph convolutional networks
GRUGated recurrent unit
CNNsConvolutional neural networks
RNNRecurrent neural networks
GNNsGraph neural networks

References

  1. Koppula, H.S.; Saxena, A. Anticipating human activities using object affordances for reactive robotic response. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 14–29. [Google Scholar]
  2. Gui, L.Y.; Zhang, K.; Wang, Y.X.; Liang, X.; Moura, J.M.; Veloso, M. Teaching robots to predict human motion. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 562–567. [Google Scholar]
  3. Li, H.; Li, X.; Zhang, Z.; Hu, C.; Dunkin, F.; Ge, S.S. ESUAV-NI: Endogenous Security Framework for UAV Perception System Based on Neural Immunity. IEEE Trans. Ind. Inform. 2023. [Google Scholar] [CrossRef]
  4. Choi, S.H.; Park, K.B.; Roh, D.H.; Lee, J.Y.; Mohammed, M.; Ghasemi, Y.; Jeong, H. An integrated mixed reality system for safety-aware human-robot collaboration using deep learning and digital twin generation. Robot. Comput.-Integr. Manuf. 2022, 73, 102258. [Google Scholar] [CrossRef]
  5. Dong, Y.; Li, X.; Dezert, J.; Zhou, R.; Zhu, C.; Wei, L.; Ge, S.S. Evidential reasoning with hesitant fuzzy belief structures for human activity recognition. IEEE Trans. Fuzzy Syst. 2021, 29, 3607–3619. [Google Scholar]
  6. Sheng, W.; Li, X. Multi-task learning for gait-based identity recognition and emotion recognition using attention enhanced temporal graph convolutional network. Pattern Recognit. 2021, 114, 107868. [Google Scholar]
  7. Kong, Y.; Wei, Z.; Huang, S. Automatic analysis of complex athlete techniques in broadcast taekwondo video. Multimed. Tools Appl. 2018, 77, 13643–13660. [Google Scholar] [CrossRef]
  8. Lehrmann, A.M.; Gehler, P.V.; Nowozin, S. Efficient nonlinear Markov models for human motion. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  9. Wang, J.M.; Fleet, D.J.; Hertzmann, A. Gaussian Process Dynamical Models for Human Motion. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 30, 283–298. [Google Scholar]
  10. Taylor, G.W.; Hinton, G.E.; Roweis, S. Modeling human motion using binary latent variables. In Proceedings of the Advances in Neural Information Processing Systems 19 (NIPS 2006), Cambridge, MA, USA, 4–7 December 2006; Volume 19. [Google Scholar]
  11. Li, C.; Zhang, Z.; Lee, W.S.; Lee, G.H. Convolutional sequence to sequence model for human dynamics. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2275–2284. [Google Scholar]
  12. Li, M.; Chen, S.; Chen, X.; Zhang, Y.; Wang, Y.; Tian, Q. Symbiotic Graph Neural Networks for 3D Skeleton-Based Human Action Recognition and Motion Prediction. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3316–3333. [Google Scholar] [CrossRef]
  13. Li, M.; Chen, S.; Zhao, Y.; Zhang, Y.; Wang, Y.; Tian, Q. Dynamic multiscale graph neural networks for 3d skeleton based human motion prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 214–223. [Google Scholar]
  14. Zhong, C.; Hu, L.; Zhang, Z.; Ye, Y.; Xia, S. Spatio-Temporal Gating-Adjacency GCN For Human Motion Prediction. In Proceedings of the Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  15. Fu, J.; Yang, F.; Dang, Y.; Liu, X.; Yin, J. Learning Constrained Dynamic Correlations in Spatiotemporal Graphs for Motion Prediction. IEEE Trans. Neural Netw. Learn. Syst. 2023. [Google Scholar] [CrossRef]
  16. Fragkiadaki, K.; Levine, S.; Felsen, P.; Malik, J. Recurrent network models for human dynamics. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 4346–4354. [Google Scholar]
  17. Jain, A.; Zamir, A.R.; Savarese, S.; Saxena, A. Structural-rnn: Deep learning on spatio-temporal graphs. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 5308–5317. [Google Scholar]
  18. Martinez, J.; Black, M.J.; Romero, J. On human motion prediction using recurrent neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2891–2900. [Google Scholar]
  19. Liu, Z.; Wu, S.; Jin, S.; Liu, Q.; Lu, S.; Zimmermann, R.; Cheng, L. Towards natural and accurate future motion prediction of humans and animals. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  20. Shu, X.; Zhang, L.; Qi, G.J.; Liu, W.; Tang, J. Spatiotemporal co-attention recurrent neural networks for human-skeleton motion prediction. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3300–3315. [Google Scholar]
  21. Liu, Z.; Wu, S.; Jin, S.; Ji, S.; Liu, Q.; Lu, S.; Cheng, L. Investigating pose representations and motion contexts modeling for 3D motion prediction. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 681–697. [Google Scholar] [CrossRef]
  22. Cai, Y.; Huang, L.; Wang, Y.; Cham, T.J.; Cai, J.; Yuan, J.; Liu, J.; Yang, X.; Zhu, Y.; Shen, X.; et al. Learning progressive joint propagation for human motion prediction. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 226–242. [Google Scholar]
  23. Aksan, E.; Kaufmann, M.; Cao, P.; Hilliges, O. A spatio-temporal transformer for 3d human motion prediction. In Proceedings of the 2021 International Conference on 3D Vision (3DV), London, UK, 1–3 December 2021; pp. 565–574. [Google Scholar]
  24. Yu, H.; Fan, X.; Hou, Y.; Pei, W.; Ge, H.; Yang, X.; Zhou, D.; Zhang, Q.; Zhang, M. Towards Realistic 3D Human Motion Prediction with A Spatio-temporal Cross-transformer Approach. IEEE Trans. Circuits Syst. Video Technol. 2023. [Google Scholar] [CrossRef]
  25. Chiu, H.K.; Adeli, E.; Wang, B.; Huang, D.A.; Niebles, J.C. Action-agnostic human pose forecasting. In Proceedings of the 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa Village, HI, USA, 7–11 January 2019; pp. 1423–1432. [Google Scholar]
  26. Guo, X.; Choi, J. Human motion prediction via learning local structure representations and temporal dependencies. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 2580–2587. [Google Scholar]
  27. Zhou, L.; Zhou, Y.; Corso, J.J.; Socher, R.; Xiong, C. End-to-end dense video captioning with masked transformer. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8739–8748. [Google Scholar]
  28. Liu, P.J.; Saleh, M.; Pot, E.; Goodrich, B.; Sepassi, R.; Kaiser, L.; Shazeer, N. Generating wikipedia by summarizing long sequences. arXiv 2018, arXiv:1801.10198. [Google Scholar]
  29. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  30. Yuan, J.; Cao, M.; Cheng, H.; Yu, H.; Xie, J.; Wang, C. A unified structure learning framework for graph attention networks. Neurocomputing 2022, 495, 194–204. [Google Scholar] [CrossRef]
  31. Velickovic, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. Stat 2017, 1050, 10–48550. [Google Scholar]
  32. Yan, S.; Xiong, Y.; Lin, D. Spatial temporal graph convolutional networks for skeleton-based action recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  33. Ma, T.; Nie, Y.; Long, C.; Zhang, Q.; Li, G. Progressively Generating Better Initial Guesses Towards Next Stages for High-Quality Human Motion Prediction. In Proceedings of the Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  34. Dang, L.; Nie, Y.; Long, C.; Zhang, Q.; Li, G. MSR-GCN: Multi-Scale Residual Graph Convolution Networks for Human Motion Prediction. In Proceedings of the International Conference on Computer Vision, New Orleans, LA, USA, 18–24 June 2021. [Google Scholar]
  35. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  36. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar]
  37. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  38. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable detr: Deformable transformers for end-to-end object detection. arXiv 2020, arXiv:2010.04159. [Google Scholar]
  39. Zhang, J.; Shi, X.; Xie, J.; Ma, H.; King, I.; Yeung, D.Y. GaAN: Gated Attention Networks for Learning on Large and Spatiotemporal Graphs. In Proceedings of the 34th Conference on Uncertainty in Artificial Intelligence 2018, UAI 2018, Monterey, CA, USA, 6–10 August 2018. [Google Scholar]
  40. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  41. Ionescu, C.; Papava, D.; Olaru, V.; Sminchisescu, C. Human3. 6m: Large scale datasets and predictive methods for 3d human sensing in natural environments. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 36, 1325–1339. [Google Scholar] [CrossRef]
  42. Loshchilov, I.; Hutter, F. Decoupled weight decay regularization. arXiv 2017, arXiv:1711.05101. [Google Scholar]
  43. Mao, W.; Liu, M.; Salzmann, M.; Li, H. Learning trajectory dependencies for human motion prediction. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 4317–4326. [Google Scholar]
Figure 1. The process of human motion prediction. The historical sequences of human motion are subjected to feature extraction in terms of both space and time. This involves the analysis of the spatial configuration of the body, as well as the timing and sequence of these movements. The derived spatial and temporal features constitute the foundation for building a comprehensive understanding of the motion patterns.
Figure 1. The process of human motion prediction. The historical sequences of human motion are subjected to feature extraction in terms of both space and time. This involves the analysis of the spatial configuration of the body, as well as the timing and sequence of these movements. The derived spatial and temporal features constitute the foundation for building a comprehensive understanding of the motion patterns.
Electronics 12 03305 g001
Figure 2. The whole architecture of our suggested solution for motion prediction, which employs an end-to-end framework. We encode human poses x 1 : T and feed them to GGTr. The GCN layer aids in understanding spatial relationships among body joints within the human motion network. Following the GRU is a transformer layer, designed to grasp global temporal dependencies. Ultimately, the transformer’s output is harnessed to predict future motion states.
Figure 2. The whole architecture of our suggested solution for motion prediction, which employs an end-to-end framework. We encode human poses x 1 : T and feed them to GGTr. The GCN layer aids in understanding spatial relationships among body joints within the human motion network. Following the GRU is a transformer layer, designed to grasp global temporal dependencies. Ultimately, the transformer’s output is harnessed to predict future motion states.
Electronics 12 03305 g002
Figure 3. Architectural elements in a GRU layer.
Figure 3. Architectural elements in a GRU layer.
Electronics 12 03305 g003
Figure 4. Architectural elements in a transformer layer.
Figure 4. Architectural elements in a transformer layer.
Electronics 12 03305 g004
Figure 5. Qualitative analysis of the Human3.6M dataset: comparing phoning, discussion, and purchases scenarios. (a) Phoning; (b) Discussion; (c) Purchases.
Figure 5. Qualitative analysis of the Human3.6M dataset: comparing phoning, discussion, and purchases scenarios. (a) Phoning; (b) Discussion; (c) Purchases.
Electronics 12 03305 g005
Table 1. The choice and configuration of related hyperparameters.
Table 1. The choice and configuration of related hyperparameters.
Hyperparameter/ConfigValue
OptimizerAdamW
Base learning rate 5 × 10 3
Weight decay 10 2
Batch size16
Warmup epochs5
Epochs800
Table 2. Prediction of 3D joint positions on Human3.6M for all actions. The best results are marked in bold.
Table 2. Prediction of 3D joint positions on Human3.6M for all actions. The best results are marked in bold.
Time(ms)801603204005601000801603204005601000
ActionWalkingEating
Res. sup. [18]29.450.876.081.581.7100.716.830.656.968.779.9100.2
LTD [43]12.323.039.846.154.159.88.416.933.240.753.477.8
ST-Trans [23]8.915.532.138.5--9.421.136.442.3--
MSR [34]12.222.738.645.252.763.08.417.133.040.452.577.1
ST-DGCN [33]10.219.834.540.348.156.47.015.130.638.151.176.0
LCDC [15]11.122.438.845.252.759.87.015.531.739.251.976.2
Ours10.319.534.941.851.154.96.915.031.636.350.271.7
ActionSmokingDiscussion
Res. sup. [18]23.042.670.182.794.8137.432.961.290.996.2121.3161.7
LTD [43]7.916.231.938.950.772.612.527.458.571.791.6121.5
ST-Trans [23]8.815.225.124.5--7.925.739.947.5--
MSR [34]8.016.331.338.249.571.612.026.857.169.788.6117.6
ST-DGCN [33]6.614.128.234.746.569.510.023.853.666.787.1118.2
LCDC [15]6.614.829.836.748.171.210.024.454.567.487.0116.3
Ours6.414.128.433.145.367.39.823.249.558.483.9106.5
ActionDirectionsGreeting
Res. sup. [18]35.457.376.387.7110.1152.534.563.4124.6142.5156.1166.5
LTD [43]9.019.943.453.771.0101.818.738.777.793.4115.4148.8
ST-Trans [23]10.217.842.548.6--13.526.154.073.2--
MSR [34]8.619.743.353.871.2100.616.537.077.393.4116.3147.2
ST-DGCN [33]7.217.640.951.569.3100.415.234.171.687.1110.2143.5
LCDC [15]6.917.441.051.769.199.114.333.572.287.3108.7142.3
Ours6.917.039.248.466.494.613.439.369.481.6103.3137.6
ActionPhoningPosing
Res. sup. [18]38.069.3115.0126.7141.2131.536.169.1130.5157.1194.7240.2
LTD [43]10.221.042.552.369.2103.113.729.966.684.1114.5173.0
ST-Trans [23]15.320.431.438.8--10.622.857.673.7--
MSR [34]10.120.741.551.368.3104.412.829.467.085.0116.3174.3
ST-DGCN [33]8.318.338.748.465.9102.710.725.760.076.6106.1164.8
LCDC [15]8.519.240.349.966.7102.210.125.460.677.3106.5163.3
Ours8.618.439.946.563.899.19.922.657.673.5103.7158.4
ActionPurchasesSitting
Res. sup. [18]36.360.386.595.9122.7160.342.681.4134.7151.8167.4201.5
LTD [43]15.632.865.779.3102.0143.510.621.946.357.978.3119.7
ST-Trans [23]17.332.560.068.3--8.522.947.866.8--
MSR [34]14.832.466.179.6101.6139.210.522.046.357.878.2120.0
ST-DGCN [33]12.528.760.173.395.3133.38.819.242.453.874.4116.1
LCDC [15]12.729.762.375.897.5137.88.819.342.954.374.9117.8
Ours12.328.561.767.991.1126.18.519.140.149.869.6115.2
ActionSitting DownTaking Photo
Res. sup. [18]47.386.0145.8168.9205.3277.626.147.681.494.7117.0143.2
LTD [43]16.131.161.575.5100.0150.29.920.945.056.677.4119.8
ST-Trans [23]9.232.758.865.9--6.816.537.948.5--
MSR [34]16.131.662.576.8102.8155.59.921.044.656.377.9121.9
ST-DGCN [33]13.927.957.471.596.7147.88.418.942.053.374.3118.6
LCDC [15]14.128.057.371.296.1147.38.418.842.053.574.5117.9
Ours13.826.155.365.394.2141.98.117.638.348.772.8110.8
ActionWaitingWalking Dog
Res. sup. [18]30.657.8106.2121.5146.2196.264.2102.1141.1164.4191.3209.0
LTD [43]11.424.050.161.579.4108.123.446.283.596.0111.9148.9
ST-Trans [23]9.220.559.862.2--22.367.8103.8122.5--
MSR [34]10.723.148.359.276.3106.320.742.980.493.3111.9148.2
ST-DGCN [33]8.920.143.654.372.2103.418.839.373.786.4104.7139.8
LCDC [15]8.720.244.355.373.2105.719.641.877.690.2109.8147.7
Ours8.519.744.650.068.6101.517.638.373.584.1103.5135.8
ActionWaiting TogetherAverage
Res. sup. [18]26.850.180.292.2107.6131.134.762.0101.1115.597.6130.5
LTD [43]10.521.038.545.255.065.612.726.152.363.581.6114.3
ST-Trans [23]7.513.727.746.2--11.024.747.757.8--
MSR [34]10.620.937.443.952.965.912.125.651.662.981.1114.2
ST-DGCN [33]8.718.634.441.051.964.310.322.747.458.576.9110.3
LCDC [15]9.119.836.342.750.561.210.423.348.859.877.8111.0
Ours9.119.236.139.848.457.710.021.846.855.174.4105.4
Table 3. Prediction of 3D joint positions on CMU-MoCap for all actions. The best results are marked in bold.
Table 3. Prediction of 3D joint positions on CMU-MoCap for all actions. The best results are marked in bold.
Time(ms)801603204005601000801603204005601000
ActionBasketballBasketball Signal
Res. sup. [18]29.553.191.2106.0128.7157.414.622.139.146.660.089.9
LTD [43]11.721.140.750.668.095.73.46.213.517.927.351.9
LPJP [22]11.621.744.457.3-90.92.64.912.718.7-75.8
MSR [34]10.318.937.747.062.086.33.05.612.516.625.550.0
LCDC [15]9.617.635.444.460.088.42.64.710.413.921.946.2
Ours9.517.532.541.556.888.42.54.611.513.120.645.8
ActionDirecting TrafficJumping
Res. sup. [18]21.838.870.585.3110.3165.130.253.089.4103.9125.6160.5
LTD [43]6.813.429.639.159.6112.817.132.159.872.594.3127.2
LPJP [22]6.212.729.139.6-149.112.927.673.592.2-176.6
MSR [34]6.112.629.439.250.5114.615.228.956.069.192.4126.2
LCDC [15]5.010.023.431.449.399.612.826.154.668.591.8126.1
Ours5.09.922.430.948.693.913.127.253.667.889.8116.9
ActionSoccerWalking
Res. sup. [18]26.547.081.596.2117.9139.114.622.936.140.951.169.5
LTD [43]13.624.344.454.373.1111.66.711.118.121.025.232.4
LPJP [22]9.218.439.249.5-93.96.710.721.727.5-37.4
MSR [34]10.919.437.447.065.3101.96.410.316.920.125.536.8
LCDC [15]10.319.036.845.762.396.96.310.416.118.623.333.6
Ours9.819.035.344.659.792.75.810.215.417.322.933.2
ActionWash WindowAverage
Res. sup. [18]19.331.856.166.083.6125.922.438.466.277.896.7129.6
LTD [43]5.911.324.131.043.466.99.317.132.940.955.985.5
LPJP [22]5.411.329.239.6-79.17.815.335.746.3-100.4
MSR [34]5.410.924.531.845.170.28.215.230.538.752.383.7
LCDC [15]4.89.522.029.042.568.97.313.928.435.950.180.0
Ours4.89.520.328.041.367.47.214.027.334.748.576.9
Table 4. The influence of the G C N N , GRU, and transformer layers (Tr) on the Human3.6M dataset, on average, is notable. These three components of our model significantly contribute to its overall accuracy. The best results are marked in bold.
Table 4. The influence of the G C N N , GRU, and transformer layers (Tr) on the Human3.6M dataset, on average, is notable. These three components of our model significantly contribute to its overall accuracy. The best results are marked in bold.
Human3.6M, MPJPE (mm)
GCN N GRUTr801603204005601000
10.323.248.157.075.7107.4
10.222.747.556.275.3107.0
10.223.547.757.877.2111.0
10.021.846.855.174.4105.4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, B.; Li, X. GGTr: An Innovative Framework for Accurate and Realistic Human Motion Prediction. Electronics 2023, 12, 3305. https://doi.org/10.3390/electronics12153305

AMA Style

Huang B, Li X. GGTr: An Innovative Framework for Accurate and Realistic Human Motion Prediction. Electronics. 2023; 12(15):3305. https://doi.org/10.3390/electronics12153305

Chicago/Turabian Style

Huang, Biaozhang, and Xinde Li. 2023. "GGTr: An Innovative Framework for Accurate and Realistic Human Motion Prediction" Electronics 12, no. 15: 3305. https://doi.org/10.3390/electronics12153305

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop