Next Article in Journal
Twenty Years of Mobile Banking Services Development and Sustainability: A Bibliometric Analysis Overview (2000–2020)
Next Article in Special Issue
Effect of a Taper Intake Port on the Combustion Characteristics of a Small-Scale Rotary Engine
Previous Article in Journal
A Study and Assessment of the Status of Energy Efficiency and Conservation at School Buildings
Previous Article in Special Issue
A Study of Evaluation Method for Turbocharger Turbine Based on Joint Operation Curve
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Encoder–Decoder-Based Velocity Prediction Modelling for Passenger Vehicles Coupled with Driving Pattern Recognition

1
College of Automotive Studies, Tongji University, Shanghai 201804, China
2
Propulsion Control and Software Engineering Department, SAIC MOTOR, Shanghai 201804, China
*
Author to whom correspondence should be addressed.
Sustainability 2022, 14(17), 10629; https://doi.org/10.3390/su141710629
Submission received: 26 July 2022 / Revised: 19 August 2022 / Accepted: 23 August 2022 / Published: 26 August 2022

Abstract

:
To improve the performance of predictive energy management strategies for hybrid passenger vehicles, this paper proposes an Encoder–Decoder (ED)-based velocity prediction modelling system coupled with driving pattern recognition. Firstly, the driving pattern recognition (DPR) model is established by a K-means clustering algorithm and validated on test data; the driving patterns can be identified as urban, suburban, and highway. Then, by introducing the encoder–decoder structure, a DPR-ED model is designed, which enables the simultaneous input of multiple temporal features to further improve the prediction accuracy and stability. The results show that the root mean square error (RMSE) of the DPR-ED model on the validation set is 1.028 m/s for the long-time sequence prediction, which is 6.6% better than that of the multilayer perceptron (MLP) model. When the two models are applied to the test dataset, the proportion with a low error of 0.1~0.3 m/s is improved by 4% and the large-error proportion is filtered by the DPR-ED model. The DPR-ED model performs 5.2% better than the MLP model with respect to the average prediction accuracy. Meanwhile, the variance is decreased by 15.6%. This novel framework enables the processing of long-time sequences with multiple input dimensions, which improves the prediction accuracy under complicated driving patterns and enhances the generalization-related performance and robustness of the model.

1. Introduction

Energy management strategies are used to determine the power distribution between different energy sources at each moment to improve the performance of hybrid vehicles while satisfying torque requirements and are one of the key technologies of hybrid vehicles [1]. Up to now, scholars have conducted a great deal of research on energy management strategies, which are mainly divided into three major categories: rule-based, optimization-based, and learning-based [2,3,4]. Rule-based energy management strategies (RB-EMSs) are designed with suitable control rules for energy distribution according to the characteristics of the controlled object. The process of design depends mainly on the designers’ knowledge and continuous exploration and experimentation. Moreover, the differences in the control rules between different models require repeated calibrations and cause difficulty with respect to obtaining optimal results. RB-EMSs are widely used due to their simple logic, easy implementation, and high stability [5]. As its research progressed, the addition of fuzzy logic improved the performance of RB-EMSs [6]; however, on the one hand, researchers continue to pursue higher requirements for all aspects of hybrid vehicle performance, and on the other hand, the disadvantages of RB-EMS will gradually appear due to its inherent poor portability, cumbersome tuning of parameters, and inability to achieve optimal control. Optimization-based energy management strategies use optimization algorithms for energy management, which are less empirically dependent than RB-EMSs and can achieve better results. Instantaneous optimization algorithms are those algorithms that do not need to know all the path-related information and can be optimized using only the current information, including the equivalent fuel consumption minimization strategy (ECMS) [7,8] and adaptive equivalent fuel consumption minimization strategy (A-ECMS) [9,10]. The optimization effect of both ECMS and A-ECMS are closely related to the equivalence factor and often have some distance from the global optimal solution. The global optimization algorithm represented by dynamic programming (DP) can obtain a global optimal solution, but at the same time, DP needs to anticipate the vehicle-state information of the entire future driving cycle in advance. The optimal strategy is obtained only for that particular cycle. Therefore, given the complex and variable driving conditions in the real world, such energy management strategies cannot be applied in practice and are often used as references for evaluating and optimizing other energy management strategies [11,12]. With the development of energy management, algorithms such as model predictive control (MPC) obtain locally optimal solutions by continuous roll-forward optimization within the prediction-sight distance, which are neither short-sighted nor sensitive compared to instantaneous optimization algorithms. An advanced knowledge of all the future driving information is not required compared to global optimization algorithms, so it has more potential for real-vehicle applications [13]. With the development of artificial intelligence technology, learning-based algorithms based on Reinforcement Learning (RL) have gained widespread attention, as they can learn energy management criteria from training samples [14], be applied to different driving conditions [15], adapt to different prediction sight distances [16], require no discretization of state nor control parameters [17], and provide the possibility of finding a theoretical global optimal solution [18]. In summary, short-term predictive energy management strategies that obtain locally optimal solutions by rolling the prediction sight distance are the focus of the current research. Among them, the short-term vehicle velocity prediction algorithm is the top priority of the predictive energy management strategies, which greatly affects the performance of these strategies [19,20], and the velocity prediction algorithm also plays a crucial role in the fields of traffic flow prediction and active vehicle safety control [21].
The mainstream of short-term vehicle velocity prediction methods is divided into stochastic and deterministic vehicle velocity prediction algorithms [22]. The Markov chain-based vehicle velocity prediction algorithm is the most representative stochastic vehicle velocity prediction algorithm [23,24]. Although the accuracy of vehicle velocity prediction can be improved by establishing a multi-level Markov chain model, it leads to an exponential expansion of the size of the probability matrix, which leads to an exponential rise in the computation time of the algorithm, and even so, it is still difficult to cover all possible Markov states [25]. Deterministic velocity prediction algorithms can be further divided into parametric and non-parametric velocity prediction algorithms. A parametric vehicle velocity prediction algorithm refers to the prediction by establishing a model with parameters, and the typical case is to establish an auto-regressive moving average model (ARMA) for a time series analysis, which is proven to have a wide range of engineering-control application cases [26]. However, for vehicles driving on roads, the randomness of the travel path and the lack of information sharing of individual vehicles leads to limitations on the ability of parametric vehicle velocity prediction models, causing larger prediction errors than non-parametric vehicle velocity prediction algorithms [27]. Instead, non-parametric velocity prediction algorithms use historical data to build a prediction model for future velocity prediction, so they are also called data-driven velocity prediction algorithms and currently represent a growing trend [28]. The neural network model is a typical data-driven algorithm due to its powerful nonlinear mapping capability and strong robustness. It has become one of the key technologies in the field of predictive modeling and optimal control of nonlinear systems [29] and is also widely applied to the field of vehicle velocity prediction. Scholars have conducted in-depth research on neural network-based vehicle velocity prediction methods and fruitful results have been achieved. Artificial neural networks (ANN) [16], long short-term memory neural networks (LSTM) [30], nonlinear autoregressive neural networks with additional inputs (NARX) [31], and RBFNN [32] are several neural networks that have been mostly studied in the field of vehicle velocity prediction. Among them, LSTM networks are widely used in time series prediction models and can maintain the internal input memory and compensate for the gradient vanishing and gradient explosion that exist in RNN during the training process [33]. In the field of vehicle velocity prediction, LSTM networks are currently used as the “Encoder” structure with a multi-step input and single-step output, and the “Decoder” structure with a single-step input and multi-step output. The former can achieve the effect of sequence prediction by combining the output as a new input, but the velocity of computation needs to be improved; the latter can achieve sequence prediction directly, but the structure lacks memory for the information of long history sequences.
Up to now, there are still few relevant studies on the sequence-to-sequence vehicle velocity prediction model established by the encoder-decoder framework. Nevertheless, with the continuous development of vehicle-networking technology, the information interaction data provided by vehicle networking can also be used jointly with neural networks in the future [34], and the relevant information used as the input of neural networks can also improve the vehicle velocity prediction accuracy and robustness. Therefore, it is necessary to provide a solution for the encoder–decoder-based time series prediction models with different input and output dimensions. In this paper, an innovative encoder–decoder-based velocity prediction modelling system for passenger vehicles coupled with driving pattern recognition is proposed. The model uses the encoder–decoder framework with LSTM units as neurons to establish a multi-input sequence-to-sequence model based on the characteristics of vehicle velocity sequences and the results of driving pattern recognition. Moreover, with the continuous development of the Internet of vehicles, an increasing amount of time-domain information can be accessed by vehicles. The multi-input feature of this model can improve the efficacy of velocity prediction by rearranging the input samples and retraining them as a “skeleton”, which is very rare in the existing similar model. Section 2 introduces the establishment of the method of the driving-pattern-recognition model and recognition performance as the foundation for the following velocity prediction model. Section 3 proposes the velocity prediction based on the Encoder–decoder model, including the revised structure under the different dimensions of input and output sequences and the training method. The effects of the history window, prediction window, and number of neurons on the model performance are analyzed in Section 4, and the performance of the optimal model is compared with that of the traditional neural network. Finally, in Section 5, conclusions are presented.

2. Driving Pattern Recognition

This section will take the accuracy of recognition, the real-time calculation speed, and the compilation feasibility into consideration and find a method that can use the current and historical driving information for driving pattern recognition. In this section, the K-means clustering algorithm is used for driving pattern recognition. Firstly, the training matrix is obtained with standard driving cycles, and the K-means clustering algorithm model is trained. Then, the model is used to recognize the driving patterns in the test data and the results are shown.

2.1. Dataset and Characteristics

One of the keys to recognizing driving patterns is to find a set of features that can represent certain characteristics of driving pattern information, such as the average velocity of the past period of time, idle ratio, vehicle acceleration, etc., and they are used as the basis for recognizing the current driving patterns of the vehicle. In recent years, the selection of features to reflect the characteristics of driving patterns, which can then be applied to the process of driving pattern recognition, has been a hot spot of research, in which the difficulty lies in the appropriate number and type of features to reflect the characteristics of driving patterns. As far as the number of features is concerned, selecting too many features or combining them with the method of principal component analysis for driving pattern recognition may lead to an increase in the difficulty of real-time online analysis, whereas selecting too few features will not accurately reflect the characteristics of driving patterns and lead to the poor accuracy of driving pattern recognition [35]. Hu et al. selected four features—to name three: the average velocity, idle ratio, and cruising time ratio—for driving pattern recognition, which ensured the accuracy of the model and greatly improved the computational real time [36].
In view of the above research results, three features have been selected in this paper to balance the real-time properties, accuracy, and compilation feasibility of driving pattern recognition, including the average velocity (v_mean), maximum velocity (v_max), and idle ratio (T_idle).
After the selection of the features, typical driving cycles need to be selected as the training set; six standard driving cycles are selected in this paper. The features of each standard driving cycle are calculated separately and used to train the K-means clustering model, so that the model can realize the real-time recognition of any driving patterns in actual driving. Standard driving cycles are mostly designed for the driving pattern s in their respective regions. In order to enhance the adaptability of the traffic conditions in different regions, this paper takes into full consideration the representative driving cycles of various regions in the world when selecting the typical driving cycles so that the regional limitations of the recognition on driving patterns is reduced. As a result, this paper refers to six representative standard driving cycles, including the WLTC, which is the most commonly used in the world; the NEDC introduced by the European Union; the CLTC-P introduced by China; and the Artemis cycles introduced by the United States. The velocity versus time of the standard driving cycles selected in this paper are shown in Figure 1.
After the typical driving cycles are selected, the division method of the driving cycle samples needs to be determined. The aim of the driving cycle division is to build a sample database, which is the premise of the driving pattern clustering analysis. In the existing work, there are two methods used to divide the driving cycle samples: one is the division method based on a certain time window, and the other is the division method based on micro-trip driving cycle samples. In this paper, we adopt the division method based on a certain time window, as shown in Figure 2. When the time window is determined (the length of the time window is selected as 60 s in this paper), the velocity sequence is intercepted every 60 s as a driving cycle sample from the starting point to the end.

2.2. K-Means Cluster Analysis

A K-means algorithm is applicable to large sample data for clustering and enabling a more reasonable classification. The optimization of K-Means entails minimizing the within-cluster sum of squared errors (SSE), as shown in Formula (1):
S S E = m = 1 k x C i d i s t X , c i 2 = m = 1 k x C i X , c i
where d i s t X , c i is the distance between object X and the clustering center ci, and dist denotes the Euclidean distance between any two objects in D.
The SSE can be analyzed by the elbow rule. The SSE decreases with the increase in the number of clustering results, but for data with a certain degree of differentiation, the SSE improves greatly when a certain point is reached and decreases slowly afterwards, and this point can be considered as the point with the best clustering performance.
For a reasonable evaluation and analysis of the clustering performance, the Silhouette Coefficient (SC) evaluation method is also a commonly used evaluation method [37]. Assuming that the average distance of the objects in the clusters containing data x and x’ is defined as a(x), b(x) is the minimum average distance from the cluster containing x to all the clusters that do not contain x′. a(x), b(x), and the Silhouette coefficient S(i) are defined as follows:
a x = x C i , x x x x C i 1
b x = min C j : 1 i K , x x x C j x x C j
S i = b i a i max b i a i
where a(i) denotes the maximum distance of the sample i from other samples in the same cluster, and b(i) denotes the minimum distance of the sample i from all samples in the other clusters.
The optimal number of clustering centers can be determined based on S(i). The closer S(i) is to 1, the more reasonable the sample clustering result; the closer S(i) is to −1, the less accurate the sample clustering result; if S(i) is close to 0, the sample is at the boundary position between two clustering results.
Combining the SC evaluation method and the elbow rule of clustering, the SSE and S(i) versus the number of driving pattern clustering centers are calculated, respectively, as shown in Figure 3.
It is evident from Figure 3 that the changing rate of SSE is greatly decreased when the number of clustering, k, is taken as 3 or 4, so k = 3 or k = 4 can be considered as the number of clustering; the Silhouette coefficient is 0.55 when k is taken as 3 and 0.50 when k is taken as 4. So, the clustering result is more reasonable when k is taken as 3 by the SC evaluation method. As a result, the number of clustering selected in this paper is 3. Combined with the actual driving scenarios, these 3 patterns are defined as urban, suburban, and highway.
The clustering results are shown in Figure 4 on training data and under the three-dimensional coordinates of the average velocity, idle ratio, and maximum velocity.
Figure 4a shows the prediction results from standard cycles as training data. The driving-pattern-recognition model can match the patterns represented by the different time axes of each standard cycle. Further analysis of the clustering results in Figure 4b shows that the K-means clustering center of red points is (18.10, 0.31, 38.00), which has a low average velocity, a high idle ratio, and a low maximum velocity, so the red points represent a congested urban driving condition. The K-means clustering center of the blue points is (100.60, 0, 111.94); the average velocity of this cluster is high, the idle ratio is 0, and the maximum velocity is also high, so the blue points represent smooth highway-driving conditions. The K-means clustering center of the green points is (48.80, 0.04, 65.74), and the average velocity, idle ratio, and maximum velocity of this cluster are between the congested urban driving condition and the smooth highway driving condition. As a result, these points represent the suburban driving conditions. Thus, the driving conditions can be broadly classified into three typical patterns, namely, urban, suburban, and highway. The classification results are consistent with the actual driving conditions mentioned above, so the recognition results of the model can be considered reasonable.

2.3. Driving Pattern Recognition

Since only three features are used for driving pattern recognition in this paper, after ensuring the real-time model calculation and the low compilation difficulty of the algorithm, in order to further illustrate the accuracy of the driving pattern recognition, actual road spectrum data are collected as a test dataset to verify the recognition accuracy of the model based on the K-means clustering algorithm. The driving-pattern-recognition model will calculate the features in real time and output the recognition results based on the historical 60 s data during the driving process. Due to some abnormal data on velocity caused by the driver’s operation and GPS device in the actual road spectrum data (as is the case for many real-world applications), the impulse noise will affect the characteristic of v_max, which may cause an instance of incorrect recognition by the model. When applying the driving-pattern-recognition model to practical predictions, the measured velocity signal should be smoothed using the moving average-filtering method to reduce the random impulse noise [38]. As shown in Formula (5), the basic idea of the moving average filtering method is to perform local averaging on a small number of points along the data of length N, thereby filtering out random noise.
y f k = 1 m k m 1 / 2 k + m 1 / 2 y i
where y · represents the original data; y f · represents the filtered data; m is the number of data points used for filtering, which is odd; and i and k represent the data point number. In this paper, m is selected as 5, which represents a 5-second moving average.
Figure 5 represents the prediction results of the driving-pattern-recognition model on the filtered test data.
As shown in Figure 5, the total length of the actual road spectrum data is about 16,000 s, which includes an about 3500 s length of the urban pattern, a 6250 s length of the suburban pattern, and a 6250 s length of highway pattern. Three patterns are shown in the figure as driving patterns one, two, and three, which are indicated by the blue dashed line. The red solid line indicates the prediction result of the driving-pattern-recognition model based on the K-means algorithm. The comparison shows that the driving-pattern-recognition model can identify urban patterns completely and accurately, and the recognition accuracy rate reaches 100%. The suburban pattern is more complicated due to the intermingling of vehicles driving at medium and low velocity or even idling, so the model identifies some of the segments with lower average velocities (e.g., 9000~9100 s segments) or longer idling ratios (e.g., 9600~9800 s segments) as an urban pattern. The prediction result of the highway patterns is also accurate, among which the velocity in the segment 12,750–12,950 s has a sudden drop, which may be due to unexpected traffic accidents or other unexpected situations in the highway pattern at the time of driving, so it cannot be used as a typical feature of a highway pattern; instead, it is more reasonable for it to be identified as a suburban or urban (congested) pattern. In summary, the driving-pattern-recognition model in this paper ensures the real-time calculation speed and low compilation difficulty of the algorithm, while its recognition accuracy is satisfactory, reaching 84.1%, which can be used as the basis for subsequent research.

3. Velocity Prediction Based on the Encoder–Decoder Model

3.1. Basic Conception of LSTM

The LSTM network is an improvement of RNN with a unique store-and-forget function compared to the traditional RNN. By learning the sequence input and extracting the hidden sequence features, it can obtain the dependency relationship between sequences accurately, and overcome the complications of gradient vanishing, gradient explosion, and long-term memory disappearance that occur in RNN during training [39].
There are three inputs to the LSTM network at moment t: the input value xt at the current moment, the output value ht−1 at the previous moment, and the cell state ct−1. The input gate it, the output gate ot, and the forget gate ft receive the same inputs [ht−1, xt], which are used to control the update process of the cell state ct after the activation function σ. The basic LSTM network is shown in Figure 6.
A gate is a fully connected layer that receives a vector as an input, while the output is a vector of real numbers between 0 and 1. W is the weight matrix of the gate and b is the bias; then, the gate can be represented as:
g x = σ W x + b
where σ denotes the activation function. In this paper, the Sigmoid function is used as the activation function to adjust the input [ht−1, xt] to the (0,1) interval.
The input gate it controls how much of the input xt is saved to the cell state ct at the current moment, and is defined as:
i t = σ W i h t 1 , x t + b i
The forget gate ft controls how much of the cell state ct−1 of the previous moment is saved to the current moment state ct, and is defined as:
f t = σ W f h t 1 , x t + b i
The output gate ot controls how much of the state ct at the current moment is output to the current output value ht, and is defined as:
o t = σ W o h t 1 , x t + b i
The cell state c ˜ t of the current input is described according to ht−1 and the current input xt, and is defined as:
c ˜ t = t a n h W c h t 1 , x t + b c
The cell state ct is adjusted by the input gate it and the forget gate ft and is defined as:
c t = f t c t 1 + i t c ˜ t
The final output ht of the network is determined by both the output gate ot and the cell state ct and is defined as:
h t = o t t a n h c t
In Formulas (7)–(10), the matrices Wi, Wf, Wo, and Wc are the gate weight matrices and the vectors bi, bf, bo, and bc are the bias terms of the gates.
The information used for updating the cell state ct, ft, and ot are determined by the gating vectors in Formulas (7)–(10). The cell state and output are updated by Formulas (11) and (12). The cell state is reset or restored by ft and the new state ct is obtained by adding partial information through the input gate it, while the hidden state ht is controlled and updated by the output gate ot.

3.2. An Encoder–Decoder Structure Coupled with Driving Pattern Recognition

The encoder–decoder framework is a structure for time series analysis. In this paper, LSTM neurons are used for establishing the encoder–decoder model and coupled with the driving pattern recognition results obtained above. This multi-input model inputs both the vehicle velocity and driving pattern recognition sequences and the constructed input sequences are encoded and decoded to achieve vehicle velocity prediction, as shown in Figure 7. The driving pattern recognition-encoder decoder (DPR-ED) model mainly contains two parts: the encoder side and the decoder side.
The operation of the ED-based model consists of two phases: training and testing. The training phase is for learning the implied knowledge and experience from the training data, and the weight matrix of the model is trained for the testing phase; the testing phase applies the trained model and the input data for testing to calculate the corresponding output data. The training and prediction samples are the preparatory work for running the model. All the above-mentioned training and prediction datasets of the DPR-ED model are closely related to the parameters of the history window (Wt), which represents the length of the history series, and the prediction window (Wn), which represents the length of the prediction series. For the traditional ED framework shown in Figure 7a, the input of the training data is a W t × M matrix and the output of the training data is a W n × M matrix; the input of the prediction data is a W t × 1 vector and the output of the prediction data is a W n × 1 vector, where M is the number of training samples. For the DPR-ED model proposed in this paper, shown in Figure 7b, the training data input is a C × W t × M tensor and the training data output is a W n × M matrix; the prediction data input is a   C × W t matrix and the prediction data output is a W n × 1 vector, where C is the number of features. The number of features in this paper is two, namely, vehicle velocity and driving pattern.
The operation process of the DPR-ED model is as follows. Using the time window sliding method, M training input samples are intercepted from the sequence of typical driving cycles and the corresponding driving pattern sequence, each of which is in the form of v 1   r 1 v 2   r 2   v t   r t , forming the training input data tensor X t r a i n = 2 × W t × M . According to the same method, M training output samples are intercepted from the sequence of typical driving cycles and the corresponding driving pattern sequence. Each of them is in the form of v t + 1   v t + 2   v t + n , forming the training output data matrix Y t r a i n = W n × M . The number of training samples obtained by the time window-sliding method in this paper is 7914, that is, M = 7914. Similarly, the validation input data tensor X v a l = 2 × W t × N and the validation output data matrix Y v a l = W n × N are obtained. In this paper, the time window-sliding method is used to randomly select one-tenth of the actual road spectrum data as the validation set. Such a small amount of extraction does not affect the persuasiveness of the model analysis on the test set, and even improves the performance of the model by using the complete input of typical driving cycles as training. The number of validation samples in this paper is 1594, that is, N = 1594. After the model is trained, the test data input matrix X t e s t = 2 × W t is intercepted from the test dataset using the time window-sliding method. The prediction result will be calculated by the trained model as the data output vector Y t e s t = W n × 1 , which is the predicted vehicle velocity sequence.
In terms of the structure of the model, the number of layers of the DPR-ED model in this paper is fixed to a single layer. Setting the number of nodes in the input layer—which represents the length of the historical velocity sequence, Wt—too small may decrease prediction accuracy for a lack of historical information, while setting it too large may increase the complexity of the model structure and thus decrease the computational speed. Regarding the number of nodes in the output layer, which represents the length of the predicted vehicle velocity sequence, Wn, it is indicated in the literature [22] that a prediction sight distance between 1~10 s is beneficial for the PEMS effect of HEVs, while [40] determines it as 5 s and obtains the best results. Considering the above, in the subsequent study of this paper, the range of Wn is set as 1~5 s, and the range of Wt is set as twice the length of Wn, which is 2~10 s. A detailed parameter study and optimization of Wn and Wt are carried out in the subsequent section. The number of neurons in the hidden layer determines the performance of the neural network. Too few neurons will result in the model’s inability to obtain enough fitting features and too many will cause the model to run slower and be prone to overfitting, so both conditions will lead to poor prediction results of the trained neural network. The number of hidden-layer neurons is also one of the important parameters to be studied and optimized in detail in this paper.
In order to evaluate the prediction accuracy, different evaluation methods have been used in the literature, the most commonly used of which is the root mean square error (RMSE), which reflects the error between the predicted vehicle velocity and the actual vehicle velocity at each step of the predicted sight distance, as shown in Formula (13):
R M S E = 1 n k = 1 n Y P k Y A k 2
where n is the length of predicted sight distance, which represents the length of the predicted vehicle velocity. It is equal to the number of output nodes of the DPR-ED model. Y P k is the predicted vehicle velocity value corresponding to the kth predicted sight distance ( k 1 , n );   Y A k is the actual vehicle velocity value corresponding to the kth predicted sight distance point.
Since the RMSE can only reflect the prediction accuracy at each forecasting step and cannot be used to evaluate the average prediction accuracy of the whole forecasting process, the introduction of an average root mean square error (RMSEA) is proposed as shown in Formula (14):
R M S E A = 1 M s t e p = 1 M R M S E
where M is the number of samples. For the sake of the conciseness of expression, the RMSE is used to denote the average root mean square error in all the subsequent research in this paper.
In this paper, to manifest the performance enhancement effect of the DPR-ED model, the traditional MLP model is also established for comparison. The specific method can be found in the literature [41] and the method for comparison draws on the experience of [42].

4. Results and Discussion

Firstly, in order to obtain the optimal model performance, the relationship between the number of neurons of the traditional MLP model and the number of LSTM neurons of the ED-based models versus the fitting accuracy needs to be investigated. In the training process of the models, both the training and validation datasets were normalized. The RMSE was used as the loss. The Adam algorithm was used to update the neural network weights and the early stopping mechanism was introduced to optimize the epochs of training to prevent the overfitting of the model. The parameter of early stopping was defined as patience = 10, which means the training is stopped when the performance of the loss in the validation set is not further improved after ten consecutive weight updates. Finally, in order to better evaluate the compatibility, the accuracy and stability of the model with respect to the parameter length, the history window, and the prediction window were selected as the maximum values (10 s and 5 s) in this study, so that the trained model would have better compatibility with simpler cases. Each model was randomly assigned initial weights, which was repeated five times during training, and the results were analyzed by box plots. The influences of the different numbers of neurons on the learning performance of the traditional MLP model, Basic ED model, and DPR-ED model for the validation set are shown in Figure 8.
Figure 8a shows the accuracy of the traditional MLP model versus different numbers of hidden-layer neurons in the validation set, where each neuron is a perceptron. The mean and standard deviation of the RMSE of the MLP model in the validation set show a trend of decreasing and then increasing as the number of hidden-layer neurons of the MLP model increases, and reach the minimum values of 0.000981 and 0.000021, respectively, when the number of hidden-layer neurons is selected as 30. The accuracy of the basic ED model and DPR-ED model versus different numbers of hidden-layer neurons in the validation set are shown in Figure 8b,c. The neurons in both models are LSTM units. The result indicates that when the number of hidden-layer neurons is taken as 10 for the basic ED model and less than 20 for the DPR-ED model, the performance of the models still cannot be improved (poor stability and accuracy) even after the iterations are completed, which is due to the fact that a simple network is unable to present the characteristics of the historical velocity and the impact of historical information on the future velocity. As the number of hidden-layer neurons increase, the model performance approaches excellence and stability. Therefore, the optimal numbers of hidden-layer neurons of the basic ED model and DPR-ED model are 20 and 30, respectively. The mean values of the RMSE are 0.000872 and 0.000862 and the standard deviations of the RMSE reach 0.000023 and 0.000032, respectively.
Through the above analysis, the structure of the traditional MLP model is determined as a network with 30 hidden-layer neurons in a single layer, and the structure of the basic ED model and the DPR-ED model is determined as a network with 20 and 30 hidden-layer neurons (LSTM units) in a single layer in the subsequent study. The reason why the DPR-ED model has more hidden-layer neurons than the Basic ED model is probably because the DPR-ED model accepts a multi series input and therefore requires more LSTM units to accurately render the corresponding features.
Figure 9 shows the accuracy of each model versus the history window and prediction window on the test set and comparison.
Figure 9a,b show the impact of different history and prediction windows on the accuracy of the traditional MLP model and the DPR-ED model on the test set. From the results of these two figures, it can be clearly seen that both the history and prediction window directly affect the prediction accuracy of the model. For the MLP model, the history window has a slight effect on the prediction accuracy of the model (as seen from the equipotential line); that is, the larger the history window for the same prediction window case, the more the RMSE decreases. For the DPR-ED model, the change of the history window has little effect on the prediction accuracy of the model, so the input sequence length of the model is also less sensitive. The prediction accuracy decreases significantly as the prediction window increases, so in the process of its application in a real vehicle, an excessively large prediction window is generally not used. In addition, it is obvious from the figure that the accuracy of the DPR-ED model is higher than that of the MLP model to different degrees under the same value of the history window and the prediction window; the performance of the DPR-ED model is slightly higher than that of the MLP model when the prediction window is small (e.g., 2 s) For instance, the RMSE of the MLP model at a history window of 6 s is 0.483 m/s, while the RMSE of the DPR-ED model is 0.476 m/s, which demonstrates a slight improvement of 1.4%. When both the history and prediction window become larger, the DPR-ED model shows a greater advantage: the RMSE of the DPR-ED model is 1.028 m/s while the RMSE of the MLP model is 1.096 m/s at the history and prediction windows of 10 s and 5 s, respectively, which is a 6.6% deterioration in performance compared to the former. Figure 9c shows the performance improvement of the DPR-ED model compared to the basic ED model in the case of long prediction sequences. In addition to the role of the ED frame, which can better handle time series problems, when the prediction window is large, the LSTM neurons of the DPR-ED model receiving multidimensional inputs are also able to parse more complex information, so the accuracy is further improved by about 2.5% compared to the basic ED model.
Figure 10 shows the prediction results of the DPR-ED model from the test set. Since the total simulation time is too long and too many data points are not conducive to graphical analysis, the results of the urban, suburban, and highway patterns are intercepted to 1000 s each. In order to analyze the prediction performance for the five steps more clearly, the local velocity prediction trajectories for two random time points with different driving pattern characteristics are compared in Figure 11.
The red curve in Figure 10 represents the velocity of the five future-prediction steps, and the blue curve represents the actual velocity data. Under the different driving patterns, the DPR-ED model can basically respond to the transient process and adjust the predicted velocity within 5 s, which shows that the DPR-ED model proposed in this paper is effective. Figure 11a shows the typical characteristics of the urban driving pattern and its prediction results. It is shown from Figure 11a that the urban driving pattern contains frequent accelerations and decelerations, and the predicted velocity is close to the actual velocity when the vehicle experiences low deceleration or a smooth large deceleration at 767 s and 774 s. If a sudden change in acceleration or deceleration occurs (reflecting the various complex road conditions that occur in the urban driving pattern during actual driving), (i.e., 757 s), the predicted velocity deviates from the actual velocity, but still follows the trajectory of the trend. Figure 11b shows the typical characteristics of the highway driving pattern and its prediction results. Figure 11b indicates that the highway driving pattern shows a relatively gentle high-speed cruise or small speed changes at most of the time. The predicted velocity of the DPR-ED model also shows a relatively gentle change trend in the highway driving pattern, which is close to the actual speed. It is evident that since the DPR-ED model additionally accepts the driving pattern sequence information, it can better predict the typical characteristics under different driving patterns.
It is worth noting that even though the predicted velocity corresponds well to the actual velocity, the predicted velocity seems quite noisy, especially in the city pattern. Firstly, the unfiltered velocity could be one of the reasons for this phenomenon, as the impulse noise of the data itself will definitely affect the prediction result. Figure 12 shows the prediction results of the DPR-ED model on the city part of the filtered test set. The filtering method is five-second moving average, which has been mentioned above. Secondly, the ability to change the speed of the applied vehicle should be examined and limited while the torque limitation exists for the power system of a certain vehicle. Figure 13 shows the acceleration distribution of the prediction results by the DPR-ED model on the raw test set.
As shown in Figure 12, after filtering the impulse and high frequency noise in the raw test data, the velocity prediction stability is improved significantly in the area where the sudden change is large. The RMSE of the prediction results from the filtered test set is an exceptional 0.79 m/s. It indicates that a low-noise vehicle speed signal plays a crucial role in the efficient operation of the model. Achieving signal noise reduction is not within the scope of this paper, so the following analysis is based on raw data. However, for real-world applications, high-precision GPS and advanced signal-processing techniques should be coupled as much as possible. Figure 13 shows that the acceleration distribution is an approximate normal distribution, while the acceleration proportion is slightly more than the deceleration proportion. A small number of data points exhibit a changing rate of velocity greater than 1 m/s, and there are individual points with a changing rate greater than 1.5 m/s, which may also add to the noise of the model. When the power limit or driving style does not allow for such an acceleration, a clipping method can be applied during the real-world application of the model to correct the prediction results within the allowed acceleration range.
Figure 14 shows the local RMSE of the MLP model and the DPR-ED model during the process of the velocity prediction for the whole test set.
From Figure 14, it is evident that the local RMSE has the lowest average value in the highway driving pattern and the highest average value in the suburban driving pattern, which is due to the fact that the velocity changes in the highway driving pattern are smoother, while the complex suburban driving pattern is full of predictive uncertainties. The local RMSE is larger when the vehicle is operating under a large rate of change in acceleration, which is due to the fact that large prediction errors usually occur when the vehicle suddenly starts to accelerate after deceleration or in a situation of emergency deceleration after acceleration; such driving conditions are challenging to predict and should be addressed with the involvement of the driver. In addition, it can be found that the DPR-ED model improves the accuracy compared to the MLP model when the vehicles are working in a large rate of change of acceleration, although the local RMSE shows a peak (especially for the highway driving pattern). For example, the local RMSE of the MLP model’s predicted velocities at times 13,538 s and 14,805 s are 5.21 m/s and 5.05 m/s, respectively, while the results predicted by the DPR-ED model at the same time are 4.88 m/s and 4.61 m/s, respectively, which shows an improvement of 6.3% and 8.7%.
Figure 15 shows the comparison of the error distribution between the MLP model and the DPR-ED model for the test data. Table 1 shows the proportion of the prediction results for the test set contained in each error segment.
As shown in Table 1, the results predicted by the DPR-ED model and the MLP model from the test data with an error of 0~0.1 m/s have similar proportions. However, the results predicted by the DPR-ED model from the test data with an error of 0.1~0.3 m/s have a proportion of 23.55%, which is about 4% higher than the result obtained by the MLP model. Therefore, more prediction results are concentrated on segments with small errors. On the other hand, the DPR-ED model has also eliminated the data with an error of more than 0.5 m/s. It reduced the proportion with an error of 0.5~0.7 m/s and over 1.1 m/s by about 1% and 1.9%, respectively; thus, the DPR-ED model filters out more of the results with larger errors. As a result, the DPR-ED model shows an excellent prediction robustness over different driving patterns.
Consequently, compared with the MLP model’s poor prediction accuracy on certain loads due to the discrete characteristics of the training data load, the optimal partially shared network shows an excellent prediction robustness over the full load range.
As shown in Figure 15, it can be found that the average error of the MLP model and the DPR-ED model prediction result is 0.81 m/s and 0.77 m/s, respectively. The DPR-ED model performs 5.2% better than the MLP model on the average prediction accuracy. Moreover, the variance is decreased greatly from 0.52 to 0.45, which is an improvement of 15.6%. This result shows that the DPR-ED model significantly improves both the accuracy and the stability.

5. Conclusions

In this paper, based on the samples from typical driving cycles, a high-precision velocity prediction model was established through the encoder–decoder framework coupled with driving pattern recognition. The main achievements and conclusions are as follows:
(1)
The driving-pattern-recognition model was established by a K-means clustering algorithm and validated based on the test data; the driving patterns were identified as urban, suburban, and highway patterns. The model achieved a satisfactory recognition accuracy of 84.1% on the total length of 16,000 s of real road spectrum data, achieving results that can be used as the basis for subsequent studies.
(2)
The MLP, basic ED, and DPR-ED models, trained using the early stopping method, were developed. The effect of different numbers of neurons on the prediction accuracy and stability of each model was investigated and the optimization of the models was completed. The results show that the DPR-ED model with 30 LSTM hidden neurons can achieve the optimal overall performance for velocity prediction, which obtains an average RMSE of 0.000862 and a standard deviation of 0.000032 after the dataset’s normalization.
(3)
Compared with the MLP model, the DPR-ED model is designed to improve the performance by implementing multidimensional inputs and applying time series analysis. In the long-time prediction series case, the DPR-ED model shows a significant advantage over the MLP model: the RMSE of the DPR-ED model applied to the validation set was 1.028 m/s, while the RMSE of the MLP model was 1.096 m/s, with a 6.6% deterioration in performance compared to the former. When the two models were applied to 16,000 s road spectrum data for testing, the proportion with a low error of 0.1~0.3 m/s was improved by 4% and a larger error proportion was filtered for the results predicted by the DPR-ED model. The DPR-ED model performed 5.2% better than the MLP model with respect to the average prediction accuracy. Meanwhile, the variance was decreased by 15.6%. This novel framework enables the processing of long-time sequences with multiple input dimensions, which improves the prediction accuracy under complicated driving patterns and enhances the generalization performance and robustness of the model.

Author Contributions

Conceptualization, D.L. and Y.Z.; methodology, Y.Z. and C.Z.; software, Y.Z.; validation, Y.Z. and L.F.; formal analysis, Y.T.; investigation, D.L. and L.F.; resources, D.L.; data curation, Y.Z.; writing—original draft preparation, D.L. and Y.Z.; writing—review and editing, Y.Z. and L.F.; visualization, D.L. and Y.T.; supervision, D.L.; project administration, D.L.; funding acquisition, D.L. and L.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Key R&D Program of China (2021YFB2500800).

Conflicts of Interest

The authors declare no conflict of interest.

Definitions/Abbreviations

v_meanAverage velocity
v_maxMaximum velocity
T_idleidle ratio
WLTCWorld Light-duty vehicle Test Cycle
NEDCNew European Driving Cycle
CLTC-PChina Light-duty vehicle Test Cycle-Passenger
SSEsum of squared errors
SCSilhouette coefficient
DPRDriving pattern recognition
LSTMLong short-term memory
EDEncoder–decoder
MLPMultilayer perceptron
RMSERoot mean square error

References

  1. Wang, Y.; Lou, D.; Xu, N.; Fang, L.; Tan, P. Energy management and emission control for range extended electric vehicles. Energy 2021, 236, 121370. [Google Scholar] [CrossRef]
  2. Karbaschian, M.A.; Söffker, D. Review and Comparison of Power Management Approaches for Hybrid Vehicles with Focus on Hydraulic Drives. Energies 2014, 7, 3512–3536. [Google Scholar] [CrossRef]
  3. Biswas, A.; Emadi, A. Energy management systems for electrified powertrains: State-of-the-art review and future trends. IEEE Trans. Veh. Technol. 2019, 68, 6453–6467. [Google Scholar] [CrossRef]
  4. Rizzoni, G.; Onori, S. Energy Management of Hybrid Electric Vehicles: 15 years of development at the Ohio State University. Oil Gas Sci. Technol. Rev. D’Ifp Energ. Nouv. 2014, 70, 41–54. [Google Scholar] [CrossRef]
  5. Tran, D.D.; Vafaeipour, M.; El Baghdadi, M.; Barrero, R.; Van Mierlo, J.; Hegazy, O. Thorough state-of-the-art analysis of electric and hybrid vehicle powertrains: Topologies and integrated energy management strategies. Renew. Sustain. Energy Rev. 2020, 119, 109596. [Google Scholar] [CrossRef]
  6. Wu, J.; Zhangc, H.; Cuin, X. Fuzzy energy management strategy for a hybrid electric vehicle based on driving cycle recognition. Int. J. Automot. Technol. 2012, 13, 1159–1167. [Google Scholar] [CrossRef]
  7. Chen, B.C.; Guan, J.C.; Li, J.H. Adaptive Power Management Control of Range Extended Electric Vehicle. Energy Procedia 2014, 61, 67–70. [Google Scholar] [CrossRef]
  8. Skugor, B.; Deur, J. Instantaneous Optimization-based Energy Management Control Strategy for Extended Range Electric Vehicle. In SAE Technical Paper Series; SAE International: Warrendale, PA, USA, 2013. [Google Scholar]
  9. Rezaei, A.; Burl, J.B.; Zhou, B.; Rezaei, M. A New Real-Time Optimal Energy Management Strategy for Parallel Hybrid Electric Vehicles. IEEE Trans. Control. Syst. Technol. 2019, 27, 830–837. [Google Scholar] [CrossRef]
  10. Zeng, Y.; Cai, Y.; Kou, G.; Gao, W.; Qin, D. Energy Management for Plug-In Hybrid Electric Vehicle Based on Adaptive Simplified-ECMS. Sustainability 2018, 10, 2060. [Google Scholar] [CrossRef]
  11. Peng, J.; He, H.; Xiong, R. Rule based energy management strategy for a series–parallel plug-in hybrid electric bus optimized by dynamic programming. Appl. Energy 2017, 185, 1633–1643. [Google Scholar] [CrossRef]
  12. Xi, L.; Zhang, X.; Sun, C.; Wang, Z.; Hou, X.; Zhang, J. Intelligent Energy Management Control for Extended Range Electric Vehicles Based on Dynamic Programming and Neural Network. Energies 2017, 10, 1871. [Google Scholar] [CrossRef]
  13. Zhang, P.; Yan, F.; Du, C. A comprehensive analysis of energy management strategies for hybrid electric vehicles based on bibliometrics. Renew. Sustain. Energy Rev. 2015, 48, 88–104. [Google Scholar] [CrossRef]
  14. Wu, J.; He, H.; Peng, J.; Li, Y.; Li, Z. Continuous reinforcement learning of energy management with deep Q network for a power split hybrid electric bus. Appl. Energy 2018, 222, 799–811. [Google Scholar] [CrossRef]
  15. Li, Y.; He, H.; Khajepour, A.; Wang, H.; Peng, J. Energy management for a power-split hybrid electric bus via deep reinforcement learning with terrain information. Appl. Energy 2019, 255, 113762.1–113762.13. [Google Scholar] [CrossRef]
  16. Li, G.; Görges, D. Energy management strategy for parallel hybrid electric vehicles based on approximate dynamic programming and velocity forecast. J. Frankl. Inst. 2019, 356, 9502–9523. [Google Scholar] [CrossRef]
  17. Li, Y.; He, H.; Peng, J.; Zhang, H. Power management for a plug-in hybrid electric vehicle based on reinforcement learning with continuous state and action spaces. Energy Procedia 2017, 142, 2270–2275. [Google Scholar] [CrossRef]
  18. Tan, H.; Zhang, H.; Peng, J.; Jiang, Z.; Wu, Y. Energy management of hybrid electric bus based on deep reinforcement learning in continuous state and action space. Energy Convers. Manag. 2019, 195, 548–560. [Google Scholar] [CrossRef]
  19. Olia, K.; Shahverdi, M.; Mazzola, M.; Sherif, A. Developing a Model Predictive Control-Based Algorithm for Energy Management System of the Catenary-Based Electric Truck; SAE International: Warrendale, PA, USA, 2016. [Google Scholar]
  20. Zhang, S.; Xiong, R.; Sun, F. Model predictive control for power management in a plug-in hybrid electric vehicle with a hybrid energy storage system. Appl. Energy 2017, 185, 1654–1662. [Google Scholar] [CrossRef]
  21. Rezaei, A.; Burl, J.B. Prediction of Vehicle Velocity for Model Predictive Control. IFAC-Paper 2015, 48, 257–262. [Google Scholar] [CrossRef]
  22. Liu, K.; Asher, Z.; Gong, X.; Huang, M.; Kolmanovsky, I. Vehicle Velocity Prediction and Energy Management Strategy Part 1: Deterministic and Stochastic Vehicle Velocity Prediction Using Machine Learning; SAE International: Warrendale, PA, USA, 2019. [Google Scholar]
  23. Shin, J.; Sunwoo, M. Vehicle Speed Prediction Using a Markov Chain with Speed Constraints. IEEE Trans. Intell. Transp. Syst. 2019, 20, 3201–3211. [Google Scholar] [CrossRef]
  24. Wang, F.; Luo, Y.T. A Research on Power Splitting Strategy for Hybrid Energy Storage System Based on Driving Condition Prediction. Automot. Eng. 2019, 41, 1251–1258, 1264. [Google Scholar]
  25. Sun, C.; Hu, X.; Moura, S.J.; Sun, F. Velocity Predictors for Predictive Energy Management in Hybrid Electric Vehicles. IEEE Trans. Control. Syst. Technol. 2015, 23, 1197–1204. [Google Scholar]
  26. Guo, J.; He, H.; Sun, C. ARIMA-Based Road Gradient and Vehicle Velocity Prediction for Hybrid Electric Vehicle Energy Management. IEEE Trans. Veh. Technol. 2019, 68, 5309–5320. [Google Scholar] [CrossRef]
  27. Jing, J.; Filev, D.; Kurt, A.; Özatay, E.; Michelini, J.; Özgüner, Ü. Vehicle speed prediction using a cooperative method of fuzzy Markov model and auto-regressive model. In Proceedings of the 2017 IEEE Intelligent Vehicles Symposium (IV), Los Angeles, CA, USA, 11–14 June 2017; pp. 881–886. [Google Scholar]
  28. Rosolia, U.; Zhang, X.; Borrelli, F. Data-Driven Predictive Control for Autonomous Systems. Annu. Rev. Control. Robot. Auton. Syst. 2018, 1, 259–286. [Google Scholar] [CrossRef]
  29. Fengqi, Z.; Xiaosong, H.; Kanghui, X.; Xiaolin, T.; Yahui, C. Current Status and Prospects for Model Predictive Energy Management in Hybrid Electric Vehicles. J. Mech. Eng. 2019, 55, 86–108. [Google Scholar]
  30. Choudhury, A.; Shanmugavadivelu, S.; Velpuri, B. Real Time Speed Trend Analysis and Hours of Service Forecasting Using LSTM Network. In Proceedings of the 2019 IEEE 5th International Conference for Convergence in Technology (I2CT), Bombay, India, 29–31 March 2019; pp. 1–6. [Google Scholar]
  31. Shi, F.; Wang, Y.; Chen, J.; Wang, J.; Hao, Y.; He, Y. Short-term Vehicle Speed Prediction by Time Series Neural Network in High Altitude Areas. IOP Conf. Ser. Earth Environ. Sci. 2019, 304, 032072. [Google Scholar] [CrossRef]
  32. Wang, Y.Q. Traffic Flow Prediction Based on Combined Model of ARIMA and RBF Neural Network. In Proceedings of the 2017 2nd International Conference on Machinery, Electronics and Control Simulation (Mecs 2017), Taiyuan, China, 24–25 June 2017; Volume 138, pp. 82–86. [Google Scholar]
  33. Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to sequence learning with neural networks. Adv. Neural Inf. Process. Syst. 2014, 27. [Google Scholar] [CrossRef]
  34. Sun, C.; Moura, S.J.; Hu, X.; Hedrick, J.K.; Sun, F. Dynamic traffic feedback data enabled energy management in plug-in hybrid electric vehicles. IEEE Trans. Control. Syst. Technol. 2014, 23, 1075–1086. [Google Scholar]
  35. Liu, G.C. Study of Energy Management Strategy of Range-Extended Electric Vehicle; Shanghai Jiao Tong University: Shanghai, China, 2017. [Google Scholar]
  36. Hu, J.; Liu, D.; Du, C.; Yan, F.; Lv, C. Intelligent energy management strategy of hybrid energy storage system for electric vehicle based on driving pattern recognition—ScienceDirect. Energy 2020, 198, 17298. [Google Scholar] [CrossRef]
  37. Rousseeuw, P. Silhouettes: A graphical aid to the interpretation and validation of cluster analysis. J. Comput. Appl. Math. 1999, 20, 53–65. [Google Scholar] [CrossRef]
  38. Cao, M.; Xu, Q.; Bian, H.; Yuan, X.; Du, P. Research on configuration strategy for regional energy storage system based on three typical filtering methods. IET Gener. Transm. Distrib. 2016, 10, 2360–2366. [Google Scholar] [CrossRef]
  39. Li, M.; Lu, F.; Zhang, H.; Chen, J. Predicting Future Locations of Moving Objects with Deep Fuzzy-LSTM Networks. Transp. A Transp. Sci. 2020, 16, 119–136. [Google Scholar] [CrossRef]
  40. Xiang, C.; Ding, F.; Wang, W.; He, W. Energy management of a dual-mode power-split hybrid electric vehicle based on velocity prediction and nonlinear model predictive control. Appl. Energy 2017, 189, 640–653. [Google Scholar] [CrossRef]
  41. Wu, Y.; Zhang, Y.; Li, G.; Shen, J.; Chen, Z.; Liu, Y. A predictive energy management strategy for multi-mode plug-in hybrid electric vehicles based on multi neural networks. Energy 2020, 208, 118366. [Google Scholar] [CrossRef]
  42. Shi, C.; Zhang, Z.; Ji, C.; Li, X.; Di, L.; Wu, Z. Potential improvement in combustion and pollutant emissions of a hydrogen-enriched rotary engine by using novel recess configuration. Chemosphere 2022, 299, 134491. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Velocity versus time of typical driving cycles for K-means clustering algorithm training.
Figure 1. Velocity versus time of typical driving cycles for K-means clustering algorithm training.
Sustainability 14 10629 g001
Figure 2. Sample division method based on 60 s time window.
Figure 2. Sample division method based on 60 s time window.
Sustainability 14 10629 g002
Figure 3. SSE and S(i) versus the number of driving pattern clustering.
Figure 3. SSE and S(i) versus the number of driving pattern clustering.
Sustainability 14 10629 g003
Figure 4. Clustering results (a) shown on training data and (b) under the three-dimensional coordinates.
Figure 4. Clustering results (a) shown on training data and (b) under the three-dimensional coordinates.
Sustainability 14 10629 g004aSustainability 14 10629 g004b
Figure 5. Prediction results of driving-pattern-recognition model based on actual road spectrum data after filtering method.
Figure 5. Prediction results of driving-pattern-recognition model based on actual road spectrum data after filtering method.
Sustainability 14 10629 g005
Figure 6. Basic LSTM Network.
Figure 6. Basic LSTM Network.
Sustainability 14 10629 g006
Figure 7. The diagram of the encoder–decoder framework: (a) Traditional encoder–decoder framework; (b) DPR-ED model.
Figure 7. The diagram of the encoder–decoder framework: (a) Traditional encoder–decoder framework; (b) DPR-ED model.
Sustainability 14 10629 g007
Figure 8. The learning outcome for validation data: (a) MLP model; (b) Basic ED model; (c) DPR-ED model.
Figure 8. The learning outcome for validation data: (a) MLP model; (b) Basic ED model; (c) DPR-ED model.
Sustainability 14 10629 g008
Figure 9. The accuracy of each model versus the history window and prediction window on test set and comparison: (a) MLP model; (b) DPR-ED model; (c) The comparison of basic ED model and DPR-ED model (history window = 10 s).
Figure 9. The accuracy of each model versus the history window and prediction window on test set and comparison: (a) MLP model; (b) DPR-ED model; (c) The comparison of basic ED model and DPR-ED model (history window = 10 s).
Sustainability 14 10629 g009
Figure 10. Velocity prediction result of the DPR-ED model: (a) City pattern; (b) Suburb pattern; (c) Highway pattern.
Figure 10. Velocity prediction result of the DPR-ED model: (a) City pattern; (b) Suburb pattern; (c) Highway pattern.
Sustainability 14 10629 g010aSustainability 14 10629 g010b
Figure 11. Local velocity prediction of DPR-ED model for 5 prediction steps: (a) City pattern; (b) Highway pattern.
Figure 11. Local velocity prediction of DPR-ED model for 5 prediction steps: (a) City pattern; (b) Highway pattern.
Sustainability 14 10629 g011
Figure 12. The prediction results of the DPR-ED model on the city part of the filtered test set.
Figure 12. The prediction results of the DPR-ED model on the city part of the filtered test set.
Sustainability 14 10629 g012
Figure 13. The acceleration distribution of the prediction results by DPR-ED model on the raw test set.
Figure 13. The acceleration distribution of the prediction results by DPR-ED model on the raw test set.
Sustainability 14 10629 g013
Figure 14. Local RMSE comparison between MLP and DPR-ED on test set.
Figure 14. Local RMSE comparison between MLP and DPR-ED on test set.
Sustainability 14 10629 g014
Figure 15. The comparison of the error distribution between the MLP model and the DPR-ED model for test data: (a) MLP model; (b) DPR-ED model.
Figure 15. The comparison of the error distribution between the MLP model and the DPR-ED model for test data: (a) MLP model; (b) DPR-ED model.
Sustainability 14 10629 g015
Table 1. The proportion of the prediction results of the test set contained in each error segment.
Table 1. The proportion of the prediction results of the test set contained in each error segment.
RMSE Segment (m/s)Proportion (MLP Model)Proportion (DPR-ED Model)
0~0.112.81%12.15%
0.1~0.319.81%23.55%
0.3~0.518.15%18.13%
0.5~0.713.22%12.34%
0.7~1.114.85%14.5%
1.1~2.115.35%14.37%
>2.15.81%4.96%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lou, D.; Zhao, Y.; Fang, L.; Tang, Y.; Zhuang, C. Encoder–Decoder-Based Velocity Prediction Modelling for Passenger Vehicles Coupled with Driving Pattern Recognition. Sustainability 2022, 14, 10629. https://doi.org/10.3390/su141710629

AMA Style

Lou D, Zhao Y, Fang L, Tang Y, Zhuang C. Encoder–Decoder-Based Velocity Prediction Modelling for Passenger Vehicles Coupled with Driving Pattern Recognition. Sustainability. 2022; 14(17):10629. https://doi.org/10.3390/su141710629

Chicago/Turabian Style

Lou, Diming, Yinghua Zhao, Liang Fang, Yuanzhi Tang, and Caihua Zhuang. 2022. "Encoder–Decoder-Based Velocity Prediction Modelling for Passenger Vehicles Coupled with Driving Pattern Recognition" Sustainability 14, no. 17: 10629. https://doi.org/10.3390/su141710629

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop