Next Article in Journal
Entanglement and Fisher Information for Atoms–Field System in the Presence of Negative Binomial States
Previous Article in Journal
Design and Investigation of a Dynamic Auto-Adjusting Ejector for the MED-TVC Desalination System Driven by Solar Energy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Remaining Useful Life Prediction Using Dual-Channel LSTM with Time Feature and Its Difference

1
School of Computer, Hunan University of Technology, Zhuzhou 412007, China
2
School of Automation, Central South University, Changsha 410083, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(12), 1818; https://doi.org/10.3390/e24121818
Submission received: 9 November 2022 / Revised: 1 December 2022 / Accepted: 5 December 2022 / Published: 13 December 2022
(This article belongs to the Topic Machine and Deep Learning)

Abstract

:
At present, the research on the prediction of the remaining useful life (RUL) of machinery mainly focuses on multi-sensor feature extraction and then uses the features to predict RUL. In complex operations and multiple abnormal environments, the impact of noise may result in increased model complexity and decreased accuracy of RUL predictions. At the same time, how to use the sensor characteristics of time is also a problem. To overcome these issues, this paper proposes a dual-channel long short-term memory (LSTM) neural network model. Compared with the existing methods, the advantage of this method is to adaptively select the time feature and then perform first-order processing on the time feature value and use LSTM to extract the time feature and first-order time feature information. As the RUL curve predicted by the neural network is zigzag, we creatively designed a momentum-smoothing module to smooth the predicted RUL curve and improve the prediction accuracy. Experimental verification on the commercial modular aerospace propulsion system simulation (C-MAPSS) dataset proves the effectiveness and stability of the proposed method.

1. Introduction

Since entering the age of industrialization, machinery and equipment are seen everywhere. However, due to the harsh environment and the impact of improper human operation, mechanical failure difficulties have also emerged one after another. It can be seen that mechanical failure has become a threat hindering social development. Ordinary people may be affected by mechanical failure at any time. Therefore, how to accurately evaluate the RUL of a machine before a failure occurs is of great significance.
The current prediction methods for RUL can be divided into four categories: physical models, statistical models, artificial intelligence models, and hybrid methods [1].
Technology based on physical model describes the degradation process of machinery by establishing a mathematical model through failure mechanisms or damage first principles [2]. Chan and Enright et al. [3] proposed a time-dependent physical crack propagation method for predicting the RUL of the turbo propulsion system. Using the enhanced risk analysis tool and material constants calibrated to IN 718 data, the effect of time-dependent crack growth on the risk of fracture in a turbo engine component demonstrated a generic rotor design and a realistic mission profile. El-Tawil et al. [4] introduced an analytic prognostic methodology based on nonlinear damage laws. It enables the assurance of high availability and productivity with less cost for industrial systems. Kacprzynski et al. [4] developed a gear health prediction model using the physical method of failure. The approach, based on a statistical model, estimates the RUL of the machinery by establishing a statistical model based on empirical knowledge [5]. Barraza-Barraza and Tercero-Gómez et al. [6] put forward a way to use exogenous variables to construct three autoregressive (AR) models to predict the RUL of aluminum plates. Three autoregressive models with exogenous variables (ARX) were constructed, and their capability to estimate the remaining useful life (RUL) of a process was evaluated following the case of the aluminum crack growth problem. An existing stochastic model of aluminum crack growth was implemented and used to assess RUL estimation performance of the proposed ARX models through extensive Monte Carlo simulations.
Physical model and statistical model methods generally require expert knowledge, which is more difficult for non-professionals to understand and implement, so they are limited in practical application. With the advancement of graphics processor technology and the improvements in computing power, artificial intelligence methods have developed. Artificial intelligence builds models by collecting data, which can reduce the difficulty of understanding professional knowledge. It is easy to apply to various industries [7]. Degradation monitoring has been carried out on life time accelerating data. Zhou et al. [8] used an LSTM network for defect prediction. Zhou et al. [9] proposed a new improved multi-scale edge-labeling graph neural network (MEGNN) to estimate the tool wear condition through the updated edge labels using a weighted voting method. Wen et al. [10] adopted a new residual convolutional neural network (ResCNN) to solve the problem of gradient disappearance and predict the RUL of turbofan engines. ResCNN applies the residual block that skips several blocks of convolutional layers by using shortcut connections, which can help to overcome vanishing/exploding gradient problems. Li et al. [11] proposed a domain-adaptive RUL prediction model by integrating adaptive batch normalization (AdaBN) into a deep convolutional neural network (DCNN). The improved AdaBN-DCNN model can not only improve the accuracy of the prediction but also adapt to the prognostic tasks under different DDs. The sliding time window (TW) and the improved piecewise linear RUL function are also used in this paper to improve the prediction capability of the model. Han Li et al. [12] and others designed a multi-scale convolutional neural network (MSCNN) to directly establish the relationship between monitoring data and RUL. The MS-DCNN has three multi-scale blocks (MS-BLOCKs) [13], where three different sizes of convolution operations are put on each block in parallel. This structure improves the network’s ability to learn complex features by extracting features of different scales. Wu et al. [14] applied an RUL prediction method based on a deep long short-term memory (DLSTM) network using multiple sensor time-series signals. Zhao et al. [15] designed a mode based long multi-head attention and LSTM. The model can select the key features in the time-series data, then input them into the LSTM layer to mine the internal connections, and finally obtain RUL predicted results through two fully connected layers. Aiming at the characteristics of high dimensionality, high lag, and complexity of engine data, a multi-scale attention-based bidirectional long short-term memory neural network model based on self-training weights was proposed by Qiu et al. [16]. Multi-scale features were extracted through bidirectional long short-term memory neural networks (BiLSTM) of different scales. To improve the accuracy of prediction, a fusion algorithm based on self-training weights was proposed, and an attention mechanism was introduced to screen features at different scales.
The hybrid method combines the two or more methods mentioned above to neutralize their characteristics. Zemouri et al. [17] used regression models and artificial neural networks to propose a hybrid method. Wang et al. [18] combined the similarity method and sparse learning to predict health indicators. Zhang et al. [19] constructed an aging model of the battery by developing a fusion technique consisting of a relevance vector machine and particle filter (PF) for RUL estimation. Yan et al. [20] presented a neoteric approach of RUL estimation of bearings, which can evaluate the degradation stage of bearings and exploit the optimal RUL prediction by a hybrid degradation tracing model. Wang et al. [21] proposed a hybrid prognostics method for the RUL prediction of rolling element bearings through using relevance vector machine regressions and exponential degradation models. Gou et al. [22] introduced a new hybrid ensemble data-driven method to accurately estimate the state of health and RUL of Li-ion batteries. Li et al. [23] presented a novel hybrid Elman-LSTM method for battery RUL estimation.
The neural network method has achieved good results, but current research in this area still has following problems:
(1)
The influence of time characteristics on the life prediction of machinery may change in different operating environments. Given this phenomenon, how we choose useful time features to avoid the appearance of feature redundancy or invalid features remains a problem to be solved.
(2)
For time characteristics, researchers frequently make the model by focusing on the size of the time features at a certain moment and ignoring the difference between the time features of two different moments. In fact, the change speed of the characteristics reflects the internal state of the machine and also reflects the health state of the machine.
(3)
Generally, the RUL of the machine is smooth and stable; that is, the remaining life of the machine in a certain period should be similar and the number of fluctuations relatively rare. However, in harsh environments, the data sent back to the system by the sensor are commonly unclean. The neural network method learns and predicts based on data, so the RUL predicted by the neural network fluctuates up and down. The RUL curve is jagged and causes a large deviation from the ground RUL.
To solve those problems, this paper proposes a dual-channel LSTM method. The proposed method constructs the direct relationship between the raw data and the ground RUL without using any prior information and improves the learning ability of neural networks by using time characteristic values and characteristic differences to predict the RUL. First of all, some features can hardly change during the entire time period, and the amount of information carried is also very small. If all the features are directly input into the neural network, the training time is longer. Therefore, we adopt a method of adaptively selecting features to eliminate features that have not changed in the life cycle and solve the problem of data redundancy. Secondly, we extract the first-order difference results of features as the input of the model to reflect the change speed of features and improve the accuracy of RUL prediction. Thirdly, in this neural network, LSTM is used to calculate the spatial dimension relationship of different features at each time, and a convolutional neural network (CNN) is used to consider the relationship in the time dimension and fuse the features of multiple time moments into one dimension. Aiming at the jagged phenomenon of the RUL curve predicted by the neural network, a momentum-smoothing method is proposed to deal with RUL curves and improve the accuracy of prediction.
The rest of the paper is organized as follows. The Section 2 introduces the structure of the dual-channel LSTM. Section 3 discusses the degradation data of aircraft turbofan engines and shows the comparative experiments of the dual-channel LSTM network with the other latest methods and so on. The Section 4 gives conclusions and further work.

2. Methodology

In this part, a new prediction structure for RUL estimation is proposed, called dual-channel LSTM. The architecture of the dual-channel LSTM network is presented in Figure 1. The whole algorithm can be divided into four parts: data preprocessing, dual-channel LSTM, RUL prediction, and momentum smoothing.

2.1. Data Preprocessing

The sample data we employed are C-MAPSS datasets from NSNA. The C-MAPSS dataset is composed of four diverse sub-datasets FD001, FD002, FD003, and FD004, which have the following related advantages and disadvantages. First, the sample size is sufficient. Each subsample contains a training set and test set, and the number of samples is acceptable. Second, there is a wide variety of samples. The whole training sample contains four sub-datasets. The data distribution of each sub-dataset is distinct, which is suitable for testing the generalization of the model. However, the sample is the data generated by simulating the real environment, and there is a certain deviation from the real sample of a turbine engine. At the same time, the sub-dataset contains a lot of noise, which is not conducive to the feature generalization of the model. Therefore, we must firstly preprocess the data, select features useful for model prediction, standardize the data, and speed up the convergence of the model. The following are segmented into the three aspects of feature selection, standardized processing, and time-window processing.

2.1.1. Feature Selection

The following uses engine unit No. 1 in the FD001 sub-dataset as an example to visualize its operational settings and some sensor measurements.
In Figure 2, the abscissa is time; the ordinate includes the operating settings of the engine and some sensors. In Figure 2 above, we can see that some time characteristics remain unchanged throughout the entire period, which has no positive effect on the prediction of RUL. It is necessary to choose the features that have a positive influence on the model prediction. Therefore, we can use prognosability to measure the variability of state indicators, and select time features that have changed significantly. The prognosability formula is as follows:
prognosability j = e ( s t d ( x j ) m e a n x j ( 1 ) x j ) , j = 1 , , M
where xj represents the measurement vector of a certain feature on the j-th system, the variable M is the number of systems to be monitored, and xj is the measurement vector at the first moment of the j-th system. The values of p r o g n o s a b i l i t y range from 0 to 1. For all features, if their prognosability is equal to 0 or NaN, these features are removed. Because these features have not changed in the whole prediction cycle, the results of each sub dataset selection are as follows:
FD001: op_setting_1, op_setting_2, sensor_2, sensor_3, sensor_4, sensor_6, sensor_7, sensor_8, sensor_9, sensor_11, sensor_12, sensor_13, sensor_14, sensor_15, sensor_17, sensor_20, sensor_21.
FD002: op_setting_1, op_setting_2, op_setting_3, sensor_1, sensor_2, sensor_3, sensor_4, sensor_5, sensor_6, sensor_7, sensor_8, sensor_9, sensor_10, sensor_11, sensor_12, sensor_13, sensor_14, sensor_15, sensor_16, sensor_17, sensor_18, sensor_19, sensor_20, sensor_21.
FD003: op_setting_1, op_setting_2, sensor_2, sensor_3, sensor_4, sensor_6, sensor_7, sensor_8, sensor_9, sensor_10, sensor_11, sensor_12, sensor_13, sensor_14, sensor_15, sensor_17, sensor_20, sensor_21.
FD004: op_setting_1, op_setting_2, op_setting_3, sensor_1, sensor_2, sensor_3, sensor_4, sensor_5, sensor_6, sensor_7, sensor_8, sensor_9, sensor_10, sensor_11, sensor_12, sensor_13, sensor_14, sensor_15, sensor_16, sensor_17, sensor_18, sensor_19, sensor_20, sensor_21.

2.1.2. Normalization

The standardization of data is to scale the data to a small specific interval, which can speed up the model convergence. The data standardization method used in this article is the z-score. The formula is as follows:
z = x u σ
where u represents the mean value of all selected features; σ represents the standard deviation of all selected features; and x represents the value of a chosen feature.
Considering that the mechanical state can be divided into a health state and a degraded state, the real RUL should be a piecewise linear function [24]. As shown in Figure 3, the ground RUL can be segmented into two parts: a constant portion and a linear degraded portion. Therefore, an RUL threshold needs to be set.

2.1.3. Time-Window Processing

The change speed of the characteristics reflects the internal state of the machine and the health state of the machine. Therefore, we can use the first-order difference results of features as the change speed of features to improve the accuracy of RUL prediction. Of course, adding higher-order difference results can also be considered to predict RUL, but the parameters of the model become more numerous. Then, there remain only features and feature differences to be considered. Two time windows need to be divided for network input. The specific operations are as follows:
First, divide the first time window. As shown in Figure 4, the window width is denoted as Nf; the length of window is Nt; the sliding stride is denoted as s (s = 1); and the feature value window Input1 = [x1, x2, …, xNf].
x i = x i w i n d o w _ s t a r t x i w i n d o w _ s t a r t + 1 x i w i n d o w _ e n d   i = 1 , , N f
The input size of Input1 is Nt  ×  Nf. Nf represents the number of features chosen by the prognosability formula. After the first time window is obtained, the time feature difference value can be obtained. The feature difference value window Input2 = [d1, d2, …, dNf].
d i = x i w i n d o w _ s t a r t + 1 x i w i n d o w _ s t a r t x i w i n d o w _ s t a r t + 2 x i w i n d o w _ s t a r t + 1 x i w i n d o w _ e n d x i w i n d o w _ e n d 1   i = 1 , N f
The input size of Input2 is (Nt − 1) ×  Nf, and each pair of inputs has its corresponding RUL value, which is used for supervised training.

2.2. Dual-Channel LSTM Predicts Lifetime

When we use recurrent neural networks, in most cases, we deal with features at different times, ignoring another dimension, the feature difference. The feature difference at unequal times can let the network know the change rate of the feature value, assist the method to consider the feature information more comprehensively, reduce the influence of noise on the model, make the model more robust, and enhance the differentiation ability. Therefore, we can use dual-channel LSTM to process the eigenvalues at different times and the feature difference separately and calculate the spatial dimension relationship of different features at each time. This is described in detail below.
As shown in Figure 1 above, after the time window is processed, we can get two inputs, Input1 and Input2. Input1 represents the value of each time feature within a window time; Input2 represents the difference value of the time feature within the same window time. Then we use two LSTM networks to process Input1 and Input2. The output after the first recurrent neural network processes Input1 is Output1; Output1 = [h1, h2, …, hhidden_size].
h i = h i w i n d o w _ s t a r t h i w i n d o w _ s t a r t + 1 h i w i n d o w _ e n d i = 1 , , h i d d e n _ s i z e
The output after the second recurrent neural network processes Input2 is Output2; Output2 = [g1, g2, …, ghidden_size].
g i = g i w i n d o w _ s t a r t + 1 g i w i n d o w _ e n d   i = 1 , , h i d d e n _ s i z e
The first line feature of Output1 can be obtained by the second line feature of Output1 and the first line feature information of Output2, so the first line feature of Output1 is not considered. Finally, we add the last (Nt−1) row vectors of Output1 and Output2 directly to get Output; Output = [o1, o2, …, ohidden_size].
o i = h i w i n d o w _ s t a r t + 1 + g i w i n d o w _ s t a r t + 1 h i w i n d o w _ e n d + g i w i n d o w _ s t a r t + 1 i = 1 , , h i d d e n _ s i z e

2.3. Life Prediction

RUL prediction can be divided into two parts. One part is to use CNN to consider the relationship in the time dimension and fuse the features of multiple time moments into one dimension after dual-channel LSTM processing, and another part is to use a fully connected neural network to predict RUL. The convolution part mainly includes convolution operation, batch normalization [25,26], and activation function. The main features of the convolutional layer are sparse interaction, parameter sharing, and equivariant representation. Among them, sparse interaction can play a regularization [27] role; parameter sharing reduces the number of parameters of the model and significantly increases the size of the network without increasing the training data.
As shown in Figure 1, the filter sizes of the three convolutional layers are selected as 32 × 1, 32 × 1, and 1 × 1. Each convolutional layer uses batch normalization to improve the performance and stability of the neural network. The activation function uses the ReLU function [28]. To avoid the gradient disappearance, the function of the first two convolutional layers is mainly to extract features, and the last one is to reduce dimensionality, which suppresses overfitting. In the portion of the fully connected layer, the neural unit mainly uses the previously extracted features to predict RUL. The activation function used in the first fully connected layer is also the ReLU function, and the dropout technology [29,30] is applied to avoid overfitting. The last fully connected layer is used to predict the RUL value.

2.4. Smooth Calibration

Taking into account the influence of many factors, such as noise, the collected data has a fixed deviation. The neural network predicts life based on the collected data, so the predicted remaining life curve fluctuates up and down throughout the cycle. In practice, the RUL of the machine should be stable. Even if it encounters the interference of some factors, the life curve fluctuates in very few local places. Because of this phenomenon, we should consider that there should be a caching relationship between the RUL at the past moment and the RUL at the current moment. Inspired by the momentum gradient descent, the article proposes a momentum smoothing method for the RUL for the test set. The formula is as follows.
p r e d i c t t = k × y t + ( 1 k ) × p r e d i c t t 1 , 0 k 1
where yt is the predicted value at time t by using dual-channel LSTM, predictt−1 is the predicted outcome after smoothing at the last time, predictt is the value after smoothing at the current time t, and k represents the proportion of yt in predictt. The larger the k is, the smaller the buffer of the previous RUL will be at the current moment; the opposite is true the smaller the value of k is.

2.5. Regularization

If the distribution of the acquired data deviates from the distribution of the actual data, the neural network model may be overfitting during training. This leads to the high accuracy of the training dataset but poor performance of the test dataset. In order to avoid this phenomenon, we can consider using regularization [31,32] methods.
The regularization methods include the dropout technology, the L2 regularization, and the early stopping [33] method of the partition verification set in this model. The dropout technology is to make the activation value of the neuron stop working with a certain probability during the forward propagation, which can make the model more generalized since it does not rely too much on some local features.
L2 regularization is also called weight attenuation, which makes the solution of the model biased towards weights with smaller norms, and limits model space by limiting the size of the weight norms, thereby avoiding overfitting to a certain extent. The formula is as follows.
C = C 0 + λ 2 n w w 2
Among them, C0 represents the loss function of the model. In the proposed method, C0 is the MSE; w represents the weight parameter of the model; n represents the number of parameters; λ is the regular term coefficient; C represents the loss function after adding L2 regularization.
In the process of training the model, the loss function keeps getting smaller and the parameters keep approaching the optimal solution, but it is possible that at a certain gradient contour, the model has reached the optimal solution in a fixed spatial range. If we continue to train at this time, the model may linger near the optimal solution or even overfit. Considering this phenomenon, we can divide the training set into a training set and a validation set, and the division ratio is p1: p2. When the model is in p training cycles, and the loss value of the validation set has not decreased, the training is halted.

2.6. Dual-Channel LSTM Model Construction Process

The flow chart of the dual-channel LSTM model construction is described step by step below.
(1)
Firstly, solve the prognosability of features and select useful features. Then the data are standardized by z-score, and the data are scaled to the same metric to speed up the training of the model. After that, set a reasonable RUL threshold to distinguish the health status of the equipment. Finally, divide the data into feature value Input1 = [x1, x2, …, xNf] and feature difference window Input2 = [d1, d2, …, dNf] to establish the input and output pair of the dual-channel LSTM model for model training and testing.
(2)
Split the training set of each sub-dataset into a training set and a validation set. The training set is used for parameter training of the model, and the validation set is used to check whether the model converges. If it converges, stop training the model.
(3)
Construct a dual-channel LSTM network model. LSTM executes parallel operations with difficulty, so training LSTM takes more time. In order to speed up the training of the model as much as possible, the number of LSTM layers is set to 1. At the same time, in order to assure that the LSTM effectively extracts the information of the sequence data, through the prediction accuracy of the model on the dataset, the dimension hidden_size of the hidden layer of the LSTM is adjusted.
(4)
On the training set, apply the mini-batch gradient descent method and Adam algorithm to train the dual-channel LSTM network model. In order to avoid model overfitting, L2 regularization constraints are applied to the model parameters and dropout technology is added to the fully connected layer.
(5)
Use the trained model to predict the RUL of the test set. This paper proposes a momentum-smoothing modular smooth neural network prediction result.
(6)
Use the scoring function (score) and the root mean square error (RMSE) to evaluate the performance of the model. The evaluation index formula is as follows.
s =   i = 1 N s i ,   s i   =   e d i 13 1 ,   for   d i   <   0 e d i 10 1 ,   for   d i 0   d i = p r e d i c t i R U L i R M S E = 1 N i = 1 N d i 2
Among them, predicti represents the predicted value, RULi represents the ground RUL, and N represents the number of all sample data.
Figure 5 visualizes the relationship between the evaluation index function. It can be seen that the score evaluation index and di are exponential. Compared with RMSE, the penalty for high deviation is raised and the penalty for low deviation is reduced in the score function.
LSTM perform parallel operations with difficulty, so training LSTM takes more time. To speed up the training of the model as much as possible, the number of LSTM layers is set to 1. At the same time, in order to ensure that the LSTM effectively extracts the information of the sequence data through the prediction accuracy of the model on the dataset, the dimension hidden size of the hidden layer of the LSTM is adjusted. Meanwhile, batch size and learning rate are adjusted according to the prediction accuracy of the model. Figure 6 depicts the flowchart of dual-channel LSTM.

3. Experiment and Result Analysis

In order to verify the effectiveness of dual-LSTM method, it is applied to predict the RUL of a turbine engine.

3.1. Description of C-MAPSS

C-MAPSS simulates a 90,000-pound thrust-type engine model [34]. The built-in control system includes a fan speed controller and a set of regulators and limiters. Figure 7 shows the main components of the engine model. Table 1 lists the 21 sensors that monitor engine conditions.
From Table 2, we can see that the C-MAPSS dataset is composed of four diverse sub-datasets: FD001, FD002, FD003, and FD004. The number of engine units in each sub-dataset is unique; the numbers of failure modes and operating settings are also distinct. Each sub-dataset is divided into a training dataset and a test dataset, recording the true RUL of the scroll engine from a healthy state to a degraded state at each moment. Each sub-dataset is divided into training datasets and test datasets, which at each moment record three operational settings and the data of the 21 sensors of the scroll engine unit. The FD001 and FD002 sub-datasets contain one failure mode (HPC degradation), while FD003 and FD004 contain two degradation modes (HPC degradation and Fan degradation). There is only one running condition for FD001 and FD003, and there are six running conditions for FD002 and FD004. Due to the complex and changeable operating environment of the FD002 and FD004 sub-dataset engine unit, it is more difficult to estimate the RUL of the FD002 and FD004 sub-datasets.

3.2. Life Prediction on C-MAPSS Datasets

Table 3 shows the detailed parameters of the proposed dual-channel LSTM. The length of the test set is determined by the size of the time window and the window moving step [12]. The parameters of the model are initialized by setting a random seed. Then, through training the model and adjusting the parameters, the parameters in Table 3 are finally determined. There are 16,324 parameters in the model, with a total size of 28.92 M.
The experiment is processed on a PC with Intel Core i5-5200U CPU, 12 GB RAM, and NVIDIA GeForce 910M GPU (on the PyTorch 1.2.0).
After training the model, we take the FD001 sub-dataset as an example to test the effect of the method. The following shows the RUL curve of the engine units 24, 34, 76, and 100 in the FD001 test set.
It can be seen from Figure 8 that the results predicted by the model are constant in the early stage. In the later stage, the predicted outcome shows a linear decline, which is in line with the overall trend of the ground RUL curve.

3.3. Compared with State-of-the-Art Methods

In this research part, we use different advanced research means to compare the dual-channel LSTM method proposed in the article.
As shown in Table 4, our proposed dual-channel LSTM method outperforms most methods in terms of RMSE and score. Compared with other methods, the dual-channel LSTM method adds a time feature difference input channel, which can extract key information from the time feature speed to reduce the error in RUL prediction. In the sub-datasets FD002 and FD004 with multiple faults and complex operating settings, the dual-channel LSTM method reduces the RMSE values to 17.63 and 17.41, respectively, and the scores to 1773.47 and 2617.45, respectively. This reflects that when the environment becomes bad, the proposed dual-channel LSTM method still performs well. The main reason is that the dual-channel recurrent neural network not only considers the time value of the time feature but also thinks about the difference of the time feature at different times, which reduces the impact of the environment on the model. Generally speaking, the dual-channel LSTM method has certain advantages compared to the other methods in the table.

3.4. Model Analysis

In this section, we use the test set of the sub-dataset FD001 as an example to analyze the model.

3.4.1. Mini-Batch Size

In the training process, the training set is divided into small batch samples for training. Table 5 shows the effect of batch processing on the prediction results of the model.
In Table 5, when mini-batch equals 64, the RMSE and score of FD001’s test set are the lowest. Therefore, the value of mini-batch in this method is 64. The results are depicted in Figure 9.

3.4.2. Momentum Smoothing

This experiment studies the influence of momentum smoothing on the predicted results under the same network structure. Table 6 describes the results of the experiment.
When k = 0.3, the RMSE of the test set of FD001 dropped from 11.79 to 11.34, and the score dropped from 280.2 to 267.1. It shows that the use of momentum smoothing can predict the RUL more accurately. At the same time, as k decreases, RMSE and score first decrease and then increase. This shows that when k is greater than or equal to 0.5, reducing the value of k (that is, increasing the buffering effect of the RUL of the previous period on the current) can lower the RMSE and score. However, when the value of k is small, continuing to reduce the value of k makes the proportion of the current RUL predicted by the model too low, resulting in the RUL being determined by the previous RUL after smoothing. So, the score rose sharply. The results are visualized in Figure 10.
In order to see the effect of momentum smoothing, take the engine unit 100 of the test set in the sub-dataset FD001 as an example to visualize the results after using momentum smoothing in Figure 11.
Due to the addition of the momentum-smoothing module in the proposed method, the historically predicted RUL affects the current predicted RUL, which reduces the randomness of the predicted results, so the residual curve is relatively smooth.

3.4.3. Stop Early

In order to verify that early stopping is beneficial for preventing the model from overfitting, we are now conducting research to set p to take a different value within 180 epochs; if the loss of the verification set does not decrease in p cycles, stop training and use the test set to check the effect. Letting p take None is equivalent to letting the model train for 180 epochs; that is, the early stopping method is not used. The results are visualized in Figure 12.
It can be summarized from Table 7 that as the p-value increases, the evaluation indicators RMSE and score are both decreasing. This is due to the use of the mini-batch gradient descent rule, which has randomness; the verification set may not drop continuously within a fixed training period. Therefore, the p-value needs to be enlarged to reduce randomness. However, it can be found that when the p-value increases to a definite value, the RMSE and score does not decrease significantly as the p-value continues to become larger because the randomness has been reduced to a negligible level. If model performance is similar, the training time should be as short as possible, and p should be set to a reasonable value, i.e., not the larger the better. The effect of the model trained in 180 epochs is not much different from the one of using the early stopping strategy, indicating that the model has reached a fixed spatial optimal solution before 180 epochs, but with the early stopping strategy, the model normally stops training before 80 epochs, which saves more than half of the training time.

3.4.4. Optimizer

It can be seen from Table 8 that the use of the SGD optimizer converges to a locally optimal solution. RMSE only dropped to 37.5, while the RMSE can be dropped to less than 15 using several other optimizers. Using the RMSprop optimizer, the RMSE can be reduced to 11.32, but the score is higher than that using the Adam optimizer, so the Adam optimizer is finally used for the gradient descent of the method. The results are visualized in Figure 13.

3.4.5. Model Generalization

In order to verify the generalization of the model, take the logarithm of RUL, let RUL = ln RUL + 1 , and train the model again. The results are as follows.
In Table 9, after taking logarithms of RUL, the RMSE value of the test set of the sub dataset FD001 is 0.15 and the score value is 1.3. It can be seen from Figure 14 that the model still fits the new RUL.

4. Discussion and Summary

This paper proposes a new network model dual-channel LSTM. In this neural network, dual-channel LSTM is used to deal with the time feature and its first-order feature and solve the problems of gradient disappearance and gradient explosion during long sequence training; CNN considers the relationship in the time dimension and integrates the features of multiple time moments into one dimension, which plays a role in dimension reduction. Aiming at the jagged phenomenon of the RUL curve predicted by the neural network, a momentum-smoothing method is proposed to deal with RUL curves and improve the accuracy of prediction. The dual-channel LSTM improves the learning ability of network because it learns the information of the two dimensions of the time features, which has good application scenarios in actual industrial environments. However, there are still some questions that need to be resolved. Firstly, in momentum smoothing, the ratio k is adjusted according to the effect. Whether there is a precise method to determine the value of k or not is not known. Secondly, the application of deep learning to industry requires a large amount of data. Then, the acquisition of industrial data is complicated. Therefore, how to apply deep learning with a small amount of data is also a current problem. Thirdly, there are many types of industrial data. Using the same model to process different types of data, the effect may not be ideal. The difficult point is whether it is possible to propose a data fusion method to improve the universality of the model, but this problem will be explored in future research.

Author Contributions

Conceptualization, C.P. and Z.T.; methodology, Q.W. and C.P.; software, J.W.; validation, J.W. and Q.W.; formal analysis, C.P. and J.W.; investigation, C.P.; resources, W.G.; writing—original draft preparation, C.P. and J.W.; writing—review and editing, C.P.; visualization, J.W.; supervision, C.P.; project administration, C.P.; funding acquisition, Z.T. and W.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Natural Science Foundation of China (No. 61871432, No. 61771492), Key project of Hunan Provincial Education Department (22A0390) and the Natural Science Foundation of Hunan Province (No. 2020JJ4275, No. 2019JJ6008, and No. 2019JJ60054).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lei, Y.; Li, N.; Guo, L.; Li, N.; Yan, T.; Lin, J. Machinery health prognostics: A systematic review from data acquisition to RUL prediction. Mech. Syst. Signal Process. 2017, 104, 799–834. [Google Scholar] [CrossRef]
  2. Cubillo, A.; Perinpanayagam, S.; Esperon-Miguez, M. A review of physics-based models in prognostics: Application to gears and bearings of rotating machinery. Adv. Mech. Eng. 2016, 8, 1687814016664660. [Google Scholar] [CrossRef] [Green Version]
  3. Chan, K.S.; Enright, M.P.; Moody, J.P.; Hocking, B.; Fitch, S.H.K. Life Prediction for Turbopropulsion Systems Under Dwell Fatigue Conditions. J. Eng. Gas Turbines Power 2012, 134, 122501. [Google Scholar] [CrossRef]
  4. El-Tawil, K.; Jaoude, A.A. Stochastic and nonlinear-based prognostic model. Syst. Sci. Control Eng. 2013, 1, 66–81. [Google Scholar] [CrossRef] [Green Version]
  5. Kacprzynski, G.J.; Roemer, M.J.; Modgil, G.; Palladino, A.; Maynard, K. Enhancement of physics-of-failure prognostic models with system level features. In Proceedings of the IEEE Aerospace Conference, Big Sky, MT, USA, 9–16 March 2002. [Google Scholar]
  6. Si, X.-S.; Wang, W.; Hu, C.-H.; Zhou, D.-H. Remaining useful life estimation—A review on the statistical data driven approaches. Eur. J. Oper. Res. 2011, 213, 1–14. [Google Scholar] [CrossRef]
  7. Strušnik, D. Integration of machine learning to increase steam turbine condenser vacuum and efficiency through gasket re-sealing and higher heat extraction into the atmosphere. Int. J. Energy Res. 2022, 46, 3189–3212. [Google Scholar] [CrossRef]
  8. Zhou, Y.; Kumar, A.; Parkash, C.; Vashishtha, G.; Tang, H.; Xiang, J. A novel entropy-based sparsity measure for prognosis of bearing defects and development of a spar-sogram to select sensitive filtering band of an axial piston pump. Measurement 2022, 203, 111997. [Google Scholar] [CrossRef]
  9. Zhou, Y.; Zhi, G.; Chen, W.; Qian, Q.; He, D.; Sun, B.; Sun, W. A new tool wear condition monitoring method based on deep learning under small samples. Measurement 2022, 189, 110622. [Google Scholar] [CrossRef]
  10. Barraza-Barraza, D.; Tercero-Gómez, V.G.; Beruvides, M.G.; Limón-Robles, J. An adaptive ARX model to estimate the RUL of aluminum plates based on its crack growth. Mech. Syst. Signal Process. 2017, 82, 519–536. [Google Scholar] [CrossRef]
  11. Wen, L.; Dong, Y.; Gao, L. A new ensemble residual convolutional neural network for remaining useful life estimation. Math. Biosci. Eng. 2019, 16, 862–880. [Google Scholar] [CrossRef]
  12. Li, J.; Li, X.; He, D. Domain Adaptation Remaining Useful Life Prediction Method Based on AdaBN-DCNN. In Proceedings of the 2019 Prognostics and System Health Management Conference (PHM-Qingdao), Qingdao, China, 25–27 October 2019; pp. 1–6. [Google Scholar] [CrossRef]
  13. Li, H.; Zhao, W.; Zhang, Y.; Zio, E. Remaining useful life prediction using multi-scale deep convolutional neural network. Appl. Soft Comput. 2020, 89, 106113. [Google Scholar] [CrossRef]
  14. Wu, J.; Hu, K.; Cheng, Y.; Zhu, H.; Shao, X.; Wang, Y. Data-driven remaining useful life prediction via multiple sensor signals and deep long short-term memory neural network. ISA Trans. 2019, 97, 241–250. [Google Scholar] [CrossRef]
  15. Zhao, S.; Pang, Y.; Chen, J.; Liu, J. Predication of remaining useful life of aircraft engines based on Multi-head Attention and LSTM. In Proceedings of the 2022 IEEE 6th Information Technology and Mechatronics Engineering Conference (ITOEC), Chongqing, China, 4–6 March 2022; Volume 6, pp. 1530–1534. [Google Scholar]
  16. Qiu, J.; Zheng, H.; Cheng, Y. Research on Prediction of Model Based on Multi-scale LSTM. J. Syst. Simul. 2022, 34, 1593–1604. [Google Scholar]
  17. Zemouri, R.; Gouriveau, R. Towards Accurate and Reproducible Predictions for Prognostic: An Approach Combining a RRBF Network and an AutoRegressive Model. IFAC Proc. Vol. 2010, 43, 140–145. [Google Scholar] [CrossRef] [Green Version]
  18. Wang, P.; Youn, B.D.; Hu, C. A generic probabilistic framework for structural health prognostics and uncertainty management. Mech. Syst. Signal Process. 2012, 28, 622–637. [Google Scholar] [CrossRef]
  19. Zhang, Y.; Xiong, R.; He, H.; Pecht, M. Validation and verification of a hybrid method for remaining useful life prediction of lithium-ion batteries. J. Clean. Prod. 2018, 212, 240–249. [Google Scholar] [CrossRef]
  20. Yan, M.; Wang, X.; Wang, B.; Chang, M.; Muhammad, I. Bearing remaining useful life prediction using support vector machine and hybrid degradation tracking model. ISA Trans. 2019, 98, 471–482. [Google Scholar] [CrossRef]
  21. Wang, B.; Lei, Y.; Li, N.; Li, N. A Hybrid Prognostics Approach for Estimating Remaining Useful Life of Rolling Element Bearings. IEEE Trans. Reliab. 2018, 69, 401–412. [Google Scholar] [CrossRef]
  22. Gou, B.; Xu, Y.; Feng, X. State-of-Health Estimation and Remaining-Useful-Life Prediction for Lithium-Ion Battery Using a Hybrid Data-Driven Method. IEEE Trans. Veh. Technol. 2020, 69, 10854–10867. [Google Scholar] [CrossRef]
  23. Li, X.; Zhang, L.; Wang, Z.; Dong, P. Remaining useful life prediction for lithium-ion batteries based on a hybrid model combining the long short-term memory and Elman neural networks. J. Energy Storage 2019, 21, 510–518. [Google Scholar] [CrossRef]
  24. Heimes, F.O. Recurrent neural networks for remaining useful life estimation. In Proceedings of the 2008 International Conference on Prognostics and Health Management, Denver, CO, USA, 6–9 October 2008. [Google Scholar]
  25. Santurkar, S.; Tsipras, D.; Ilyas, A.; Madry, A. How does batch normalization help optimization? In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018.
  26. Bjorck, N.; Gomes, C.P.; Selman, B.; Weinberger, K.Q. Understanding batch normalization. In Proceedings of the 32nd Conference on Neural Information Processing Systems (NeurIPS 2018), Montréal, QC, Canada, 3–8 December 2022. [Google Scholar]
  27. Moradi, R.; Berangi, R.; Minaei, B. A survey of regularization strategies for deep models. Artif. Intell. Rev. 2019, 53, 3947–3986. [Google Scholar] [CrossRef]
  28. Glorot, X.; Bordes, A.; Bengio, Y. Deep sparse rectifier neural networks. In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, JMLR Workshop and Conference Proceedings, Lauderdale, FL, USA, 11–13 April 2011. [Google Scholar]
  29. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  30. Labach, A.; Salehinejad, H.; Valaee, S. Survey of dropout methods for deep neural networks. arXiv 2019, arXiv:1904.13310. [Google Scholar]
  31. Ghiasi, G.; Lin, T.-Y.; Le Quoc, V. Dropblock: A regularization method for convolutional networks. In Proceedings of the 32nd Conference on Neural Information Processing Systems (NeurIPS 2018), Montréal, Canada, 3–8 December 2022. [Google Scholar]
  32. Hernández-García, A.; König, P. Do deep nets really need weight decay and dropout? arXiv 2018, arXiv:1802.07042. [Google Scholar]
  33. Prechelt, L. Early stopping-but when? In Neural Networks: Tricks of the Trade; Springer: Berlin/Heidelberg, Germany, 1998; pp. 55–69. [Google Scholar]
  34. Saxena, A.; Goebel, K.; Simon, D.; Eklund, N. Damage propagation modeling for aircraft engine run-to-failure simulation. In Proceedings of the 2008 International Conference on Prognostics and Health Management, Denver, CO, USA, 6–9 October 2008. [Google Scholar]
  35. Sayah, M.; Guebli, D.; Zerhouni, N.; Masry, Z.A. Towards Distribution Clustering-Based Deep LSTM Models for RUL Pre-diction. In Proceedings of the 2020 Prognostics and Health Management Conference (PHM-Besançon), Besancon, France, 4–7 May 2020; pp. 253–256. [Google Scholar] [CrossRef]
  36. Ellefsen, A.L.; Bjørlykhaug, E.; Æsøy, V.; Ushakov, S.; Zhang, H. Remaining useful life predictions for turbofan engine degradation using semi-supervised deep architecture. Reliab. Eng. Syst. Saf. 2018, 183, 240–251. [Google Scholar] [CrossRef]
  37. Li, X.; Ding, Q.; Sun, J.-Q. Remaining useful life estimation in prognostics using deep convolution neural networks. Reliab. Eng. Syst. Saf. 2018, 172, 1–11. [Google Scholar] [CrossRef] [Green Version]
  38. Zheng, S.; Ristovski, K.; Farahat, A.; Gupta, C. Long short-term memory network for remaining useful life estimation. In Proceedings of the 2017 IEEE International Conference on Prognostics and Health Management (ICPHM), Dallas, TX, USA, 19–21 June 2017. [Google Scholar]
  39. Zhang, C.; Lim, P.; Qin, A.K.; Tan, K.C. Multiobjective Deep Belief Networks Ensemble for Remaining Useful Life Estimation in Prognostics. IEEE Trans. Neural Netw. Learn. Syst. 2017, 28, 2306–2318. [Google Scholar] [CrossRef]
  40. ElDali, M.; Kumar, K.D. Fault Diagnosis and Prognosis of Aerospace Systems Using Growing Recurrent Neural Networks and LSTM. In Proceedings of the 2021 IEEE Aerospace Conference (50100), Big Sky, MT, USA, 6–13 March 2021. [Google Scholar] [CrossRef]
Figure 1. Dual-channel LSTM network architecture.
Figure 1. Dual-channel LSTM network architecture.
Entropy 24 01818 g001
Figure 2. Partial time features of engine unit 1.
Figure 2. Partial time features of engine unit 1.
Entropy 24 01818 g002
Figure 3. The curve of the ground RUL.
Figure 3. The curve of the ground RUL.
Entropy 24 01818 g003
Figure 4. Time window.
Figure 4. Time window.
Entropy 24 01818 g004
Figure 5. Evaluation index function.
Figure 5. Evaluation index function.
Entropy 24 01818 g005
Figure 6. Flowchart of dual-channel LSTM.
Figure 6. Flowchart of dual-channel LSTM.
Entropy 24 01818 g006
Figure 7. Simplified diagram of engine simulated in C-MAPSS.
Figure 7. Simplified diagram of engine simulated in C-MAPSS.
Entropy 24 01818 g007
Figure 8. The remaining service life curve of some FD001 engines. (a) Engine Unit 24 RUL prediction curve; (b) Engine Unit 34 RUL prediction curve; (c) Engine Unit 76 RUL prediction curve; (d) Engine Unit 100 RUL prediction curve.
Figure 8. The remaining service life curve of some FD001 engines. (a) Engine Unit 24 RUL prediction curve; (b) Engine Unit 34 RUL prediction curve; (c) Engine Unit 76 RUL prediction curve; (d) Engine Unit 100 RUL prediction curve.
Entropy 24 01818 g008
Figure 9. The effect of mini-batch size.
Figure 9. The effect of mini-batch size.
Entropy 24 01818 g009
Figure 10. The effect of the k value. (a) The influence of k value on RMSE; (b) the influence of k value on score.
Figure 10. The effect of the k value. (a) The influence of k value on RMSE; (b) the influence of k value on score.
Entropy 24 01818 g010
Figure 11. Effect before and after momentum smoothing. (a) RUL curve before smoothing; (b) RUL curve after smoothing; (c) model standardized residual before smoothing; (d) model standardized residual after smoothing; (e) histogram of the standardized residual before smoothing; (f) histogram of the standardized residual after smoothing.
Figure 11. Effect before and after momentum smoothing. (a) RUL curve before smoothing; (b) RUL curve after smoothing; (c) model standardized residual before smoothing; (d) model standardized residual after smoothing; (e) histogram of the standardized residual before smoothing; (f) histogram of the standardized residual after smoothing.
Entropy 24 01818 g011
Figure 12. The effect of the p value. (a) The influence of p value on RMSE; (b) the influence of p value on score.
Figure 12. The effect of the p value. (a) The influence of p value on RMSE; (b) the influence of p value on score.
Entropy 24 01818 g012
Figure 13. Effect of optimizer. (a) The influence of optimizer on RMSE; (b) the influence of optimizer on score.
Figure 13. Effect of optimizer. (a) The influence of optimizer on RMSE; (b) the influence of optimizer on score.
Entropy 24 01818 g013
Figure 14. The remaining service life curve of some FD001 engines. (a) Engine Unit 34 RUL prediction curve; (b) Engine Unit 76 RUL prediction curve; (c) Engine Unit 81 RUL prediction curve; (d) Engine Unit 100 RUL prediction curve.
Figure 14. The remaining service life curve of some FD001 engines. (a) Engine Unit 34 RUL prediction curve; (b) Engine Unit 76 RUL prediction curve; (c) Engine Unit 81 RUL prediction curve; (d) Engine Unit 100 RUL prediction curve.
Entropy 24 01818 g014
Table 1. Sensor introduction.
Table 1. Sensor introduction.
IndexDescriptionSymbol
1Total temperature at fan inlet˚R
2Total temperature at LPC outlet˚R
3Total temperature at HPC outlet˚R
4Total temperature at LPT outlet˚R
5Pressure at fan inletpsia
6Total pressure in bypass-ductpsia
7Total pressure at HPC outletpsia
8Physical fan speedrpm
9Physical core speedrpm
10Engine pressure ratio (P50/P2)-
11Static pressure at HPC outletpsia
12Ratio of fuel flow to Ps30pps/psi
13Corrected fan speedrpm
14Corrected core speedrpm
15Bypass ratio-
16Burner fuel–air ratio-
17Bleed enthalpy-
18Demanded fan speedrpm
19Demanded corrected fan speedrpm
20HPT coolant bleedlbm/s
21LPT coolant bleedlbm/s
This dataset is simulated using the commercial modular aerospace propulsion system simulation. Details about the dataset are as follows.
Table 2. Dataset introduction.
Table 2. Dataset introduction.
Sub-DatasetsFD001FD002FD003FD004
Engine units in the training dataset100260100249
Engine units in the test dataset100259100248
Fault modesOne (HPC Degradation)One (HPC Degradation)Two (HPC Degradation, Fan Degradation)Two (HPC Degradation, Fan Degradation)
ConditionsOne (Sea Level)SixOne (Sea Level)Six
Training samples17,73148,81921,82057,522
Test samples100259100248
Table 3. Parameter details.
Table 3. Parameter details.
ParameterValueParameterValue
Nt for FD001 to FD00430/20/30/15Batch size64
Nf for FD001 to FD00417/24/18/24Learning rate0.001
hidden_size32RUL threshold125
k for FD001 to FD0040.3/0.5/0.4/0.5p8
L2 normalization0.0001Activate functionReLU
Table 4. Comparison of method effects.
Table 4. Comparison of method effects.
MethodFD001FD002FD003FD004
RMSEScoreRMSEScoreRMSEScoreRMSEScore
Dual-LSTM11.34267.117.631773.4710.8334.8617.412617.45
Ensemble
ResCNN [10]
12.16212.4820.852087.7712.01180.7624.973400.44
AdaBN-DCNN [11]13.1727920.87202014.9781724.573690
MS-DCNN [12]11.44196.2219.35374711.67241.8922.224844
MA-LSTM [15]13.5225822.57374012.9825623.884560
Deep
LSTM [15]
16.1433824.49445016.1885228.175550
BiLSTM [15]13.6529523.18413013.7431724.865430
L(12.10.7.2)N(2) [35]14.0830818.59188012.1522120.912633
Semi-supervised [36]12.5623122.73336612.125122.662840
DCNN [37]12.61273.722.3610,41212.64284.123.3112,466
MODBNE [38]15.04334.2325.055585.3412.51421.9128.666557.62
DBN [39]15.21417.5927.129031.6414.71442.4329.887954.51
LSTM [38]16.1433824.49445016.1885228.175550
MLP [39]16.78560.5928.7814,026.7218.47479.8530.9610,444.35
Growing RNN [40]14.31302.823.714105.516.42457.4927.954906.04
VarSeq
LSTM [40]
15.2325022.87453214.53152326.115627
Table 5. Performance of mini-batch size.
Table 5. Performance of mini-batch size.
Mini-Batch1632641282565121024
RMSE11.5111.9611.3411.6511.4211.7812.09
Score283.22323.53267.1297.3286.4319.46344.24
Table 6. The effect of the k value.
Table 6. The effect of the k value.
k0.10.20.30.40.50.60.70.80.91
RMSE11.5911.2811.3411.4211.4911.5511.6111.6711.7211.79
Score326.44273.48267.1267.07268.45270.29272.37274.67277.24280.2
Table 7. Performance of different p-values.
Table 7. Performance of different p-values.
p1248None
RMSE15.5812.8312.0411.3412.17
Score359.26348.66307.39267.1343.03
Table 8. Performance of different optimizers.
Table 8. Performance of different optimizers.
RMSpropAdamAdagradAdadeltaAdamaxSGD
RMSE11.3211.3414.3613.6311.3637.8
Score324.26267.1318.92392.83278.513607.59
Table 9. Performance of different RULs.
Table 9. Performance of different RULs.
RULRULln(RUL + 1)
RMSE15.580.15
Score359.261.3
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Peng, C.; Wu, J.; Wang, Q.; Gui, W.; Tang, Z. Remaining Useful Life Prediction Using Dual-Channel LSTM with Time Feature and Its Difference. Entropy 2022, 24, 1818. https://doi.org/10.3390/e24121818

AMA Style

Peng C, Wu J, Wang Q, Gui W, Tang Z. Remaining Useful Life Prediction Using Dual-Channel LSTM with Time Feature and Its Difference. Entropy. 2022; 24(12):1818. https://doi.org/10.3390/e24121818

Chicago/Turabian Style

Peng, Cheng, Jiaqi Wu, Qilong Wang, Weihua Gui, and Zhaohui Tang. 2022. "Remaining Useful Life Prediction Using Dual-Channel LSTM with Time Feature and Its Difference" Entropy 24, no. 12: 1818. https://doi.org/10.3390/e24121818

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop