Next Article in Journal
Geometric Method: A Novel, Fast and Accurate Solution for the Inverse Problem in Risley Prisms
Previous Article in Journal
Self-Heating of Annealed Ti/Al/Ni/Au Contacts to Two-Dimensional Electron Gas in AlGaN/GaN Heterostructures
Previous Article in Special Issue
Special Issue on Intelligent Diagnostic and Prognostic Methods for Electronic Systems and Mechanical Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Uncertainty-Controlled Remaining Useful Life Prediction of Bearings with a New Data-Augmentation Strategy

1
College of Logistics Engineering, Shanghai Maritime University, Shanghai 201306, China
2
AVIC Shanghai Aero Measurement & Controlling Research Institute, Shanghai 201601, China
3
State Key Laboratory of Mechanical System and Vibration, Institute of Vibration, Shock and Noise, Shanghai Jiao Tong University, Shanghai 200240, China
4
Center for Applied Mathematics, Tianjin University, Tianjin 300072, China
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2022, 12(21), 11086; https://doi.org/10.3390/app122111086
Submission received: 23 September 2022 / Revised: 23 October 2022 / Accepted: 28 October 2022 / Published: 1 November 2022

Abstract

:
The remaining useful life (RUL) of bearings based on deep learning methods has been increasingly used. However, there are still two obstacles in deep learning RUL prediction: (1) the training process of the deep learning model requires enough data, but run-to-failure data are limited in the actual industry; (2) the mutual dependence between RUL predictions at different time instants are commonly ignored in existing RUL prediction methods. To overcome these problems, a RUL prediction method combining the data augmentation strategy and Wiener–LSTM network is proposed. First, the Sobol sampling strategy is implemented to augment run-to-failure data based on the degradation model. Then, the Wiener–LSTM model is developed for the RUL prediction of bearings. Different from the existing LSTM-based bearing RUL methods, the Wiener–LSTM model utilizes the Wiener process to represent the mutual dependence between the predicted RUL results at different time instants and embeds the Wiener process into the LSTM to control the uncertainty of the result. A joint optimization strategy is applied in the construction of the loss function. The efficacy and superiority of the proposed method are verified on a rolling bearing dataset obtained from the PRONOSTIA platform. Compared with the conventional bearing RUL prediction methods, the proposed method can effectively augment the bearing run-to-failure data and, thus, improve the prediction results. Meanwhile, fluctuations of the bearing RUL prediction result are significantly suppressed by the proposed method, and the prediction errors of the proposed method are much lower than other comparative methods.

1. Introduction

Rolling bearings as key components are widely used in various machines, such as gearboxes, wind turbines, and so on. The micropitting [1], internal clearance [2], and lubricant contamination [3] are factors that affect the healthy operation of rolling bearings. These factors affect the bearing life through load distribution, frictional torque, generated vibration, and heat. Predicting the remaining useful life (RUL) of bearings is critical to keeping a machine in healthy working order [4]. Current methods for RUL prediction can be divided into several major categories: physics model-based methods, statistical model-based methods, and artificial intelligence (AI)-based methods [5,6]. Among these methods, statistical model-based methods and AI-based methods are widely used in bearing RUL prediction.
The goal of statistical model-based methods is to accurately explain the machine’s degrading process using a statistical model [7]. There are some commonly used statistical models, such as Wiener processes [8], Gamma processes [9,10], and so on. Then, the parameters in the statistical model are estimated by particle filtering [11], expectation maximization algorithm [12], recursive filtering algorithm [13], and so on. The Gamma process and the Inverse Gaussian process, however, are only appropriate for describing monotonous degradation processes. Because of self-healing and small machine repairs, non-monotonous processes are more common in actual industrial settings. The Wiener process is more popular because it can be used both in the monotonous and non-monotonous degradation process. Although statistical model-based methods have implemented several well-established studies in RUL prediction, statistical model-based methods are often developed on a case-specific basis, with the limitation of requiring explicit prior knowledge [14,15].
The AI-based methods can deal with a large amount of monitoring data by machine learning (ML) techniques without much prior knowledge. Traditional machine learning techniques, such as Support Vector Machine (SVM) [16,17] and Relevance Vector Machine (RVM) [18], have been widely used in RUL prediction. Compared with the traditional machine learning techniques, deep learning methods show greater potential in dealing with high non-linearity and data complexity. There are some commonly used deep learning (DL) models, such as Auto-Encoder (AE) [19], Convolutional Neural Network (CNN) [20,21,22,23,24], Recurrent Neural Network (RNN) [25], and so on. Zhao et al. [26] used CNN in bearing RUL prediction. Raw signals are transformed into the time-frequency domain features and then mapped with the RUL by the 2D-CNN model. The Long Short-Term Memory Neural Network (LSTM) network is one of the RNN variations that can address the issue of gradient disappearance and gradient explosion. Xiang et al. [27] proposed an LSTM with attention mechanism for gear RUL prediction, calculating the contribution of the input during the RUL prediction by using the attention mechanism to improve the performance of the network. However, the AI-based methods still have two limitations: (1) the training process of AI-based methods requires enough data; Thus, lacking run-to-failure data limits the training process. (2) The RUL prediction by the existing deep learning model at different times is mutually independent. The RUL prediction sequence is always ignored in this way.
In order to get the sufficient data for deep learning model, several data augmentation techniques have been applied in fault diagnosis [28] and RUL prediction [29]. Generative Adversarial Networks (GANs) have been widely used to solve the problem of data imbalance in fault diagnosis [30,31]. Behera et al. [32] proposed a RUL prediction method based on conditional generative adversarial network (CGAN) and deep gated recurrent unit (DGRU) network. Hundreds of run-to-failure data were used to train the CGAN to generate more data. Then, the DGRU was applied to predict the RUL. However, GAN models still require enough data to train, which is not suitable when only several run-to-failure samples are offered.
The RUL prediction sequence is rarely considered in the existing DL-based models. Thus, the uncertainty between the predicted result at different times is always ignored [33]. The accuracy of the predicted result is the only object in the training process. Both the uncertainty and accuracy of the predicted result are important for maintaining the health of the machine. In the existing literature, limited research about the uncertainty of the predicted RUL was offered. In [34], Recurrent Convolutional Neural Network (RCNN) was used for RUL prediction and the variational inference was used to quantify the uncertainty of the RCNN model in RUL prediction. However, only quantifying the uncertainty is not enough. Large fluctuations of the predicted RUL would appear unless uncertainty is controlled. Thus, the uncertainty of the predicted RUL should not only be quantified but also controlled in the actual application to get a reliable and stable prognosis.
This paper proposes a RUL prediction method based on data augmentation strategy and Wiener–LSTM network to solve these two problems. During the data augmentation process, the RMS of the monitored vibration signal is calculated to fit the degradation model. Then, the range of parameters in the degradation model can be acquired when fitting different data. Sobol sampling is implemented in each range to reconstruct different combinations of the parameters. Finally, these different combinations of the parameters are substituted into the degradation model to get more generated run-to-failure data.
During the process of the RUL prediction by Wiener–LSTM network optimized by PSO algorithm, the Wiener process is used to model the predicted RUL, which can control the uncertainty propagation. Then the LSTM is applied to fit the assumed Wiener process during the iterations. After the training of the Wiener–LSTM model, the predicted RUL is closed to the assumed Wiener process. Thus, a smooth curve without large fluctuations of the predicted RUL can be acquired by the proposed model. In order to embed the Wiener process into the LSTM, a joint optimization strategy is implemented to derive a new loss function. Meanwhile, the PSO algorithm is used here to optimize the hyperparameters in the Wiener–LSTM model. The main contributions of this paper are described as follows:
(1)
A data augmentation method based on degradation process modeling and Sobol sampling augments the run-to failure training data;
(2)
A new loss function for the Wiener–LSTM model is proposed, and the Wiener process is introduced into the LSTM network to control the uncertainty.
The rest of this paper is organized as follows: In Section 2, the problem of the RUL prediction methods based on conventional LSTM is proposed. In Section 3, the framework of the proposed method is introduced in detail. In Section 4, the proposed methods are validated on a rolling bearing dataset. In Section 5, the paper is sealed with a conclusion.

2. Problem Statement

The signal of a working machine was inspected by sensors placed in the specified location. x 0 , x 1 , x 2 , , x t is the health indicator inspected from time 0 to t, which is recorded as x 0 : t for convenience. r 0 , r 1 , r 2 , , r t is the corresponding RUL from time 0 to t, which is recorded as r 0 : t for convenience. The failure time of the machine is recorded as τ . The RUL at time t is recorded as r t = τ t . The LSTM network is typically used in the RUL prediction process in order to learn potential relationships between the input data and the RUL. The LSTM network is assumed as a predictor ϕ . The RUL predicted by the predictor at time t is recorded as α t . All the parameters of the LSTM network are assumed as Θ . The process of the RUL prediction based on the predictor ϕ can be described as
ϕ ( x 0 : t ; Θ ) = α t .
However, the lack of run-to-failure data hinders the training process of the LSTM network. Unless the LSTM model ϕ is well-trained and has enough data, it is impossible to build a mapping between health metrics and corresponding RUL. In the actual industry, fault data are limited because the machine always works in a healthy condition most of the time. Thus, run-to-failure data are hard to acquire. Data augmentation is essential for RUL prediction. In this paper, the proposed data augmentation method based on degradation modeling and Sobol sampling are introduced in detail in Section 3.1.
Another problem is that the RUL prediction based on the conventional LSTM ignored the relationship between predictions at different times with the same data stream. The uncertainty of the predicted result can not be controlled over the time horizon. The training process of the conventional LSTM network aims to minimize the differences between the real RUL and the predicted RUL at a specific time. The loss function of the conventional LSTM network is
L o s s ( x 0 : t , r 0 : t ; Θ ) = i = 0 t r i ϕ ( x 0 : i ; Θ ) 2 ,
where r i denotes the real RUL at time i between time 0 to t, x i denotes the input signal at time i. In the training process, the optimal parameter Θ * of the LSTM can be acquired when minimizing the loss function as
Θ * = arg min Θ L o s s ( x 0 : t , r 0 : t ; Θ ) ,
the shortcoming of the conventional LSTM prediction is that the predictions at different times are mutually independent. The time-dependence among the RUL prediction sequence is not considered in the conventional LSTM model. Unexpected fluctuation of the predicted RUL may occur because the mutual dependence of the RUL prediction is not concerned according to the loss function described in Equation (2).
In order to get a reliable prediction without controllable fluctuations, the uncertainty among the RUL prediction over the time horizon should be taken into consideration. The increment of the predicted RUL α t and α s is assumed to follow the normal distribution [35] as
α t α s N s t , ( t s ) / c ,
where the s t is the mean of the normal distribution. ( t s ) / c is the variance of the normal distribution. The parameter c can be used to control the uncertainty of the variance. This property of the predicted RUL contributes to the assumption of the Wiener process on the predicted RUL [35]. When considering the relationship among the consecutive predicted RUL at different times, the RUL prediction combining the LSTM network and the Wiener assumption, as described in Equation (2), has two objects now: (1) the error between the predicted RUL and the real RUL can be reduced by optimizing the parameter Θ of the LSTM network; (2) the uncertainty of the increment can be controlled by optimizing the parameter c of the Wiener process. When the historical predicted RUL is considered, the training loss function is transformed into a joint optimization with a new uncertainty control parameter c
( Θ * , c * ) = arg min Θ , c L o s s ( x 0 : t , r 0 : t ; Θ , c ) ,
where the optimization processes of the parameters c and Θ in Equation (5) are mutually dependent. How to jointly optimize c and Θ in the function and construct the loss function of Wiener–LSTM model will be specified in Section 3.2.

3. Uncertainty-Controlled Remaining Useful Life Prediction with a Data Augmentation Strategy

Figure 1 shows the proposed method’s structure, including two training and testing processes. In the training process, RMS of the vibration data is calculated firstly as the health indicator. Then, the Fault Occurrence Time (FOT) of each run-to-failure data is determined by the 3 σ rule. RMS data of the degradation stage is prepared for the data augmentation. The data augmentation based on the degradation modeling and Sobol sampling are introduced in detail in Section 3.1. The run-to-failure data generated in the next step is used to train the Wiener–LSTM model. Then the PSO algorithm is applied to optimize the hyperparameters in the Wiener–LSTM model. The detailed descriptions of the Wiener–LSTM model are introduced in Section 3.2. In the testing process, the procedure is similar to the training process. The RMS of the testing datasets is calculated, and the FOT is determined by the 3 σ rule. Then the testing datasets are sent to the trained Wiener–LSTM model to verify the effectiveness of the proposed method.

3.1. Data Augmentation Based on Degradation Modeling and Sobol Sampling

This study proposes a data generation method for producing enough data for training based on degradation modeling and Sobol sampling. The flowchart of the data generation method can be divided into four steps. Detailed descriptions of each step are given as follows.
(1)
Degradation modeling is the first step. The RMS is a commonly used time-domain feature, which can reflect the degradation process. Thus, the RMS of the monitored bearing is chosen as the health indicator. The RMS of the whole life bearing signal is shown in Figure 2a, which can be divided into two different stages [36]. The bearing is in the health stage at an early time. RUL prediction is not necessary for this stage because RMS value in the health stage shows a smooth trend. At the Fault Occurrence Time (FOT), the bearing gets into the degradation stage. The degradation can be modeled as
X t = exp ( m + σ 1 ε ( t ) ) t γ a exp ( e 2 2 + b ( t γ ) + e ε ( t γ ) ) t > γ ,
where the γ is the FOT, it is a random variable because different bearing has different FOT. When t γ , the bearing is in the healthy stage. The m is a normal random variable representing a stable level of normal bearing health. The σ 1 is a noise variance in this healthy stage. ε ( 1 ) , ε ( 2 ) in ε ( t ) is independent at different time. RUL prediction often starts from FOT to the failure time, which means the stage t > γ is always used for RUL prediction. At this stage, a is a log-normal random variable representing the magnitude of the exponential degradation trend. b is a normal random variable representing the slope of the exponentially degenerate trend. e is the noise variance in the degradation process. If there are N number of vibration signal data at each sampling point as D = [ d 1 , d 2 , d 3 , , d N ] . In [37], commonly used health indicators in residual useful life prediction are root mean square (RMS), kurtosis, peak value, and so on. The RMS is a commonly used indicator to describe the degradation trends of bearings, due to its ability to reflect the vibration energy characteristics with robustness. In [38], it is also stated that the RMS indicator is a commonly used indicator in the machinery fault diagnosis and remaining useful life prediction. The RMS is chosen as the indicator in the proposed method, and the unit of RMS is ‘g’. The RMS of the vibration signal is calculated firstly as the health indicator as
x i r m s = 1 N i = 1 N d i d m 2 ,
where d i is the signal in the sampling point. d m is the mean of the signal recorded from 0–N. The FOT of the RMS is then calculated using the laida criterion, also known as the 3 σ rules. The RMS health stage mean μ and standard deviation σ are calculated. The first value, which exceeds 3 σ is assumed as the FOT. Then the data from FOT to failure are collected and prepared for the next step.
(2)
Since the RUL prediction is not necessary for the stage t γ . The model fitting is implemented only in the stage t > γ to acquire the parameters a, b, and e in Equation (6). The ε ( t ) follows a normal distribution, which can not be fitted in this process. All the training samples are fitted by the function of the degradation model. Different parameters are acquired after fitting different samples. If there are M run-to-failure data, M sets of parameters of a , b , and e can be acquired after the model fitting on the M run-to-failure datasets. Among these sets of parameters, the maximum and minimum values should be chosen. Ranges of each parameter can be recorded as [ a m i n , a m a x ] , [ b m i n , b m a x ] and [ e m i n , e m a x ] .
(3)
The Sobol sequence sampling is used to get different combinations of parameters. Random sampling algorithms are quasi-random and limited to one period. When the cycle is exceeded the period, they are repeated and are no longer mutually independent random numbers. Sobol sequences sampling method focus on producing uniform distributions in the probability space compared with the random sampling method. Localized clustering can be avoided in this way. As one of the low deviation sequences, Sobol sequence sampling is superior to other low deviation sequences. The random numbers generated afterward will be distributed to the areas that were not previously sampled. A set of independent parameters can be acquired after one Sobol sequence sampling among the ranges of [ a m i n , a m a x ] , [ b m i n , b m a x ] and [ e m i n , e m a x ] . The algorithm of Sobol sequence sampling is as follows.
Consider i random data Z i are generated in the range of [ a m i n , a m a x ] . Z i follows the normal distribution. A non-integrable polynomial f ( Z ) can be constructed as
f ( Z ) = Z n + m 1 Z n 1 + m 2 Z n 2 + + m n 1 Z + 1 ,
where the n is the highest order. The m 1 , m 2 , m 2 , is the coefficients of the polynomial. The direction value v i can be acquired as
v i = m 1 v i 1 m 2 v i 2 m s v i n v i n 2 n ,
where the ⊕ is binary logic operation. ⌊⌋ is the rounding down operation. The random data Z i can be based on the direction value as
Z i = b 1 v 1 b 2 v 2 b 3 v 3 b i v i ,
where the b 1 , b 2 , b 3 , is the binary representation of i. After the Sobol sequence sampling operation is implemented on the ranges of parameters a , b , and e. i sets of parameters of the degradation model can be acquired.
(4)
Now, i sets of parameters sampled in the last step are substitute to the degradation model function in Equation 6 again without ε ( t ) to get i new run-to-failure data as
X t i = a exp ( e 2 2 + b ( t γ ) ) ,
the rand noise ε ( t γ ) is added in the new sample to imitate the random shock in the working situation in the last step. The i run-to-failure data are prepared and can be sent to the Wiener–LSTM model for the training process.
Figure 2. (a) The RMS of the whole life data of bearing (b) Degradation model fitting of the degradation stage. Two stages are included in the run-to-failure data of bearing. The RMS of the bearing shows a steady state in the health stage. After Failure Occurrence Time (FOT), the RMS of the bearing shows an exponential degradation in the degradation stages.
Figure 2. (a) The RMS of the whole life data of bearing (b) Degradation model fitting of the degradation stage. Two stages are included in the run-to-failure data of bearing. The RMS of the bearing shows a steady state in the health stage. After Failure Occurrence Time (FOT), the RMS of the bearing shows an exponential degradation in the degradation stages.
Applsci 12 11086 g002

3.2. Wiener–LSTM Bearing RUL Prediction Model

3.2.1. Forward Propagation of the LSTM ϕ

The LSTM network is established as the RUL predictor in the proposed method, which is one of the variants of the Recurrent Convolutional Neural (RNN). The gradient vanishing and explosion problems of the RNN can be addressed by the LSTM network memory cell. Figure 3 shows the memory neuron structure of the LSTM network. The memory neuron consists of three kinds of gate units called forget gate f t , input gate i t , and output gate O t . The forward propagation process of the automatic predictor ϕ is introduced in detail now.
The forget gate is used to determine what information should be discarded. The input of the forget gate is the output of the hidden state in last time h t 1 and the current time input data x t . The x t is the signal of the monitored machine at time t. The important information of the last time would be reserved and the redundant information would be discarded. The calculation process of the forget gate can be followed as
f t = σ W f · h t 1 + W f · x t + b f ,
where W f and b f are the weights and bias of forget gate. σ (·) is the sigmoid activation function.
Two steps are involved in this part. The first step is the calculation of the input gate. The second step is adding the candidate values C ˜ t to the memory cell C t . C t 1 is the old state value of the memory cell at time t 1 , C t is the new state value of cell at time t which is updated by the process as
i t = σ W i · h t 1 + W i · x t + b i , C ˜ t = tanh W i · h t 1 + W i · x t + b C , C t = f t C t 1 + i t C ˜ t ,
where W i and b i are the weights and bias of input gate. i t is the output of the input gate. * denotes dot product.
The output gate is used to determine what information should be exported as
O t = σ W O · h t 1 + W O · x t + b O , h t = O t tanh C t ,
where W O and b O are the weights and bias of the output gate. h t denotes the output of the LSTM cell at time t. The LSTM layer is normally constructed with the dense layer, the output h t of the LSTM is the input of the dense layer. The output of the dense layer is the predicted RUL, which can be calculated as
α t = σ W d · h t ( x t ) + b d ,
where W d and b d are the weights and bias of the dense layer. All the parameters of the LSTM layer and dense layer, such as W f , b f , W i , b i , and so on, are recorded as Θ for convenience.

3.2.2. The Wiener–LSTM Model with Joint Optimization Loss Function

In order to control the uncertainty of the increment of the prediction result, the uncertainty of the predicted RUL is modeled by the Wiener process. The object of the training of the LSTM network is shown in Equation (5) now. The possibility function of r t , α t , and α s should be acquired to solve the parameters in Equation (5) The possibility of joint prediction for two times is recorded as P ( r t , α t , α s ) . The joint equations can be get to solve the possibility of joint prediction as
ϕ x 0 : t ; Θ = α t ϕ x 0 : s ; Θ = α s P ( r t , α t , α s ) = P ( α s ) P ( α t r s ) P ( r t α t ) ,
where the α t and α s at different times are dependent because of the Markov property of the stochastic process. The possibility of the joint prediction P ( r t , α t , α s ) can be transformed as three possibility as shown in Equation (16). r t denotes the real RUL at time t, the α t , and α s denotes the RUL predicted by the predictor at time t and s ( t > s ).
The modeling of P ( α t α s ) , P ( r t α t ) and P ( α s ) is given in [35] As for two predicted RUL α t and α s at two times t > s , the prediction increment is assumed to follow the normal distribution
α t α s N ( s t , ( t s ) / c ) ,
where the mean of the normal distribution is s t , the variance of the normal distribution is ( t s ) / c . The P ( α s ) is assumed to follow the normal distribution [35]
α s N r s , r s / c .
The P ( r t α t ) follows the Inverse Gaussian distribution [39]
I G ( α s , c ) = c α s 2 2 π x 3 1 / 2 exp c ( x α s ) 2 2 x ,
where I G denotes the Inverse Gaussian distribution. The means of the distribution is α s , the variance of the distributed is α s / c .
Thus, inspired by the joint likelihood possibility proposed in Equation (16), the loss function is reconstructed to guide the uncertainty control and accuracy improvement. As for a run-to-failure sample, the failure time of the machine is recorded as τ . α 0 , α 1 , α 2 denotes the predicted RUL at different times. If the historical predictions are all considered in the joint likelihood possibility function, which can be represented as
ϕ x 0 ; Θ = α 0 ϕ x 0 : 1 ; Θ = α 1 ϕ x 0 : τ ; Θ = α τ P ( r τ , α τ , α τ 1 α 0 ) = P ( α 0 ) i = 1 τ P ( α i α i 1 ) P ( r τ α τ ) .
All the samples should be considered in the joint likelihood function as the new loss function of the Wiener–LSTM network. Then the joint likelihood function can be maximized automatically by several iterations of the LSTM network. The new loss function can be expressed as
L o s s n e w ( x 0 : τ n , r 0 : t n ; Θ , c ) = 2 N n = 1 N P ( α 0 n ) i = 1 τ P ( α i n α i 1 n ) P ( r τ n α τ n ) ,
where n denotes the number of the running-to-failure samples.
The parameters of the LSTM network include Θ and c now. Θ are the parameters of the LSTM network. c is the parameter of the Wiener process which can control the uncertainty of the predicted RUL. However, the parameter c cannot be updated by the iteration of the LSTM network. It is a parameter out of the structure of the LSTM network. The c and the Θ can not be updated simultaneously by the iteration of the LSTM network. Therefore, the optimal value of c should be calculated first. The variable Θ is assumed as the constant value in the process of solving the optimal valve c
c * = n + 1 n = 1 N r τ n ϕ n ( x 0 : τ ; Θ ) 2 r τ n + r 0 n ϕ n ( x 0 ; Θ ) 2 r 0 n + i = 1 τ ϕ n ( x 0 : i ; Θ ) ϕ n ( x 0 : i 1 ; Θ ) + 1 2 .
The optimal c * can be expressed only including parameter Θ now. The expression of the parameter c * can be substituted into the new loss function. All the parameters of the loss function can be updated by iteration of the LSTM network. The proposed loss function of the LSTM network is the logarithmic function as
log Loss n e w x 0 : τ n , r 0 : τ n ; Θ = 1 N n = 1 N 3 log ( 2 π ) log c * + log r 0 n + c * r 0 n ϕ n x 0 , r 0 ; Θ 2 r 0 i + + c * i = 1 τ ϕ n x 0 : i , r 0 : i ; Θ ϕ n x 0 : i 1 , r 0 : i 1 ; Θ + 1 2 + 2 log ϕ n x 0 : τ , r 0 : τ ; Θ + 3 log r τ n + c * r τ n ϕ n x 0 : τ , r 0 : τ ; Θ 2 r τ n
The log-likelihood loss function, as shown in Equation (23) is convenient for calculating. The new loss function is the training object of the Wiener–LSTM model. During backpropagation, the parameter Θ is updated using the Adam algorithm. The Adam algorithm is a first-order optimization technique that, like a conventional stochastic gradient descent procedure, iteratively adjusts the weights of the neural network based on training data [40]. The process Θ * = arg min Θ L o s s n e w ( x 0 : τ n , r 0 : τ n ; Θ ) can be acquired by the Adam algorithm.

3.2.3. Optimization of the Hyperparameters by PSO Algorithm

It is well known that the performance of the network is influenced by the hyperparameters, such as the learning rate, number of neurons, maximum epochs, and so on. However, these hyperparameters are always tested by so many trials to seek the optimal structure of the network. As one of the optimization algorithms, the PSO algorithm has been widely used in several optimization tasks of the network [41] because of the fast convergence and robust applicability. The structure of the Wiener–LSTM model with PSO optimization is shown in Figure 4. The Wiener–LSTM model consists of one LSTM layer and one dense layer. Thus, the hyperparameters in this model include the number of hidden layer neurons ( h n ), initial learning rate ( l r ), and drop rate factor ( d r ). The process of the optimization can be divided into four steps:
  • Step 1: Parameter initialization. The particle dimension, population size, iterations, learning factors, inertia weight, velocity, and position are determined.
  • Step 2: Initialize the particle positions and velocities, then generate population particle ( h n , l r , d r ) at random.
  • Step 3: The loss function in Equation (23) is chosen to be the fitness function in the PSO algorithm here. The particle position and velocity are updated by epoch. The extreme individual value and extreme global value are then updated by computing the fitness value in accordance with the new situation.
  • Step 4: Judge whether the termination conditions are met. If satisfied, the algorithm ends and outputs the optimization result ( h n , l r , d r ); otherwise, return to Step 1.
Figure 4. The structure of the Wiener–LSTM model with PSO optimization. The signals measured by sensors are x 1 , x 2 , x 3 , x t at time 1 , 2 , 3 , t . The corresponding RUL is recorded as r 1 , r 2 , r 3 , r t . The predicted RUL α 1 , α 2 , α 3 , α t is predicted by the LSTM network ( ϕ ). The Wiener process Y t of the predicted RUL is introduced to control the uncertainty of the predicted result. W t denotes the standard Wiener process. c is the parameter used to control the uncertainty propagation rate. The Wiener process can be introduced into LSTM in the back propagation of the LSTM network.
Figure 4. The structure of the Wiener–LSTM model with PSO optimization. The signals measured by sensors are x 1 , x 2 , x 3 , x t at time 1 , 2 , 3 , t . The corresponding RUL is recorded as r 1 , r 2 , r 3 , r t . The predicted RUL α 1 , α 2 , α 3 , α t is predicted by the LSTM network ( ϕ ). The Wiener process Y t of the predicted RUL is introduced to control the uncertainty of the predicted result. W t denotes the standard Wiener process. c is the parameter used to control the uncertainty propagation rate. The Wiener process can be introduced into LSTM in the back propagation of the LSTM network.
Applsci 12 11086 g004

4. Experiment

This paper uses a popular dataset to verify the effectiveness of the proposed Wiener–LSTM network. All code was run on a PC with an Intel Core i7-5557U CPU, and 4 GB RAM. The codes are based on the Keras framework using the Tensorflow backend.

4.1. Data Description

In Figure 5, the accelerated degradation experiments are taken in the PRONOSTIA platform [42]. Three parts, such as the rotating part, the degradation part, and the measurement part, are included in the platform. The degradation part is used to accelerate the bearing degradation. This platform of the PHM2012 dataset is composed of two high-frequency accelerometers of type 3035B DYTRAN placed horizontally and vertically on each bearing to pick up the horizontal and vertical vibration signals. The accelerometer is a single axis with a range of 50 g, and the response frequency is 0.34 to 10,000 Hz. Two acceleration sensors are mounted on the vertical and horizontal axes in order to measure the bearing’s vibration signal. Compared to the vertical vibration signal, the horizontal vibration signal has a higher and more noticeable amplitude. As a result, the experiment solely uses the horizontal vibration signal. The sampling frequency is 25.6 kHz, and the signal is sampled for 0.1 s every 10 s. The experiments contain 17 run-to-failure bearings. All the datasets are shown in Table 1. The raw vibration acceleration signal of the test bearing1_3 is shown in Figure 6. The signal to the left of FOT is the healthy bearing signal, and the one to the right of FOT is the fault signal. The overall time domain waveform shows that the vibration acceleration amplitude is relatively low in the front, starts to gradually increase after the early failure point FOT, and increases dramatically after the failure stage. The lower left figure shows the signal in the normal phase, with lower amplitude and a smoother signal. The lower right figure shows the signal in the failure phase, and the vibration amplitude is larger than the normal state, which can be seen due to the shock component caused by the failure.

4.2. Data Generation Based on the Degradation Model and Sobol Sampling

Five run-to-failure training data are used to fit the degradation model here. Bearing3_1 is discarded because the RMS of this bearing is not suitable for the degradation model in Equation (6). Then five sets of parameters a , b , e can be acquired by the proposed data augmentation method. 250 run-to-failure data are generated for the preparation of the Wiener–LSTM training. Ten examples of the generated run-to-failure data are shown in Figure 7. The lifetime of the generated run-to-failure data ranges from 1000 s to 3000 s. It should be noted that the RUL of the bearing is calculated from the Failure Occurrence Time (FOT), while the PHM2012 is an accelerated degradation experiment.
In Figure 7, the number 1 , 2 , 3 , , 10 denotes the generated run to failure data of bearing 1 to 10. The horizontal axis shows the inspection time. The vertical l axis shows the value of RMS. The RMS of the generated run to failure bearing data is consistent with the exponential model. To simulate the run to failure data with a different lifetime, the generated data are different from each other.
As for the testing data. Bearing 1_3 and Bearing 1_4 are chosen to verify the effectiveness of the proposed method. The RMS values of the Bearing 1_3 and Bearing 1_4 are calculated and the data of 1000 s before failure are used for RUL prediction on these two testing bearings.

4.3. Wiener–LSTM Training and Optimization

A Wiener–LSTM model with PSO optimization is proposed to control the uncertainty and improve the accuracy of the prediction RUL simultaneously. In the training process, the time-domain feature RMS of the vibration signal is calculated first. Then, the data are generated according to the data augmentation model for model training. In the testing process, the RMS of the testing vibration signal is sent to the predictor which has trained in the training process. Then, the predicted RUL of the testing data can be given by the predictor automatically. The prediction uncertainty model of the predicted RUL is introduced in the LSTM network. When the loss function (introduced in Equation (23)) is used in the LSTM model. The training sample of the Wiener–LSTM should be a sequence of data such as x 0 : t , which is different from the point prediction that one training sample is one data at time t. Thus, the relationship of the difference of predicted RUL α t + 1 α t , α t α t 1 can be established and introduced in the LSTM model.
A randomly under-sample method based on mini-batch is implemented in the training process, which is different from the normal LSTM training process with the same training sample in every epoch. Fixed-length sub-sequence from every input sequence is used to form a data batch. The fixed length of the sub-sequence is set for the minimum of the length of the input sequences. The under-sampling method in each minibatch can guarantee the independence of the statistical process and make full use of each sequence of data. The fixed length of the sub-sequence is set for 99 (to make full use of each sequence data among under-sample, the fixed-length equals the minimum value of the lifetime data minus one). The process of the under-sample method based on mini-batch follows three steps: (1) under-sampling method is implemented to get the training sample X i ; (2) c value is updated by Equation (22); and (3) parameters Θ are updated by the Adam algorithms. The algorithm of the under-sample method based on mini-batch is in Algorithm 1.
Algorithm 1: Down-sample algorithm based on mini-batch
Input: Training data X, corresponding RUL r, epoch of training process I
 1: initialization Θ and c of the network.
 2: If i < I
 3:    X i are under-sampled on training data X.
 4:   Update parameter c by Equation (22)
 5:   Update parameter Θ by random gradient descent algorithm by the loss function in Equation (23)
 6: End
 7: return ( Θ * , c * )=argmin(log(loss( Θ , c, X))).
Output: Θ * , c *
The structure of the LSTM model contains two layers: one LSTM layer and one fully-connected layer. After the optimization by the PSO algorithm, the optimal number of neurons in LSTM is set for 64. The optimal dropout rate is 0.23, and the optimal learning rate is 0.02; the activation function is ReLU. In the fully-connected layer, the number of neurons is set for 1, and the activation function is sigmoid. The input of the Wiener–LSTM model is (250, 99, 1). The number of the training sample is 250, the length of each sub-sequence under-sampled in each epoch is 99, and the feature of input data is 1. The label of the training data is (250, 99), which is the corresponding RUL of training samples.
The loss functions of Wiener–LSTM model among epochs are shown in Figure 8. The X-axis represents the number of epoch. The Y-axis represents the magnitude of the loss function. As shown in Figure 8 from the first to the 20th epoch, the Wiener–LSTM model’s training loss function dramatically decreased. From the 20th to the 60th epoch, the loss function continues slowly degrade. The loss function gradually finds convergence from the 60th to the 80th epoch.

4.4. Results and Discussions

4.4.1. Comparison 1

To verify the effectiveness of the data augmentation strategy, the comparison experiments between the Wiener–LSTM model with data augmentation and without data augmentation are taken. The predicted RUL by 6 training samples without data augmentation on test bearing1_3 and bearing1_4 are shown in Figure 9. The blue line is the real RUL from time 1000–0 s. The orange line represents the outcome that the Wiener–LSTM model anticipated. In the training process, only six training samples are used to train the Wiener–LSTM model such as bearing1_1, bearing1_2, bearing2_1, bearing2_2, bearing3_1, bearing3_2. The result of the test bearing1_3 is shown in Figure 9a. The result of the test bearing1_4 is shown in Figure 9b. The predicted RUL are far from the real RUL in the Figure 9 both in test bearing1_3 and test bearing1_4. Thus, enough training samples are necessary for the training process of the Wiener–LSTM model.

4.4.2. Comparison 2

To verify the effectiveness of the Wiener–LSTM model, the comparison experiments between the Wiener–LSTM model and the LSTM model are taken. In the training process of the LSTM model, the loss function MSE is shown in Figure 10. The training loss function of normal MSE declines rapidly from the beginning to the 10th epoch and shows a steady trend from the 10th epoch to the 80th epoch. The results of the loss function curve show that each model has reached convergence at the 80th epoch. The same number of epoch should be ensured for the two models for comparison. Thus the training epoch is set for 80.
Predicted RUL results by Wiener–LSTM and LSTM models are shown in Figure 11. The blue line in Figure 11a,b represents the real RUL of the bearing1_3 and bearing1_4. The orange line is the result predicted by the Wiener–LSTM model with the loss function in Equation (23). The yellow line indicates the result predicted by the normal LSTM using the MSE loss function. Both models have the same structure: the same layers and the same number of neurons in each layer. Predicted results of the test bearing1_3 are shown in Figure 11a. As for the test bearing1_3, the curve of the predicted RUL based on the Wiener–LSTM model is closer to the real RUL than the results based on the normal LSTM model at the beginning 0–400 s. After 400 s, the predicted RUL based on the Wiener–LSTM model is farther from the real RUL than the results based on the normal LSTM model. Although the Wiener–LSTM model has worse accuracy than the normal LSTM model at the end of the time in Figure 11. The predicted result based on the Wiener–LSTM model has a stable result without large fluctuation. On the contrary, the unstable RUL result with fluctuations is predicted by the LSTM model. The large fluctuation of the predicted RUL would affect the decision-making of the maintain. However, the predicted RUL results by the Wiener–LSTM model are smooth. The assumed Wiener process is proved that can be well fitted by the LSTM model to control the uncertainty of the predicted RUL. The uncertainty of the RUL between continuous time can be controlled by the Wiener–LSTM model. The predicted result based on Wiener–LSTM shows better accuracy than the LSTM model for test bearing1_4 as shown in Figure 11b. The predicted RUL curve is closer to the real RUL (the blue line) than the LSTM model all the time. The curve of the RUL result predicted by the Wiener–LSTM model is smoother than the RUL result predicted by the LSTM model. Fluctuations are obvious in the result of the LSTM model. The uncertainty of the result can be well controlled no matter which test bearing. The reliability of maintenance decisions is improved by the proposed Wiener–LSTM model.
A normal indicators Mean Absolute Error (MAE) is implemented in this paper to assess the predicted result quantitatively. The functions of the MAE is
M A E = i = 1 N r i α i / N ,
where the r i denotes the real RUL at time i, the α i denotes the predicted RUL at time i. N is the number of the samples.
The MAE of the predicted results by Wiener–LSTM and LSTM models are shown shown in Figure 12. The MAE of the predicted results of the test bearing1_3 is shown in Figure 12a. The MAE of the predicted results of the test bearing1_4 is shown in Figure 12b. The blue line is the MAE result by using the Wiener–LSTM model. The orange line is the MAE result by using the LSTM model. In Figure 12a, The MAE of the Wiener–LSTM model is lower than the LSTM model on the test bearing1_3 at the beginning time 0–600 s. Although the MAE of the LSTM model is mostly the same as the Wiener–LSTM model. The MAE of the LSTM model has large fluctuation. On the contrary, the MAE of the Wiener–LSTM model is smooth all the time. The prediction uncertainty of the result can be proved that controlled by the Wiener–LSTM model in Figure 12. The error of the predicted result is in a stable state, which means the uncertainty of the predicted result is in a stable state without large fluctuations. It is important for the maintenance decisions in the industry. The MAE of the Wiener–LSTM model is lower than the LSTM model on bearing 1_4, as shown in Figure 12b, and it is more smooth than the LSTM model. As shown in Table 2, the proposed method is applied to the test bearing1_3 and bearing1_4 data. The MAE error between the remaining useful life predicted by the proposed method and the true remaining useful life is much lower than that of the traditional LSTM method. The larger MAE error in the testing process indicates a lower accuracy of the model, and conversely, a minor MAE error indicates a higher accuracy. It is evident that the uncertainty of the result can be well controlled by the Wiener–LSTM model by the experiments on the test bearing1_3 and bearing1_4. The large fluctuations can be removed from the RUL prediction by using the proposed Wiener–LSTM model.

5. Conclusions

In this paper, a bearing RUL prediction method with data augmentation strategy and Wiener–LSTM network with PSO optimization is proposed. The data augmentation method is designed to offer enough training data for the training process. The Wiener–LSTM model is constructed to improve the accuracy of the predicted RUL and control the uncertainty of the result simultaneously. The difference in the predicted RUL result is modeled by the Wiener process to control the uncertainty propagation rate of the result. Wiener process is applied as the connection between the parametric probabilistic model and the LSTM model. The parameters in the Wiener process model and the LSTM network can be updated iteratively by optimizing the joint probability loss function. The proposed Wiener–LSTM method is verified on a popular dataset. The predicted RUL shows better accuracy on test bearings than the normal LSTM model, and the uncertainty of the RUL result can be proved well controlled in the experiments. The stable RUL results with fewer fluctuations predicted by the proposed Wiener–LSTM model are more suitable for decision-making in industrial applications. Two error metrics are applied to verify the validity of the proposed method. The larger MAE error in the testing process indicates a lower accuracy of the model, and conversely, a minor MAE error indicates a higher accuracy. On the test bearing1_3 and bearing1_4, the MAE error of the proposed method is 6.34 and 8.03, which is much lower than that of the traditional LSTM model. The PSO algorithm method is applied here to solve the problem of hyperparameter optimization. An uncertainty propagation modeling framework is proposed for automatic predictors, such as deep learning models. The great potential of the combination of the statistical model and the DL model has been shown in the experiments. Future research on interactive training between the surrogate model and the automatic RUL predictor should be proposed and realized.

Author Contributions

Conceptualization, R.W., L.Y. and Y.D.; methodology, R.W., L.Y., F.Y., R.S. and Y.D.; formal analysis, F.Y. and R.S.; writing—original draft, F.Y., R.W., L.Y., R.S. and Y.D.; writing—review and editing, R.W., L.Y., Y.D. and R.S.; supervision, R.W. and L.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (Grant No. 51505277, 12074254, 71701143).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The bearings data from the PHM2012 dataset (https://hal.archives-ouvertes.fr/hal-00719503, accessed on 22 September 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ueda, M.; Wainwright, B.; Spikes, H.; Kadiric, A. The effect of friction on micropitting. Wear 2022, 488–489, 204130. [Google Scholar] [CrossRef]
  2. Ambrożkiewicz, B.; Syta, A.; Gassner, A.; Georgiadis, A.; Litak, G.; Meier, N. The influence of the radial internal clearance on the dynamic response of self-aligning ball bearings. Mech. Syst. Signal Process. 2022, 171, 108954. [Google Scholar] [CrossRef]
  3. Sahu, P.K. Grease Contamination Detection in the Rolling Element Bearing Using Deep Learning Technique. Int. J. Mech. Eng. Robot. 2022, 11, 275–280. [Google Scholar] [CrossRef]
  4. Xie, Z.; Jiao, J.; Yang, K.; He, T.; Chen, R.; Zhu, W. Experimental and numerical exploration on the nonlinear dynamic behaviors of a novel bearing lubricated by low viscosity lubricant. Mech. Syst. Signal Process. 2023, 182, 109349. [Google Scholar] [CrossRef]
  5. Lei, Y.; Li, N.; Guo, L.; Li, N.; Yan, T.; Lin, J. Machinery health prognostics: A systematic review from data acquisition to RUL prediction. Mech. Syst. Signal Process. 2018, 104, 799–834. [Google Scholar] [CrossRef]
  6. Zhu, J.; Chen, N.; Shen, C. A new data-driven transferable remaining useful life prediction approach for bearing under different working conditions. Mech. Syst. Signal Process. 2020, 139, 106602. [Google Scholar] [CrossRef]
  7. Zhang, Z.X.; Si, X.S.; Hu, C.; Lei, Y. Degradation Data Analysis and Remaining Useful Life Estimation: A Review on Wiener-Process-Based Methods. Eur. J. Oper. Res. 2018, 271, 775–796. [Google Scholar] [CrossRef]
  8. Liao, G.; Yin, H.; Chen, M.; Lin, Z. Remaining useful life prediction for multi-phase deteriorating process based on Wiener process. Reliab. Eng. Syst. Saf. 2021, 207, 107361. [Google Scholar] [CrossRef]
  9. Wang, H.; Liao, H.; Ma, X.; Bao, R. Remaining Useful Life Prediction and Optimal Maintenance Time Determination for a Single Unit Using Isotonic Regression and Gamma Process Model. Reliab. Eng. Syst. Saf. 2021, 210, 107504. [Google Scholar] [CrossRef]
  10. Lin, C.P.; Ling, M.H.; Cabrera, J.; Yang, F.; Yu, D.Y.W.; Tsui, K.L. Prognostics for lithium-ion batteries using a two-phase gamma degradation process model. Reliab. Eng. Syst. Saf. 2021, 214, 107797. [Google Scholar] [CrossRef]
  11. Du, W.; Hou, X.; Wang, H. Time-Varying Degradation Model for Remaining Useful Life Prediction of Rolling Bearings under Variable Rotational Speed. Appl. Sci. 2022, 12, 4044. [Google Scholar] [CrossRef]
  12. Song, K.; Cui, L. A common random effect induced bivariate gamma degradation process with application to remaining useful life prediction. Reliab. Eng. Syst. Saf. 2022, 219, 108200. [Google Scholar] [CrossRef]
  13. Peng, Y.; Wang, Y.; Zi, Y. Switching State-Space Degradation Model With Recursive Filter/Smoother for Prognostics of Remaining Useful Life. IEEE Trans. Ind. Inform. 2019, 15, 822–832. [Google Scholar] [CrossRef]
  14. Wang, Y.; Wu, J.; Cheng, Y.; Wang, J.; Hu, K. Memory-enhanced hybrid deep learning networks for remaining useful life prognostics of mechanical equipment. Measurement 2022, 187, 110354. [Google Scholar] [CrossRef]
  15. Wang, B.; Lei, Y.; Li, N.; Li, N. A Hybrid Prognostics Approach for Estimating Remaining Useful Life of Rolling Element Bearings. IEEE Trans. Reliab. 2020, 69, 401–412. [Google Scholar] [CrossRef]
  16. Soualhi, A. Bearing Health Monitoring Based on Hilbert–Huang Transform, Support Vector Machine, and Regression. IEEE Trans. Instrum. Meas. 2014, 64, 52–62. [Google Scholar] [CrossRef] [Green Version]
  17. Chelmiah, E.T.; McLoone, V.I.; Kavanagh, D.F. Remaining Useful Life Estimation of Rotating Machines through Supervised Learning with Non-Linear Approaches. Appl. Sci. 2022, 12, 4136. [Google Scholar] [CrossRef]
  18. Di Maio, F.; Tsui, K.L.; Zio, E. Combining Relevance Vector Machines and exponential regression for bearing residual life estimation. Mech. Syst. Signal Process. 2012, 31, 405–427. [Google Scholar] [CrossRef]
  19. Wang, Q.; Kun, X.; Kong, X.; Huai, T. A linear mapping method for predicting accurately the RUL of rolling bearing. Measurement 2021, 176, 109127. [Google Scholar] [CrossRef]
  20. Ye, Z.; Zhang, Q.; Shao, S.; Niu, T.; Zhao, Y. Rolling Bearing Health Indicator Extraction and RUL Prediction Based on Multi-Scale Convolutional Autoencoder. Appl. Sci. 2022, 12, 5747. [Google Scholar] [CrossRef]
  21. Wang, C.; Jiang, W.; Yang, X.; Zhang, S. RUL Prediction of Rolling Bearings Based on a DCAE and CNN. Appl. Sci. 2021, 11, 1516. [Google Scholar] [CrossRef]
  22. Yoo, Y.; Baek, J.G. A Novel Image Feature for the Remaining Useful Lifetime Prediction of Bearings Based on Continuous Wavelet Transform and Convolutional Neural Network. Appl. Sci. 2018, 8, 1102. [Google Scholar] [CrossRef] [Green Version]
  23. Jiang, G.; Zhou, W.; Chen, Q.; He, Q.; Xie, P. Dual residual attention network for remaining useful life prediction of bearings. Measurement 2022, 199, 111424. [Google Scholar] [CrossRef]
  24. Yang, B.; Liu, R.; Zio, E. Remaining Useful Life Prediction Based on A Double-Convolutional Neural Network Architecture. IEEE Trans. Ind. Electron. 2019, 66, 9521–9530. [Google Scholar] [CrossRef]
  25. Chen, D.; Qin, Y.; Wang, Y.; Zhou, J. Health indicator construction by quadratic function-based deep convolutional auto-encoder and its application into bearing RUL prediction. ISA Trans. 2020, 114, 44–56. [Google Scholar] [CrossRef] [PubMed]
  26. Zhao, B.; Yuan, Q. A novel deep learning scheme for multi-condition remaining useful life prediction of rolling element bearings. J. Manuf. Syst. 2021, 61, 450–460. [Google Scholar] [CrossRef]
  27. Xiang, S.; Qin, Y.; Zhu, C.; Wang, Y.; Chen, H. LSTM networks based on attention ordered neurons for gear remaining life prediction. ISA Trans. 2020, 106, 343–354. [Google Scholar] [CrossRef] [PubMed]
  28. Ma, L.; Ding, Y.; Wang, Z.; Wang, C.; Ma, J.; Lu, C. An interpretable data augmentation scheme for machine fault diagnosis based on a sparsity-constrained generative adversarial network. Expert Syst. Appl. 2021, 182, 115234. [Google Scholar] [CrossRef]
  29. Pan, Y.; Jing, Y.; Wu, T.; Kong, X. Knowledge-based data augmentation of small samples for oil condition prediction. Reliab. Eng. Syst. Saf. 2022, 217, 108114. [Google Scholar] [CrossRef]
  30. Liu, S.; Chen, J.; He, S.; Xu, E.; Lv, H.; Zhou, Z. Intelligent fault diagnosis under small sample size conditions via Bidirectional InfoMax GAN with unsupervised representation learning. Knowl. Based Syst. 2021, 232, 107488. [Google Scholar] [CrossRef]
  31. Liu, J.; Zhang, C.; Jiang, X. Imbalanced fault diagnosis of rolling bearing using improved MsR-GAN and feature enhancement-driven CapsNet. Mech. Syst. Signal Process. 2022, 168, 108664. [Google Scholar] [CrossRef]
  32. Behera, S.; Misra, R. Generative adversarial networks based remaining useful life estimation for IIoT. Comput. Electr. Eng. 2021, 92, 107195. [Google Scholar] [CrossRef]
  33. Liu, K.; Shang, Y.; Ouyang, Q.; Widanage, W.D. A Data-Driven Approach With Uncertainty Quantification for Predicting Future Capacities and Remaining Useful Life of Lithium-ion Battery. IEEE Trans. Ind. Electron. 2021, 68, 3170–3180. [Google Scholar] [CrossRef]
  34. Wang, B.; Lei, Y.; Yan, T.; Li, N.; Guo, L. Recurrent convolutional neural network: A new framework for remaining useful life prediction of machinery. Neurocomputing 2020, 379, 117–129. [Google Scholar] [CrossRef]
  35. Deng, Y.; Bucchianico, A.D.; Pechenizkiy, M. Controlling the accuracy and uncertainty trade-off in RUL prediction with a surrogate Wiener propagation model. Reliab. Eng. Syst. Saf. 2020, 196, 106727. [Google Scholar] [CrossRef]
  36. Chen, N.; Tsui, K.L. Condition Monitoring and Remaining Useful Life Prediction Using Degradation Signals: Revisited. TIE Trans. 2012, 45, 939–952. [Google Scholar] [CrossRef]
  37. Cerrada, M.; Sánchez, R.V.; Li, C.; Pacheco, F.; Cabrera, D.; Valente de Oliveira, J.; Vásquez, R.E. A review on data-driven fault severity assessment in rolling bearings. Mech. Syst. Signal Process. 2018, 99, 169–196. [Google Scholar] [CrossRef]
  38. Liao, L. Discovering Prognostic Features Using Genetic Programming in Remaining Useful Life Prediction. IEEE Trans. Ind. Electron. 2014, 61, 2464–2472. [Google Scholar] [CrossRef]
  39. Jeanblanc, M.; Yor, M.; Chesney, M. Mathematical Methods for Financial Markets. Finance 2010, 31, 81. [Google Scholar] [CrossRef]
  40. Fei, Z.; Wu, Z.; Xiao, Y.; Ma, J.; He, W. A new short-arc fitting method with high precision using Adam optimization algorithm. Optik 2020, 212, 164788. [Google Scholar] [CrossRef]
  41. Ren, X.; Liu, S.; Yu, X.; Dong, X. A method for state-of-charge estimation of lithium-ion batteries based on PSO-LSTM. Energy 2021, 234, 121236. [Google Scholar] [CrossRef]
  42. Nectoux, P.; Gouriveau, R.; Medjaher, K.; Ramasso, E.; Chebel-Morello, B.; Zerhouni, N.; Varnier, C. PRONOSTIA: An experimental platform for bearings accelerated degradation tests. In Proceedings of the IEEE International Conference on Prognostics and Health Management, PHM’12, Denver, CO, USA, 18 June 2012; pp. 1–8. [Google Scholar]
Figure 1. The flowchart of the proposed method. Three stages can be divided in the method. The generated data in the data augmentation process are sent to the training process of the Wiener–LSTM model. What is more, the hyperparameters of the Wiener–LSTM model are optimized by PSO algorithm. At the end, the testing data are sent to the trained Wiener–LSTM model to verify the effectiveness of the proposed method.
Figure 1. The flowchart of the proposed method. Three stages can be divided in the method. The generated data in the data augmentation process are sent to the training process of the Wiener–LSTM model. What is more, the hyperparameters of the Wiener–LSTM model are optimized by PSO algorithm. At the end, the testing data are sent to the trained Wiener–LSTM model to verify the effectiveness of the proposed method.
Applsci 12 11086 g001
Figure 3. The structure of the LSTM network cell. The cell consists of three kinds of gates units called forget date, input gate, and output gate. f t is the output of the forget gate, i t is the output of the input gate, O t is the output of the output gate. x t is the input of the LSTM cell at time t, h t is the output of the LSTM cell. c t is the state of the LSTM at time t.
Figure 3. The structure of the LSTM network cell. The cell consists of three kinds of gates units called forget date, input gate, and output gate. f t is the output of the forget gate, i t is the output of the input gate, O t is the output of the output gate. x t is the input of the LSTM cell at time t, h t is the output of the LSTM cell. c t is the state of the LSTM at time t.
Applsci 12 11086 g003
Figure 5. The PRONOSTIA platform for bearing accelerated degradation tests. The rotating part is driven by an AC motor. Acceleration sensors are placed on the tested bearing both on the horizontal axis and vertical axis. Cylinder pressure is placed to accelerate the degradation of the tested bearing.
Figure 5. The PRONOSTIA platform for bearing accelerated degradation tests. The rotating part is driven by an AC motor. Acceleration sensors are placed on the tested bearing both on the horizontal axis and vertical axis. Cylinder pressure is placed to accelerate the degradation of the tested bearing.
Applsci 12 11086 g005
Figure 6. Time domain waveform and the Failure Occurrence Time (FOT) of the raw signal. The middle one is the time domain waveform of test bearing1_3. The signal to the left of FOT is the healthy bearing signal, and the one to the right of FOT is the fault signal.
Figure 6. Time domain waveform and the Failure Occurrence Time (FOT) of the raw signal. The middle one is the time domain waveform of test bearing1_3. The signal to the left of FOT is the healthy bearing signal, and the one to the right of FOT is the fault signal.
Applsci 12 11086 g006
Figure 7. Ten examples of the generated data by the exponential model. Different color curves represent different run-to-failure data. The number 1 , 2 , 3 , , 10 denotes the generated run to failure data of bearing 1 to 10. The horizontal axis shows the inspection time. The vertical axis shows the value of RMS.
Figure 7. Ten examples of the generated data by the exponential model. Different color curves represent different run-to-failure data. The number 1 , 2 , 3 , , 10 denotes the generated run to failure data of bearing 1 to 10. The horizontal axis shows the inspection time. The vertical axis shows the value of RMS.
Applsci 12 11086 g007
Figure 8. The loss function curve of the Wiener–LSTM model training process. Three stages can be divided among the epoch of the loss function, which include the rapid the decline stage, and the slow decline stage, convenience stage.
Figure 8. The loss function curve of the Wiener–LSTM model training process. Three stages can be divided among the epoch of the loss function, which include the rapid the decline stage, and the slow decline stage, convenience stage.
Applsci 12 11086 g008
Figure 9. The comparison between Wiener–LSTM model with data augmentation and without data augmentation. (a) The predicted results on test bearing1_3. (b) The predicted results on test bearing1_4. The blue line is the real RUL from 1000–0 s. The red line is the predicted result by the Wiener–LSTM model without data augmentation. The orange line is the predicted result by the Wiener–LSTM model with data augmentation.
Figure 9. The comparison between Wiener–LSTM model with data augmentation and without data augmentation. (a) The predicted results on test bearing1_3. (b) The predicted results on test bearing1_4. The blue line is the real RUL from 1000–0 s. The red line is the predicted result by the Wiener–LSTM model without data augmentation. The orange line is the predicted result by the Wiener–LSTM model with data augmentation.
Applsci 12 11086 g009
Figure 10. The loss function curve of the LSTM model training process. Two stages can be divided among the epoch of the loss function, which include the rapid decline stage and the convergence stage.
Figure 10. The loss function curve of the LSTM model training process. Two stages can be divided among the epoch of the loss function, which include the rapid decline stage and the convergence stage.
Applsci 12 11086 g010
Figure 11. The predicted RUL by the Wiener–LSTM model and LSTM model. (a) The predicted results on test bearing1_3. (b) The predicted results on test bearing1_4. The blue line is the real RUL from 1000–0 s. The yellow line is the predicted result by the normal LSTM model using the MSE loss function. The orange line is the predicted result by the Wiener–LSTM model using the proposed loss function.
Figure 11. The predicted RUL by the Wiener–LSTM model and LSTM model. (a) The predicted results on test bearing1_3. (b) The predicted results on test bearing1_4. The blue line is the real RUL from 1000–0 s. The yellow line is the predicted result by the normal LSTM model using the MSE loss function. The orange line is the predicted result by the Wiener–LSTM model using the proposed loss function.
Applsci 12 11086 g011
Figure 12. The MAE of predicted RUL results by the Wiener–LSTM model and LSTM model. (a) The MAE of predicted results on test bearing1_3. (b) The MAE of predicted results on test bearing1_4. The blue line is the MAE of results by using the Wiener–LSTM model. The orange line is the MAE of results by using the LSTM model.
Figure 12. The MAE of predicted RUL results by the Wiener–LSTM model and LSTM model. (a) The MAE of predicted results on test bearing1_3. (b) The MAE of predicted results on test bearing1_4. The blue line is the MAE of results by using the Wiener–LSTM model. The orange line is the MAE of results by using the LSTM model.
Applsci 12 11086 g012
Table 1. Datasets of IEEE PHM 2012 prognostic challenge.
Table 1. Datasets of IEEE PHM 2012 prognostic challenge.
Data SetOperation Conditions
Conditions 1Conditions 2Conditions 3
Load (N)400042005000
Speed (rpm)480016501500
Training setBearing 1 _ 1Bearing 2 _ 1Bearing 3 _ 1
Bearing 1 _ 2Bearing 2 _ 2Bearing 3 _ 2
Testing setBearing 1 _ 3Bearing 2 _ 3Bearing 3 _ 3
Bearing 1 _ 4Bearing 2 _ 4
Bearing 1 _ 5Bearing 2 _ 5
Bearing 1 _ 6Bearing 2 _ 6
Bearing 1 _ 7Bearing 2 _ 7
Table 2. Numerical prognostic performance comparisons of different methods.
Table 2. Numerical prognostic performance comparisons of different methods.
ProposedLSTM
Test BearingMAEMAE
Bearing1_36.3443.6
Bearing1_48.0323.8
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, R.; Yan, F.; Shi, R.; Yu, L.; Deng, Y. Uncertainty-Controlled Remaining Useful Life Prediction of Bearings with a New Data-Augmentation Strategy. Appl. Sci. 2022, 12, 11086. https://doi.org/10.3390/app122111086

AMA Style

Wang R, Yan F, Shi R, Yu L, Deng Y. Uncertainty-Controlled Remaining Useful Life Prediction of Bearings with a New Data-Augmentation Strategy. Applied Sciences. 2022; 12(21):11086. https://doi.org/10.3390/app122111086

Chicago/Turabian Style

Wang, Ran, Fucheng Yan, Ruyu Shi, Liang Yu, and Yingjun Deng. 2022. "Uncertainty-Controlled Remaining Useful Life Prediction of Bearings with a New Data-Augmentation Strategy" Applied Sciences 12, no. 21: 11086. https://doi.org/10.3390/app122111086

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop