Next Article in Journal
EggBlock: Design and Implementation of Solar Energy Generation and Trading Platform in Edge-Based IoT Systems with Blockchain
Next Article in Special Issue
Compression Reconstruction and Fault Diagnosis of Diesel Engine Vibration Signal Based on Optimizing Block Sparse Bayesian Learning
Previous Article in Journal
Heuristic Greedy Scheduling of Electric Vehicles in Vehicle-to-Grid Microgrid Owned Aggregators
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Performance Degradation Prediction Using LSTM with Optimized Parameters

1
College of Electrical Engineering and Automation, Anhui University, Hefei 230601, China
2
Anhui NARI Jiyuan Electric Co., Ltd., Hefei 230601, China
3
China North Vehicle Research Institute, Beijing 100071, China
4
JJL College of Engineering, South Dakota State University, Brookings, SD 57007, USA
*
Authors to whom correspondence should be addressed.
Sensors 2022, 22(6), 2407; https://doi.org/10.3390/s22062407
Submission received: 5 February 2022 / Revised: 1 March 2022 / Accepted: 10 March 2022 / Published: 21 March 2022
(This article belongs to the Special Issue Machine Health Monitoring and Fault Diagnosis Techniques)

Abstract

:
Predicting the degradation of mechanical components, such as rolling bearings is critical to the proper monitoring of the condition of mechanical equipment. A new method, based on a long short-term memory network (LSTM) algorithm, has been developed to improve the accuracy of degradation prediction. The model parameters are optimized via improved particle swarm optimization (IPSO). Regarding how this applies to the rolling bearings, firstly, multi-dimension feature parameters are extracted from the bearing’s vibration signals and fused into responsive features by using the kernel joint approximate diagonalization of eigen-matrices (KJADE) method. Then, the between-class and within-class scatter (SS) are calculated to develop performance degradation indicators. Since network model parameters influence the predictive accuracy of the LSTM model, an IPSO algorithm is used to obtain the optimal prediction model via the LSTM model parameters’ optimization. Finally, the LSTM model, with said optimal parameters, was used to predict the degradation trend of the bearing’s performance. The experiment’s results show that the proposed method can effectively identify the trends of degradation and performance. Moreover, the predictive accuracy of this proposed method is greater than that of the extreme learning machine (ELM) and support vector regression (SVR), which are the algorithms conventionally used in degradation modeling.

1. Introduction

Studies have shown that more than 45% of equipment failures in rotating machinery are caused by bearing failure [1]. The financial losses and major safety accidents that this causes in the industry can be avoided by assessing the deterioration status of equipment, which would bolster an organization’s ability to detect faults in machine bearings [2].
According to previous studies on the topic, data-driven modeling has, over time, gradually become the most effective forecasting method [3,4]. In order to predict the remaining useful life (RUL) of bearings, a large number of studies, focusing on data-driven modeling, have been carried out, including the support vector machine (SVM) and artificial neural network (ANN) [5,6]. Zheng et al. proposed the ensemble SVM for the fault detection and diagnosis of rolling bearings, in which composite multiscale fuzzy entropy was used to discern health indicators [7]. However, traditional machine learning methods like SVM require a priori knowledge of feature engineering, which is extremely difficult to implement with regard to bearings due to the complex working conditions they operate under. Deep learning-based algorithms provide an alternative solution to this problem [8,9,10]. Chen et al. proposed a method based on neuro-fuzzy systems (NFSs) and Bayesian algorithms, which use trained NFSs as predictors to discern the degradation of a given machine’s fault state over time [11]. Ren et al. proposed the use of the spectral principal energy vector method in obtaining bearing signal feature vectors. They adopt the deep convolutional neural network to predict the RUL of rolling bearings [12]. The recurrent neural network (RNN), as an important subfield of deep learning, performs well with regarding time series processing because the RNN can forecast using all available historical data [13]. Malhi et al. made further strides towards putting forth a long-term prediction model for machine condition monitoring based on RNN [14]. However, the disappearance, or even the explosion, of gradients during network training seems to restrict this method’s applicability [15].
To solve the issue, Hochreiter and Schmidhuber presented a variant of the RNN network in 1997, namely, the long short-term memory (LSTM) neural network, which addressed the problem by adding a gating mechanism [16]. At present, LSTMs are widely used in a variety of different fields, such as speech recognition, time series modeling, video classification, traffic flow prediction, and so on. Besides this, the LSTM method has also been used to predict bearing degradation, exploring the correlation between bearing degradation data and time [17]. Liu et al. proposed the use of an end-to-end model in predicting the degradation trends of bearings. His model used CNN for data reduction and feature extraction and a LSTM for time series processing [18]. Elsheikh et al. proposed bidirectional rocking long-term and short-term memory to predict the RUL of turbofan engines [19]. Tang et al. used a stacked automatic encoder (SAE) to obtain the bottleneck characteristics of bearing signals and predicted bearing performance degradation with a LSTM [20].
In application, the choice of the network structure, the number of hidden layers, and the learning rate setting will significantly influence the predictive capability of LSTMs [21]. Typically, the complex structure and parameters of LSTM neurons are mostly determined by experience, or by multiple parameter adjustments with expensive time, which involves a lot of randomness and reduces the model’s predictive capability. Therefore, a set of hyper-parametric optimization algorithms were developed to select the parameters automatically. The particle swarm optimization (PSO) algorithm is commonly used for model parameter optimization in the field of bearing performance degradation assessment [8,22,23]. However, the traditional PSO algorithm suffers from slow convergence as well as local optimization problems, which affects the performance of the model. Hence, a modified PSO algorithm is suggested for the purpose of optimizing the LSTM model’s parameters. The modified IPSO-LSTM module was applied to predict bearing performance degradation trends.

2. Methodology

2.1. LSTM

The mechanical degradation process, for example, on a rolling bearing, is a process of accumulation and continuous fault development [24]. Its degradation is determined by assessing its currently observable state as compared with its state in the recent past. The traditional neural network only uses the most recently documented state for its model, making it difficult to characterize deterioration and performance degradation over time. The LSTM is a type of RNN. An RNN is a neural network that handles sequential data and can be used to connect information from the recent past to the current task. However, as the distance between relevant information and the information taken from the past increases, the RNN loses its ability to learn and use distant details. Multiple control gates have been designed to replace the RNN in order to solve this problem. Thus, the LSTM network is constructed [16].
The LSTM solves the problem of gradient disappearance and explosion through the use of the aforesaid gates. In the LSTM structure, f t , i t , and o t are three gates, which are designed to control the flow of information. f t controls the information of memory cells from time t-1 to time t.   i t controls the information input to the memory cells at time t, and   o t controls the information of memory cells at time t to the hidden state of h t .
f t = σ ( w f c C t 1 + w f h h t 1 + w f x x t + b f )
i t = σ ( w i c C t 1 + w i h h t 1 + w i x x t + b i )
o t = σ ( w o c C t 1 + w o h h t 1 + w o x x t + b o )
where w f c , w i h , and w o h are the weight matrix between gate f t and memory cell C t 1 . b f is the bias of the gate f t . Other weight matrices are derived from the following: C t and C t 1 represent the values of memory cells at time t and time t-1. b f ,   b i , and b o represent the bias. σ is the activation function. The hiding unit structure of the long and short-term memory network is shown in Figure 1.
The LSTM can predict degradation due to the time-varying characteristics of performance degradation and the advantages of LSTMs in modeling and forecasting time series. However, the structure of the LSTM model is complex. Some parameters need to be set synthetically, such as the time frame, the batch size, the number of hidden layer units, etc., which makes it difficult to meet the highly precise requirements for predicting time series degradation. Thus, it is necessary to find the optimal model parameters for each iteration in order to maintain strong predictive accuracy. This optimal model is realized through the use of a swarm intelligence algorithm, which auto-selects and optimizes the LSTM model’s parameters to improve the prediction.

2.2. IPSO

A particle swarm optimization (PSO) algorithm is a population intelligent optimization algorithm used to simulate birds’ foraging behavior. Kennedy and Eberhart first proposed it in 1995 [25]. A standard particle swarm optimization algorithm sets the particle swarm size as m, and each particle has an n dimension search region. x i = x i 1 , x i 2 , x i 3 , , x i n represents the search position of particle I in space.   v i = v i 1 , v i 2 , v i 3 , , v i n is the velocity of the particle, i, which represents the moving distance of the particle in each position update. p i = p i 1 , p i 2 , p i 3 , , p i n records the search optimal bit value of the particle, i. p g = p g 1 , p g 2 , p g 3 , , p g n is the optimal particle location in the current population. In a traditional PSO algorithm, the positions and velocities of particles are updated through Equations (4) and (5).
v i j ( t + 1 ) = w v i j ( t ) + c 1 R 1 ( p i j ( t ) x i j ( t ) ) + c 2 R 2 ( p g j ( t ) x i j ( t ) )
x i j ( t + 1 ) = x i j ( t ) + v i j ( t + 1 )
where w is the inertia weight factor, and the range of w is (0, 1.4); c 1 and c 2   are learning factors; R 1 and R 2 are random numbers between 0 and 1; v i j t + 1 is the j dimension velocity component of the particle, i, in the t+1 iteration; x i j t + 1 is the j dimension position component of the particle, i, in the t+1 iteration; p i j t is the j dimension optimal position component of the particle, i, in the t iteration; p g j t is the j dimension position component of the optimal solution in the population in the t iteration; 1 i , g m , 1 j n .
However, the disadvantages of a PSO algorithm include low convergence accuracy and premature results. The diversity of the population decreases while the iteration times increase, the algorithm falls into the local optimal solution and the algorithm prematurely solves an incomplete problem. To solve this problem, the position updating model and parameter adjustment strategy of the particle swarm are modified.
  • Parameter adjustment strategy
The inertia weight, w , has a significant influence on the performance of particle swarm optimization. In the early stages, a strong search ability is needed to search for the best information quickly. And in the late stages, a fine selection is required to search for accuracy.
w t = w max t * ( w max w min ) / t max
where w max and w min are the upper and lower limits of the preset inertia weight, and in general, w max = 0.9 and w min = 0.4 , respectively; t max is the maximum number of iterations.
In the traditional PSO algorithm, c 1 and c 2 are fixed values. The improved c 1 and c 2 can adaptively adjust learning factors and inertia weight. To find the optimal solution, independent and team learning abilities are adjusted in different search times.
c 1 = c max + ( c max c min ) ( 1 ( e w 1 ) 2 )
c 2 = c max ( c max c min ) ( 1 ( e w 1 ) 2 )
2.
Particle swarm position updating model [26]
In the early stages of the searching process, particles have a strong self-learning ability, and the search iterative step size should be set to large. With the search time increasing, the space range of solutions becomes smaller. To search for accurate solutions, the search iteration step size of particles should be reduced accordingly. Therefore, an adaptive adjustment factor, μ , is added to the particle position updating algorithm.
μ = 1 / ( 1 + e t / t max ) + 1 / 2
where t is the iterations.
The improved particle position update formula is then as follows.

2.3. IPSO-LSTM

As can be seen in Section 2.1, due to the advantage of processing time sequences, a two-layer LSTM is used as the backbone network for the high dimensional degradation feature extraction in this paper. The hidden state of each time step in the first layer is retained to serve as the input of the second layer, which only returns the hidden state of the last time step. To avoid model overfitting, a dropout regularization strategy is employed after each LSTM layer. Then, the learned representation features are fed into the fully connected layer to be mapped into a one-dimensional degeneration metric.
First, the hyper-parameters that need to be determined for the LSTM’s backbone network include the number of hidden nodes in the first second layer. The hidden layers play a vital role in extracting high-dimensional features and internal laws. The model’s performance is affected mainly by the number of hidden nodes. Too many nodes will increase the training time and may lead to overfitting. Too few will reduce the model’s learning ability to the extent that the sparse adequate information extracted will not suffice in solving the problem. Therefore, the model structure’s complexity and predictive accuracy should be considered comprehensively in selecting the number of nodes when designing the network.
In addition, most neural networks are usually optimized by a gradient descent algorithm. The gradient descent is calculated as follows:
g = 1 m θ ˜ i L ( f ( x ( i ) ; θ ˜ ) , y ( i ) )
where m is the batch size; y i is the target corresponding to m; θ is the updated parameter; f is the random target function with the parameter, θ .
As seen from Equation (10), increasing the batch size reduces the gradient and makes the gradient more accurate. This indicates that t stability of the convergence is enhanced by increasing batch size in the correct range.
As described above, it is clear that the three hyper-parameters, namely the number of hidden nodes in the first LSTM layer, the number of hidden nodes in the second LSTM layer, and batch size, are the key factors affecting the performance of the model. The specific representations and ranges are shown in Table 1. In this paper, the IPSO algorithm is used to optimize and automatically select the parameters of the LSTM model.
h 1 h 2 The flowchart of LSTM parameters optimized by IPSO is shown in Figure 2. The steps are as follows:
  • Initialize the parameters. Determine the population size range, iteration times, learning factors, location, and velocity;
  • Initialize the position and velocity of the particles. Generate the population particles X i , 0 h 1 , h 2 , S m randomly. Where h 1 and h 2 denote the number of neurons in the first and second hidden layer, respectively, and S m represents the batch size;
  • Determine the evaluative function of the particles. The particle   X i , 0 , in step 2 above, is assigned to the LSTM parameter. The data are partitioned into the training samples, validation samples, and test samples. The fitness value,   f i t , of individual X i is defined as the target function, which is set as:
    f i t = 1 n i = 1 n ( y ^ i y i ) 2
    where y ^ i is the predicted value; y i is the actual observation;
  • Calculate the fitness value of each particle position, X i . Individual extreme value and the population extreme value are determined according to the initial particle’s fitness value, and each particle’s best position is taken as its historical best position;
  • Update the velocity and position of the particle;
  • Determine whether the end condition of the iteration has been met. If it has, output the optimal parameter; Otherwise, go to step 4 to continue the iteration.
Figure 2. Parameter optimization flowchart.
Figure 2. Parameter optimization flowchart.
Sensors 22 02407 g002

3. Results IPSO-LSTM Method for Bearing Performance Degradation Prediction

In most practical industrial applications, the actual working conditions of mechanical equipment are complex and dynamic. Bearing vibration signals collected by sensors contain rich information. A single feature cannot fully describe the state of bearing vibration signals. The kernel joint approximate diagonalization of eigen-matrices (KJADE) is to map the observation data X = x 1 , x 2 , , x m to a high-dimensional feature space F through a nonlinear function Φ , and the mapped feature space is F = Φ x 1 , Φ x 2 , , Φ x M . The inner product of two vectors in the feature space is calculated following the kernel function, and an m × m kernel matrix K is established as follows:
K i j = k x i , x j = Φ ( x i ) Φ ( x j )
where x i and x j are the sample vectors. Therefore, the KJADE algorithm is employed to characterize the bearing degradation state.
The step flow chart of the method is shown in Figure 3. The operations are described as follows:
  • Original feature extraction. The full life vibration signal of bearing is analyzed in both the time and frequency domains to avoid the insufficiency of single feature evaluation ability. Eight features in time-domain and frequency-domain are extracted to form a high-dimensional feature vector, as shown in Table 2. T1–T8 are the mean value, root mean square (RMS), absolute average, skewness, waveform index, impulsion index, and kurtosis index, respectively. Among others are frequency domain features, where s i is a spectrum for i = 1 , 2 , , N (N is the number of spectrum lines) and f i is the frequency value of the i-th spectrum line, indicating the degree of dispersion or concentration of the spectrum and the change of the dominant frequency band;
  • KJADE features fusion. Considering the nonlinear characteristics of bearing vibration, the redundancy of the original multi-domain degradation characteristics, and some features that are not sensitive to the bearing degradation state, it is necessary to fuse multi-domain features. Therefore, the KJADE algorithm based on kernel function is employed to extract more effective, but low-dimensional, degradation characteristic indexes [27]. KJADE has better nonlinear processing capabilities for bearing vibration signals. It maps the observation data to a high-dimensional feature space through a nonlinear function. Then the JADE can be used in this feature space to change the nonlinear separable problem into a linear one;
  • Degradation assessment index calculation. The vibration signal collected at the beginning of the bearing operation is taken as the health state, corresponding to the health sample after dimension reduction. The subsequent signal is selected as the monitoring sample. To quantify the dispersion between monitoring and health samples and the aggregation between different classes, the evaluation factor, SS, composed of between- and within-class scatter matrix, is used as the performance degradation index [12];
  • IPSO-LSTM model construction. The number of hidden layer nodes and the batch size of the LSTM neural network are taken as optimization objects. The LSTM is constructed according to the corresponding parameters of each particle. The IPSO algorithm is used to acquire the optimal hyper-parameter set for each iteration automatically;
  • Predicting the performance degradation. The LSTM model is constructed with the optimal value of hyper-parameters, and the bearing data is used as input for training and prediction.
Figure 3. Performance degradation prediction by IPSO.
Figure 3. Performance degradation prediction by IPSO.
Sensors 22 02407 g003

4. Case Analysis

4.1. Case 1

The Intelligent Maintenance System (IMS) Center of the University of Cincinnati’s full-life vibration signals of bearings are used to confirm the proposed method [28]. The experimental platform is shown in Figure 4.
The bearing type is ZA-2115, and the experimental conditions were as follows: output speed was 2000 rpm, the radial load was 6000 lbs, and the sampling frequency was 20,480 Hz. A total of 984 sets of vibration signal data were recorded. The whole experiment was completed in three groups. By the end of the experiment, an inner fault in bearing 3 and a rolling fault in bearing 4 were observed in the first group. An outer fault in bearing 1 in the second group and an outer fault in bearing 3 in the third group were also observed. Among them, the rolling fault and inner fault in the first group, along with the outer fault in the second group, were selected as objects for analysis. The corresponding vibration data of life is shown in Figure 5.
Based on the method in Section 2, the IPSO algorithm is used to optimize the LSTM model’s predictive parameters. The initial parameters of the IPSO are as follows: the number of particles is 10, the dimension of particle swarm is 3, the maximum velocity of the particle is 1, and the maximum iteration number is 50. The range of particle locations, namely the number of hidden layer nodes, is set to (100, 300), and the batch size is (30, 200). The upper and lower limits of the inertia weight are w_max = 0.9 and w_min = 0.5, while the upper and lower limits of the initial learning factors c_max and c_min are 2 and 1, respectively. These are the optimal parameters obtained by comparative experiments. In this study, the first 60% of the performance data is used as the training set, and 20% of the rest is saved as a validation set. Besides this, the model is optimized by an Adam algorithm, and the root mean square error (RMSE) is applied as the target criteria.
To demonstrate the superiority of the proposed method, the performance of conventional LSTMs and PSO-LSTMs have been compared. The resulting real degradation trends, which can be expressed as a degradation index, are obtained via feature fusion using the KJADE algorithm. Additionally, the comparison results of the degradation trends predicted by each model are shown in Figure 6, Figure 7 and Figure 8, where the y-axis is the degradation index. In addition, RMSE is used as an additional metric to measure the performance of the model, with the results shown in Table 3. The RMSE calculation is shown in Equation (13).
RMSE = 1 n i = 1 n y ^ i y i 2
where y ^ i is the predicted value; y i is the actual observation; n is the total number of samples in the faulty bearing.
From Figure 6, Figure 7 and Figure 8, it can be seen that our proposed IPSO-LSTM method tracks the degenerate states significantly better than the other two methods in all three failure modes, especially the LSTM method without the hyper-parameter optimization process. In terms of quantitative metrics, the RMSE results in Table 3 also illustrate the superiority of the proposed method.
The above results show that the IPSO algorithm is effective in optimizing the hyper-parameters of the LSTM based network, which can automatically and accurately search for the optimal parameters. To further illustrate the advantages of the IPSO algorithm in optimizing speed and avoiding local extremum, we visualize the parameter search processes, which are shown in Figure 9.
Overall, the convergence speed and fitness of the IPSO algorithm are better than the traditional PSO algorithm. Specifically, as Figure 9b,c demonstrate, IPSO has good optimization ability and can quickly find the optimal global point. Compared with the PSO, the IPSO algorithm has a faster convergence speed. Figure 9a shows that although the final fitness error is the same, the IPSO algorithm converge is faster.
Furthermore, extreme learning machines (ELM) and support vector regression (SVR), which have been widely used with good performance degradation prediction [29,30], are compared with the proposed IPSO-LSTM for effectiveness. The comparison results are shown in Figure 10, Figure 11 and Figure 12.
The results show that the prediction results of the IPSO-LSTM method are more in line with the original curve, with greater predictive accuracy. This is demonstrated in the RMSE values in Table 4. Predictive errors in the proposed method are minimal, which verifies the effectiveness of the proposed IPSO-LSTM method.

4.2. Case 2

The lab experiments used four HRB6305 bearings. They were fixed on the same shaft and connected with the motor. A radial load of 750 kg was applied to all bearings to accelerate the bearing damage process, and the bearing speed was 3000 rpm. Full-life vibration signals were obtained by the NI PXI acquisition system. The vibration signals acquisition frequency was 20 kHz, the data were collected every 5 min. The experimental platform is shown in Figure 13.
The fault in the rolling element is taken as the experimental object. Figure 14 shows the full-life original vibration signal of the rolling element. The mixed-domain features are extracted from the bearing data. KJADE is used for feature fusion to acquire an optimal feature parameter set, and the SS is calculated from fusion features to obtain the degradation index. The proposed method is used to predict the performance degradation and compared with the LSTM and PSO-LSTM methods. The prediction curve is shown in Figure 15.
The results demonstrate that the predictive accuracy of the proposed method is greater than that of the other two methods. The RMSE results of LSTM, PSO-LSTM, and IPSO-LSTM are shown in Table 5. The iteration results of IPSO and PSO optimization are shown in Figure 16. It demonstrates that the IPSO algorithm converges earlier and is less likely to succumb to the local minimum problem, which is an advantage over the performance of the PSO.
Similar to case 1, extreme learning machines (ELM) and support vector regression (SVR) are compared with the proposed method.
The results of the comparison are shown in Figure 17 and Table 6. It can be seen that the proposed method is more effective than the other two methods in predicting the degradation trend of bearings. The RMSE values also reflect that the proposed IPSO-LSMT’s predictive accuracy is higher than the ELM and SVR methods.

5. Conclusions

This paper proposes a method based on an improved PSO optimized LSTM (IPSO-LSTM) to analyse bearing performance degradation. The proposed method can effectively resolve the problem of online parameter selection and the low predictive accuracy of the LSTM method. The KJADE method is used to fuse the bearing vibration signal to form an effective feature vector, and SS is calculated to acquire a performance degradation index. Then, the improved PSO algorithm is used to optimize the LSTM parameters to obtain an optimal performance degradation prediction model. In this study, the proposed method is compared with the LSTM, PSO-LSTM, ELM, and SVR through lab experiments. The experiments’ results have verified the effectiveness and superiority of the proposed method over others. This method has good prospective applications in predicting bearing performance degradation, and it can also be tailored and applied to other mechanical systems for online health and prognosis management.

Author Contributions

Conceptualization, Y.H. and Y.L.; methodology, Y.Y. and R.W.; software, R.W.; validation, X.L., Z.H. and Y.H.; formal analysis, Y.L.; investigation, C.H. and R.W.; resources, Y.Y.; data curation, Z.H.; writing—original draft preparation, Y.H. and R.W.; writing—review and editing, H.L.; visualization, H.L.; supervision, Y.L.; project administration, Y.Y.; funding acquisition, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China [52075001, 52105082, 52105040, 52075002]; and the Key Basic Research Project [MKF20210008].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All data generated, or that appeared in this study, are available upon request by contact with the corresponding author. Furthermore, the models and codes used during the study cannot be shared at this time as the data also forms part of an ongoing study.

Acknowledgments

The authors are grateful to the editors and anonymous reviewers for their helpful comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Rai, A.; Upadhyay, S.H. The use of MD-CUMSUM and NARX neural network for anticipating the remaining useful life of bearings. Measurement 2017, 111, 397–410. [Google Scholar] [CrossRef]
  2. Nohal, L.; Vaculka, M.; Iop. Experimental and computational evaluation of rolling bearing steel durability. In Proceedings of the 4th International Conference Recent Trends in Structural Materials (COMAT), Pilsen, Czech Republic, 9–11 November 2017. [Google Scholar]
  3. Saravanan, N.; Ramachandran, K.I. Incipient gear box fault diagnosis using discrete wavelet transform (DWT) for feature extraction and classification using artificial neural network (ANN). Expert Syst. Appl. 2010, 37, 4168–4181. [Google Scholar] [CrossRef]
  4. Zhang, J.; Sun, Y.; Guo, L.; Gao, H.; Hong, X.; Song, H. A new bearing fault diagnosis method based on modified convolutional neural networks. Chin. J. Aeronaut. 2020, 33, 439–447. [Google Scholar] [CrossRef]
  5. Zhu, J.; Chen, N.; Peng, W. Estimation of Bearing Remaining Useful Life Based on Multiscale Convolutional Neural Network. Ieee Trans. Ind. Electron. 2019, 66, 3208–3216. [Google Scholar] [CrossRef]
  6. Tian, Z. An artificial neural network method for remaining useful life prediction of equipment subject to condition monitoring. J. Intell. Manuf. 2012, 23, 227–237. [Google Scholar] [CrossRef]
  7. Zheng, J.; Pan, H.; Cheng, J. Rolling bearing fault detection and diagnosis based on composite multiscale fuzzy entropy and ensemble support vector machines. Mech. Syst. Signal Processing 2017, 85, 746–759. [Google Scholar] [CrossRef]
  8. Zhang, B.; Zhang, S.; Li, W. Bearing performance degradation assessment using long short-term memory recurrent network. Comput. Ind. 2019, 106, 14–29. [Google Scholar] [CrossRef]
  9. Li, X.; Jiang, H.; Wang, R.; Niu, M. Rolling bearing fault diagnosis using optimal ensemble deep transfer network. Knowl.-Based Syst. 2021, 213, 106695. [Google Scholar] [CrossRef]
  10. Hu, M.; Wang, G.; Ma, K.; Cao, Z.; Yang, S. Bearing performance degradation assessment based on optimized EWT and CNN. Measurement 2021, 172, 108868. [Google Scholar] [CrossRef]
  11. Chen, C.; Zhang, B.; Vachtsevanos, G. Prediction of Machine Health Condition Using Neuro-Fuzzy and Bayesian Algorithms. Ieee Trans. Instrum. Meas. 2012, 61, 297–306. [Google Scholar] [CrossRef]
  12. Ren, L.; Sun, Y.Q.; Wang, H.; Zhang, L. Prediction of Bearing Remaining Useful Life With Deep Convolution Neural Network. IEEE Access 2018, 6, 13041–13049. [Google Scholar] [CrossRef]
  13. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  14. Malhi, A.; Yan, R.Q.; Gao, R.X. Prognosis of Defect Propagation Based on Recurrent Neural Networks. IEEE Trans. Instrum. Meas. 2011, 60, 703–711. [Google Scholar] [CrossRef]
  15. Zhao, R.; Wang, D.Z.; Yan, R.Q.; Mao, K.Z.; Shen, F.; Wang, J.J. Machine Health Monitoring Using Local Feature-Based Gated Recurrent Unit Networks. IEEE Trans. Ind. Electron. 2018, 65, 1539–1548. [Google Scholar] [CrossRef]
  16. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  17. Ma, M.; Mao, Z. Deep-Convolution-Based LSTM Network for Remaining Useful Life Prediction. IEEE Trans. Ind. Inform. 2021, 17, 1658–1667. [Google Scholar] [CrossRef]
  18. Liu, G.; Zhao, J.; Zhang, X. Bearing degradation trend prediction under different operational conditions based on CNN-LSTM. IOP Conf. Ser. Mater. Sci. Eng. 2019, 612, 032042. [Google Scholar] [CrossRef]
  19. Elsheikh, A.; Yacout, S.; Ouali, M.S. Bidirectional handshaking LSTM for remaining useful life prediction. Neurocomputing 2019, 323, 148–156. [Google Scholar] [CrossRef]
  20. Tang, G.; Zhou, Y.; Wang, H.; Li, G. Prediction of bearing performance degradation with bottleneck feature based on LSTM network. In Proceedings of the 2018 IEEE International Instrumentation and Measurement Technology Conference, I2MTC 2018, Houston, TX, USA,, 14–17 May 2018; pp. 1–6. [Google Scholar]
  21. Zhao, Z.; Wu, J.; Li, T.; Sun, C.; Yan, R.; Chen, X. Challenges and Opportunities of AI-Enabled Monitoring, Diagnosis & Prognosis: A Review. Chin. J. Mech. Eng. 2021, 34, 56. [Google Scholar] [CrossRef]
  22. Ding, N.; Li, H.; Yin, Z.; Zhong, N.; Zhang, L. Journal bearing seizure degradation assessment and remaining useful life prediction based on long short-term memory neural network. Measurement 2020, 166, 108215. [Google Scholar] [CrossRef]
  23. Rathore, M.S.; Harsha, S.P. Prognostics Analysis of Rolling Bearing Based on Bi-Directional LSTM and Attention Mechanism. J. Fail. Anal. Prev. 2022, 1–20. [Google Scholar] [CrossRef]
  24. Wang, B.; Lei, Y.; Yan, T.; Li, N.; Guo, L. Recurrent convolutional neural network: A new framework for remaining useful life prediction of machinery. Neurocomputing 2020, 379, 117–129. [Google Scholar] [CrossRef]
  25. Marini, F.; Walczak, B. Particle swarm optimization (PSO). A tutorial. Chemom. Intell. Lab. Syst. 2015, 149, 153–165. [Google Scholar] [CrossRef]
  26. Kang, Y.; Jiang, C.; Qin, Y.; Ye, C. Robot Path Planning and Experiment with an Improved PSO Algorith. Robot 2020, 42, 8. [Google Scholar] [CrossRef]
  27. Liu, Y.B.; He, B.; Liu, F.; Lu, S.L.; Zhao, Y.L. Feature fusion using kernel joint approximate diagonalization of eigen-matrices for rolling bearing fault identification. J. Sound Vib. 2016, 385, 389–401. [Google Scholar] [CrossRef]
  28. Gousseau, W.; Antoni, J.; Girardin, F.; Griffaton, J. Analysis of the Rolling Element Bearing data set of the Center for Intelligent Maintenance Systems of the University of Cincinnati. In Proceedings of the CM 2016, Charenton, France, 10 October 2016. [Google Scholar]
  29. Fang, L.; Yongbin, L.; Fenglin, C.; Bing, H. Residual life prediction for ball bearings based on joint approximate diagonalization of eigen matrices and extreme learning machine. Proc. Inst. Mech. Eng. Part C J. Mech. Eng. Sci. 2017, 231, 1699–1711. [Google Scholar]
  30. Liu, F.; Li, L.; Liu, Y.; Cao, Z.; Lu, S. HKF-SVR Optimized by Krill Herd Algorithm for Coaxial Bearings Performance Degradation Prediction. Sensors 2020, 20, 660. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Structure of long short-term memory hidden unit.
Figure 1. Structure of long short-term memory hidden unit.
Sensors 22 02407 g001
Figure 4. Experimental setup.
Figure 4. Experimental setup.
Sensors 22 02407 g004
Figure 5. Exemplary diagram of bearing vibration data: (a) rolling fault; (b) inner fault; (c) outer fault.
Figure 5. Exemplary diagram of bearing vibration data: (a) rolling fault; (b) inner fault; (c) outer fault.
Sensors 22 02407 g005
Figure 6. Performance degradation predictions for outer bearings for the: (a) LSTM; (b) PSO-LSTM; (c) IPSO-LSTM.
Figure 6. Performance degradation predictions for outer bearings for the: (a) LSTM; (b) PSO-LSTM; (c) IPSO-LSTM.
Sensors 22 02407 g006
Figure 7. Performance degradation predictions of roller bearings for the: (a) LSTM; (b) PSO-LSTM; (c) IPSO-LSTM.
Figure 7. Performance degradation predictions of roller bearings for the: (a) LSTM; (b) PSO-LSTM; (c) IPSO-LSTM.
Sensors 22 02407 g007
Figure 8. Performance degradation predictions of inner bearings for the: (a) LSTM; (b) PSO-LSTM; (c) IPSO-LSTM.
Figure 8. Performance degradation predictions of inner bearings for the: (a) LSTM; (b) PSO-LSTM; (c) IPSO-LSTM.
Sensors 22 02407 g008
Figure 9. Optimization iteration results for the: (a) outer bearing; (b) roller bearing; (c) inner bearing.
Figure 9. Optimization iteration results for the: (a) outer bearing; (b) roller bearing; (c) inner bearing.
Sensors 22 02407 g009
Figure 10. Performance degradation predictions of the outer bearings for the: (a) ELM; (b) SVR; (c) IPSO-LSTM.
Figure 10. Performance degradation predictions of the outer bearings for the: (a) ELM; (b) SVR; (c) IPSO-LSTM.
Sensors 22 02407 g010
Figure 11. Performance degradation predictions of roller bearings for the: (a) ELM; (b) SVR; (c) IPSO-LSTM.
Figure 11. Performance degradation predictions of roller bearings for the: (a) ELM; (b) SVR; (c) IPSO-LSTM.
Sensors 22 02407 g011
Figure 12. Performance degradation predictions of the inner bearings for the: (a) ELM; (b) SVR; (c) IPSO-LSTM.
Figure 12. Performance degradation predictions of the inner bearings for the: (a) ELM; (b) SVR; (c) IPSO-LSTM.
Sensors 22 02407 g012
Figure 13. Experimental setup.
Figure 13. Experimental setup.
Sensors 22 02407 g013
Figure 14. The full-life original vibration signal.
Figure 14. The full-life original vibration signal.
Sensors 22 02407 g014
Figure 15. Performance degradation predictions of the roller bearings for the: (a) LSTM; (b) PSO-LSTM; (c) IPSO-LSTM.
Figure 15. Performance degradation predictions of the roller bearings for the: (a) LSTM; (b) PSO-LSTM; (c) IPSO-LSTM.
Sensors 22 02407 g015
Figure 16. Optimization iteration results of roller bearing.
Figure 16. Optimization iteration results of roller bearing.
Sensors 22 02407 g016
Figure 17. Performance degradation predictions of the roller bearing for the: (a) ELM; (b) SVR; (c) IPSO-LSTM.
Figure 17. Performance degradation predictions of the roller bearing for the: (a) ELM; (b) SVR; (c) IPSO-LSTM.
Sensors 22 02407 g017
Table 1. Optimized Parameters.
Table 1. Optimized Parameters.
DescriptionNotion
Number of nodes in the first LSTM layer h 1
Number of nodes in the second LSTM layer h 2
Batch sizeSm
Table 2. Original features.
Table 2. Original features.
Time-domain T 1 = 1 N i = 1 N x i ,       T 2 = 1 N i = 1 N x i 2 ,         T 3 = 1 N i = 1 N x i 2 ,     T 4 = 1 N i = 1 N x i
T 5 = 1 N i = 1 N x i 3 ,     T 6 = 1 N i = 1 N x i 2 F 4 ,         T 7 = max x 1 N i = 1 N x i ,         T 8 = 1 N i = 1 N x i 4 1 N j = 1 N x j 2 4
Frequency-domain F 1 = 1 N i = 1 N s i ,   F 2 = 1 N j = 1 N ( s j 1 N i = 1 N s i ) 2 ,   F 3 = i = 1 N f i s i j = 1 N s j ,   F 4 = 1 N j = 1 N ( s j 1 N i = 1 N s i ) 3 F 10 3 ,
F 5 = i = 1 N f i 2 s i j = 1 N s j ,   F 6 = 1 N i = 1 N s i f i F 12 2 ,   F 7 = i = 1 N f i 4 s i j = 1 N f j 2 s j ,   F 8 = i = 1 N f i 2 s i j = 1 N s j k = 1 N f k 4 s k
Table 3. The RMSE of LSTM with different optimization methods.
Table 3. The RMSE of LSTM with different optimization methods.
RMSEOuterRollerInner
LSTM0.0420.0390.042
PSO-LSTM0.0250.0130.018
IPSO-LSTM0.0120.0110.013
Table 4. The RMSE of different methods.
Table 4. The RMSE of different methods.
RMSEOuterRollerInner
ELM0.0550.0310.029
SVR0.0310.0270.029
IPSO-LSTM0.0120.0110.013
Table 5. The RMSE of LSTM with different optimization methods.
Table 5. The RMSE of LSTM with different optimization methods.
LSTMPSO-LSTMIPSO-LSTM
RMSE0.0650.0540.048
Table 6. The RMSE of different methods.
Table 6. The RMSE of different methods.
ELMSVRIPSO-LSTM
RMSE0.0730.1010.048
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hu, Y.; Wei, R.; Yang, Y.; Li, X.; Huang, Z.; Liu, Y.; He, C.; Lu, H. Performance Degradation Prediction Using LSTM with Optimized Parameters. Sensors 2022, 22, 2407. https://doi.org/10.3390/s22062407

AMA Style

Hu Y, Wei R, Yang Y, Li X, Huang Z, Liu Y, He C, Lu H. Performance Degradation Prediction Using LSTM with Optimized Parameters. Sensors. 2022; 22(6):2407. https://doi.org/10.3390/s22062407

Chicago/Turabian Style

Hu, Yawei, Ran Wei, Yang Yang, Xuanlin Li, Zhifu Huang, Yongbin Liu, Changbo He, and Huitian Lu. 2022. "Performance Degradation Prediction Using LSTM with Optimized Parameters" Sensors 22, no. 6: 2407. https://doi.org/10.3390/s22062407

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop