Next Article in Journal
Sound Event Detection by Pseudo-Labeling in Weakly Labeled Dataset
Next Article in Special Issue
A Deep Autoencoder-Based Convolution Neural Network Framework for Bearing Fault Classification in Induction Motors
Previous Article in Journal
Can Formal Security Verification Really Be Optional? Scrutinizing the Security of IMD Authentication Protocols
Previous Article in Special Issue
Internal Damage Detection of Composite Structures Using Passive RFID Tag Antenna Deformation Method: Basic Research
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Auto-Encoder and Deep Forest-Assisted Failure Prognosis for Dynamic Predictive Maintenance Scheduling

1
Integrated System Integration Department, No. 38 Research Institute of CETC, Hefei 230088, China
2
College of Automation Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 211106, China
3
Institute of Intelligent Manufacturing, Nanjing Tech University, Nanjing 210009, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(24), 8373; https://doi.org/10.3390/s21248373
Submission received: 27 November 2021 / Revised: 12 December 2021 / Accepted: 14 December 2021 / Published: 15 December 2021

Abstract

:
Prognostics and health management (PHM) with failure prognosis and maintenance decision-making as the core is an advanced technology to improve the safety, reliability, and operational economy of engineering systems. However, studies of failure prognosis and maintenance decision-making have been conducted separately over the past years. Key challenges remain open when the joint problem is considered. The aim of this paper is to develop an integrated strategy for dynamic predictive maintenance scheduling (DPMS) based on a deep auto-encoder and deep forest-assisted failure prognosis method. The proposed DPMS method involves a complete process from performing failure prognosis to making maintenance decisions. The first step is to extract representative features reflecting system degradation from raw sensor data by using a deep auto-encoder. Then, the features are fed into the deep forest to compute the failure probabilities in moving time horizons. Finally, an optimal maintenance-related decision is made through quickly evaluating the costs of different decisions with the failure probabilities. Verification was accomplished using NASA’s open datasets of aircraft engines, and the experimental results show that the proposed DPMS method outperforms several state-of-the-art methods, which can benefit precise maintenance decisions and reduce maintenance costs.

1. Introduction

The rapid development of industrial Internet of Things technology greatly promotes the complexity, integration, and intelligence of modern engineering systems [1,2,3,4,5]. It also raises significant challenges in the safety and reliability of systems’ operation. Due to the unavoidable degradation of various components caused by wearing, aging, fatigue, functional design defects, and complicated environmental factors, the failure probability of the whole system is high and the consequences may be intolerable [6,7,8,9,10]. Precise predictive maintenance is an urgent need in those systems, especially for applications such as nuclear power plants, missile weapons, and aerospace vehicles, which have extremely high reliability and safety requirements. Therefore, prolonging the effective service life and ensuring the reliability with less maintenance cost is of great significance in practice. The key is to realize failure prognosis-based maintenance decision-making, that is, to predict the failure probabilities and carry out just-in-time maintenance activities [11,12,13,14,15].
In recent decades, failure prognosis has received extensive attention, and many research results have been achieved. For instance, an integrated feature-based failure prognosis method was developed in [16], where the dynamics of various failures were detected using signal processing technology, and the adaptive Bayesian algorithm was used to forecast the remaining useful life (RUL) of the faulty bearings. In [17], a multi-stream deep recurrent neural network was developed to handle the features extracted from the vibration signals for failure prognosis. In [18], a Kalman filter and particle filter were combined to predict the failure of satellite reaction wheels. A new hybrid fault prognosis method for multi-functional spoiler systems was proposed in [19], where distributed neural networks were used to estimate the failure parameters, and the recursive Bayesian algorithm was employed to anticipate the system RUL with the estimated failure parameters. Also, some time series-based forecasting methods have been combined with statistically based classification techniques to forecast the failures in cyber-physical systems [20].
However, the above studies only focused on the accuracy of failure prognosis and took no consideration on how to use these valuable failure prognostic results for maintenance decisions. Indeed, there are obvious advantages to jointly considering the two aspects of failure prognosis and maintenance decision-making. For example, in practice, the failure prognosis and maintenance decision-making can be regarded as a whole process affecting the safe operation of the system, where the accuracy of failure prognostics directly affects the effectiveness of maintenance decision-making. Accordingly, their joint research can provide important technical support for the integration of the control, decision-making, and management of complex engineering systems. For such a purpose, this paper develops a deep auto-encoder and deep forest-assisted failure prognosis method for dynamic predictive maintenance scheduling (DPMS). The proposed DPMS method involves a complete process from performing failure prognosis to making maintenance decisions. Firstly, representative features that can reflect system degradation are extracted from raw data by using a deep auto-encoder. Secondly, the features are processed by a deep forest network to compute the system failure probabilities in moving time horizons. Finally, these failure probabilities are used to compute the costs of possible maintenance decisions, and the maintenance activities are scheduled accordingly based on two rules.
The main contributions of this paper are highlighted as follows:
  • A unified and powerful deep auto-encoder and deep forest integrated algorithm is proposed to handle the raw condition monitoring data. It can automatically extract the representative features reflecting system degradation and construct the mapping between the features and discrete degradation states for failure prognosis.
  • Two decision rules are designed to deal with the DPMS. With the prognostic failure probabilities, the maintenance and inventory decisions can be made through quickly evaluating the costs of different decisions.
  • With NASA’s open datasets of aircraft engines, the proposed DPMS method outperforms several state-of-the-art methods, which can benefit precise maintenance decisions and reduce maintenance costs.
The remainder of this paper is structured as follows. In Section 2, the proposed deep auto-encoder and deep forest-assisted failure prognosis and maintenance decision-making method will be described in detail. Section 3 will validate the effectiveness of the proposed methodology, and its performance will be highlighted by comparing it with several state-of-the-art methods. The last section concludes this paper.

2. Methodology

2.1. Key Idea

The proposed methodology is based on real-time condition monitoring data (such as temperature, pressure, and rotational speed) collected by multiple sensors installed in the system. It contains two parts, failure prognosis and maintenance decision-making, and the complete process from performing failure prognosis to making maintenance decisions is shown in Figure 1.
In the failure prognosis stage, the use of a deep auto-encoder is proposed to extract the representative features reflecting system degradation from sensor data. Once the representative features are obtained, the mapping between features and degradation states can be constructed by using the deep forest algorithm, where the degradation states are determined according to the requirements of operational planners. As an illustration, if the operational planners require the failure information in two different time windows ( w 1 and w 2 ), the degradation process of the system will be divided into three discrete states (Deg1, Deg2, and Deg3). Deg1 denotes the case where the system RUL is greater than w 2 , i.e., R U L > w 2 , whereas Deg3 represents the case where the system RUL is less than w 1 , i.e., R U L w 1 . Deg2 refers to the case where the system RUL is in the period ( w 1 , w 2 ] , i.e., w 1 < R U L w 2 . Compared with Deg1 and Deg2, the degradation of Deg3 is more serious.
In the maintenance decision-making stage, the online condition monitoring data are first fed into the well-trained deep auto-encoder model to produce the representative features of the system degradation. With the representative features as the input, the well-trained deep forest model will output the failure probabilities belonging to different degradation states. Finally, the failure probabilities are used to compute the costs of different decisions, and the maintenance and inventory activities are scheduled according to two decision cost-based rules.

2.2. Degradation Feature Extraction Using a Deep Auto-Encoder

Auto-encoder technology is an important branch of deep learning theory and can be regarded as a feature extraction method of isodimensional mapping [21]. A basic auto-encoder consists of two main parts: an encoder and a decoder. The function of the encoder is to encode the high-dimensional input x R D into a low-dimensional implicit variable h R d ( d < D ), so as to force the neural network to learn the most informative features. The function of the decoder is to restore the hidden variable h of the code layer to the initial dimension. The best state is that the output x ˜ R D of the decoder can perfectly or approximately restore the original input x R D , i.e., x ˜ x . Therefore, the implicit variable h R d can be considered as the representative feature reflecting system degradation.
In this paper, a deep structure of a basic auto-encoder, known as a deep auto-encoder, was employed to extract deeper representative degradation features. As shown in Figure 2, the deep auto-encoder contains several hidden layers. For an L -layer deep auto-encoder, the encoding process of the original data x from the input layer to the l -th hidden layer can be expressed as follows:
{ h ( 0 ) = x h ( l ) = f e ( l ) ( W e ( l ) h e ( l 1 ) + b e ( l ) )
where f e ( l ) : R D R d refers to the encoding function and W e ( l ) and b e ( l ) are the weight and bias of the encoding layer. The decoding process of the implicit variable h ( L ) from the code layer to the output layer is
{ x ˜ ( L + 1 ) = h ( L ) x ˜ ( l ) = f d ( l ) ( W d ( l ) x ˜ ( l + 1 ) + b d ( l ) )
where f d ( l ) : R d R D refers to the decoding function and W d ( l ) and b d ( l ) are the weight and bias of the decoding layer. Commonly, the sigmoid function can be used for the encoding function f e ( l ) and decoding function f d ( l ) . It is given by
s i g m o i d ( x ) = 1 1 + e x
To determine the deep auto-encoder parameters { W e ( l ) , W d ( l ) , b e ( l ) , b d ( l ) } , a loss function is designed as follows:
J ( W , b ) = 1 2 M i = 1 M x ˜ i x i 2 2 + λ 2 l = 1 L 1 i = 1 s l + 1 j = 1 s l ( W i j ( l ) ) 2
where M is the number of training samples, λ refers to the weight decay parameter, s l represents the number of nodes in layer l , and W i j ( l ) is the weight coefficient of the propagation path of the j -th neuron in layer l and the i -th neuron in layer l + 1 . In Equation (4), the first term refers to the reconstruction error of the deep auto-encoder, while the second term is used to prevent over-fitting. Finally, Equation (4) can be minimized by using the stochastic gradient descent method.

2.3. Failure Prognosis Using Deep Forest

To capture the nonlinear relationship between the representative features and discrete degradation states, an integrated learning method called deep forest [22] is employed. One of the main advantages of the deep forest is the ability to handle unbalanced observation data. As mentioned in Section 2.1, the degradation process of a system can be divided into three discrete states (Deg1, Deg2, and Deg3) according to its RUL. Considering that the system operates under normal conditions most of the time, the amount of data of Deg1 will be significantly more than those of Deg2 and Deg3. Therefore, when applying the deep forest for system failure prognosis, it can improve the prognostic accuracy by handling the unbalanced observation data.
The schematic diagram of the deep forest prognosis model constructed in this paper is shown in Figure 3. Similar to the layer expansion of the deep neural network (DNN), the deep forest is formed by the automatic cascading expansion of a multi-layer forest structure (from level 1 to level N), without the need to set the number of forest layers in advance. Each layer of forest usually receives the feature vector generated from the previous layer, and then outputs the generated feature vector to the next layer. In the construction of each layer of the model, two kinds of random forests are contained, i.e., complete random forest (noted as random forest A) and ordinary random forest (noted as random forest B). The complete random forests consist of multiple decision trees, and each tree contains all the features. A feature is randomly selected as the split node of the split tree, and the splitting process is not terminated until each leaf node contains only one category or no more than 10 samples. The ordinary random forests randomly select d ( d is the input feature dimension) candidate features and then employs the Gini index [23] to select split nodes for the growth of the tree.
Given a sample set V , the Gini index can be calculated by
G i n i ( V ) = 1 g = 1 G ( V g V ) 2
where V g is a subset of samples belonging to the g category, and G is the number of categories. If the sample set V is divided into two parts, V 1 and V 2 ( V = V 1 + V 2 ), according to whether the feature A takes a certain possible value a , then under the condition of the feature A , the Gini index of the sample set V (denoted by G i n i ( V , A ) ) can be defined as
G i n i ( V , A ) = | V 1 | | V | G i n i ( V 1 ) + | V 2 | | V | G i n i ( V 2 )
G i n i ( V , A ) represents the uncertainty of the sample set V through the feature A = a segmentation. Similar to entropy, the greater the G i n i ( V , A ) value, the greater the uncertainty. Thus, the feature value with the minimum G i n i ( V , A ) will be chosen, i.e.,
a = arg a A min G i n i ( V , A = a )
According to the optimal segmentation feature value a , each decision tree can continuously divide subspaces in the feature space, and each subspace is labeled. Then, the leaf nodes can obtain the probabilities of different categories in the training sample, and the proportions of the various types in the entire forest will be produced by averaging the various proportions of all decision trees in each forest, as shown in Figure 4. In the training process of the cascade forest, the generated category distribution vector will be connected with the original input vector to feed the lower cascade layer through the cascade channel. When the classification accuracy of the verification set converges or reaches the expected value, the model training is terminated. It should be noted that the deep forest prognosis model constructed in this paper is different from many decision tree algorithms and does not require a pruning step. This is mainly because the deep forest is formed through the integration of random forests, and the random sampling and random feature selection of the random forest strengthen the diversity of the integrated learning and the generalization ability of the overall model. In other words, even if the deep forest is not pruned, there may be no over-fitting.

2.4. Maintenance-Related Decision Rules Based on Prognostic Information

Within the proposed predictive maintenance framework, the basic decisions that need to be made at each monitoring point are whether to place a spare part order or maintain the system. Actually, due to the technical and logistical constraints, it is difficult to perform maintenance on the spot in real time. In this paper, maintenance actions are considered perfect and decisions are only made at each inspection moment. Given the failure prognostic information ( P ( R U L > w 2 ) , P ( w 1 < R U L w 2 ) , and P ( R U L w 1 ) ), the following two maintenance-related decision rules [24] are designed:
  • Inventory decision: The inventory decision aims to determine whether to place a spare part order at the current inspection time (h-th inspection period for example). For an option to order the spare part, the cost is computed by
C o = C i n s i = [ w 2 / Δ T ] + + P ( ( i 1 ) Δ T < R U L i Δ T ) ( i Δ T w 2 ) P ( R U L > w 2 ) C i n s [ max ( T - f h Δ T w 2 , Δ T ) Δ T ] + Δ T
where C i n s represents the inventory cost of the spare part per unit time, w 2 = [ L / Δ T ] + Δ T is the time window associated with the lead time L , Δ T is a fixed inspection interval, T - f is the mean time of system failure, and [ ] + means taking an upper integer. On the contrary, the cost incurred by not ordering the spare part is given by
C n o = C o s P ( w 2 < R U L w 2 + Δ T )
where C o s refers to the out-of-stock cost. Notably, the cost of not ordering a spare part is not the actual charged cost. It is considered as a damage estimation with wrong decision. With the computed C o and C n o , the inventory decision can be made by
O r d e r = { 0 , C n o < C o 1 , o t h e r s
where O r d e r = 0 means that there is no need to order a spare part at the current inspection time, whereas O r d e r = 1 requires a spare part order to be placed.
  • Maintenance decision: The maintenance decision aims to determine whether to maintain the system at the current inspection time. For an option to maintain the system, the cost rate is computed by
C m = C p + I ( S h = 0 ) C o s h Δ T
where C p refers to the preventive cost, S h denotes the storage state, and I ( ) is an indicator function, where I ( ) = 1 when the condition is met and I ( ) = 0 otherwise. On the contrary, the cost rate incurred by not maintaining the system is given by
C n m = P ( R U L w 1 ) ( C c + I ( S h = 1 ) C i n s Δ T + I ( S h + 1 = 0 ) C o s ) ( h + 1 ) Δ T
where C c refers to the corrective cost and w 1 = Δ T is the time window associated with the inspection interval Δ T . With the computed C m and C n m , the maintenance decision can be made by
M a i n t e n a n c e = { 0 , C n m < C m 1 , o t h e r s
where M a i n t e n a n c e = 0 means that there is no need to perform maintenance activities at the current inspection time, whereas M a i n t e n a n c e = 1 requires maintenance to the system.

2.5. Implementation Process of Predictive Maintenance

For an in-service system under consideration, the proposed DPMS method can be implemented by the following procedures:
  • Obtain the real-time condition monitoring data from multiple sensors installed in the system;
  • Obtain the representative features that can reflect system degradation using the deep auto-encoder;
  • Produce the failure probabilities in moving time horizons using deep forest;
  • Compute the costs of different decisions, and schedule maintenance and inventory activities according to two decision cost-based rules.
To evaluate the performance of the proposed DPMS method, the prognostic accuracy and maintenance cost rate reflecting the economic benefits of the system are adopted. The prognostic accuracy can be mathematically stated as
A c c u r a c y = 1 N i = 1 N C i
where C i = 1 if y ^ i = y i and C i = 0 otherwise, N is the number of test samples, y ^ i is the predicted category of the i -th sample, and y i is the true category. The maintenance cost rate is defined as the ratio of total maintenance cost to total running time [25], i.e.,
M C R = { C p + I ( S h = 0 ) C o s h Δ T , preventive maintenance is performed . C c + I ( S h = 0 ) C o s h Δ T , corrective maintenance is performed .
The strategy with the lower maintenance cost rate is considered to have a better performance.

3. Results

The described DPMS method was implemented on Python 2.7 software, where the “gcForest” toolbox available at GitHub platform [26] was used. To validate its performance, the C-MAPSS dataset available in NASA’s data repository [27] was considered.

3.1. Description of the C-MAPSS Dataset

The C-MAPSS dataset is a popular dataset simulating various scenarios of aircraft engine degradation and has been widely used to test the performance of various data-driven failure prognosis methods [28,29,30,31,32]. Figure 5 shows the schematic diagram of the simulated engine. It is composed of several main components, such as a fan, low-pressure compressor (LPC), combustor, high-pressure compressor (HPC), low pressure turbine (LPT), high pressure turbine (HPT), and nozzle. In the initial phase of each scenario, the engine operated normally. As the service time increased, its performance degraded gradually, and the complete run-to-failure data have been recorded in the “train-FD001.txt” subset. In this paper, to determine the structure of deep forest and test the decision accuracy of the engine at different inspection moments, the “train-FD001.txt” subset including 100 trajectories will be divided into three parts. The first part consisting of the first 70 trajectories is used to train the proposed feature extraction model and failure prognosis model. The second part that contains the next 10 trajectories is used as the verification set to estimate the training results, and the deep forest structure will be preserved when the classification accuracy is greater than 90%. The third part including the remaining 20 trajectories is used as the test set for simulating the real-time condition monitoring processes.
The training, verification, and test sets are composed of 26 columns describing the characteristics of the engine units. The first five columns correspond to the basic information of engine units, such as the engine number, degradation time steps, and constant operational settings, while the remaining 21 columns provide the sensor measurements. Table 1 shows the detailed description of 21 sensor variables, and Table 2 presents part of the condition monitoring data of an engine case. One can see that from the second column of Table 2, the sensor measurements (total temperature at fan inlet) do not change with the increase of operating cycle, and this means that the measurements cannot really reflect the degradation of the engine. In other words, before training the failure prognosis model, it is necessary to extract the representative features reflecting the engine degradation from the measurements of 21 sensors.

3.2. Accuracy of Failure Prognosis Model

To extract the representative degradation features from the 21-sensor data, in the deep auto-encoder modeling, the feature dimensions are set to 4, 8, 12, and 16, respectively. Table 3 reports the prognostic accuracies of different feature dimensions in the cross-validation set. It is observed that when the feature dimension is set to 12, the highest prognostic accuracy can be achieved. In other words, 12 is regarded as a satisfactory feature dimension.
Then, the 12 dimensional degradation features extracted from the deep auto-encoder are fed into the deep forest for training. During the training process, if it is found that the prognostic accuracy within 10 running cycles is not improved, the training will stop. Figure 6 presents the average confusion matrices of different prognostic models on the test engines with three respective classes: Deg1 ( R U L > 20 ), Deg2 ( 10 < R U L 20 ), and Deg3 ( R U L 10 ). From this figure, no matter which kind of prognostic model (“LSTM network” [24], “Bi-LSTM network” [34], “deep forest” [35], or “deep auto-encoder + deep forest”) is used, the prognostic accuracy for Deg1 is the highest. This can be explained as Deg1 refers to the low degradation, and accordingly there are more training data. Regarding Deg2, the use of the integrated model of a deep auto-encoder and deep forest allows improving the prognostic accuracy. This is mainly because the deep auto-encoder can extract the representative features of engine degradation, while the deep forest can handle the uneven classification sample data. As for Deg3, each model can obtain a prognostic accuracy of more than 90%. Actually, the high prognostic accuracy at this stage is very important for improving the flight safety of the aircraft. In summary, the prognostic model of this paper outperforms the LSTM network, Bi-LSTM network, and single deep forest model.
Remark 1.
It should be noted that in practice, the time windows w 1 and w 2 can be determined according to the requirements of operational planners. In this paper, the values of w 1 and w 2 are assigned to 10 and 20, respectively, to be consistent with the case in [24]. If the proposed failure prognosis method is applied to other cases, these values can still work.

3.3. Performance of the Dynamic Predictive Maintenance Strategy

With the available failure prognostic probabilities, the inventory and maintenance decisions can be made according to the decision rules presented in Section 2.4. Table 4 lists the decision results of some test engines at different inspection moments, given C i n s = 0.1 , C o s = 10 , C p = 100 , and C c = 500 [24]. The first column represents the running cycle of the engine, while the second column provides the real RUL according to the “RUL_FD001.txt” text file from the C-MAPSS dataset. The next three columns represent the failure probabilities belonging to different degradation states, while the last three columns represent the order, stock, and maintenance states, respectively.
Remark 2.
It should be noted that these cost parameters, C i n s , C o s , C p , and C c , are closely related to the performance of the method. Considering that the failure prognosis-based maintenance strategy presented in [24] has been verified to be flexible under different cost structures, this paper will not repeat the verification. To facilitate the comparative analysis, these parameter values for every case in this paper are set to be the same as those in [24].
Considering the engine ID91, one can see that before t = 100 , the failure probabilities belonging to different degradation states are 96.64%, 3.19%, and 0.17%, which means that the engine is in a low degradation state during operation. Correspondingly, the proposed strategy does not require any maintenance or spare part management activities, and the order, stock, and maintenance Boolean variables are set to be zeros. When t = 110 , according to Equation (10), it is suggested to order a spare part, and the available spare part will be delivered at t = 130 . At t = 130 , due to the high failure probability of Deg3, the optimal decision is to maintain the engine. Regarding the engine ID92, nothing needs to be done before t = 310 . When t = 320 , it is suggested to order a spare part, and the available spare part will arrive after two decision periods. However, the engine is required to be maintained at t = 330 . In this case, one must pay the out-of-stock cost. For the engine ID93, it can be found that there are decision results similar to those of engine 91, and this can be explained by the fact that their life lengths are very close.
Next, the average maintenance cost rates will be used to evaluate the performances of maintenance strategies using different failure prognosis methods, as shown in Figure 7. In this figure, the 20 test engines (ID from 81 to 100) are equally divided into two groups. It is observed that, regardless of the group, the use of deep forest can reduce the average maintenance cost rate of the engines. Further, the inclusion of the deep auto-encoder allows improving the accuracy of decision-making. Accordingly, the strategy of this paper (i.e., “deep auto-encoder + deep forest”-based predictive maintenance) can obtain the lowest maintenance cost rates.

4. Conclusions

In this paper, a deep auto-encoder and deep forest-assisted failure prognosis method was proposed for dynamic predictive maintenance scheduling (DPMS). The representative features reflecting system degradation were extracted from raw sensor data by using a deep auto-encoder, while the inclusion of the deep auto-encoder allowed the nonlinear relationship between the representative features and discrete degradation states to be learned. The integration of these two models led to an effective failure prognosis model. With the prognostic failure probabilities, the maintenance and inventory activities are scheduled through quickly evaluating the costs of different decisions. The verification results using NASA’s open datasets of aircraft engines reveal the feasibility of the proposed DPMS method. According to the prognostic accuracies and average maintenance cost rates, the proposed method is better than several state-of-the-art methods. Future work will focus on the investigation of predictive maintenance strategy for more sophisticated systems with multiple components.

Author Contributions

Conceptualization, C.C. and N.L.; methodology, H.Y. and C.W.; validation, H.Y.; formal analysis, H.Y.; investigation, H.Y. and C.C.; writing—original draft preparation, H.Y.; writing—review and editing, N.L. and C.C.; visualization, H.Y.; supervision, N.L.; project administration, N.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China (Grant Nos. 61873122, 61973288, 62020106003), Research Fund of State Key Laboratory of Mechanics and Control of Mechanical Structures, Nanjing University of Aeronautics and Astronautics (Grant No. MCMS-I-0521G02), Aeronautical Science Foundation of China (Grant No. 20200007018001), and China Scholarship Council (Grant No. 202006830060).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the researchers who built the public dataset and provided the public code.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Olesen, J.F.; Shaker, H.R. Predictive Maintenance for Pump Systems and Thermal Power Plants: State-of-the-Art Review, Trends and Challenges. Sensors 2020, 20, 2425. [Google Scholar] [CrossRef] [PubMed]
  2. Yin, A.; Yan, Y.; Zhang, Z.; Li, C.; Sánchez, R.-V. Fault Diagnosis of Wind Turbine Gearbox Based on the Optimized LSTM Neural Network with Cosine Loss. Sensors 2020, 20, 2339. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Lee, Y.; Lee, Y.S. A low-cost surge current detection sensor with predictive lifetime display function for maintenance of surge protective devices. Sensors 2020, 20, 2310. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Chen, C.; Lu, N.; Jiang, B.; Xing, Y.; Zhu, Z.H. Prediction interval estimation of aero-engine remaining useful life based on bidirectional long short-term memory network. IEEE Trans. Instrum. Meas. 2021, 70, 3527213. [Google Scholar] [CrossRef]
  5. Chen, C.; Lu, N.; Jiang, B.; Wang, C. A Risk-Averse Remaining Useful Life Estimation for Predictive Maintenance. IEEE/CAA J. Autom. Sin. 2021, 8, 412–422. [Google Scholar] [CrossRef]
  6. Chen, C.; Lu, N.; Jiang, B.; Xing, Y. Condition-based maintenance optimization for continuously monitored degrading systems under imperfect maintenance actions. J. Syst. Eng. Electron. 2020, 31, 841–851. [Google Scholar]
  7. Chen, C.; Wang, C.; Lu, N.; Jiang, B.; Xing, Y. A data-driven predictive maintenance strategy based on accurate failure prognostics. Eksploat. I Niezawodn. 2021, 23, 387–394. [Google Scholar] [CrossRef]
  8. Hu, C.-H.; Pei, H.; Si, X.-S.; Du, D.-B.; Pang, Z.-N.; Wang, X. A Prognostic Model Based on DBN and Diffusion Process for Degrading Bearing. IEEE Trans. Ind. Electron. 2019, 67, 8767–8777. [Google Scholar] [CrossRef]
  9. Berri, P.C.; Vedova, M.D.D.; Mainini, L. Computational framework for real-time diagnostics and prognostics of aircraft actuation systems. Comput. Ind. 2021, 132, 103523. [Google Scholar] [CrossRef]
  10. Wang, C.; Lu, N.; Cheng, Y.; Jiang, B. A Data-Driven Aero-Engine Degradation Prognostic Strategy. IEEE Trans. Cybern. 2019, 51, 1531–1541. [Google Scholar] [CrossRef]
  11. Leukel, J.; González, J.; Riekert, M. Adoption of machine learning technology for failure prediction in industrial maintenance: A systematic review. J. Manuf. Syst. 2021, 61, 87–96. [Google Scholar] [CrossRef]
  12. Dalzochio, J.; Kunst, R.; Pignaton, E.; Binotto, A.; Sanyal, S.; Favilla, J.; Barbosa, J. Machine learning and reasoning for predictive maintenance in Industry 4.0: Current status and challenges. Comput. Ind. 2020, 123, 103298. [Google Scholar] [CrossRef]
  13. Dawood, T.; Elwakil, E.; Novoa, H.M.; Delgado, J.F.G. Pressure data-driven model for failure prediction of PVC pipelines. Eng. Fail. Anal. 2020, 116, 104769. [Google Scholar] [CrossRef]
  14. Zakikhani, K.; Nasiri, F.; Zayed, T. Availability-based reliability-centered maintenance planning for gas transmission pipelines. Int. J. Press. Vessel. Pip. 2020, 183, 104105. [Google Scholar] [CrossRef]
  15. Fernandes, S.; Antunes, M.; Santiago, A.R.; Barraca, J.P.; Gomes, D.; Aguiar, R.L. Forecasting Appliances Failures: A Machine-Learning Approach to Predictive Maintenance. Information 2020, 11, 208. [Google Scholar] [CrossRef] [Green Version]
  16. Rezamand, M.; Kordestani, M.; Carriveau, R.; Ting, D.S.-K.; Saif, M. An Integrated Feature-Based Failure Prognosis Method for Wind Turbine Bearings. IEEE/ASME Trans. Mechatron. 2020, 25, 1468–1478. [Google Scholar] [CrossRef]
  17. Su, Y.; Tao, F.; Jin, J.; Wang, T.; Wang, Q.; Wang, L. Failure Prognosis of Complex Equipment With Multistream Deep Recurrent Neural Network. J. Comput. Inf. Sci. Eng. 2020, 20, 1–13. [Google Scholar] [CrossRef] [Green Version]
  18. Rahimi, A.; Kumar, K.D.; Alighanbari, H. Failure Prognosis for Satellite Reaction Wheels Using Kalman Filter and Particle Filter. J. Guid. Control. Dyn. 2020, 43, 585–588. [Google Scholar] [CrossRef]
  19. Kordestani, M.; Samadi, M.F.; Saif, M. A New Hybrid Fault Prognosis Method for MFS Systems Based on Distributed Neural Networks and Recursive Bayesian Algorithm. IEEE Syst. J. 2020, 14, 5407–5416. [Google Scholar] [CrossRef]
  20. Ruiz-Arenas, S.; Rusák, Z.; Mejía-Gutiérrez, R.; Horváth, I. Implementation of System Operation Modes for Health Management and Failure Prognosis in Cyber-Physical Systems. Sensors 2020, 20, 2429. [Google Scholar] [CrossRef] [PubMed]
  21. Chen, C.; Zhu, Z.H.; Shi, J.; Lu, N.; Jiang, B. Dynamic Predictive Maintenance Scheduling Using Deep Learning Ensemble for System Health Prognostics. IEEE Sensors J. 2021, 21, 26878–26891. [Google Scholar] [CrossRef]
  22. Zhou, Z.H.; Feng, J. Deep forest. Natl. Sci. Rev. 2019, 6, 74–86. [Google Scholar] [CrossRef] [PubMed]
  23. Liu, F.T.; Ting, K.M.; Yu, Y.; Zhou, Z.H. Spectrum of Variable-Random Trees. J. Artif. Intell. Res. 2008, 32, 355–384. [Google Scholar] [CrossRef] [Green Version]
  24. Nguyen, K.T.; Medjaher, K. A new dynamic predictive maintenance framework using deep learning for failure prognostics. Reliab. Eng. Syst. Saf. 2019, 188, 251–262. [Google Scholar] [CrossRef] [Green Version]
  25. Zhang, L.; Zhang, J. A Data-Driven Maintenance Framework Under Imperfect Inspections for Deteriorating Systems Using Multitask Learning-Based Status Prognostics. IEEE Access 2021, 9, 3616–3629. [Google Scholar] [CrossRef]
  26. Saxena, A.; Goebel, K. Turbofan Engine Degradation Simulation Data Set. 2008. Available online: https://ti.arc.nasa.gov/tech/dash/groups/pcoe/prognostic-data-repository/ (accessed on 20 July 2021).
  27. Zhou, Z.H.; Feng, J. gcForest v1.1.1. Available online: https://github.com/kingfengji/gcForest (accessed on 25 July 2021).
  28. Huang, C.-G.; Huang, H.-Z.; Li, Y.-F. A Bidirectional LSTM Prognostics Method under Multiple Operational Conditions. IEEE Trans. Ind. Electron. 2019, 66, 8792–8802. [Google Scholar] [CrossRef]
  29. Wang, C.; Zhu, Z.; Lu, N.; Cheng, Y.; Jiang, B. A data-driven degradation prognostic strategy for aero-engine under various operational conditions. Neurocomputing 2021, 462, 195–207. [Google Scholar] [CrossRef]
  30. Shi, Z.; Chehade, A. A dual-LSTM framework combining change point detection and remaining useful life prediction. Reliab. Eng. Syst. Saf. 2021, 205, 107257. [Google Scholar] [CrossRef]
  31. Ellefsen, A.L.; Bjørlykhaug, E.; Æsøy, V.; Ushakov, S.; Zhang, H. Remaining useful life predictions for turbofan engine degradation using semi-supervised deep architecture. Reliab. Eng. Syst. Saf. 2019, 183, 240–251. [Google Scholar] [CrossRef]
  32. Chui, K.T.; Gupta, B.B.; Vasant, P. A Genetic Algorithm Optimized RNN-LSTM Model for Remaining Useful Life Prediction of Turbofan Engine. Electronics 2021, 10, 285. [Google Scholar] [CrossRef]
  33. Saxena, A.; Goebel, K.; Simon, D.; Eklund, N. Damage Propagation Modeling for Aircraft Engine Run-to-Failure Simulation. In Proceedings of the 2008 International Conference on Prognostics and Health Management, Denver, CO, USA, 6–9 October 2008; pp. 1–9. [Google Scholar]
  34. Zhang, Y. Aeroengine Fault Prediction Based on Bidirectional LSTM Neural Network. In Proceedings of the 2020 International Conference on Big Data, Artificial Intelligence and Internet of Things Engineering (ICBAIE), Fuzhou, China, 12–14 June 2020; pp. 317–320. [Google Scholar]
  35. Liu, X.; Tian, Y.; Lei, X.; Liu, M.; Wen, X.; Huang, H.; Wang, H. Deep forest based intelligent fault diagnosis of hydraulic turbine. J. Mech. Sci. Technol. 2019, 33, 2049–2058. [Google Scholar] [CrossRef]
Figure 1. Proposed predictive maintenance framework.
Figure 1. Proposed predictive maintenance framework.
Sensors 21 08373 g001
Figure 2. Deep auto-encoder architecture.
Figure 2. Deep auto-encoder architecture.
Sensors 21 08373 g002
Figure 3. Structure diagram of deep forest prognosis model.
Figure 3. Structure diagram of deep forest prognosis model.
Sensors 21 08373 g003
Figure 4. Structure diagram of class vector generation in deep forest.
Figure 4. Structure diagram of class vector generation in deep forest.
Sensors 21 08373 g004
Figure 5. Structure diagram of the simulated engine [33].
Figure 5. Structure diagram of the simulated engine [33].
Sensors 21 08373 g005
Figure 6. Confusion matrices (%) of different prognostic models on the test engines: (a) LSTM network; (b) Bi-LSTM network; (c) deep forest; and (d) deep auto-encoder + deep forest.
Figure 6. Confusion matrices (%) of different prognostic models on the test engines: (a) LSTM network; (b) Bi-LSTM network; (c) deep forest; and (d) deep auto-encoder + deep forest.
Sensors 21 08373 g006
Figure 7. Average maintenance cost rates using different failure prognosis methods for 20 test engines.
Figure 7. Average maintenance cost rates using different failure prognosis methods for 20 test engines.
Sensors 21 08373 g007
Table 1. Description of 21 sensor variables [33].
Table 1. Description of 21 sensor variables [33].
IDSymbolDescriptionUnits
1T2Total temperature at fan inletºR
2T24Total temperature at LPC outletºR
3T30Total temperature at HPC outletºR
4T50Total temperature at LPT outletºR
5P2Pressure at fan inletpsia
6P15Total pressure in bypass-ductpsia
7P30Total pressure at HPC outletpsia
8NfPhysical fan speedrpm
9NcPhysical core speedrpm
10eprEngine pressure ratio (P50/P2)--
11Ps30Static pressure at HPC outletpsia
12phiRatio of fuel flow to Ps30pps/psi
13NRfCorrected fan speedrpm
14NRcCorrected core speedrpm
15BPRBypass ratio--
16farBBurner fuel–air ratio--
17htBleedBleed enthalpy--
18Nf_dmdDemanded fan speedrpm
19PCNfR_dmdDemanded corrected fan speedrpm
20W31HPT coolant bleedlbm/s
21W32LPT coolant bleedlbm/s
Table 2. Sample run-to-failure data from an engine case.
Table 2. Sample run-to-failure data from an engine case.
Operating CycleSensor #1 (ºR)Sensor #2 (ºR)Sensor #3 (ºR) Sensor #21 (lbm·s−1)
1518.67641.821589.70 23.42
2518.67642.151591.82 23.42
3518.67642.351587.99 23.34
192518.67643.541601.41 22.96
Table 3. Prognostic accuracies of different feature dimensions in cross-validation set.
Table 3. Prognostic accuracies of different feature dimensions in cross-validation set.
Feature DimensionPrognostic Accuracy (%)
497.23
897.73
1298.13
1697.55
Table 4. Decision results of some test engines at different inspection moments.
Table 4. Decision results of some test engines at different inspection moments.
Running CycleReal RULDeg1 (%)Deg2 (%)Deg3 (%)OrderStockMaintenance
Engine   ID 91   T f = 135 )
904596.643.190.17000
1003597.012.850.14000
1102576.5122.471.01100
1201537.3158.544.15100
13052.2915.1082.61111
Engine   ID 92   ( T f = 341 )
3004194.755.020.23000
3103191.338.280.39000
3202126.8165.997.20100
330116.5739.5853.85101
Engine   ID 93   ( T f = 155 )
1104599.950.050.00000
1203599.900.090.01000
1302572.2926.241.47100
1401542.4853.873.65100
15051.7211.2487.04111
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yu, H.; Chen, C.; Lu, N.; Wang, C. Deep Auto-Encoder and Deep Forest-Assisted Failure Prognosis for Dynamic Predictive Maintenance Scheduling. Sensors 2021, 21, 8373. https://doi.org/10.3390/s21248373

AMA Style

Yu H, Chen C, Lu N, Wang C. Deep Auto-Encoder and Deep Forest-Assisted Failure Prognosis for Dynamic Predictive Maintenance Scheduling. Sensors. 2021; 21(24):8373. https://doi.org/10.3390/s21248373

Chicago/Turabian Style

Yu, Hui, Chuang Chen, Ningyun Lu, and Cunsong Wang. 2021. "Deep Auto-Encoder and Deep Forest-Assisted Failure Prognosis for Dynamic Predictive Maintenance Scheduling" Sensors 21, no. 24: 8373. https://doi.org/10.3390/s21248373

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop