Next Article in Journal
Active Fault-Tolerant Control for Near-Space Hypersonic Vehicles
Previous Article in Journal
A Simulated Annealing Algorithm with Tabu List for the Multi-Satellite Downlink Schedule Problem Considering Waiting Time
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fault Detection of Aero-Engine Sensor Based on Inception-CNN

1
College of Energy and Power Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 210016, China
2
Ningbo Institute of Materials Technology & Engineering, Chinese Academy of Sciences, Ningbo 315201, China
*
Author to whom correspondence should be addressed.
Aerospace 2022, 9(5), 236; https://doi.org/10.3390/aerospace9050236
Submission received: 1 March 2022 / Revised: 17 April 2022 / Accepted: 21 April 2022 / Published: 25 April 2022

Abstract

:
The aero-engine system is complex, and the working environment is harsh. As the fundamental component of the aero-engine control system, the sensor must monitor its health status. Traditional sensor fault detection algorithms often have many parameters, complex architecture, and low detection accuracy. Aiming at this problem, a convolutional neural network (CNN) whose basic unit is an inception block composed of convolution kernels of different sizes in parallel is proposed. The network fully extracts redundant analytical information between sensors through different size convolution kernels and uses it for aero-engine sensor fault detection. On the sensor failure dataset generated by the Monte Carlo simulation method, the detection accuracy of Inception-CNN is 95.41%, which improves the prediction accuracy by 17.27% and 12.69% compared with the best-performing non-neural network algorithm and simple BP neural networks tested in the paper, respectively. In addition, the method simplifies the traditional fault detection unit composed of multiple fusion algorithms into one detection algorithm, which reduces the complexity of the algorithm. Finally, the effectiveness and feasibility of the method are verified in two aspects of the typical sensor fault detection effect and fault detection and isolation process.

1. Introduction

As the measuring element of the aero-engine control system, the function of the sensor is fundamental and essential for the system. However, with the increasing control demand and the increasing complexity of the control system, new requirements are put forward for the number and reliability of sensors, and in the poor working environment of aero-engine sensors, faults are difficult to avoid [1,2]. According to statistics, sensor faults cover more than 80% of faults in the aero-engine control system [3], so the fault detection and isolation (FDI) of aero-engine sensors are vital to improving the reliability of aero-engine control systems.
At present, the FDI methods of aero-engine sensors can be divided into two types: model-based and signal-based. The main idea of the model-based method is to establish the mapping relationship between the actual input and output of the controlled object by analyzing the physical characteristics of the controlled object and then analyzing the residuals between the output of the actual system and the output of the model for diagnosis and isolation [4]. The prominent representatives are the Kalman filter [5,6,7,8,9,10], particle filter [11], etc. Based on the model method, sensor fault detection becomes very simple under the premise of accurate model mapping. However, the complexity of actual controlled objects is the difficulty of its practical application.
Signal-based methods directly analyze signals through reliability analysis or machine learning, among which machine learning has made significant progress in the field of sensor FDI. Regarding fault identification based on one-dimensional sensor signals, literature [12] expands the one-dimensional signals of a single sensor into two-dimensional data. It inputs them into a CNN for fault detection through time series misalignment. However, this approach has poor randomness and reliability for aero-engines with complex control systems and diverse working conditions. Therefore, the most common solution is to use the analytically redundant space information between multiple sensors to detect faults of single or multiple sensors. In literature [13,14,15], the support vector machine (SVM) algorithm is used to build a nonlinear mapping relationship between faulty sensors and healthy sensors. Finally, an appropriate threshold is selected for the predicted value of the algorithm to determine whether the sensor is faulty. The literature [16,17,18] used the Levenberg–Marquardt algorithm, Artificial Bee Colony algorithm, and adaptive method to optimize a BP neural network to construct the mapping relationship between sensors for banning SVM. All of the above methods have high accuracy for single-sensor fault diagnosis. However, for multi-sensor diagnosis, it is necessary to build multiple algorithm units and match them with a complex logical judgment method or fusion judgment algorithm to make its accuracy usable. As mentioned in reference [19], assuming that no more than three sensors fail simultaneously among the nine sensors, 84 multi-input multi-output generalized regression neural networks (MIMOGRNNs) should be constructed for fusion diagnosis. In literature [20,21], the structure of the self-associative neural network and the extreme learning machine algorithm are optimized and changed by the genetic algorithm, respectively, to achieve multi-sensor detection, but its detection accuracy is still far from the 95% specified in the literature [22].
In recent years, CNN, as an essential part of deep learning algorithms, has had many applications in different engineering fields. Especially in data-based prediction, CNN shows good performance due to its robust feature extraction ability. Literature [23] proposed a prediction model based on CNN and lion algorithm fusion; the model was improved by niche immunity and finally used in the short-term load prediction of electric vehicles; the optimized prediction model can allow the existence of deformed data and has good prediction performance. Literature [24] proposed an innovative hyperspectral remote sensing image classification method based on CNN. The parameter update of the model adopts an extreme learning machine; finally, the model’s accuracy has been verified on the remote sensing image Jiuzhaigou. Literature [25] proposed a prediction model for detecting power theft by combining the bidirectional gated recurrent unit and CNN. Compared with the multilayer perceptron, the long short-term memory network (LSTM), and the single gated recurrent unit, the model performs better in this application scenario. Moreover, in the field of power protection, the author proposes a deep hybrid neural network model that combines CNN and particle swarm optimization (PSO) in literature [26]. The model uses CNN as the feature extractor of the PSO algorithm, which aims to condense useful feature information in the original time series. Literature [27] proposed a one-dimensional CNN prediction model for the real-time detection of motor faults. The model has a simple structure, low hardware requirements, and a fast detection speed. Literature [28] uses CNN to predict the sound source of plate-like structures. Compared with stacked autoencoders, CNN can accept more information input and has a flexible information input method. In the literature [29], the author converts the vibration sensor signal of the motor into a three-dimensional feature matrix as the input of the CNN, and different sensors occupy different channel dimensions in the feature matrix. This paper provides ideas for the situation in which there are different kinds of feature data as CNN input. In the literature [30], the author used the output of the virtual sensor of the aero-engine physical model and the Kalman filter as the input of a one-dimensional CNN and realized the life prediction of the aero-engine. Compared with feed-forward neural networks (FNN) and LSTM, this method has smaller variance and higher accuracy in multiple predictions. In literature [31], the object detection deep learning algorithm YOLOv3 based on CNN is used. The prediction model built by this algorithm realizes the detection of the number of people in an air-conditioned room and reduces the energy consumption of the air conditioning. Literature [32] built a predictive model for power transformer and cyberattack fault diagnosis.
This paper proposes a convolutional neural network based on the inception block; the Inception-CNN uses the convolution layer and pooling layer to extract the analytical redundancy information between each sensor. Next, this information is used to adjust network parameters and predict the probability of individual sensor failures.
The innovations of this paper are as follows.
(1)
The construction of the forecasting model
Compared to the various signal-based methods mentioned above, Inception-CNN can take advantage of the characteristics of the inception block to incorporate more interphase information of sensors of different scales into the fault detection model, which is undoubtedly beneficial to the signal-based method. In addition, the method is based on the self-iterative characteristics of neural networks; thus, it does not require a tedious parameter optimization process, and can focus more on the construction of the network model itself.
(2)
Generation of fault datasets
Actual aero-engine sensor failure data are very sparse, which contradicts the large amount of training data required for deep learning algorithms. This paper generates sufficient data for training by the Monte Carlo simulation method.
(3)
The optimization of the forecasting model input
In this paper, the one-dimensional data of the sensor are converted into a two-dimensional feature matrix by the pan and regroup method. This method makes the extraction of sensor phase information more sufficient, and the prediction model’s accuracy is improved.
(4)
The optimization of the forecasting model details
The activation function and output mode of the predictive model are optimized so that the model can perform fault detection on all target sensors simultaneously, and the accuracy of the model is improved.
The rest of the paper is organized as follows: Section 2 introduces the research object of the prediction model in this paper and the construction of training data and validation data. Section 3 shows the various parts of the CNN and how they work and then shows the specific architecture of the Inception-CNN prediction model. Section 4 verifies the feasibility and effectiveness of Inception-CNN in terms of fault detection effect and FDI process, respectively. Section 5 concludes this paper.

2. Data Preparation

This paper takes a Geared Turbofan Engine (GTF) as the research object, which can be expressed as Equation (1):
x s t , x v t = F w t , θ t
Among them, w is the operating condition, including height (alt), flight speed (Mach), power lever angle (PLA), etc. θ is a health parameter. x s is a measurable physical property, including the values of each sensor, and x v is an unobservable physical quantity.
Due to the limitation of calculation conditions, this paper only studies the ground operation of an aero-engine, namely height = 0 and speed = 0. Ten sensors are selected as the research object of this paper, as shown in Equation (2), and the cross-section position of the aero-engine is shown in Figure 1.
x s e n s o r = NL , NH , Pt 25 , Ps 3 , Pt 5 , Pt 7 , Tt 25 , Tt 36 , Tt 45 , Tt 5
In this paper, the method is tested and verified by the GTF model of NASA/T-MATS master [33] in the simulation environment of Matlab/Simulink2021a.
By referring to the verification method of the civil turbofan engine fault diagnosis system proposed by Donald et al. [34,35], the Monte Carlo simulation method is used to generate the training and verification data set of the neural network. Firstly, the power lever angle (PLA) of the healthy running engine model is given, as shown in Figure 2, and sensor data x s e n s o r are derived. Then, ten sensor data x s e n s o r are randomly disturbed by the normal distribution, as shown in Equation (3).
y s e n s o r i = x s e n s o r i + α i X X N 0 , 1
The coefficient α i is determined by the average M e a n x s e n s o r i of the input sensor value.
In the training data set and validation data set, the data point that exceeds 5% of the original value of the sensor is set as the fault point. For example, if | y s e n s o r i j x s e n s o r i j | > 0.05 × x s e n s o r i j , sensor i is considered a failure at j data points, whereas sensor i is considered to have no failure. Since class imbalance can adversely affect network training, this paper adjusts α i to reduce the gap between faulty and healthy data points and loops to generate data multiple times. Since the probability of failure of more than three sensors at the same time is extremely small, and the redundant information between the sensors decreases with the increase in the number of simultaneously failed sensors, the data points where more than three sensors fail at the same time i = 1 10 φ x s e n s o r i j > 3 are deleted. The scatter plot of a small part of the value of NL, NH, Pt25, and Tt25 sensors after adding disturbance is shown in Figure 3, where the abscissa represents the label of the data points.
In the final method verification link, we adjust the coefficient α i to ensure that the sensor data y s e n s o r are equivalent to the actual engine data collected and filtered, add the typical faults shown in Table 1, and change the PLA to verify the neural network’s generalization performance.

3. Introduction to Inception-CNN

3.1. Introduction to CNN

As a transformed form of multilayer perceptron, the convolutional neural network (CNN) was developed based on studies on the visual cortex of cats [36]. It was initially applied in the field of image recognition. Now, CNN has become a hot spot in many research fields. A typical CNN consists of convolutional layers, pooling layers, and fully connected layers, as shown in Figure 4.
Convolutional layers extract local features of the input data by a perceptual structure and reduce the number of parameters of the CNN by sharing weights. The pooling layer merges adjacent data into a single datum, reduces the dimensionality of the data, speeds up the calculation, and prevents the overfitting of the parameters. The fully connected layer is the basic unit of the BP neural network, which generates output based on the feature data extracted by the previous layer.

3.2. The Basic Module of CNN

The input of each node of the convolutional layer is only the local features of the previous layer, and the convolution kernel is used to convert the subnode matrix of the current layer into a unit node matrix with unlimited channel dimensions in the next layer. Under normal circumstances, the convolution layer generally only converts the channel dimension of the input data and adopts the method of edge zero-padding to ensure that the length and width of the input data remain unchanged. For example, the convolution kernel transforms the matrix of m 1 × n 1 × k 1 into 1 × 1 × k 2 as Equation (4).
g i = f Σ x = 1 m 1 Σ y = 1 n 1 Σ z = 1 k 1 a x , y , z × w x , y , z i + b i , 0 < i k 2
Among them, g i represents the value of the ith identity matrix of this node, and a x , y , z represents a certain length and width of the convolution kernel sampling in the input matrix of this node. The interception matrix with the channel dimension x , y , z . w x , y , z i represents the convolution kernel weight value matrix corresponding to the ith unit matrix, b i represents the bias value corresponding to the ith unit matrix, and f represents the activation function. In this paper, Scaled Exponential Linear Units (SELU) is selected as the activation function, which normalizes the data distribution and ensures that the gradient will not explode or disappear during the training process. The SELU activation function can be expressed as Equation (5):
s e l u z = λ z z > 0 α e z α z 0
where z represents the output value of the convolution operation, and λ and α are constants, λ 1.051 , α 1.673 .
Pooling layers sample the data by sliding the pooling kernel. Unlike convolutional layers, pooling layers generally only change the length and width of the input matrix. The most common types of pooling layers are max pooling and average pooling. The process of converting the m 1 × n 1 × k 1 matrix to 1 × 1 × k 1 by max layer pooling is as Equation (6):
g i = S u b s a m p l i n g a x , y i , 0 < x m 1 , 0 < y n 1 , 0 < i k 1
Among them, g i represents the value of the ith unit matrix of this node, and a x , y i represents the interception matrix of a length and width x , y sampled by the pooling kernel when the input matrix of this node corresponds to the ith unit matrix. The pooling layer reduces the data size, speeds up computation, and prevents parameter overfitting without losing data features.
Finally, the network expands the feature matrix extracted by the convolutional and pooling layers into a one-dimensional vector, which is input to the fully connected layer and produces an output based on these features.

3.3. The Calculation Process of CNN

Minimizing the loss function L W , b as much as possible is the goal of neural network training, where W and b represent the weights and bias parameters in the neural network, respectively. The loss function consists of two parts: the first part is the residual between the output value and the expected value, and the second part is the regularization loss caused by overfitting, which is regulated by the parameter θ . The loss function can be expressed as Equation (7):
L W , b = E W , b + θ 2 W T W
This paper uses the Stochastic Gradient Descent plus Momentum (SGDm) as the solver to update the CNN parameters. Through the back-propagation of the loss function, the trainable parameters of each layer in the CNN can be updated layer by layer. The mathematical representations are as Equations (8) and (9):
W i = W i 1 + m i , m i = η L W i , b i W i + β m i 1
b i = b i 1 + m i , m i = η L W i , b i b i + β m i 1
where η is the learning rate, and β is the momentum coefficient.

3.4. Inception Block

Although a deep neural network can be abstractly considered the superposition of several neural networks, different network architectures and choices of hyperparameters will make a huge difference in the performance of deep neural networks. Furthermore, with the deepening of the network, the interpretability of the deep neural network becomes worse. A deep neural network with good performance often needs to explore the design intuition formed in this field for many years, so it is vital to master and use the previous exploration results.
The inception block is the basic module of the GoogLeNet network architecture proposed in 2014 [37], and GoogLeNet prevailed in the ImageNet competition that year.
Earlier convolutional neural networks are often series architectures. The size of the convolution kernel is also very different, such as from 1 × 1 to 11 × 11 , etc. However, the inception block parallels convolutional layers with convolution kernels of different sizes in order to have the ability to extract correlated features at different scales. It has been shown that using convolution kernels of different sizes is beneficial for feature extraction.
In addition, deeper neural networks are often helpful to improve prediction accuracy. However, deep networks also bring the risk of gradient disappearance, which will cause the model to fail to converge eventually. Furthermore, due to the exponential increase in computing parameters in deeper networks, the demands on computing equipment are further increased. The inception block can reduce the network computing parameters effectively.
As shown in Figure 5, the inception block consists of four parallel paths. The first three paths use convolutional layers with kernel sizes 1 × 1 , 3 × 3 , and 5 × 5 to extract information from different spatial sizes. The two paths in the middle first perform 1 × 1 -convolution on the input to reduce the number of channels and reduce the complexity of the model. The fourth path uses a pooling layer with a pooling kernel size 3 × 3 and then uses a 1 × 1 convolution kernel to vary the number of channels. All four paths use appropriate padding to match the height and width of the input and output and, finally, join the output of each line in the channel dimension and form the output of the inception block. In the inception block, the adjusted hyperparameters are mainly the number of output channels per layer.
The specific representation of the first inception block used in this paper is shown in Figure 6. The input is an 10 × 10 matrix with a channel number of 48, which passes through four paths, respectively. After the different convolution pooling operations, which are shown in Figure 5, the number of channels is changed and merged into the output matrix in the channel dimension, as the following input to one layer of the neural network.

3.5. Architecture of Inception-CNN

Considering that the convolution kernel has the characteristics of the local receptive field, for one-dimensional data, it is difficult for the convolution kernel to extract the interphase features of the data with a large separation distance. In order to fully extract the relevant features between sensor signals, this paper performs two-dimensional expansion and recombination of the original data before the data are input into the neural network. That is, the size of the data 1 × n is expanded to n × n .
As shown in Figure 7, first, through the method described in Section 1, a one-dimensional sensor data set is obtained by collecting data from the engine model. Different colored and numbered squares represent data from different sensors. Then, the one-dimensional data group is expanded into a two-dimensional data group while rearranging the data through the data translation operation. For example, the second row of the two-dimensional data group in the figure is obtained by shifting the original one-dimensional data group by two units.
Suppose that a convolution kernel of size three is used to extract the interphase features of the sensor data. In this case, it is impossible to extract the interphase features of the sensor data with an interval greater than 3 in the original one-dimensional data set. However, it can be done in the two-dimensional data set after translation expansion.
Since the neural network model requires a large amount of data for parameter updating, this paper adopts the operating mode of offline training and online detection. The Inception-CNN architecture is shown in Figure 8. A large amount of perturbed and labeled sensor data is used as the network input during offline training, and the network parameters are updated. During online diagnosis, the data format is kept unchanged, the sensor data that match the actual situation are input to the network, and the prediction is made based on the updated parameters during offline training.
As shown in Figure 8, the two-dimensional sensor data set expanded by translation is input into the network and passes through one convolution layer, four inception blocks, two max pooling layers, and one average pooling layer. The first convolutional layer has 48 convolution kernels of size 3 × 3 , which converts the original data of size 1 × 10 × 10 into a feature response matrix of size 48 × 10 × 10 . The pooling kernel size of all pooling layers in the network is 2 × 2 . After the last average pooling, the three-dimensional matrix is converted to a one-dimensional vector with no information loss. Finally, the extracted feature response vector is input to the fully connected layer and produces an output with 20 nodes.
The parameters of the four inception block convolution kernels and pooling kernels are shown in Figure 5. The changed hyperparameters are only the number of channels output by each layer. The channel parameters are shown in Table 2.

3.6. Loss Function

In the traditional multiclassification problem using neural networks, if the classification category is n, it is necessary to construct an n-dimensional one-hot encoding vector to represent the category, and network output nodes are required. For example, in the three-classification problem, the three one-hot vectors 1 , 0 , 0 0 , 1 , 0 0 , 0 , 1 represent three different categories. It can be seen that one element in the vector is 1, and the others are 0.
y i = e a i k = 1 C e a k i 1 C
Equation (10) is the softmax function expression, where a is the output value of the last fully connected layer. By calculating the natural logarithm, the softmax function converts the network’s output into the probability of each category on the one-hot encoding, i.e., p 1 , p 2 , p 3 . Among them, p 1 + p 2 + p 3 = 1 . The softmax function is widely used in multi-classification problems due to its fit with one-hot encoding and its ability to widen the difference between categories in most cases.
However, in the sensor fault detection problem in this paper, due to the assumption that multiple sensors fail simultaneously, the detection vector output by the network will present multiple 1s at the same time. For example, 1 , 0 , 1 , 0 , 0 , 0 , 0 , 0 , 0 , 0 means that the NL sensor and the Pt25 sensor fail simultaneously, so the softmax function is no longer applicable in the network, and the sensor failure probability is only calculated during model validation.
For multi-classification problems, a combination of the softmax function and cross-entropy loss function is often used. The expression of the cross-entropy loss function is as Equation (11).
E W , b = 1 N i c = 1 M y i c log p i c
Among them, M is the number of categories, y is the symbolic function, i is the batch data size, and p is the predicted probability of the category, which is the output of softmax. Since softmax is no longer applicable, the output of the fully connected layer a is replaced with p.
There are two types of individual sensors, faulty and healthy, and one-hot encoding is constructed for each sensor separately, so the number of final output nodes of Inception-CNN is 10 × 2 , a total of 20 output nodes. Every two nodes constitute a prediction unit of the sensor. The output value of the cross-entropy loss function of each prediction unit is accumulated as a total loss value for the back-propagation of the network.
In summary, the loss function expression used in this paper is as Equation (12):
E W , b = j = 1 10 1 N i c = 1 M y i c log a i c
where j is the number of the prediction unit, and a is the output of the last fully connected layer of the network.

4. Validation and Analysis

4.1. Training the Networks

The Inception-CNN and other comparative neural networks in this paper are trained in the simulation environment of Python3.9/Pytorch1.10.1. The rest of the non-neural network machine learning algorithms are trained in the simulation environment of Matlab2021a. For the data set constructed in Section 1, the accuracy of some neural networks during the training process is shown in Figure 9, and the accuracy of each algorithm when it converges is shown in Table 3:
Each neural network sets the same training hyperparameters. The learning rate is 0.0001, and the batch size is 256.
It can be seen from Table 3 that the accuracy of the neural network is significantly improved compared to other non-neural network algorithms. For example, a simple BP neural network improves the accuracy by 4.58% compared to the best-performing Gaussian SVM in a non-neural network. Compared with the simple BP neural network, the accuracy of the simple four-layer CNN network is increased by 2.69%, so the CNN fits this data set better than other algorithms; thus, the improved CNN was chosen as the fault detection algorithm in this paper.
Figure 9 takes Epoch as the abscissa and the ordinate as the accuracy of each network on the test data set. The first three CNN networks use the mean square error (MSE) loss function, and the fourth network uses the improved cross-entropy loss function in the paper. It can be seen from Figure 9 and Table 3 that the CNN based on the inception block has a rapid increase in accuracy and a higher upper limit than the simple four-layer CNN. After replacing the traditional ReLU activation function with the SELU activation function, the convergence time of the network is reduced, and the training time is reduced by approximately 60%. After replacing MSE with the improved cross-entropy loss function in this paper, the network’s sensitivity to wrong predictions is improved. The accuracy rate rises faster, and the accuracy during convergence is 3.28% higher than that of Inception-CNN using the MSE loss function. The final accuracy of Inception-CNN used in this paper is 95.41%.

4.2. Typical Sensor Failure Validation

The typical sensor fault verification link is divided into two parts. On the one hand, it tests the detection effect of Inception-CNN itself for typical faults, and on the other hand, it verifies the effectiveness of the network in the FDI process.
Excellent generalization performance is an important indicator to measure the performance of neural network models. For the research content of this paper, even if the data set for training Inception-CNN is only composed of a simple slope response after adding disturbance, it is still expected to have good prediction results for all ground operating conditions of the engine model.
Therefore, in a typical sensor failure verification process, the steady-state or transition-state conditions used for each verification are different.

4.2.1. Validation of Fault Detection Effect

First, we verify the case of a single fault and simultaneous multiple faults in the steady state of the engine.
A single sensor failure occurs at a steady state, as shown in Figure 10. Under different steady-state conditions, the impulse, drift, and bias fault are added to the NL, Pt25, and Tt36 sensor, respectively. It can be seen from the figure that when the fault probability is 80% as the threshold, the injected impulse fault, drift fault, and bias fault are identified by the fault detection system after delays of around 0.1 s, 4.2 s, and 0 s, respectively. The drift fault has a relatively long detection delay due to the small offset in the early stage when the fault occurs. In addition, after injecting faults, the failure probabilities of other sensors without faults are all below 6% and fluctuate much less than the threshold.
As described in Section 1, when constructing the neural network training data, this paper removes data points where more than three sensors fail simultaneously. Thus, the simultaneous failure of up to three sensors is the assumption of this paper.
The simultaneous failure of multiple sensors in a steady state is shown in Figure 11. Under a specific steady-state condition, the bias, impulse, and drift fault are injected into the NH, Ps3, and Tt25 sensor at approximately 5 s, 6 s, and 9 s, respectively. The detection delay of each fault is around 0 s, 1.6 s, and 2.5 s, respectively. Among them, the delay of drift fault detection is still caused by insufficient offset in the early stage of fault occurrence. As can be seen from the figure on the right, although multiple faults occur simultaneously, it does not affect the accurate identification of a single fault by the fault detection system, and the detection delay is hardly affected by multi-sensor faults. Among the non-faulty sensors, the failure probability of the Pt25 sensor fluctuates up to approximately 19%, but it is still far below the failure probability threshold.
Next, we verify the case of a single fault and simultaneous multiple faults in the transition state of the engine.
A single sensor failure occurs at a transition state, as shown in Figure 12. Under different transition state conditions, the impulse, drift, and bias fault were injected into the NH, Pt5, and Tt45 sensor, respectively. The situation is similar to a single sensor failure in a steady state. The detection system identifies each fault after delaying by 0.1 s, 2.6 s, and 0 s, respectively. It can be seen that the performance of Inception-CNN in the transition state and steady state is comparable.
The simultaneous failure of multiple sensors in a transition state is shown in Figure 13. Under a specific steady-state condition, the drift, impulse, and bias fault are injected into the NL, Pt7, and Tt5 sensor at approximately 5 s, 3 s, and 9 s, respectively. The detection delay of each fault is approximately 5.9 s, 0.2 s, and 0 s, respectively. The delay in detecting drift faults is due to the fact that the injected drift fault offset is small, and the drift fault itself has the characteristic of slowly increasing the offset.

4.2.2. Validation in the FDI Process

As shown in Figure 14, the FDI process consists of a fault detection module (FDM) and a fault isolate module (FIM). The FDM contains fault detection estimators (FDE) running in real time, and the FIM contains fault isolate estimators (FIE) that require the output of the FDM to activate. The essence of each estimator is a method of fault detection or isolation. The Inception-CNN proposed in this paper is an FDE in the FDI process.
The Gaussian process regression algorithm is used to build the FIE module.
Due to space limitations, the FDI process verification is only carried out for the simultaneous failure of multiple sensors. First, we verify the FDI flow of the engine at a steady state.
As shown in Figure 15, when multiple sensors fail simultaneously in a steady state, FDM effectively activates FIM through the built-in Inception-CNN. After the FDI process, the injected fault offsets of drift, impulse, and bias faults are reduced by around 81%, 61%, and 100%, respectively, which correspond to different faults’ detection delays. The lower impulse fault offset reduction percentage is the lower injected impulse fault offset.
When multiple sensors fail simultaneously in the transition state, it is similar to the steady state, as shown in Figure 16. After the FDI process, the drift, impulse, and bias fault offsets were reduced by 76%, 56%, and 100%, respectively. The effectiveness of Inception-CNN in the FDI process is verified.

5. Conclusions

In this paper, a convolutional neural network based on an inception block is proposed and utilized for aero-engine sensor fault detection. The traditional detection unit composed of multiple fusion algorithms is simplified into one detection algorithm, which mainly solves the problems of traditional sensor FDI methods with many parameters and complex systems.
The effectiveness and feasibility of the method are verified by the detection effect and the FDI process. On the data set of this paper, the detection accuracy of Inception-CNN is 95.41%, which improves the prediction accuracy by 17.27% and 12.69% compared with the best-performing non-neural network algorithm and simple BP neural networks tested in the paper, respectively.
In addition, this paper constructs the training data set and validation data set of the signal-based sensor FDI method through the Monte Carlo simulation method, which solves the problem wherein the experiment cannot be carried out due to insufficient fault data.
The sensor fault detection method based on Inception-CNN proposed in this paper is a data-driven algorithm. Its accuracy and applicability are positively related to the quality of the data. Thus, in the future, this method can be combined with the mechanism study of the engine to improve the algorithm’s performance through higher-quality data. In addition, the research content of this paper can be combined with the research on the safety control strategy of aero-engines based on sensor value judgment. Taking the research [38] of Cao et al. as an example, if FDI is used as the front module of the safety protection control module in the control loop, the possibility of control strategy failure due to sensor failure can be reduced, and the robustness of the system can be improved.

Author Contributions

Conceptualization, X.D. and J.W.; Data curation, X.D.; Investigation, X.D.; Methodology, X.D., J.C. and J.W.; Project administration, X.D.; Resources, X.D., H.Z. and J.W.; Software, X.D.; Supervision, J.W.; Validation, X.D.; Visualization, X.D.; Writing—original draft, X.D. and J.C.; Writing—review and editing, X.D., J.C., H.Z. and J.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Central Military Commission Foundation to Strengthen Program Technology Fund (No. 2019-JCJQ-JJ-347); Central Military Commission Special Fund for Defense Science, Technology and Innovation (20-163-00-TS-009-096-01); Aviation Science Fund of China-Xi’an 631 Research Institute (No. 201919052001); Kunpeng Talent Plan of Zhejiang Province (E10938DL03); and Startup Foundation of Ningbo Institute of Materials Technology and Engineering, CAS (E10921RA06).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the first author or corresponding author.

Acknowledgments

The authors would like to acknowledge the open-source developers of the deep learning framework and graphing tools used in this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Garg, S. Controls and Health Management Technologies for Intelligent Aerospace Propulsion Systems. In Proceedings of the 42nd AIAA Aerospace Sciences Meeting and Exhibit, Reno, Nevada, 5–8 January 2004; American Institute of Aeronautics and Astronautics: Reno, Nevada, 2004. [Google Scholar] [CrossRef] [Green Version]
  2. Jaw, L.; Mattingly, J. Aircraft Engine Controls; American Institute of Aeronautics and Astronautics, Inc.: Washington, DC, USA, 2009. [Google Scholar] [CrossRef]
  3. He, B.; Yu, D.; Shi, X. Application of simulation model of sensors to analysis fault of control system of turbojet engine. J. Propuls. Technol. 2001, 5, 364–367. [Google Scholar] [CrossRef]
  4. Giantomassi, A.; Ferracuti, F.; Iarlori, S.; Ippoliti, G.; Longhi, S. Electric Motor Fault Detection and Diagnosis by Kernel Density Estimation and Kullback-Leibler Divergence Based on Stator Current Measurements. IEEE Trans. Ind. Electron. 2015, 62, 1770–1780. [Google Scholar] [CrossRef]
  5. Dai, X.; Gao, Z. From Model, Signal to Knowledge: A Data-Driven Perspective of Fault Detection and Diagnosis. IEEE Trans. Ind. Inf. 2013, 9, 2226–2238. [Google Scholar] [CrossRef] [Green Version]
  6. Yuan, C.; Yao, H. Aero-engine adaptive model re-construction under sensor failure. J. Aerosp. Power 2006, 21, 195–199. [Google Scholar] [CrossRef]
  7. Wei, X.; Yingqing, G. Multiple Sensors Soft Failure Diagnosis for Aircraft Engine Control System. Comput. Meas. Control 2007, 15, 585–587. [Google Scholar] [CrossRef]
  8. Zhao, C.; Ye, Z.; Wang, J.; Yin, B. Rapid prototype real-time simulation of turbo-fan engine sensor fault diagnosis. J. Aerosp. Power 2014, 29, 451–457. [Google Scholar] [CrossRef]
  9. Han, B.; Gou, L.; Mao, N.; Wang, X. A method of multiple fault diagnosis based on fault matching. Aeronaut. Comput. Tech. 2015, 45, 96–99. [Google Scholar]
  10. Lu, F.; Chen, Y.; Huang, J.; Zhang, D.; Liu, N. An Integrated Nonlinear Model-Based Approach to Gas Turbine Engine Sensor Fault Diagnostics. Proc. Inst. Mech. Eng. Part G J. Aerosp. Eng. 2014, 228, 2007–2021. [Google Scholar] [CrossRef]
  11. Tadić, P.; Ðurović, Ž. Particle Filtering for Sensor Fault Diagnosis and Identification in Nonlinear Plants. J. Process Control 2014, 24, 401–409. [Google Scholar] [CrossRef]
  12. Liu, R.; Meng, G.; Yang, B.; Sun, C.; Chen, X. Dislocated Time Series Convolutional Neural Architecture: An Intelligent Fault Diagnosis Approach for Electric Machine. IEEE Trans. Ind. Inf. 2017, 13, 1310–1320. [Google Scholar] [CrossRef]
  13. Cai, K.; Sun, Y.; Yao, W. Fault diagnosis and adaptive reconfiguration control for sensors in aeroengine. Electron. Opt. Control 2009, 23, 1118–1126. [Google Scholar]
  14. Duan, S.; Li, Q.; Zhao, Y. Fault diagnosis for sensors of aero-engine based on improved least squares support vector regression. In Proceedings of the 2011 Eighth International Conference on Fuzzy Systems and Knowledge Discovery (FSKD), Shanghai, China, 26–28 July 2011; Volume 3, pp. 1962–1966. [Google Scholar] [CrossRef]
  15. Lu, F.; Huang, J.; Chen, Y.; Song, Y. Research on sensor fault diagnosis of aero-engine based on data fusion of spso-svr. J. Aerosp. Power 2009, 24, 1856–1865. [Google Scholar] [CrossRef]
  16. Liu, Y.; Yang, D.; Liu, Y.; Kang, J. Failure diagnose research for the plane engine of basic neural nework. Electron. Des. Eng. 2012, 20, 89–92. [Google Scholar] [CrossRef]
  17. Xian, N. Aero-engine fault diagnosis method based on abc-bp neural network. Equip. Manuf. Technol. 2018, 5, 173–175. [Google Scholar]
  18. Jianliang, A.; Yang, X. Fault diagnosis of aero-engine based on self-adaptive neural network. Sci. Sin. Technol. 2018, 48, 326–335. [Google Scholar]
  19. ChangZheng, L.; Zhang, Y. Sensor fault detection based on general regression neural network. J. Propuls. Technol. 2017, 38, 2130–2137. [Google Scholar] [CrossRef]
  20. Lv, S.; Guo, Y.; Sun, H. Aero-engine sensor data preprocessing based on sdq algorithm of ga-aann neural network. J. Propuls. Technol. 2018, 39, 1142–1150. [Google Scholar] [CrossRef]
  21. Cui, J.; Liu, H.; Tao, S.; Yu, M.; Gao, Y.; Automation, S. Aeroengine fault diagnosis method based on elm. Fire Control Command Control 2018, 43, 113–121. [Google Scholar]
  22. Yao, H. Full Authority Digital Electronic Control System for Aero-Engine; Aviation Industry Press: Beijing, China, 2014. [Google Scholar]
  23. Li, Y.; Huang, Y.; Zhang, M. Short-Term Load Forecasting for Electric Vehicle Charging Station Based on Niche Immunity Lion Algorithm and Convolutional Neural Network. Energies 2018, 11, 1253. [Google Scholar] [CrossRef] [Green Version]
  24. Ye, A.; Zhou, X.; Miao, F. Innovative Hyperspectral Image Classification Approach Using Optimized CNN and ELM. Electronics 2022, 11, 775. [Google Scholar] [CrossRef]
  25. Duarte Soares, L.; de Souza Queiroz, A.; López, G.P.; Carreño-Franco, E.M.; López-Lezama, J.M.; Muñoz-Galeano, N. BiGRU-CNN Neural Network Applied to Electric Energy Theft Detection. Electronics 2022, 11, 693. [Google Scholar] [CrossRef]
  26. Ullah, A.; Javaid, N.; Samuel, O.; Imran, M.; Shoaib, M. CNN and GRU Based Deep Neural Network for Electricity Theft Detection to Secure Smart Grid. In Proceedings of the 2020 International Wireless Communications and Mobile Computing (IWCMC), Limassol, Cyprus, 15–19 June 2020; IEEE: Limassol, Cyprus, 2020; pp. 1598–1602. [Google Scholar] [CrossRef]
  27. Ince, T.; Kiranyaz, S.; Eren, L.; Askar, M.; Gabbouj, M. Real-Time Motor Fault Detection by 1-D Convolutional Neural Networks. IEEE Trans. Ind. Electron. 2016, 63, 7067–7075. [Google Scholar] [CrossRef]
  28. Ebrahimkhanlou, A.; Salamone, S. Single-Sensor Acoustic Emission Source Localization in Plate-Like Structures Using Deep Learning. Aerospace 2018, 5, 50. [Google Scholar] [CrossRef] [Green Version]
  29. Wang, H.; Li, S.; Song, L.; Cui, L.; Wang, P. An Enhanced Intelligent Diagnosis Method Based on Multi-Sensor Image Fusion via Improved Deep Learning Network. IEEE Trans. Instrum. Meas. 2020, 69, 2648–2657. [Google Scholar] [CrossRef]
  30. Arias Chao, M.; Kulkarni, C.; Goebel, K.; Fink, O. Fusing Physics-Based and Deep Learning Models for Prognostics. Reliab. Eng. Syst. Saf. 2022, 217, 107961. [Google Scholar] [CrossRef]
  31. Elsisi, M.; Tran, M.-Q.; Mahmoud, K.; Lehtonen, M.; Darwish, M.M.F. Deep Learning-Based Industry 4.0 and Internet of Things towards Effective Energy Management for Smart Buildings. Sensors 2021, 21, 1038. [Google Scholar] [CrossRef]
  32. Elsisi, M.; Tran, M.; Mahmoud, K.; Mansour, D.-E.A.; Lehtonen, M.; Darwish, M.M.F. Effective IoT-Based Deep Learning Platform for Online Fault Diagnosis of Power Transformers against Cyberattacks and Data Uncertainties. Measurement 2022, 190, 110686. [Google Scholar] [CrossRef]
  33. Chapman, J.W.; Lavelle, T.M. Toolbox for the Modeling and Analysis of Thermodynamic Systems (T-MATS) User’s Guide; NASA: Washington, DC, USA, 2014; Volume 50, Available online: https://ntrs.nasa.gov/api/citations/20140012486/downloads/20140012486.pdf (accessed on 10 September 2021).
  34. Simon, D.L.; Bird, J.; Davison, C.; Volponi, A.; Iverson, R.E. Benchmarking Gas Path Diagnostic Methods: A Public Approach. In Volume 2: Controls, Diagnostics and Instrumentation; Cycle Innovations; Electric Power; ASMEDC: Berlin, Germany, 2008; pp. 325–336. [Google Scholar] [CrossRef] [Green Version]
  35. Simon, D.L. Propulsion Diagnostic Method Evaluation Strategy (ProDiMES) User’s Guide; NASA: Washington, DC, USA, 2010; Volume 47. Available online: https://ntrs.nasa.gov/api/citations/20100005639/downloads/20100005639.pdf (accessed on 15 September 2021).
  36. Hubel, D.H.; Wiesel, T.N. Receptive Fields, Binocular Interaction and Functional Architecture in the Cat’s Visual Cortex. J. Physiol. 1962, 160, 106–154. [Google Scholar] [CrossRef] [PubMed]
  37. Szegedy, C.; Liu, W.; Ji, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; IEEE: Boston, MA, USA, 2015; pp. 1–9. [Google Scholar] [CrossRef] [Green Version]
  38. Chen, C.; Wang, M.; Dimirovski, G.M.; Zhao, J. Safety Protection Control for Aeroengines Based on Finite Times of Controller Switches. In Proceedings of the 31st Chinese Control Conference, Hefei, China, 25–27 July 2012; IEEE: Beijing, China, 2012; pp. 2071–2076. [Google Scholar]
Figure 1. Schematic diagram of the engine structure and the position of the sensors used for detection. Abbreviations: low-pressure compressor (LPC), high-pressure compressor (HPC), high-pressure turbine (HPT), low-pressure turbine (LPT).
Figure 1. Schematic diagram of the engine structure and the position of the sensors used for detection. Abbreviations: low-pressure compressor (LPC), high-pressure compressor (HPC), high-pressure turbine (HPT), low-pressure turbine (LPT).
Aerospace 09 00236 g001
Figure 2. The PLA used by the model to generate the data.
Figure 2. The PLA used by the model to generate the data.
Aerospace 09 00236 g002
Figure 3. Scatter plot of some sensor data after adding perturbation.
Figure 3. Scatter plot of some sensor data after adding perturbation.
Aerospace 09 00236 g003
Figure 4. Schematic diagram of a typical convolutional neural network.
Figure 4. Schematic diagram of a typical convolutional neural network.
Aerospace 09 00236 g004
Figure 5. Schematic diagram of the inception block.
Figure 5. Schematic diagram of the inception block.
Aerospace 09 00236 g005
Figure 6. Concrete representation of an inception block.
Figure 6. Concrete representation of an inception block.
Aerospace 09 00236 g006
Figure 7. Expansion and recombination of sensor data.
Figure 7. Expansion and recombination of sensor data.
Aerospace 09 00236 g007
Figure 8. The overall architecture of Inception-CNN.
Figure 8. The overall architecture of Inception-CNN.
Aerospace 09 00236 g008
Figure 9. Accuracy changes during some neural networks’ training.
Figure 9. Accuracy changes during some neural networks’ training.
Aerospace 09 00236 g009
Figure 10. Detection of single sensor failure in steady state.
Figure 10. Detection of single sensor failure in steady state.
Aerospace 09 00236 g010
Figure 11. Detection of simultaneous failure of multiple sensors in steady state.
Figure 11. Detection of simultaneous failure of multiple sensors in steady state.
Aerospace 09 00236 g011
Figure 12. Detection of single sensor failure in transition state.
Figure 12. Detection of single sensor failure in transition state.
Aerospace 09 00236 g012
Figure 13. Detection of simultaneous failure of multiple sensors in transition state.
Figure 13. Detection of simultaneous failure of multiple sensors in transition state.
Aerospace 09 00236 g013
Figure 14. Schematic diagram of FDI process.
Figure 14. Schematic diagram of FDI process.
Aerospace 09 00236 g014
Figure 15. Validation of the FDI process with simultaneous multi-sensor failures in steady state.
Figure 15. Validation of the FDI process with simultaneous multi-sensor failures in steady state.
Aerospace 09 00236 g015
Figure 16. Validation of the FDI process with simultaneous multi-sensor failures in transition state.
Figure 16. Validation of the FDI process with simultaneous multi-sensor failures in transition state.
Aerospace 09 00236 g016
Table 1. Typical sensor failure.
Table 1. Typical sensor failure.
Type of FailureCause of Failure
ImpulseRandom disturbances, surges, sparks, etc.
DriftSensor component deterioration, etc.
BiasBias current, bias voltage, etc.
Table 2. Channel parameters for each inception block.
Table 2. Channel parameters for each inception block.
Serial NumberPath 1Path 2Path 3Path 4Number of Output Channels
11624.324.848.864
23232.488.2464.16120
34824.524.12120.16128
44028.566.16128.16128
Table 3. Performance comparison of Inception-CNN and other algorithms.
Table 3. Performance comparison of Inception-CNN and other algorithms.
AlgorithmActivation FunctionLoss FunctionAccuracy at Convergence
Gaussian Naive Bayes\\67.81%
Cubic SVM\\73.35%
Cosine KNN\\75.54%
Gaussian SVM\\78.14%
BP Neural NetworksReLU (Rectified Linear Unit)MSE82.72%
Simple 4-Tier CNNReLUMSE85.41%
Inception-CNNReLUMSE91.82%
Inception-CNNSELUMSE92.13%
Inception-CNNSELUCross-Entropy95.41%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Du, X.; Chen, J.; Zhang, H.; Wang, J. Fault Detection of Aero-Engine Sensor Based on Inception-CNN. Aerospace 2022, 9, 236. https://doi.org/10.3390/aerospace9050236

AMA Style

Du X, Chen J, Zhang H, Wang J. Fault Detection of Aero-Engine Sensor Based on Inception-CNN. Aerospace. 2022; 9(5):236. https://doi.org/10.3390/aerospace9050236

Chicago/Turabian Style

Du, Xiao, Jiajie Chen, Haibo Zhang, and Jiqiang Wang. 2022. "Fault Detection of Aero-Engine Sensor Based on Inception-CNN" Aerospace 9, no. 5: 236. https://doi.org/10.3390/aerospace9050236

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop