Next Article in Journal
DisKnow: A Social-Driven Disaster Support Knowledge Extraction System
Next Article in Special Issue
Bearing Fault Diagnosis Based on Improved Convolutional Deep Belief Network
Previous Article in Journal
Speech Enhancement for Hearing Aids with Deep Learning on Environmental Noises
Previous Article in Special Issue
Bearing Fault Identification Using Machine Learning and Adaptive Cascade Fault Observer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Stacked Auto Encoders Sparse Filter Rotating Component Comprehensive Diagnosis Network for Extracting Domain Invariant Features

1
College of Energy and Power Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 210000, China
2
College of Mechanical and Electronic Engineering, Shandong University of Science Technology, Qingdao 266000, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(17), 6084; https://doi.org/10.3390/app10176084
Submission received: 14 July 2020 / Revised: 7 August 2020 / Accepted: 28 August 2020 / Published: 2 September 2020
(This article belongs to the Special Issue Bearing Fault Detection and Diagnosis)

Abstract

:
In recent years, the method of deep learning has been widely used in the field of fault diagnosis of mechanical equipment due to its strong feature extraction and other advantages such as high efficiency, portability, and so on. However, at present, most kinds of intelligent fault diagnosis algorithms mainly focus on the diagnosis of a single fault component, and few intelligent diagnosis models can simultaneously carry out comprehensive fault diagnosis for a rotating system composed of a shaft, bearing, gear, and so on. In order to solve this problem, a novel stacked auto encoders sparse filter rotating component comprehensive diagnosis network (SAFC) was proposed to extract domain invariant features of various health conditions at different speeds. The model clusters domain invariant features at different speeds through the self-coding network, and then classifies fault types of various parts through sparse filtering. The SAFC model was validated by the vibration data collected, and the results show that this model has higher diagnostic performance than other models.

1. Introduction

With the coming of the IoT (Internet of Things) era, the reliability of mechanical equipment is more and more demanding [1]. The rotating components of mechanical equipment such as shafts, bearings, and gears are prone to shaft cracks, bearing wear, and gear breakage, which have great safety hazards [2]. In order to ensure the safe and reliable operation of equipment, accurate and efficient fault diagnosis technology is particularly important. As a powerful tool for processing big data, the deep learning [3,4] model has been successfully applied to mechanical intelligent fault diagnosis.
In the field of intelligent fault diagnosis of rotating parts, many scholars have studied related diagnosis by using various deep learning methods. Deng et al. [5] used the features extracted from the time domain, frequency domain, and time–frequency domain of vibration signals as deep Boltzmann machine (DBM) inputs to realize the accurate classification of seven different bearing faults. Ding et al. [6] adopted the wavelet packet energy map as the input signal and designed the depth convolutional neural networks (CNN) for bearing fault diagnosis. Meanwhile, in order to better display the multi-layer representation of the neural network, a multi-scale layer was added behind the final convolutional layer as the connection between the output of the final convolutional layer and the previous pooling layer. Li et al. [7] first implemented wavelet packet transform (WPT) on the original vibration signal, and then input the extracted statistical characteristics into a two-layer DBM to realize the fault diagnosis of the gear box, and the classification rate of 11 operating modes reached 97.68%. Liu et al. [8] proposed a method based on variational mode decomposition (VMD) and CNN for feature extraction and fault diagnosis of planetary gear for local weak feature information, and realized the identification and classification of planetary gear weak fault state, where VMD based partition extraction method was better than EEMD, which could achieve 100% total CNN recognition rate with less training time (14 times). Chen et al. [9] not only adopted the statistical features in the time domain and frequency domain, but also added the test values of load and speed to form a set of feature vectors as the input of the deep belief network (DBN), realizing the fault diagnosis of the gear box.
Even more exciting, stacked auto encoders (SAE) [10], as a widely used deep learning algorithm, has attracted attention in the field of fault diagnosis. Feng Jia et al. [11] established a five-layer SAE and used frequency domain signals as input to diagnose the rotating bearing faults under different loads and speeds, with the diagnosis accuracy reaching almost 100%. Ngiam et al. [12] proposed an unsupervised feature learning method: sparse filtering (SF). This algorithm only focuses on optimizing the sparse performance of learning features while ignoring the distribution of learning data. Ya Guo Lei et al. [13] introduced sparse filtering into mechanical intelligent fault diagnosis. First, sparse filtering was used to extract unsupervised features of the time-domain signals of bearings, and then softmax was used as a classifier to realize fault diagnosis and good results were obtained.
For various fault diagnosis components, the main intelligent diagnosis algorithm was to detect the fault of the bearing and gear, respectively. Chen et al. [14] selected the method of adding Gaussian noise to the training data for data enhancement, and then input it into SAE for bearing fault diagnosis. Li Heng et al. [15] obtained spectrum samples by short-time Fourier transform (STFT) operation of the bearing vibration signals, and then input them into CNN for network training. The results showed that this method had high identification accuracy for different bearing fault types and could improve the robustness of this method by increasing the type and number of fault samples. Wang Weifeng et al. [16] proposed a fault diagnosis method based on Bi LSTM (binary long short term memory) network in order to solve problems such as large amounts of data and difficulty in extracting features in gear fault diagnosis. The results showed that the Bi LSTM network method for gear fault diagnosis was better than the CNN and LSTM network.
From what has been discussed above, it can be seen that the method of deep learning has been widely used in the field of fault diagnosis of mechanical equipment due to its strong feature extraction and other advantages such as its high efficiency, portability, and so on [17]. However, at present, all kinds of intelligent fault diagnosis algorithms mainly focus on the diagnosis of a single fault component, and few intelligent diagnosis models can simultaneously carry out comprehensive fault diagnosis for a rotating system composed of a shaft, bearing, gear, and so on. In order to solve this problem, a novel stacked auto encoders sparse filter rotating component comprehensive diagnosis model (SAFC) was proposed to extract domain invariant features of various health conditions at different speeds and classify them.
The proposed model can be used to extract domain invariant features of various health conditions at different speeds. Using the spectrum of fault signal as input, the deep neural network (DNN) was established through SAE, and the back propagation (BP) algorithm was used to fine-adjust the network. Meanwhile, the batch standardization (BN) algorithm was introduced into DNN to realize the rapid training of intelligent fault diagnosis model of rotating parts. After the clustering and dimension reduction of the data, the data was input into the sparse filtering model, and the unsupervised feature extraction of the input signal was carried out by using the sparse filtering. Finally, softmax was used as the classifier to realize fault diagnosis. The SAFC model was validated by the vibration data collected, and the results show that the proposed model has high diagnostic performance.
The main innovations are as follows:
(1)
A novel SAFC model that can extract and classify fault signal domain invariant features is proposed.
(2)
The proposed method is able to process multiple rotating parts at the same time and has the ability to process large amounts of data.
(3)
The domain invariant features of the same fault at different speeds can be extracted and clustered.
The organizational structure of the rest of this paper is as follows. Section 2 briefly introduces the SAE and SF theories, respectively; Section 3 introduces the architecture and training methods of SAFC model; Section 4 verifies the superiority of the SAFC model through two groups of experiments; Section 5 is the conclusion.

2. Introduction to Theory

2.1. Stacked Auto Encoders

The basic component unit of a SAE is an automatic encoder (AE) structure, and an AE is a symmetrical three-layer neural network consisting of an input layer, hidden layer, and output layer. The structure is shown in Figure 1. The learning process includes encoding and decoding: encoding is used to map input signals into hidden layer expression, while decoding is a reconstruction of hidden layer expression [18]. In the process of reconstruction, a reconstruction error is set to minimize the reconstruction error through iterative training to obtain the best hidden layer data expression.
To build an AE, three things need to be done: building an encoder, building a decoder, and setting up a loss function measuring the information lost due to compression. The process of encoding and decoding usually adopts a parametric equation, and the parameter optimization process can be realized by minimizing the loss function.
If the input is a tableless dataset { x n } n = 1 N , in which x n R m × l , h n is the hidden layer encoding vectors, and x ^ n is the output layer decoding vector. Therefore, the encoding process is as follows:
h n = f ( W x n + b 1 )
where f ( ) is the coded activation function [19]; W is the coded weight matrix; and b 1 is the offset vector.
The decoding process is:
x ^ n = g ( W T h n + b 2 )
where g ( ) is the coded activation function; W T is the coded weight matrix; and b 2 is the offset vector.
The network parameter set is optimized by minimizing the reconstruction error:
Ø ( Θ ) = arg min 1 n i = 1 n L ( x i , x ^ i )
where L is loss function: L ( x i , x ^ i ) = x i - x ^ i 2 .
AE was superimposed layer by layer, that is, the first hidden layer was the second input, thus forming the DNN model structure based on SAE, as shown in Figure 2. The constituted DNN first carries out the layer by layer training of the forward propagation realization network, and realizes the network weight updating and parameter fine-tuning by the BP algorithm. When given an input signal, the input layer and the first hidden layer of DNN are regarded as the coding network of the first AE. The parameters obtained after the first AE training are then used to initialize the second hidden layer of DNN. The calculation process of the first coding vector of the input signal is as follows:
h n 1 = f ( x n )
Then, the encoding vector was used as input data to train the second AE, that is, the first hidden layer and the second hidden layer of the DNN were used as the encoding network of the second AE. Accordingly, the second hidden layer of DNN is initialized by the second AE, step by step until the last AE completes the training:
h n M = f ( h n M 1 )
After pre-training the superimposed AE, the softmax classification layer containing sample tags was used as the DNN output layer, followed by SAE, and then the BP algorithm was used to reverse fine-tune the DNN parameters layer by layer. Therefore, the DNN output calculated from the input signal was:
y n = f ( h n M )

2.2. Batch Normalization

The batch normalization can reparameterize almost all DNN in an elegant way. The process can be applied to each activation layer without parameter adjustment. It can be standardized in an independent way for each row of a matrix with zero mean and unit variance. Suppose an n-dimensional input x = ( x 1 x n ) in order to improve training and reduce the internal covariant transfer problem, the batch standardization algorithm adopts two necessary simplification steps:
First, normalize each unit:
x ^ i = x i E [ x i ] Var [ x i ]
where E [ x i ] is the mean value of each unit and Var [ x i ] is the standard deviation. Then, translate and scale the normalized value:
f i = γ i x ^ i + β i
where γ i represents the scaling parameter and β i represents the translation parameter.
Second, assume that there are m values in x i , namely the minimum batch is ϕ = x 1 m , x ^ 1 m represents the normalized value, and the corresponding linear transformation is y 1 m . Therefore, the batch standardization transformation is B N γ , β : x 1 m y 1 m , namely:
E [ x ϕ ] = 1 m j = 1 m x j
Var [ x ϕ ] = 1 m j = 1 m ( x j E [ x ϕ ] ) 2
x ^ j = x j E [ x ϕ ] Var [ x ϕ ] + ε
y ^ j = γ x ^ j + β
where ε is an imposed constant to avoid undefined gradient problems of s when s = 0.
Furthermore, gradient loss is back propagated in batch standardized training:
x ^ j = y j γ
Var [ x ϕ ] = j = 1 m x ^ j ( x j E [ x ϕ ] ) [ 1 2 ( Var [ x ϕ ] + ε ) 3 2 ]
E [ x ϕ ] = j = 1 m x ^ j 1 Var [ x ϕ ] + ε
x j = x ^ j 1 Var [ x ϕ ] + ε + Var [ x ϕ ] 2 ( x j E [ x ϕ ] ) m + E [ x ϕ ] 1 m
γ = j = 1 m y j x ^ j
β = j = 1 m y j
Batch normalized transformation introduces standardized activation into each layer of the network to ensure that all layers of the network can continuously learn the input distribution to reduce the problem of internal covariate transfer, so an easy initial environment can be established at the beginning of network training to speed up network training.
Applying the batch normalization to each activation layer of the SAE. Batch normalization transformations are applied to normalization W x n + b 1 in Equation (1). The bias is considered to be subtracted in the subsequent mean subtraction process, so the bias was not used here. Therefore, Equation (1) can be changed to:
h n = f ( BN ( W x n ) )
In addition, the saturation problem can lead to the disappearance of the gradient by using linear rectifier unit function (rectified linear units, ReLU) to solve.
Re LU ( x ) = max ( x , 0 )
The ReLU function can make it easier for the network to learn optimization during training. Compared with the traditional sigmoid and tanh functions, its advantages are as follows:
(1)
Due to the nonlinear characteristics of sigmoid and tanh functions, both of them carry out exponential operations, leading to an increase in the calculation amount when taking the derivative of back propagation, while the ReLU function can save a lot of calculation time.
(2)
When sigmoid and tanh functions carry out back propagation in deep network, it is very easy to generate a gradient disappearance problem, that is, when approaching the saturation zone, the derivative approaches 0, transformation slows down, and information is easily lost. Therefore, deep network training cannot be realized.
(3)
The ReLU function can set part of the output to 0, making the model more sparse, thus reducing the dependency relationship between network parameters and reducing the overfitting phenomenon.

2.3. Sparse Filtering

As a simple and effective unsupervised feature learning algorithm, the sparse filtering algorithm only needs one parameter adjustment, namely the number of features to be learned [20]. Its working principle is mainly to bypass the estimation of data distribution and optimize only a simple cost function, the L2 norm, to directly analyze the distribution of optimized features.
The structure of sparse filtering is a two-layer neural network including input layer, output layer, and weight matrix. The input is the collected signal and the output is the learned feature. The collected signals are divided into many identical samples to form a training set { x i } i = 1 M , where x i = N × 1 is a sample containing N data points and M is the number of samples. The sample is mapped to the eigenvector f i L × 1 by the usual-value matrix W R N × L , and f j i was used to represent the jth eigenvalue of the ith sample of the eigenmatrix. The features learned are linear:
f j i = W j T x i
First, all row vectors of the eigenmatrix are normalized, then all column vectors of the eigenmatrix are normalized, and finally the sum of the absolute values of all elements of the matrix is obtained. The details are as follows:
First, each feature is normalized to an equal activation value, that is, the L2 norm of each feature divided by all the samples in which it is located:
f ˜ j = f j / f j 2
Then, the characteristics of each sample are normalized, so that the characteristics of all samples fall on the unit sphere of L2 norm:
f ^ i = f ˜ i / f ˜ i 2
Finally, the L1 norm penalty is used to carry out the sparsity constraint to realize all the normalized features before optimization. Assuming that a dataset has M samples, the objective function of sparse filtering is:
minimize W i = 1 M f ^ i 1 = i = 1 M f ˜ i f ˜ i 2

3. Proposed Smart Diagnosis Method

The SAFC model is composed of SAE and SF, where SAE has five layers of network and SF has two layers of network, as shown in Figure 3. The specific steps of the extracting features are as follows:
(1)
All the collected time domain signals are converted into frequency domain signal through FFT. According to different health conditions, the spectrum samples of vibration signals at different speeds are composed into a training set X i , l i i = 1 M , where M is the number of samples, X j N × 1 means that the ith sample contains N Fourier coefficients, and l i is the health label of the ith sample.
(2)
The no label training set X i i = 1 M was used to pre-train the SAE layer by layer. The pre-training process was to superimpose N AE into N hidden layers, and the characteristics learned in the previous layer were used as the input of the next layer.
(3)
BP algorithm combined with label data X i , l i i = 1 M was used to realize weight updating and parameter fine-tuning of SAE, and back propagation training was carried out by minimizing the error between extracting features and health labels.
(4)
BN was applied to each activation layer of SAE, and the characteristic difference at different speeds was reduced by its translation and scaling.
(5)
The features initially extracted after dimension reduction through two-layer SAE were taken as new training samples and a new training sample set { x i , l i } i = 1 m was formed. m is the number of samples [21], x j R N i n × 1 means the ith sample contains Nin data points and l i is the label of the sample.
(6)
The sample set was formed into a matrix form T R N i n × M , and then input into the sparse filtering model for training. The weight matrix W is obtained through the minimization formula (24).
(7)
The learning feature f i R N o u t × 1 can be calculated by W and x i . ReLU was used as the activation function here to extend the sparse filtering to nonlinear mapping. Therefore, Equation (20) is extended to:
f j i = ReLU ( W j T x i )
(8)
When the learning feature set { f i } i = 1 M is obtained, it is combined with the tag set { l i } i = 1 M and input into the softmax regression classifier for training. Then, the remaining samples are used as test samples to test the accuracy of the proposed method [22].
The testing process is as follows. First, each test sample is converted into spectrum after FFT transformation, and all spectrum samples are combined into matrix form as input; second dimension reduction and preliminary feature extraction are carried out in superposition self-coding to cluster the same type of signals with different speeds. Then, the trained sparse filter model is used to study the characteristics of the test samples. Finally, the learned features are combined with the real fault type label, and then input to the softmax regression model that has been trained for classification, thus obtaining the test accuracy rate.

4. Experimental Verification

To verify the feasibility of the proposed method, a rotating parts fault diagnosis bench was designed based on two groups of tests. The first group classified the test data according to different parts, while the second group classified the test data on the basis of different fault locations.

4.1. Case Study I: Depending on Different Fault Components

4.1.1. Test Equipment and Data Introduction

As shown in Figure 4, the designed experimental equipment mainly included the motor, coupling, planetary gear box, bearing seat, rotor, and so on. Aside from the normal fault-free vibration signal, six different fault types were set respectively for different fault positions of the shaft, bearing, and gear. Depending on the location and depth of the shaft crack, the six fault types were three cracks on the left side of the larger rotor (1.2 mm, 2.4 mm, and 3.6 mm) and three cracks on the right side of the smaller rotor (1.2 mm, 2.4 mm, and 3.6 mm), as shown in Figure 5. The six faults were referred simply to as L1, L2, L3, R1, R2, and R3, where L represents left, and R represents right. According to the location and depth of the bearing fault, the six fault types were three inner ring faults (0.2 mm, 0.6 mm, and 1.2 mm) and three outer ring faults (0.2 mm, 0.6 mm, and 1.2 mm), as shown in Figure 6. These six faults were also simply referred to as I1, I2, I3, O1, O2, and O3, where I represents inner, and O represents outer. On the basis of the location and depth of the gear failure, the six fault types were three kinds of sun wheel failures and three kinds of planetary wheel failures: crack, pitting, and wear, as is shown in Figure 7. In the same way, these six faults were simply referred to as P1, P2, P3, S1, S2, and S3, where P represents planetary, and S represents sun. Therefore, there were three groups of seven health conditions each, and a total of 21 sets of test data needed to be processed. However, the 21 sets of test data to be processed actually only represent 19 health conditions since the failure-free data in the three sets was the same.
The data settings are shown in Table 1. During the test, the motor speed was respectively set as 900 r/min, 1100 r/min, 1300 r/min, and 1500 r/min. The output speed was three times that of the motor speed since the speed ratio of planetary gearbox was 1:3. It is necessary to collect test data at all speeds for the all 19 cases above-mentioned. Therefore, 78 sets of test data needed to be collected. For the collection, a three-way vibration acceleration sensor (PCB315A) was installed on the surface of the bearing seat. Vibration signals on the upper surface of the bearing seat were collected by the LMS data acquisition instrument, and the sampling frequency was set at 12.8 Khz. A total of 200 samples were collected for each health condition, and 600 data points were taken for each sample at each rotation speed. By stitching together signals at different speeds for the same health condition, 200 samples of each health condition were obtained, each with 2400 data points.
The 2400 data points of each sample were processed by FFT to obtain 1200 Fourier coefficients. The reason why frequency spectrum was chosen was due to the time shifting characteristics of the time-domain samples. It is difficult to ensure the consistent position of the fault feature points in different samples of the same fault, which produces great difficulty in feature extraction. Additionally, the length of each signal input to the classifier will affect the accuracy rate. The spectrum information can not only easily obtain richer information of different health conditions from the test signal by FFT, but its more regular characteristic distribution also overcomes the time-shift characteristic of the time-domain signal. Taking the bearing data, for example, the corresponding signal wave-forms in the time domain and frequency domain are shown in Figure 8. It can be seen that it is difficult to distinguish the different fault types from both the time domain and frequency domain.

4.1.2. Parameter Selection

The number of neurons in the input layer of the SAFC model was the same as the sample dimension, which was 200, the dimension of the sample spectrum. In the three hidden layers, the number of neurons was set as 600, 200, and 100. The number of neurons in the output layer was the same as the total number of health conditions. There was a total number of 20 training iterations for each layer. The learning rate was determined to be 1 E-4, while batch size was determined by parameter selection. After the data was output through the SAE network, the second hidden layer of SAE was used as the input layer of SF. The number of neurons in the input layer was also the same as the sample dimension. The number of neurons in the output layer was determined by parameter selection [23]. The test accuracy of the SAE part was taken as clustering accuracy, and the test accuracy of SF part as classification accuracy.
First of all, the proportion of training samples in the SAE part was selected for research. Gear data was selected for testing, and the training batch size was tentatively set at 10. Meanwhile, the percentage of training samples in SF was set at 10%, and the output dimension was set at 200. In order to eliminate the effect of randomness, each group of tests were carried out 20 times. The diagnostic accuracy is shown in Figure 9. The results show that with the increase in the training sample size, the diagnostic accuracy increased continuously, and the time spent increased almost linearly. When the percentage of SAE training samples increased to 20%, the clustering accuracy reached 99.98% ± 0.03%, and the classification accuracy reached 99.22 ± 0.11%, after which both the two accuracy rates remained almost unchanged. Therefore, after weighing the accuracy of classification against the time spent, 20% was used as a percentage of the training sample in the following experiment.
Then, after considering the selection of SAE training batch size, the diagnostic results of different batch sizes are as shown in Figure 10. It can be seen that when the batch size was reduced, the clustering accuracy and classification accuracy were slightly improved, but the average training time increased sharply with the decrease in batch size. When the batch size was 5, the training time was up to 153.27 s, which processed only one-third of the data that needed to be processed. When the batch size was 40, the time taken was 28.74 s, which was far less than 82.01 s when the batch size was 10, but the accuracy and variance were almost the same. Therefore, 40 was selected as the batch size.
The parameters of the SF part were also taken into consideration after confirming the parameters of the SAE part. In order to study the effect of output dimension, 10% samples were still selected for testing. The diagnostic accuracy is shown in the figure. The results show that with the increase in output dimension, the diagnostic accuracy increases and the corresponding standard deviation decreases. The results show that with the increase of the output dimension, the diagnostic accuracy increases and the corresponding standard deviation decreases, while the average training time increases. Therefore, a trade-off was made between classification accuracy and time spent. Considering that the increase in accuracy was no longer significant after the output dimension reached 200, 200 was selected as the number of neurons in the output layer.
The diagnostic results with different proportions of training samples in the SF part are shown in Figure 10. It can be seen that with the increase in sample percentage, the diagnostic accuracy was constantly improved and the standard deviation was reduced. When the sample percentage was 10%, the classification accuracy reached 99.13% ± 0.15%. However, as the percentage of training samples increased further, the diagnostic accuracy hardly improved, and the diagnostic time fluctuated only within a small range. Therefore, 10% was chosen as the proportion of training sample.

4.1.3. The Results of the Diagnosis

The first 20% of the sample was selected in the SAE section, and after the dimension reduction, the first 10% of the sample was selected in the SF section. Using the same training dataset to train the three models, the diagnostic accuracy of the data of the three parts is shown in Figure 11. It can be seen that SAFC, which contained both methods, was more accurate and stable than SAE or SF alone. In the process of diagnosis, the diagnostic accuracy of the three methods for gear fault was the lowest among the three parts, so took gear as an example. The average accuracy and time of the three methods were calculated, the results of which showed that the SAFC accuracy was above 99.13%, and the standard deviation was no more than 0.11 while the mean accuracy rate of SAE and SF was only 86.37% and 93.80%, and the standard deviation exceeded 3.10 and 1.05. This shows that the proposed method had higher accuracy and stability. The average training time of the sample was also calculated, and the specific calculation results are shown in Table 2. It can be seen that the proposed method SAFC is not only more accurate than SF, but also faster in diagnosis, which benefits from the rapid speed of SAE in training data and its dimension reduction of the data. It can also be seen that although SAFC is not as fast as SAE in diagnostic speed, its accuracy is far better than SAE, which is due to the powerful feature extraction capability of SF.
In order to prove the classification effect of the proposed method better, the three methods were used to extract features from the data of three parts, and the t-SNE dimension reduction algorithm was used to transform the obtained feature vectors into two-dimensional views.
The results of feature mapping of the three methods are respectively shown in Figure 12, Figure 13 and Figure 14 [24]. It can be seen that in the proposed method, characteristic samples of the same health status were clustered in the corresponding clusters, and test samples of different health status were well separated. Only one sample of Sun 3 was divided into the cluster of planet 2 in gear failure. However, in SAE, not only in the bearing data, a sample of inner ring 2 was not well aggregated, but also in the shaft data, three groups of crack characteristics on the left side of the larger rotor were not very well separated. In addition, in the process of gear fault diagnosis, four groups of characteristics were staggered, while in the SF, although the characteristics of different health conditions were generally separated, uniform health conditions at different speeds were also separated, which obviously did not meet the classification requirements.
Figure 15, Figure 16 and Figure 17 show the visualization of the feature distribution of the three components under the three methods. It can be seen that in SAFC, the characteristics of different fault types were different with obvious differences, while in SAE, the difference was relatively fuzzy. In SF, subdivision occurred in each of the same characteristics, that is, the same health condition showed different characteristics, which was consistent with the results in t-SNE. In conclusion, the proposed method can well extract domain invariant features and classify them.

4.2. Case Study II: According to Different Fault Location

4.2.1. Data Introduction

To further prove the effectiveness of the proposed method in distinguishing different fault parts, the previously measured test data were regrouped according to the different fault locations. The general idea is that different degrees of failure of one of the two fault locations of each part at different speeds were combined with normal trouble-free data, which resulted in two datasets (Table 3) with 10 health conditions each. For example, we can make a group of three sets of crack data on the right side of the smaller rotor, three sets of fault data of the inner ring of the bearing, and three sets of fault data of the planet wheel and normal fault-free data. The remaining three sets of crack data on the left side of the larger rotor, three sets of failure data of the bearing outer ring, three sets of failure data of the sun wheel and normal trouble-free data are made into the other group. Again, the normal failure-free data is Shared by both sets of data, so there are still 19 different health conditions. The selection of parameters is the same as before.

4.2.2. The Result of the Diagnosis

The diagnostic accuracy of the three methods for the two sets of data is shown in Figure 18. It can be seen that SAFC is more accurate and stable than SAE an SF. The specific calculation results are shown in the Table 4. In the diagnosis of the two sets of data, the diagnostic accuracy is generally higher than the diagnosis of each component, which is because the characteristics of different components have more difference than the characteristics of different positions of the same component. Similarly as before, the diagnostic time of SAFC is between SAE and SF, but SAFC still has the highest diagnostic accuracy.
Again, these three methods were used to extract features from the two sets of data, and t-SNE dimension reduction was used to obtain a two-dimensional view. The feature mapping results of the three methods are shown in Figure 19 and Figure 20. It can be seen that in the proposed method, characteristic samples of the same health status are clustered in the corresponding clusters, and test samples of different health status are well separated. However, in SAE, not only in the group 1, some samples of planet wheel 3 were not well aggregated, but also in the group 2, some groups of characteristics were not very well separated. While in the SF, although the characteristics of different health conditions were generally separated, uniform health conditions at different speeds were also separated, which obviously did not meet the classification requirements.
Figure 21 and Figure 22 show the visualization of the feature distribution of the two groups under the three methods. It can be seen that in SAFC, the characteristics of different fault types were different with obvious differences, while in SAE, the difference was relatively fuzzy. In SF, subdivision occurred in each of the same characteristics, that is, the same health condition showed different characteristics, which was consistent with the results in t-SNE. In conclusion, the proposed method can well extract domain invariant features and classify them.

4.3. A Comparison with the Other Two Methods

In the literature [25,26], the deep learning methods of BNAE and L1/2-SF are provided respectively, which also have the function of feature extraction. With the previous dataset in Section 4.2, the three models were tested separately, and the proposed method was compared with the two methods, as shown in Figure 23. It can be seen that the proposed SAFC is superior to the compared method in feature extraction and classification, which is also presented in the obfuscation matrix, as shown in Figure 24, which again proves the feasibility of the proposed method.

5. Conclusions

A SAFC architecture was proposed to solve the problem of fault diagnosis. This model has a strong feature extraction capability in which SAE is used to reduce and cluster the data, while SF is used to extract and classify the features. A good diagnostic performance is demonstrated under different workloads.
The results in Section 4 show that SAFC can not only extract domain invariant features at different speeds, but can also well classify the extracted features according to different parts and fault locations. SAFC had a higher diagnostic accuracy and stability than SAE or SF alone. When dealing with large data volumes, SAFC is able to extract domain invariant features quickly thanks to the rapid training data capability of SAE. When classifying extracted features, SAFC has a high diagnostic accuracy as a benefit of the strong feature extraction capability of SF.
In future work, we will further study the classification of different parts and consider the diagnostic impact of sample imbalance.

Author Contributions

Conceptualization, R.D. and K.X.; methodology R.D. and J.W.; software, R.D.; validation, R.D. and J.L.; formal analysis, R.D. and S.L.; investigation S.L.; resources S.L.; data curation K.X.; writing—original draft preparation R.D.; writing—review and editing R.D.; visualization R.D.; supervision J.L.; project administration S.L.; funding acquisition R.D., J.L and S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Major national Science and Technology Projects (2017-IV-0008-0045), the Advance Research Field Fund Project of China (61400040304), and the National Natural Science Foundation of China (51975276).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lei, Y. Intelligent Fault Diagnosis and Remaining Useful Life Prediction of Rotating Machinery; Xi′an Jiaotong University Press: Xi′an, China, 2017. [Google Scholar]
  2. Shen, C. Research on Fault Diagnosis and Prediction Methods of Key Components of Rotating Machinery Equipment. Ph.D. Thesis, University of Science and Technology of China, Changsha, China, 2014. [Google Scholar]
  3. Hinton, G.; Salakhutdinov, R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Luciano, L.; Hamza, A.B. Deep learning with geodesic moments for 3D shape classification. Pattern Recognit. Lett. 2018, 105, 182–190. [Google Scholar]
  5. Deng, S.; Cheng, Z.; Li, C.; Yao, X.; Chen, Z.; Sanchez, R.-V. Rolling Bearing Fault Diagnosis Based on Deep Boltzmann Machines. In Proceedings of the IEEE Prognostics and System Health Management Conference, Chengdu, China, 19–21 October 2017; pp. 1–6. [Google Scholar]
  6. Ding, X.; He, Q. Energy-fluctuated multiscale feature learning with deep convnet for intelligent spindle bearing fault diagnosis. IEEE Trans. IM 2017, 66, 1926–1935. [Google Scholar] [CrossRef]
  7. Li, C.; Sanchez, R.; Zurita, G.; Cerrada, M.; Cabrera, D.; Vásquez, R. Gearbox fault diagnosis based on deep random forest fusion of acoustic and vibratory signals. Mech. Syst. Sign. Process. 2016, 76, 283–293. [Google Scholar] [CrossRef]
  8. Liu, C.; Cheng, G.; Chen, X.; Pang, Y. Planetary Gears Feature Extraction and Fault Diagnosis Method Based on VMD and CNN. Sensors 2018, 18, 1523. [Google Scholar] [CrossRef] [Green Version]
  9. Chen, Z.; Li, C.; Sánchez, R. Multi-layer neural network with deep belief network for gearbox fault diagnosis. J. Vibroeng. 2015, 17, 2379–2392. [Google Scholar]
  10. Bengio, Y. Learning deep architectures for AI. Found. Trends Mach. Learn. 2009, 2, 1–127. [Google Scholar] [CrossRef]
  11. Jia, F.; Lei, Y.; Lin, J.; Zhou, X.; Lu, N. Deep neural networks: A promising tool for fault characteristic mining and intelligent diagnosis of rotating machinery with massive data. Mech. Syst. Signal Process. 2016, 72, 303–315. [Google Scholar] [CrossRef]
  12. Ngiam, J.; Pang, W.K.; Chen, Z.; Bhaskar, S.A.; Koh, P.W.; Ng, A.Y. Sparse Filtering. In Proceedings of the International Conference on Neural Information Processing Systems, Granada, Spain, 12–17 December 2011; pp. 1125–1133. [Google Scholar]
  13. Lei, Y.; Jia, F.; Lin, J.; Xing, S.; Ding, S.X. An Intelligent Fault Diagnosis Method Using Unsupervised Feature Learning Towards Mechanical Big Data. IEEE Trans. Ind. Electron. 2016, 63, 3137–3147. [Google Scholar] [CrossRef]
  14. Chen, R.; Chen, S.; He, M.; He, D. Rolling bearing fault severity identification using deep sparse auto-encoder network with noise added sample expansion. Proc. Inst. Mech. Eng. Part O. J. Risk Reliability 2017, 231, 666–679. [Google Scholar] [CrossRef]
  15. Li, H.; Zhang, H.; Qin, X.R.; Zhou, P. Bearing fault diagnosis method based on SHORT-time Fourier Transform and convolutional neural network. Vib. Impact 2018, 37, 124–131. [Google Scholar]
  16. Wang, W.; Qiu, X.; Sun, J.; Wang, F. Fault diagnosis method of Gear based on double-layer Long-length memory network. J. Acad. Armored Force Eng. 2018, 2, 81–86. [Google Scholar]
  17. Xu, K.; Li, S.; Wang, J.; An, Z.; Qian, W.; Ma, H. A novel convolutional transfer feature discrimination network for imbalanced fault diagnosis under variable rotational speed. Meas. Sci. Technol. 2019, 30, 10. [Google Scholar] [CrossRef]
  18. Han, B.; Wang, X.; Ji, S.; Zhang, G.; Jia, S.; He, J. Data-enhanced Stacked Autoencoders for Insufficient Fault Classification of Machinery and Its Understanding Via Visualization. IEEE Access 2020, 8, 67790–67798. [Google Scholar] [CrossRef]
  19. Chakraborty, S. Learning representation for multi-view data analysis: Models and applications. Comp. Rev. 2019, 60, 284. [Google Scholar]
  20. Han, Y.; Lee, K. Detecting fingering of overblown flute sound using sparse feature learning. EURASIP J. Audio Speech Music Proc. 2016, 2016, 1–10. [Google Scholar] [CrossRef] [Green Version]
  21. Zhang, Z.; Li, S.; Wang, J.; Xin, Y.; An, Z. General normalized sparse filtering: A novel unsupervised learning method for rotating machinery fault diagnosis. Mech. Syst. Signal Proc. 2019, 124, 596–612. [Google Scholar] [CrossRef]
  22. Lei, Y. Individual intelligent method-based fault diagnosis. Intell. Fault Diagn. Remain. Useful Life Predict. Rotating Mach. 2017, 10, 167–174. [Google Scholar]
  23. Zhang, Z.; Chen, H.; Li, S.; Wang, J. A new method for intelligent fault diagnosis based on time-frequency feature extraction and Softmax regression. J. Central S. Univ. 2019, 26, 1607–1618. [Google Scholar] [CrossRef]
  24. George, S.; Hamza, A.B. Fault Detection Using Robust Multivariate Control Chart. Expert Syst. Appl. 2009, 36, 5888–5894. [Google Scholar]
  25. Jinrui, W.; Shunming, L.; Zenghui, A.; Xingxing, J.; Weiwei, Q.; Shanshan, J. Batch-normalized deep neural networks for achieving fast intelligent fault diagnosis of machines. Neurocomputing 2019, 329, 53–65. [Google Scholar]
  26. Jinrui, W.; Shunming, L.; Han, B.; An, Z.; Xin, Y.; Qian, W.; Wu, Q. Construction of a batch-normalized autoencoder network and its application in mechanical intelligent fault diagnosis. Meas. Sci. Technol. 2019, 30, 015106. [Google Scholar]
Figure 1. The automatic encoder AE architecture.
Figure 1. The automatic encoder AE architecture.
Applsci 10 06084 g001
Figure 2. The stacked auto encoders SAE architecture.
Figure 2. The stacked auto encoders SAE architecture.
Applsci 10 06084 g002
Figure 3. The overall architecture of the sparse filter rotating component comprehensive diagnosis model (SAFC) model.
Figure 3. The overall architecture of the sparse filter rotating component comprehensive diagnosis model (SAFC) model.
Applsci 10 06084 g003
Figure 4. Rotating components fault diagnosis test bench.
Figure 4. Rotating components fault diagnosis test bench.
Applsci 10 06084 g004
Figure 5. Shaft crack.
Figure 5. Shaft crack.
Applsci 10 06084 g005
Figure 6. Bearing fault.
Figure 6. Bearing fault.
Applsci 10 06084 g006
Figure 7. Bearing fault.
Figure 7. Bearing fault.
Applsci 10 06084 g007
Figure 8. Time and frequency domain wave-forms of bearing signals with different health conditions.
Figure 8. Time and frequency domain wave-forms of bearing signals with different health conditions.
Applsci 10 06084 g008
Figure 9. SAE parameter selection.
Figure 9. SAE parameter selection.
Applsci 10 06084 g009
Figure 10. SF parameter selection.
Figure 10. SF parameter selection.
Applsci 10 06084 g010
Figure 11. Diagnostic accuracy of the three methods for each component.
Figure 11. Diagnostic accuracy of the three methods for each component.
Applsci 10 06084 g011
Figure 12. Signal characteristic dimension reduction of shaft crack using the three methods.
Figure 12. Signal characteristic dimension reduction of shaft crack using the three methods.
Applsci 10 06084 g012
Figure 13. Signal characteristic dimension reduction of bearing fault using the three methods.
Figure 13. Signal characteristic dimension reduction of bearing fault using the three methods.
Applsci 10 06084 g013
Figure 14. Signal characteristic dimension reduction of gear fault using the three methods.
Figure 14. Signal characteristic dimension reduction of gear fault using the three methods.
Applsci 10 06084 g014
Figure 15. Signal characteristic of shaft crack using the three methods.
Figure 15. Signal characteristic of shaft crack using the three methods.
Applsci 10 06084 g015
Figure 16. Signal characteristic of bearing fault using the three methods.
Figure 16. Signal characteristic of bearing fault using the three methods.
Applsci 10 06084 g016
Figure 17. Signal characteristic of gear fault using the three methods.
Figure 17. Signal characteristic of gear fault using the three methods.
Applsci 10 06084 g017
Figure 18. Diagnostic accuracy of three methods for two groups of data.
Figure 18. Diagnostic accuracy of three methods for two groups of data.
Applsci 10 06084 g018
Figure 19. Signal feature dimension reduction for group 1 using the three methods.
Figure 19. Signal feature dimension reduction for group 1 using the three methods.
Applsci 10 06084 g019
Figure 20. Signal feature dimension reduction for group 2 using the three methods.
Figure 20. Signal feature dimension reduction for group 2 using the three methods.
Applsci 10 06084 g020
Figure 21. Signal characteristics for group 1.
Figure 21. Signal characteristics for group 1.
Applsci 10 06084 g021
Figure 22. Signal characteristics for group 2.
Figure 22. Signal characteristics for group 2.
Applsci 10 06084 g022
Figure 23. Signal feature dimension reduction using the three methods.
Figure 23. Signal feature dimension reduction using the three methods.
Applsci 10 06084 g023
Figure 24. Training sample distribution and mixed matrix comparison.
Figure 24. Training sample distribution and mixed matrix comparison.
Applsci 10 06084 g024
Table 1. Description of datasets.
Table 1. Description of datasets.
Group Number1234567
ShaftFault locationNormalCrack on left side of the larger rotorCrack on right side of the smaller rotor
abbreviationNL1L2L3R1R2R3
Depth (mm)01.22.43.61.22.43.6
BearingFault locationNormalFailure of bearing inner ringFailure of bearing outer ring
abbreviationNI1I2I3O1O2O3
Depth (mm)00.20.61.20.20.61.2
GearFault locationNormalFailure of planetary wheelFailure of sun wheel
abbreviationNP1P2P3S1S2S3
Fault typenonecrackpittingwearcrackpittingwear
Table 2. Time and accuracy of diagnosis.
Table 2. Time and accuracy of diagnosis.
SAFCSAESF
Accuracy (%)Time (s)Accuracy (%)Time (s)Accuracy (%)Time (s)
Shaft99.61 ± 0.0628.85 ± 0.6589.86 ± 1.3317.63 ± 0.2995.03 ± 0.2442.27 ± 1.05
Bearing99.87 ± 0.0428.23 ± 0.6793.86 ± 0.5717.58 ± 0.2497.99 ± 0.1042.93 ± 1.10
Gear99.13 ± 0.1128.39 ± 0.4886.37 ± 3.1017.33 ± 0.2593.80 ± 1.0542.46 ± 1.24
Table 3. Description of datasets.
Table 3. Description of datasets.
Group Number12345678910
Group 1locationNormalRight side of the smaller rotorBearing inner raceplanet gear
acronymNR1R2R3I1I2I3P1P2P3
Group 2locationNormalLeft side of the larger rotorBearing outer racesun gear
acronymNL1L2L3O1O2O3S1S2S3
Table 4. Time and accuracy of diagnosis.
Table 4. Time and accuracy of diagnosis.
SAFCSAESF
Accuracy (%)Time (s)Accuracy (%)Time (s)Accuracy (%)Time (s)
Group 199.91 ± 0.0443.35 ± 1.4696.84 ± 0.4925.16 ± 0.7698.32 ± 0.1262.37 ± 2.57
Group 299.62 ± 0.0843.64 ± 1.8192.03 ± 0.6625.84 ± 0.8997.27 ± 0.6062.49 ± 2.67

Share and Cite

MDPI and ACS Style

Ding, R.; Li, S.; Lu, J.; Xu, K.; Wang, J. A Novel Stacked Auto Encoders Sparse Filter Rotating Component Comprehensive Diagnosis Network for Extracting Domain Invariant Features. Appl. Sci. 2020, 10, 6084. https://doi.org/10.3390/app10176084

AMA Style

Ding R, Li S, Lu J, Xu K, Wang J. A Novel Stacked Auto Encoders Sparse Filter Rotating Component Comprehensive Diagnosis Network for Extracting Domain Invariant Features. Applied Sciences. 2020; 10(17):6084. https://doi.org/10.3390/app10176084

Chicago/Turabian Style

Ding, Rui, Shunming Li, Jiantao Lu, Kun Xu, and Jinrui Wang. 2020. "A Novel Stacked Auto Encoders Sparse Filter Rotating Component Comprehensive Diagnosis Network for Extracting Domain Invariant Features" Applied Sciences 10, no. 17: 6084. https://doi.org/10.3390/app10176084

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop