Next Article in Journal
Estimating the Measurement Uncertainty of the Number of Vehicles in a Car Park Using an Indirect Method
Next Article in Special Issue
Identification of Aerodynamic Tonal Noise Sources of a Centrifugal Compressor of a Turbocharger for Large Stationary Engines
Previous Article in Journal
Mathematical Modeling and Digital Simulation of Teeth Dynamics for the Approximation of Orthodontic Treatment Duration
Previous Article in Special Issue
A Novel Method for Fault Diagnosis of Bearings with Small and Imbalanced Data Based on Generative Adversarial Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Bearing Fault Diagnosis Method Based on Wavelet Denoising and Machine Learning

1
Information and Computing Science, China University of Geosciences, Wuhan 430074, China
2
Computer Science and Technology, China University of Geosciences, Wuhan 430074, China
3
Mathematics and Physics, China University of Geosciences, Wuhan 430074, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(10), 5936; https://doi.org/10.3390/app13105936
Submission received: 18 April 2023 / Revised: 3 May 2023 / Accepted: 9 May 2023 / Published: 11 May 2023

Abstract

:
There are a lot of interference factors in the operating environment of machinery, which makes it ineffective to use traditional detection methods to judge the fault location and type of fault of the machinery, and even misjudgment of the fault location and type may occur. In order to solve these problems, this paper proposes a bearing fault diagnosis method based on wavelet denoising and machine learning. We use sensors to detect the operating conditions of rolling bearings under different working conditions to obtain datasets of different types of bearing failures. On the basis of using the wavelet denoising algorithm to reduce noise, we comprehensively evaluated five machine learning models, including K-means clustering, decision tree, random forest, and support vector machine to classify bearing faults and compare their results. By designing the fault classification evaluation prediction criteria, the following conclusions are drawn. The model proposed in this paper is significantly better than other traditional diagnostic models for bearing faults. In order to solve the problem of weak signal strength and background noise interference, this paper selects a better noise reduction algorithm under different quantitative evaluation indicators for wavelet denoising, which can better restore the true characteristics of the fault signal. Using unsupervised learning and supervised machine learning classification algorithms, the evaluation indicators before and after denoising are compared to make the classification results more accurate and reliable. This article will help researchers to intelligently diagnose the faults of rolling bearing equipment in rotating machinery.

1. Introduction

With the rise of global industrialization and the widespread use of machinery, fault data have become increasingly dense and unrelated. The goal of fault diagnosis technology for instruments and equipment is to monitor the operational state of machines and determine their overall or partial functionality. By employing data mining and machine learning techniques, faults and their underlying causes can be identified, and future fault developments can be predicted.
The technical characteristics of fault type identification are as follows: reducing the impact of environmental noise or abnormal data on original data, extracting reliable waveform feature criteria, and selecting or improving existing machine learning methods. The work of [1,2,3] has received long-term attention in the industry [4]. As a key component of machinery, bearings can fix objects and provide loads [5]. If the bearing fails, it will lead to a decline in the quality of the product and cause serious economic losses [6]. So bearing fault detection is the focus and difficulty of current research [7]. Abdul et al. summarize recently published work on the application of artificial intelligence to bearing fault detection in rotating machinery [8,9]. Scholars have proposed many intelligent fault detection methods for the pulse vibration signals generated by bearings. Among these methods, the bearing detection method of deep learning is used the most. Compared with the traditional bearing detection method, deep learning can accurately identify fault information in the context of big data, and reduces the requirement for professional knowledge [10]. Zhang et al. improved the feature extraction method of the convolutional neural network and performed fault diagnosis on bearings.
Although the above methods have achieved great success, there is a strong dependence on the fault data, and more labels are required when training the model. Such research is contrary to the real industrial environment: in the real industrial environment, there are far more normal data than fault data. Such a dataset will greatly reduce the training effect of the deep learning model. Ref. [11] has attracted widespread attention in the academic community.
In order to solve the above-mentioned problems in the industry, Ref. [12] Sun et al. used traditional machine learning, an artificial neural network, deep learning and transfer learning to carry out a typical fault diagnosis method of a wind power system based on the ML method from the theoretical basis and industrial application. An overview is given and its advantages and disadvantages are discussed [13]. He et al. worked on improving the model structure and classification loss function, proposing to optimize support vector machines with dynamic penalty factors [14]. Praneeth Chandra et al. used unsupervised machine learning and the DBSCAN algorithm to detect defects on rail fixtures [15]. Praneeth Chandra et al. diagnosed defects using supervised machine learning. The learning data are acquired by sensors, and six machine learning models are used. The result values are best when using the the AdaBoost model [16]. Yoon et al. found unexpected test accuracy in cross-validation for machine learning [17]. Yang et al. proposed a new machine learning model for automatic labeling, feature extraction, model building, model fusion, and evaluation of railway data [18]. Nason et al. described a stationary wavelet transform and used it in astronomy as well as veterinary anatomy [19]. Baydar et al. used wavelet transform to compare acoustic signals and vibration signals, and the results showed that acoustic signals have a strong influence on early fault detection [20]. Firas et al. used an optimized stationary wavelet packet transform to provide identification for detecting IM bearing faults at an early stage and verified the effectiveness of the method experimentally [21]. Lu et al. proposed a higher-density wavelet packet neighborhood coefficient denoising algorithm, and experiments showed that the algorithm can significantly improve the signal-to-noise ratio [22]. Aslam et al. proposed a fault detection method based on SNR [23]. Shim et al. used the improved CNN LeNet-5 model to detect railway anomalies. Ref. [24] M. F. Palangar et al. trained a classifier based on two indicators of experimental data and confirmed the reliable performance of ADD [25]. Stępień K et al. present a method for evaluating the surface texture of machine parts using wavelet transform of three selected indices.
This paper proposes a fault type classification and identification model based on wavelet denoising and machine learning. There are a lot of interference factors in the operating environment of machinery, which makes it ineffective to use traditional detection methods to judge the fault location and type of fault of the machinery, and even misjudgment of the fault location and type may occur. Therefore, how to realize intelligent fault diagnosis in an environment full of a large number of interference factors is an urgent problem to be solved. We use two methods of wavelet soft threshold denoising and wavelet packet analysis to denoise the original fault signal data. Eliminate the interference factors in the environment, check the correctness of the denoising results, and select the denoising method with better effect on the basis of quantitative indicators. Based on the wavelet denoising algorithm, five of the most widely used machine learning models are applied: K-means clustering, quadratic support vector machine (SVM), random forest (RF), decision tree (DT) and AdaBoost algorithm. These machine learning algorithms are used for binary classification of bearing faults and to compare their results. By designing the fault classification evaluation prediction criteria, the following conclusions are drawn. If unsupervised machine learning methods are used, K-means clustering can achieve good classification results; if supervised machine learning methods are used, decision tree classification can achieve the best results. In addition, the focus of our future work is as follows: First, increase the types of rolling bearing faults, narrow the characteristic differences of fault signals, and improve the adaptability and flexibility of the model. Second, more types of faults and larger datasets are considered to improve the generalization ability of the model. Finally, the wavelet denoising algorithm can be further optimized by selecting different wavelet bases, decomposition levels and threshold functions according to different types of signals. Additionally, machine learning models can further improve their performance by tuning parameters, selecting features, or combining multiple models. In this paper, it is of great significance to intelligently diagnose the faults of instruments and equipment.

2. Methods

2.1. Wavelet Soft Thresholding Denoising

Wavelet transform can process the signal in a more subtle way, and can also better express some characteristics of the signal. There are also many denoising methods in wavelet analysis, and these different denoising methods will also have a great impact on the denoising effect of each signal. The Wavelet threshold denoising method is the most widely used method in engineering. The basic idea is as follows: a noise-containing one-dimensional signal model can be expressed as f ( t ) = s ( t ) + n ( t ) , where s ( t ) is the original signal, and n ( t ) is Gaussian white noise with variance σ 2 , obeying N 0 , σ 2 . Wavelet transform belongs to linear transformation, and the wavelet coefficient obtained after discrete wavelet transform of f ( t ) is divided into two parts: signal and noise. The wavelet coefficients with larger amplitudes are useful signals. The wavelet coefficient with smaller amplitude is the energy of the noise signal. Therefore, an appropriate threshold λ can be found for distinguishing noise and signal, and the useful signal is retained with reference to this threshold. For this problem, we use the soft threshold method [26] for denoising, first transforming the actual signal into the wavelet domain using wavelet transform. Then the wavelet coefficients are processed by the nonlinear shrinkage rule; finally, the thresholded wavelet coefficients are subjected to inverse wavelet transform to obtain the denoising signal. The soft threshold function is defined as follows:
w ¯ i , j = sign w j , k w j , k λ | w j , k | λ 0 | w j , k | < λ

2.2. Wavelet Packet Analysis Denoising

The idea of wavelet packet analysis and wavelet soft threshold denoising [27] is basically the same, but the analysis method of wavelet packet is more complex and flexible. When analyzing, it decomposes the low-frequency part and high-frequency part of the previous level at the same time, making the local analysis capability more accurate. The wavelet packet decomposition also needs to choose the wavelet packet base and level of decomposition. For the selection of the wavelet packet base, it is necessary to choose the wavelet packet base with better symmetry and regularity. The wavelet with better symmetry does not produce phase distortion, and it is easy to obtain a smooth reconstructed signal with the wavelet with better regularity. In the wavelet base family, the Symlets family is similar to the Daubechies family, which also has orthogonality, biorthogonality, compact support and approximate symmetry, and can perform discrete wavelet transform. Therefore, in the part of wavelet packet analysis and denoising, the wavelet packet base of this family is used for denoising.

2.3. Wavelet Denoising Index

We used wavelet soft threshold denoising and wavelet packet analysis denoising to perform denoising respectively and selected six indicators: mean square error (MSE), sum square error (SSE), root mean square error (RMSE), coefficient of determination ( R 2 ), mean absolute error (MAE) and signal-to-noise ratio (SNR) to evaluate the denoising effect. Among these six evaluation indexes, except for the coefficient of determination ( R 2 ) and SNR which are maximum indexes, the other four indexes are minimum indexes. In addition to quantitative evaluation indicators, the denoising effect can also be intuitively displayed by comparing the original signal with the denoised signal image.
Let y i be the sampling point of the original signal, y ^ i be the sampling point of the signal after denoising, and m be the number of sampling points.
M S E = 1 m i = 1 m y i y ^ i 2
S S E = i = 1 m y i y ^ i 2
R M S E = 1 m i = 1 m y i y ^ i 2
R 2 = 1 i = 1 m y i y ^ i 2 i = 1 m y i y ¯ i 2
M A E = 1 m i = 1 m y i y ^ i
S N R = 10 × lg i = 1 m y i 2 i = 1 m y i y ^ i 2

2.4. K-Means Clustering Algorithm

The strategy of K-means clustering is to select the optimal partition by minimizing the loss function. The optimal partition is selected by minimizing the loss function.
First use the square of the Euclidean distance as the distance between samples:
d x i , y i = k = 1 m x k i x k j 2 = x i x j 2
Then the sum of the distances between the sample and the center of the class to which it belongs is the loss function, namely,
W ( C ) = i = 1 k c ( i ) = 1 x i x ¯ l 2
where x ¯ l = x ¯ 1 l , x ¯ 2 l x ¯ n l T is the mean center of classes, namely,
n l = i = 1 n I ( C ( i ) = l )
It is an indicator function, which takes a value of 1 or 0. The function W(C), also known as energy, represents the similarity of samples in the same class.
Therefore, K-means clustering can be transformed into solving an optimization problem:
C * = arg min C W ( C )
The flow chart of the steps of the K-means clustering algorithm is shown in Figure 1.

2.5. Support Vector Machine (SVM)

Support vector machines map vectors into a higher dimensional space and, for linearly separable tasks, find a hyperplane with the largest margin. Two parallel hyperplanes are built on both sides of the hyperplane separating the data, and the separating hyperplane maximizes the distance between the two parallel hyperplanes. The larger the distance or gap between parallel hyperplanes, the smaller the overall error of the classifier. The steps are as follows:
  • Import data;
  • Normalize the data;
  • Execute the support vector machine to find the optimal hyperplane;
  • Draw classification hyperplane kernel support vector;
  • Using polynomial features to perform support vector machines in high-dimensional space;
  • Select an appropriate kernel function and execute a nonlinear support vector machine.

2.6. Decision Tree (DT)

Decision tree is a typical supervised classification method that approximates discrete function values to classify data. First, it processes data and uses an inductive algorithm to generate readable rules and decision trees. Then, it uses the decision to analyze new data. Essentially, decision trees classify data through a series of rules.

2.7. Random Forest (RF)

As an idea of integrated learning, random forest inputs the data obtained by random sampling into many decision trees, performs voting, and obtains the final output result.
Constructing a random forest requires the following four steps:
  • A sample with a sample size of N is drawn N times with replacement, and one sample is drawn each time, then N samples can be formed. These N samples are used to train a decision tree as samples of the root node of the decision tree;
  • If each sample has M attributes, when each node of the decision tree needs to be split, randomly select m attributes from the M attributes to meet the condition m M . From these attributes, information gain is used to select an attribute as the splitting attribute of the node;
  • During the formation of the decision tree, each node is split according to step 2;
  • Build a large number of decision trees according to steps 1–3, and then construct the entire random forest.

2.8. AdaBoost Algorithm

The AdaBoosting algorithm [28] is an important ensemble learning technique, which can enhance a weak learner whose prediction accuracy is only slightly higher than random guessing into a strong learner with high prediction accuracy, which provides a new idea for solving the difficulty of directly constructing a strong learner. AdaBoost is one of the most successful algorithms.
A brief description of the AdaBoost algorithm is as follows:
  • First initialize the weight distribution D1 of the training data. When there are N training sample data, each training sample is given the same weight at the beginning: w 1 = 1 N ;
  • Then, train the weak classifier H. In the specific training process, if a certain training sample point is accurately classified by the weak classifier H, then its corresponding weight should be reduced when constructing the next training set. Conversely, if a training sample point is misclassified, its weight should be increased. The sample set whose weights have been updated is used to train the next classifier, and the entire training process will continue to iterate.
  • Finally, combine the weak classifiers obtained from each training into a strong classifier. After the training process of each weak classifier is over, the weight of the weak classifier with a small classification error rate is increased so that it plays a greater decisive role in the final classification function. The weight of the weak classifier with a large classification error rate is reduced so that it plays a less decisive role in the final classification function.

2.9. Clustering Evaluation Index

When evaluating the prediction results, select the following six indicators: accuracy rate, recall rate, time-consuming, contour coefficient, F value and entropy value. The calculation formulas of precision rate, recall rate and contour coefficient index are as follows:
There are four situations in the data test results: TP: predicted positive, actually positive; TN: predicted negative, actually negative; FP: predicted positive, actually negative; FN: predicted negative, actually positive.
The calculation formula of the indicator is as follows:
  • Accuracy:
    A c c u r a c y = T P + T N T P + T N + F N + F P
  • Recall rate:
    R = T P T P + F N
  • Silhouette coefficient:
    S ( i ) = 1 a ( i ) b ( i ) a ( i ) < b ( i ) 0 a ( i ) = b ( i ) b ( i ) a ( i ) 1 a ( i ) > b ( i )
    where a(i) represents the cohesion of the sample point, and the calculation method is as follows:
    a ( i ) = 1 n 1 j i n distance ( i , j )
    where j represents other sample points in the same class as sample i, and distance represents the distance between i and j. So the smaller a(i) is, the closer the class is. b(i) is calculated similarly to a(i). However, it is necessary to traverse other clusters to obtain multiple values b 1 ( i ) , b 2 ( i ) , b 3 ( i ) b m ( i ) and select the smallest value as the final result.
  • The entropy value of the entire cluster division is:
    e = i = 1 K m i m e i
    where K is the number of clusters, m is the number of members involved in the overall cluster division, m i is the number of all members in cluster i, and e i is the entropy value of each cluster:
    e i = j = 1 L P i j log 2 P i j
    In this formula, P i j refers to the probability that a member in cluster i belongs to class j,
    P i j = m i j m i
    m i j is the number of members in cluster i belonging to class j.
  • The F value is the weighted proportion of different categories of data in the total number of samples. Its formula is as follows:
    F = n i n F i + n j n F j
    where n i , n j is the number of samples of each category, and F i , F j is the F value of each type of sample:
    F i = 2 · P i · R i P i + R i
    P i is the precision rate, and its calculation formula is:
    P = T P T P + F P
    Use the above formula to calculate each evaluation index.

3. Results and Discussion

3.1. Dataset Introduction

We have measured the fault data of rolling bearing wheel sets in a factory in Wuhan. The operating status of rolling bearings in rotating machinery includes normal, inner ring faults and outer ring faults. The inner ring fault and the outer ring fault are recorded as type A and type B faults, respectively. Consider the FFT transformation, choose 2048 as the sample sampling length, and make sample sets in an overlapping manner. In the training set, the normal data contain 200 samples, and each fault dataset of class A and class B contains 100 samples.

3.2. Display of Various Index Results of Different Noise Reduction Methods

3.2.1. Index Analysis of Wavelet Soft Threshold Denoising

When performing wavelet soft threshold denoising, it is necessary to set multiple parameters, including threshold selection criteria, wavelet basis functions, and wavelet levels of decomposition. Among the wavelet basis functions of Haar, Daubechies, Coiflets and other wavelet basis functions of the wavelet basis family, the wavelet basis of the Daubechies family has orthogonality, biorthogonality, compact support, and approximate symmetry, and can perform discrete wavelet transform. Therefore, wavelet transformation is performed using this wavelet basis function. At the same time, wavelet bases with different vanishing moments in the Daubechies family are used to process the denoised signal, and the processing results are compared. In practice, the basic wavelet is often not only required to meet the admissible condition, but also to impose the so-called vanishing moment condition, so that as many wavelet coefficients as possible are zero or as few non-zero wavelet coefficients as possible are generated, which is conducive to data compression and noise elimination. The larger the vanishing moment, the more wavelet coefficients are zeroed out. But in general, the higher the vanishing moment is, the longer the support length is, and the smoother the wavelet is. Therefore, in terms of support length and vanishing moment, we must make a compromise. Considering the general situation, the selected level of decomposition is 4. Change the vanishing moments of the Daubechies wavelet family, and take db3, db6, and db9 three vanishing moments for noise reduction, and analyze the influence of different vanishing moments on noise reduction.
Table 1 and Table 2 show the index analysis results of the first fault signal of type A and type B after wavelet soft threshold denoising.

3.2.2. Wavelet Packet Analysis Denoising Index Analysis

The level of decomposition has a great influence on the denoising effect. If there are too many levels of decomposition, all the wavelet space coefficients of each level will be processed. It is easy to cause serious information signal loss, decrease in signal-to-noise ratio after denoising, and slow processing due to increased computation. If the level of decomposition is too small, the denoising effect is not ideal. For each signal, there is a level of decomposition that denoises best or is close to best. Therefore, when dealing with this problem, on the premise that the wavelet packet base and its vanishing moment are fixed, different levels of decomposition are selected, and the denoising effects are compared to find the most suitable level of decomposition for the denoising signal. Select the Symlets wavelet family, fix its vanishing moment to 6, change its level of decomposition, and perform four-level of decomposition and five-level of decomposition, respectively. Table 3 and Table 4 show the index analysis results of the first fault signal of type A and type B after denoising by wavelet packet analysis.

3.3. Comparative Analysis of Noise Reduction Results

For type A faults, the comparative analysis of wavelet threshold denoising shows that under the condition of constant level of decomposition, with the increase of vanishing moment, the indicators after denoising basically become better. However, when the vanishing moment increases to 9, the mean absolute error MAE increases sharply. It is considered that the signal after denoising changes greatly compared with the original signal, and some characteristics of the original signal may be lost, so it is not used. For wavelet packet analysis denoising, when the wavelet base is determined to be sym6, the denoising effect of wavelet packet analysis with four-level of decomposition is better than that with five-level of decomposition. Then we compare the denoising effects of wavelet soft thresholding and wavelet packet analysis. It is judged that the noise reduction effect of four-level of decomposition with wavelet base sym6 in wavelet packet analysis is better than that of four-level of decomposition with wavelet base db6 in wavelet soft threshold analysis.
Taking the four-level of decomposition whose wavelet base is sym6 in the first four signals of type A fault as an example, the denoising results of wavelet packet analysis are shown in Figure 2.
For the four signals of type B fault, the result analysis is basically consistent with the four signals of type A fault. However, in the comparison between wavelet soft threshold and wavelet packet analysis, it is found that the noise reduction effect of 4-level of decomposition with wavelet base db9 in wavelet packet analysis is better than that of four-level of decomposition with wavelet base sym6 in wavelet soft threshold analysis.
Taking the four-level of decomposition with wavelet base as db9 in the four signals of Type B fault as an example, the denoising results of wavelet packet analysis are shown in Figure 3.
It should be noted that, from the wavelet packet analysis results of the first signal of the Type B fault, it can be seen that the excessive level of decomposition has a serious impact on the denoising effect. From the SNR point of view, using the five-level of decomposition of the sym6 wavelet base, the signal-to-noise ratio is significantly lower than that of the four-level of decomposition, indicating that this denoising method loses a lot of useful signals. This can also be illustrated from the signal images before and after denoising, as shown in Figure 4.

3.4. Applications of Unsupervised Learning

For unsupervised clustering requirements, we use different scales of Shannon entropy E5, E6, E7, E8 and seven indicators of kurtosis, margin, and center of gravity frequency as clustering indicators for clustering. Among them, indicators E5, E6, E7, and E8 are energy indicators.
During data processing, it was observed that direct K-means clustering analysis on the original data is not effective, because K-means clustering only considers the absolute distance between each sample point and is not sensitive to data of different orders of magnitude. Therefore, it is considered to perform logarithmic transformation on the data to expand the influence of the order of magnitude on the distance, so that it meets the requirements of cluster analysis. It is found that the clustering effect is better and the clustering result is stable. At this time, the effect diagram of K-means cluster analysis is shown in Figure 5.
Table 5 shows the evaluation results of 100 K-means clustering.
For this result, the mean accuracy was 94%, and each item achieved 90%. The clustering result is stable, and the standard deviation of accuracy is 1.63.

3.5. Applications of Supervised Learning

In order to solve the binary classification problem based on the supervised learning method, quadratic support vector machine (SVM), random forest (RF), decision tree (DT) and AdaBoost algorithm are used for binary classification.
On the basis of the evaluation index of unsupervised learning, the evaluation index of entropy value and contour coefficient is removed, and the evaluation index AUC value and precision rate are added. The AUC value is an evaluation index dedicated to the binary classification, which is used to evaluate the quality of the binary classification classifier. Its advantage is that it can avoid converting probability predictions into categories. A classifier with a larger AUC value has a higher correct rate.
Figure 6 shows the clustering results of the double SVM.
Table 6 shows the results of classification evaluation metrics using various supervised machine learning algorithms.
It can be seen from the table that the accuracy of double SVM classification is the lowest, and when the accuracy of other classification methods is the same, the time-consumption of the decision tree is the shortest and the AUC value is the highest. From this analysis, it can be concluded that for the binary classification algorithm using supervised learning, the effect of equipment fault diagnosis using decision tree classification is the most ideal.
Using the decision tree algorithm to conduct 100 classification experiments, the following evaluation Table 7 can be obtained:
For this result, the mean accuracy rate is 96.8%, and each item is above 95%. Therefore, the clustering result is stable, and the standard deviation of the accuracy rate is 0.330.

3.6. The Effect of Denoising on the Effect of Clustering Experiments

In order to compare and analyze the impact of denoising on the test results, we counted the changes in the classification indicators of unsupervised machine learning and supervised machine learning, respectively.
It can be seen from Table 8 that if unsupervised learning is used for binary classification experiments, the accuracy and recall rate of the data after denoising will be greatly improved and the time spent will be shortened compared with that before denoising. It proves that the data features after denoising are more obvious than those before denoising. After noise reduction, the redundant data that need to be processed by the algorithm are reduced, so the usage time is reduced. The silhouette coefficient before noise reduction has a slight increase compared with that after noise reduction, which proves that the clustering effect is better after noise removal. The F value in the evaluation index indicates that the results of the test sample can represent the true degree of the whole. Because the data before and after denoising are clustered with all the signal samples, no test samples are proposed for learning, so the F value before and after denoising has not changed. The entropy value reflects the reliability of the information. A small decrease in the entropy value after denoising indicates that the original signal tends to be stable after removing the messy noise signal.
For the classification effect of supervised learning, the signal features before and after denoising are shown in Table 9 and Table 10.
From the results obtained in these two tables, it can be seen that for the decision tree algorithm and the AdaBoost algorithm, the accuracy and recall rates are slightly improved. However, for the double SVM algorithm, the accuracy and recall rate did not increase significantly, which shows that for the supervised clustering algorithm, whether denoising has an impact on the clustering accuracy, but the impact is not significant. There is no significant difference in time consumption, indicating that noise does not affect the time complexity of supervised clustering algorithms. The same as the unsupervised clustering algorithm, the F value in the evaluation index indicates that the results of the test sample can represent the true degree of the whole. However, because the supervised algorithm obtains the classification label in advance, and the denoised samples can better show the characteristics of their own category, the F value increases after denoising.
Combining the tables of unsupervised learning and supervised learning, we can conclude that whether using the clustering method of unsupervised learning or the clustering method of supervised learning, signal denoising can significantly improve classification performance, making the results more accurate and reliable.

4. Conclusions

This paper proposes a machine learning fault signal diagnosis and classification model based on wavelet denoising, which can be used for intelligent diagnosis of rolling bearing faults in rotating machinery in practical mechanical applications. Experimental results show that the proposed model is significantly better than traditional diagnostic models on instrument faults. Aiming at weak signal strength and background noise interference, under the quantitative evaluation index, a better noise reduction algorithm is selected for wavelet denoising, which can better restore the true characteristics of the fault signal. Using unsupervised learning and supervised machine learning classification algorithms, the evaluation indicators before and after denoising are compared to make the classification results more accurate and reliable. By designing the fault classification evaluation prediction criteria, the following conclusions are drawn. In equipment fault diagnosis, if the unsupervised machine learning method is used, K-means clustering can achieve good classification results. If the supervised machine learning method is used, the effect of classification by decision tree is the most ideal. This article will help researchers to intelligently diagnose the faults of instruments and equipment.Therefore, the classification model is widely used in industrial production.
In addition, increasing the types of rolling bearing faults, narrowing the characteristic differences of fault signals and improving the adaptability and flexibility of the model will be the focus of our future work. The dataset used in this paper is limited to one hundred groups of two types of faults under different working conditions. More types of faults and larger datasets can be considered in future research to improve the generalization ability of the model. The wavelet denoising algorithm can be further optimized by selecting different wavelet bases, levels of decomposition and threshold functions according to different types of signals. Machine learning models can further improve their performance by tuning parameters, selecting features, or combining multiple models.

Author Contributions

Conceptualization, Y.W. and S.F.; methodology, S.F.; software, S.F.; validation, S.F. and R.W.; formal analysis, R.W., Y.W. and S.F.; investigation, Y.W.; resources, R.W.; data curation, R.W.; writing—original draft preparation, S.F.; writing—review and editing, M.M.; visualization, Y.W.; supervision, M.M.; project administration, Y.W. and R.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Kulshrestha, A.; Mahela, O.P.; Gupta, M.K. A Discrete Wavelet Transform and Rule Based Decision Tree Based Technique for Identification of Fault in Utility Grid Network with Wind Energy. In Proceedings of the 2021 International Conference on Advances in Electrical, Computing, Communication and Sustainable Technologies (ICAECT), Bhilai, India, 19–20 February 2021; pp. 1–6. [Google Scholar]
  2. Lei, Y.; Jia, F.; Lin, J.; Xing, S.; Ding, S.X. An Intelligent Fault Diagnosis Method Using Unsupervised Feature Learning Towards Mechanical Big Data. IEEE Trans. Ind. Electron. 2016, 63, 3137–3147. [Google Scholar] [CrossRef]
  3. Peng, H.-W.; Chiang, P.-J. Control of mechatronics systems: Ball bearing fault diagnosis using machine learning techniques. In Proceedings of the 2011 8th Asian Control Conference (ASCC), Kaohsiung, Taiwan, 15–18 May 2011. [Google Scholar]
  4. Ambrożkiewicz, B.; Syta, A.; Gassner, A.; Georgiadis, A.; Litak, G.; Meier, N. The influence of the radial internal clearance on the dynamic response of self-aligning ball bearings. Mech. Syst. Signal Process. 2022, 171, 108954. [Google Scholar] [CrossRef]
  5. Chen, H.; Jiang, B.; Ding, S.X.; Huang, B. Data-Driven Fault Diagnosis for Traction Systems in High-Speed Trains: A Survey, Challenges, and Perspectives. Mech. Syst. Signal Process. 2022, 23, 1700–1716. [Google Scholar] [CrossRef]
  6. Zhang, T.; Chen, J.; Li, F.; Zhang, K.; Lv, H.; He, S.; Xu, E. Intelligent fault diagnosis of machines with small & imbalanced data: A state-of-the-art review and possible extensions. ISA Trans. 2022, 119, 152–171. [Google Scholar] [PubMed]
  7. Almounajjed, A.; Sahoo, A.K.; Kumar, M.K.; Alsebai, M.D. Investigation Techniques for Rolling Bearing Fault Diagnosis Using Machine Learning Algorithms. In Proceedings of the 2021 5th International Conference on Intelligent Computing and Control Systems (ICICCS), Madurai, India, 6–8 May 2021; pp. 1290–1294. [Google Scholar]
  8. Randall, R.B.; Antoni, J. Rolling element bearing diagnostics—A tutorial. Mech. Syst. Signal Process. 2011, 25, 485–520. [Google Scholar] [CrossRef]
  9. Georgoulas, G.; Loutas, T.; Stylios, C.D.; Kostopoulos, V. Bearing fault detection based on hybrid ensemble detector and empirical mode decomposition. Mech. Syst. Signal Process. 2013, 41, 510–525. [Google Scholar] [CrossRef]
  10. Zhang, W.; Li, C.; Peng, G.; Chen, Y.; Zhang, Z. A deep convolutional neural network with new training methods for bearing fault diagnosis under noisy environment and different working load. Mech. Syst. Signal Process. 2018, 100, 439–453. [Google Scholar] [CrossRef]
  11. Xing, S.; Lei, Y.; Yang, B.; Lu, N. Adaptive Knowledge Transfer by Continual Weighted Updating of Filter Kernels for Few-Shot Fault Diagnosis of Machines. IEEE Trans. Ind. Electron. 2021, 69, 1968–1976. [Google Scholar] [CrossRef]
  12. Sun, T.; Yu, G.; Gao, M.; Zhao, L.; Bai, C.; Yang, W. Fault Diagnosis Methods Based on Machine Learning and its Applications for Wind Turbines: A Review. IEEE Access 2021, 9, 147481–147511. [Google Scholar] [CrossRef]
  13. He, Z.; Shao, H.; Cheng, J.; Zhao, X.; Yang, Y. Support tensor machine with dynamic penalty factors and its application to the fault diagnosis of rotating machinery with unbalanced data. Mech. Syst. Signal Process. 2019, 141, 106441. [Google Scholar] [CrossRef]
  14. Chandran, P.; Thiery, F.; Odelius, J.; Lind, H.; Rantatalo, M. Unsupervised Machine Learning for Missing Clamp Detection from an In-Service Train Using Differential Eddy Current Sensor. Sustainability 2022, 14, 1035. [Google Scholar] [CrossRef]
  15. Chandran, P.; Thierry, F.; Odelius, J.; Famurewa, S.M.; Lind, H.; Rantatalo, M. Supervised Machine Learning Approach for Detecting Missing Clamps in Rail Fastening System from Differential Eddy Current Measurements. Appl. Sci. 2021, 11, 4018. [Google Scholar] [CrossRef]
  16. Yoon, H.J. Finding Unexpected Test Accuracy by Cross Validation in Machine Learning. IJCSNS Int. J. Comput. Sci. Netw. Secur. 2021, 21, 549–555. [Google Scholar]
  17. Yang, C.; Létourneau, S. Learning to Predict Train Wheel Failures. In Proceedings of the 11th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD 2005), Chicago, IL, USA, 21–24 August 2005; pp. 516–525. [Google Scholar]
  18. Nason, G.P.; Silverman, B.W. The stationary wavelet transform and some statistical applications. Science 1995, 103, 281–299. [Google Scholar]
  19. Baydar, N.; Ball, A. Detection of gear failures via vibration and acoustic signals using wavelet transform. Mech. Syst. Signal Proc. 2001, 17, 787–804. [Google Scholar] [CrossRef]
  20. Abid, B.; Sallem, M.; Braham, A. Optimized SWPT and Decision Tree for Incipient Bearing Fault Diagnosis. In Proceedings of the 2019 19th International Conference on Sciences and Techniques of Automatic Control and Computer Engineering (STA), Sousse, Tunisia, 24–26 March 2019. [Google Scholar]
  21. Lu, X.; Zhao, Y.; Li, F. On the use of higher-density wavelet packet neighboring coefficient to improve SNR for signal denoising. In Proceedings of the 2020 International Conference on Intelligent Computing, Automation and Systems (ICICAS), Chongqing, China, 11–13 December 2020; pp. 337–340. [Google Scholar]
  22. Aslam, M.; Rehan, M.S.; Albogamy, F.R.; Murawwat, S.; Basit, A.; Hafeez, G. Prognostication of Failures Using Signal-to-Noise Ratio to Determine Partial Discharges Activities in Power Transformers. IEEE Access 2022, 10, 87500–87507. [Google Scholar] [CrossRef]
  23. Shim, J.; Koo, J.; Park, Y.; Kim, J. Anomaly Detection Method in Railway Using Signal Processing and Deep Learning. Appl. Sci. 2022, 12, 12901. [Google Scholar] [CrossRef]
  24. Palangar, M.F.; Mohseni, S.; Mirzaie, M.; Mahmoudi, A. Designing an Automatic Detector Device to Diagnose Insulator State on Overhead Distribution Lines. IEEE Trans. Ind. Inform. 2022, 18, 1072–1082. [Google Scholar] [CrossRef]
  25. Stępień, K.; Makiela, W.; Stoić, A.; Samardžić, I. Defining the criteria to select the wavelet type for the assessment of surface quality. Teh. Vjesn. Gaz. 2015, 22, 781–784. [Google Scholar] [CrossRef]
  26. Li, C.; Cao, D.; Yuan, Y. Research on Improved Wavelet Denoising Method for sEMG Signal. In Proceedings of the 2019 Chinese Automation Congress (CAC), Hangzhou, China, 22–24 November 2019; pp. 5221–5225. [Google Scholar]
  27. Yang, G.; Song, Y.; Du, J. Speech Signal Denoising Algorithm and Simulation Based on Wavelet Threshold. In Proceedings of the 2022 4th International Conference on Natural Language Processing (ICNLP), Xi’an, China, 25–27 March 2022; pp. 304–309. [Google Scholar]
  28. Liu, C.; Zhang, J.; Reng, G. Fault Diagnosis of Analogue Circuits Using Hyper-sphere SVM and AdaBoost. In Proceedings of the 2009 Second International Symposium on Computational Intelligence and Design, Changsha, China, 12–14 December 2009; pp. 11–14. [Google Scholar]
Figure 1. 5-layer decomposition of original signal and denoised signal.
Figure 1. 5-layer decomposition of original signal and denoised signal.
Applsci 13 05936 g001
Figure 2. The original signal and denoised signal of the first four signals of Type A fault.
Figure 2. The original signal and denoised signal of the first four signals of Type A fault.
Applsci 13 05936 g002
Figure 3. The original signal and denoised signal of the four signals of Type B fault.
Figure 3. The original signal and denoised signal of the four signals of Type B fault.
Applsci 13 05936 g003
Figure 4. 5-level of decomposition of original signal and denoised signal.
Figure 4. 5-level of decomposition of original signal and denoised signal.
Applsci 13 05936 g004
Figure 5. K-means cluster analysis effect diagram.
Figure 5. K-means cluster analysis effect diagram.
Applsci 13 05936 g005
Figure 6. Double SVM clustering result graph.
Figure 6. Double SVM clustering result graph.
Applsci 13 05936 g006
Table 1. Analysis results of the first signal index of Type A fault.
Table 1. Analysis results of the first signal index of Type A fault.
Type of WaveletMSESSERMSE R 2 MAESNR
db31.20 × 10 3 5.04 × 10 0 3.50 × 10 2 9.88 × 10 1 4.46 × 10 6 1.96 × 10 1
db66.40 × 10 4 2.61 × 10 0 2.52 × 10 2 9.94 × 10 1 2.59 × 10 6 2.25 × 10 1
db94.60 × 10 4 1.89 × 10 0 2.15 × 10 2 9.96 × 10 1 2.30 × 10 5 2.39 × 10 1
Table 2. Analysis results of the first signal index of Type B fault.
Table 2. Analysis results of the first signal index of Type B fault.
Type of WaveletMSESSERMSE R 2 MAESNR
db33.87 × 10 3 1.59 × 10 1 6.22 × 10 2 8.64 × 10 1 3.56 × 10 6 8.69 × 10 0
db61.91 × 10 3 7.82 × 10 0 4.37 × 10 2 9.33 × 10 1 4.72 × 10 6 1.18 × 10 1
db91.33 × 10 3 5.47 × 10 0 3.65 × 10 2 9.53 × 10 1 5.06 × 10 6 1.33 × 10 1
Table 3. Analysis results of the first signal index of Type A fault.
Table 3. Analysis results of the first signal index of Type A fault.
LevelsMSESSERMSE R 2 MAESNR
46.25 × 10 4 2.56 × 10 0 2.50 × 10 2 9.94 × 10 1 4.46 × 10 6 2.26 × 10 1
51.10 × 10 2 4.50 × 10 1 1.05 × 10 1 8.95 × 10 1 4.46 × 10 6 1.01 × 10 1
Table 4. Analysis results of the first signal index of Type B fault.
Table 4. Analysis results of the first signal index of Type B fault.
levelsMSESSERMSE R 2 MAESNR
41.88 × 10 3 7.74 × 10 0 4.35 × 10 2 9.34 × 10 1 5.06 × 10 6 1.18 × 10 1
52.67 × 10 2 1.09 × 10 2 1.63 × 10 1 6.50 × 10 2 1.32 × 10 6 3.00 × 10 1
Table 5. Evaluation index table of clustering results.
Table 5. Evaluation index table of clustering results.
Number of ExperimentsAccuracy/%Recall Rate/%Time Cost/msSilhouette CoefficientF ValueEntropy Value
19697.11.90.930.960.24
29495.71.80.910.940.26
 ⋮ ⋮ ⋮ ⋮ ⋮ ⋮ ⋮
1009291.61.90.920.950.27
average9494.81.870.920.950.26
standard deviation1.632.330.050.010.010.01
Table 6. Evaluation index table of clustering results of different classifiers.
Table 6. Evaluation index table of clustering results of different classifiers.
Classification MethodAccuracy/%Recall Rate/%Time Cost/msSilhouette CoefficientF ValueEntropy Value
decision tree96.796.7180.9680.9660.987
random forest96.796.710400.9680.9660.91
Adaboost96.796.78670.9680.9670.986
double SVM9695.816900.960.960.987
Table 7. Evaluation table for 100 classification experiments.
Table 7. Evaluation table for 100 classification experiments.
Number of ExperimentsAccuracy/%Recall Rate/%Time Cost/msSilhouette CoefficientF ValueEntropy Value
196.796.7180.9680.9660.987
297.297.2180.9730.9740.987
 ⋮ ⋮ ⋮ ⋮ ⋮ ⋮ ⋮
10096.496.4180.9650.9620.987
average96.896.8180.9690.9670.987
standard deviation0.330.3300.0030.0050
Table 8. Unsupervised clustering features of signals before and after denoising.
Table 8. Unsupervised clustering features of signals before and after denoising.
 Accuracy/%Time Cost/msRecall Rate/%Silhouette CoefficientF ValueEntropy Value
Before denoising773.5710.8340.960.242
After denoising961.9297.140.9280.960.237
Table 9. Supervised clustering features of signals before denoising.
Table 9. Supervised clustering features of signals before denoising.
Classification MethodAccuracy/%Time Cost/msRecall RateSilhouette CoefficientF ValueEntropy Value
decision tree93.393.3200.9330.9330.987
Adaboost93.393.38990.9330.9330.986
double SVM96.095.816900.9600.9600.987
Table 10. Supervised clustering features of signals after denoising.
Table 10. Supervised clustering features of signals after denoising.
Classification MethodAccuracy/%Time Cost/msRecall RateSilhouette CoefficientF ValueEntropy Value
decision tree96.796.7180.9680.9660.987
Adaboost96.796.78670.9680.9670.986
double SVM96.195.816900.9600.9600.987
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fu, S.; Wu, Y.; Wang, R.; Mao, M. A Bearing Fault Diagnosis Method Based on Wavelet Denoising and Machine Learning. Appl. Sci. 2023, 13, 5936. https://doi.org/10.3390/app13105936

AMA Style

Fu S, Wu Y, Wang R, Mao M. A Bearing Fault Diagnosis Method Based on Wavelet Denoising and Machine Learning. Applied Sciences. 2023; 13(10):5936. https://doi.org/10.3390/app13105936

Chicago/Turabian Style

Fu, Shaokun, Yize Wu, Rundong Wang, and Mingzhi Mao. 2023. "A Bearing Fault Diagnosis Method Based on Wavelet Denoising and Machine Learning" Applied Sciences 13, no. 10: 5936. https://doi.org/10.3390/app13105936

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop