Next Article in Journal
Effect of Spatial and Spectral Scaling on Joint Characterization of the Spectral Mixture Residual: Comparative Analysis of AVIRIS and WorldView-3 SWIR for Geologic Mapping in Anza-Borrego Desert State Park
Previous Article in Journal
Comparison of Simulated and Measured Power of the Earth-Space Link for Satellite-Based AIS Signals
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on Wind Turbine Fault Detection Based on the Fusion of ASL-CatBoost and TtRSA

School of Information Science and Technology, Qingdao University of Science and Technology, Qingdao 266061, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(15), 6741; https://doi.org/10.3390/s23156741
Submission received: 25 May 2023 / Revised: 30 June 2023 / Accepted: 10 July 2023 / Published: 28 July 2023
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

:
The internal structure of wind turbines is intricate and precise, although the challenging working conditions often give rise to various operational faults. This study aims to address the limitations of traditional machine learning algorithms in wind turbine fault detection and the imbalance of positive and negative samples in the fault detection dataset. To achieve the real-time detection of wind turbine group faults and to capture wind turbine fault state information, an enhanced ASL-CatBoost algorithm is proposed. Additionally, a crawling animal search algorithm that incorporates the Tent chaotic mapping and t-distribution mutation strategy is introduced to assess the sensitivity of the ASL-CatBoost algorithm toward hyperparameters and the difficulty of manual hyperparameter setting. The effectiveness of the proposed hyperparameter optimization strategy, termed the TtRSA algorithm, is demonstrated through a comparison of traditional intelligent optimization algorithms using 11 benchmark test functions. When applied to the hyperparameter optimization of the ASL-CatBoost algorithm, the TtRSA-ASL-CatBoost algorithm exhibits notable enhancements in accuracy, recall, and other performance measures compared with the ASL-CatBoost algorithm and other ensemble learning algorithms. The experimental results affirm that the proposed algorithm model improvement strategy effectively enhances the wind turbine fault detection classification recognition rate.

1. Introduction

With the rapid development of the global economy, the scale of demand for energy continues to expand. Traditional thermal power generation methods are highly prone to causing environmental pollution and they do not meet the requirements of sustainable development. Green power generation methods such as wind power generation are in line with the future development direction of the energy industry. With the development of Industry 4.0, the global installed capacity of wind turbines is expected to reach two billion kilowatts by 2030 [1]. Promoting the development of renewable energy will not only meet the energy needs of economic development, but also reduce the proportion of traditional thermal power generation methods, and accelerate the construction of a clean, low-carbon, energy-efficient system [2]. However, with the continuous expansion of wind power generation, and an increase in operational lifespan, the issue of turbine faults has become increasingly prominent, posing a series of challenges to the wind power industry. The conventional maintenance approach for wind turbine units is typically based on scheduled inspections and fault responses, which presents several issues. Firstly, scheduled inspections often fail to accurately predict the occurrence time and types of turbine faults, which can result in unnecessary maintenance and high costs. Secondly, maintenance carried out in response to faults is often conducted after the occurrence of the faults, potentially leading to the prolonged downtime of the units and a reduction in production capacity. Additionally, the remote geographical locations where wind turbine units are typically installed pose significant challenges and high costs for maintenance, limiting their reliability and maintainability. Therefore, the development and implementation of wind turbine fault detection systems are of great significance. By leveraging advanced sensor technologies, data analysis, and machine learning algorithms, real-time monitoring of the operational status and performance parameters of the units can be achieved, enabling the timely detection of potential faults and abnormalities [3]. Accurate fault detection can help reduce maintenance costs, improve maintenance efficiency, and optimize operations and equipment performance [4,5]. Moreover, by enhancing the accuracy and timeliness of fault detection, the safety of wind turbine units can be enhanced, reducing the risk of accidents and promoting the overall sustainable development of the wind power industry.
Artificial intelligence and machine learning have unique advantages in the field of wind turbine fault detection [6,7,8,9]. In order to improve the detection speed and accuracy of wind turbine faults, a novel dynamic model sensor method is proposed for SCADA data-based wind turbine fault detection. A dynamic model representing the relationship between the generator temperature, wind speed, and ambient temperature is constructed using the first principles, and it is used as the basic structure of the model sensor. When the model sensor is applied for fault detection, its parameters are updated regularly using the generator temperature, wind speed, and ambient temperature data from the SCADA system. Then, from the updated model, the fault sensitive features of the wind turbine system are extracted by performing system frequency analysis for use in turbine fault detection [10]. Aziz, U [11] used a realistic framework for SCADA data simulation by critically comparing power-based wind turbine fault-detection methods. Song [12] proposed the use of an improved denoising autoencoder to detect wind turbine rolling bearing faults. Liu [13] proposed a twin neural network method that allows the algorithm to achieve wind turbine fault detection with only a small amount of training data. In order to solve the problem of inaccurate and untimely fault detection caused by wind turbine data features, Liu [14] proposed a new deep network called the Deep Residual Network (DRN) for the fault detection of wind turbines. The results indicate that the proposed DRN achieves a better performance and outperforms some published fault detection methods. However, none of the aforementioned studies considered the impact of the imbalanced positive and negative samples in wind turbine fault detection datasets on the accuracy of algorithms, leaving room for further improvement in the field concerning the detection accuracy of faults.
Fault detection models typically contain many hyperparameters that are not learned from data, but are manually set by the user. These hyperparameters play a crucial role in determining the performance and behavior of the model. However, finding the optimal combination of hyperparameters can be a challenging and time-consuming task. Swarm intelligence optimization algorithms can search for the optimal solution distributed in a certain range of space, and they have good parallelism and autonomous exploration, which is of significance for the hyperparameter optimization of fault detection models [15,16,17]. Lei [18] analyzed a fault detection model based on long short-term neural networks and Bayesian optimization algorithms, and they applied the model to the fault warning of the induced draft fan of a coal-fired power plant, achieving good results. Zhang [19] used the Whale Optimization Algorithm in order to find the global optimal solution, realize hyperparameter optimization of the BiLSTM network, improve the prediction accuracy of wind power generation, and save a significant amount of time on debugging. Huang [20] proposed a crawling animal search algorithm based on the interactive cross strategy of Levy flight, and they verified its effectiveness in practical engineering via innovations such as the welding beam design.
Based on the above research results and their shortcomings, the following work was conducted in this paper. Firstly, the accuracy and recall rate of the CatBoost algorithm to achieve fault detection is not high enough, and there is an imbalance between positive and negative samples in the wind turbine fault detection dataset. The cross-entropy loss function of the CatBoost algorithm was thus replaced by an asymmetric loss function, and the ASL-CatBoost algorithm was proposed to implement wind turbine icing fault detection and capture the state information of wind turbine faults. To verify the effectiveness of the improved method, experiments were conducted using integrated learning algorithms (CatBoost, XGBoost, LightGBM, etc.), machine learning algorithms (SVM), and deep learning algorithms (LSTMAE) as comparison algorithms for the wind turbine icing fault dataset. The experimental results show the superior performance of the improved algorithms. Secondly, due to the difficulty in setting appropriate hyperparameters in the ASL-CatBoost algorithm, in this paper, we propose an improved reptile search algorithm based on the tent chaotic map and t-distribution mutation strategy to optimize hyperparameters such as learning rate, iteration number, and the tree depth of the ASL-CatBoost algorithm. The improved algorithm was also proposed in order to introduce the optimized hyperparameters into the ASL-CatBoost algorithm for model training. To verify the optimization ability of the improved reptile search algorithm, the TtRSA algorithm was compared with classical population intelligence optimization algorithms such as PSO, WOA, and SSA regarding 11 benchmark functions. The experimental results show that the improved reptile search algorithm has better performance, convergence speed, and accuracy. Finally, the TtRSA-optimized ASL-CatBoost algorithm has a higher detection accuracy and detection efficiency than the original ASL-CatBoost algorithm.

2. Materials and Methods

2.1. CatBoost Algorithm

In 2017, Yandex proposed a new integrated learning algorithm called CatBoost [21]. This algorithm is an improvement on Gradient Boosting Decision Trees, and it outperforms other algorithms in the same GBDT [22] framework, such as XGBoost [23] and LightGBM [24], in terms of model accuracy. The main innovation of the CatBoost algorithm is the use of Ordered Boosting instead of the traditional gradient estimation method, which solves the problems of Gradient Bias and Prediction Shift. In addition, the algorithm uses Oblivious trees as base models, which improves the model’s ability to classify correctly, and it takes into account its generalization ability, effectively preventing algorithm over fitting.
The Gradient Boosting Decision Tree [25] algorithm uses One-hot encoding during the category encoding process. However, when the data dimension is high, the problem of dimension explosion may arise. To address this issue, CatBoost has designed a method called Ordered Target Statistics. This method first randomly arranges all data samples S = {(X_1, Y_1), (X_2, Y_2), (X_3, Y_3), ···, (X_n, Y_n)} to generate multiple sets of random sequences. During the training process, the average label value replaces the category for a particular feature sequence. Assuming that σ = (σ_1, σ_2, σ_3,..., σ_n) is the reordered sequence of the dataset, the k feature xik of the i sample in the original dataset can be represented by σ, as shown in Equation (1) [26]. This method can convert categorical features into numerical features, reduce computational complexity, and minimize information loss.
x σ p k = j = 1 p 1 [ x σ j k = x M σ p k ] Y σ j + a P j = 1 p 1 [ x σ j k = χ σ p k ] + a
After converting categorical features into numerical features using the Ordered Target Statistics method, feature interactions may be affected because numerical features cannot be effectively cross-matched. CatBoost uses a greedy strategy to perform feature interactions. During the first split of the tree generation, CatBoost does not use any cross-features. In subsequent splits, CatBoost uses all of the original features and cross-features that were used to generate the tree, as well as all categorical features in the dataset, to perform feature interactions.
During the model training process using XGBoost and LightGBM algorithms, we found that the model can fit well with F1 wind turbine data, but its fitting effect was poor when testing F2 wind turbine data. Based on this, the CatBoost algorithm proposed the idea of Ordered Boosting, which can effectively reduce the error of gradient estimation and alleviate the problem of prediction shift.
The CatBoost algorithm uses a symmetric binary tree as the base model, and this tree’s structural constraint has a certain regularity effect. For the prediction process of the CatBoost algorithm, the splitting of each feature is independent and not sequential. Multiple samples can be predicted together, improving the prediction speed of the CatBoost algorithm.

2.2. Introduction to ASL-CatBoost Algorithm

The icing fault detection of wind turbines is a typical imbalanced data classification problem. During the entire lifecycle of a wind turbine’s operation, fault data only accounts for a very small portion of that operation, which can easily cause the model to be greatly affected by normal data, and it can make it difficult to improve the detection accuracy of fault data. The default cross entropy loss function of the CatBoost algorithm is not good at dealing with the problem of unbalanced positive and negative samples in the dataset. To solve the problem of unbalanced positive and negative samples in the dataset, He Kaiming and others proposed Focal Loss [27], as shown in Equation (2), where Pt is the probability that the prediction sample is a positive sample and γ is a weight parameter. However, in the actual scenario application process, the author found that the accuracy of the loss function was not high enough.
Focal   Loss = { ( 1 P t ) γ log P t , y = 1 P t γ log ( 1 P t ) , y = 0
Therefore, this paper proposes an improved asymmetric loss function based on the focal loss function and considers the application of ASL for the CatBoost algorithm. The main innovations of the asymmetric loss function are as follows:
(1) As shown in Equation (3), the asymmetric loss function focuses on the γ Parameter decoupling to γ+ and γ−. The loss weights of positive samples and negative samples are adjusted using the asymmetric focusing method to reduce the impact of negative samples and simple samples on the loss function and help the model better learn meaningful features in positive samples and difficult to detect samples.
A S L * = { ( 1 P t ) γ + log P t , y = 1 P t γ log ( 1 P t ) , y = 0
(2) In order to reduce the contribution of negative samples to the loss function as much as possible, with high confidence, Asymmetric Loss proposed a probability transfer mechanism to process the hard threshold of negative samples with high confidence. As shown in Equation (4), m ≥ 0 is an adjustable hyperparameter, which is generally set to 0.2. When the predicted probability Pt of positive samples is less than the set hyperparameter m, it indicates that the current sample has a high probability of being a negative sample. Therefore, the probability of predicting the sample as a positive sample can be directly set to 0, and the probability result of predicting the sample as a positive sample can be returned in Pm.
P m = M A X ( P t m , 0 )
After completing the above two improved methods, the Asymmetric Loss expression was finally obtained, as shown in (5).
ASL = { ( 1 P t ) γ + log P t ,   y = 1 ( P m ) γ log ( 1 P m ) , y = 0
In conclusion, this paper replaces the default cross entropy loss function of the CatBoost algorithm with Asymmetric Loss, and it proposes the ASL-CatBoost algorithm, which makes the algorithm more sensitive to fault data. The Asymmetric Loss function has three adjustable parameters, namely, γ + , γ , and m. During the process of detecting the icing fault of wind turbines using the ASL CatBoost algorithm, the author found that it is more appropriate to set γ + as 2, γ as 3, and m as 0.3. The model training was conducted under this super parameter. Since the prediction probability is a value between 0–1, the value after the 3rd power is less than the value after the 2nd power, therefore, the impact of negative samples on the Loss function will be reduced. At the same time, for a negative sample, if the predicted result is 0.1, the confidence level of the negative sample is very high. The Loss function will conclude that it is a negative sample, and the weight influence of the sample on the Loss function is 0. Therefore, the improved ASL CatBoost algorithm will focus on training difficult to detect samples and positive samples to improve the detection accuracy of fault data. The feasibility of this algorithm was verified in Section 4.3 of the article.

2.3. Introduction to the Reptile Search Algorithm

Setting the ASL-CatBoost algorithm training hyperparameter, as previously proposed, has a great impact on the accuracy and efficiency of the algorithm’s fault detection abilities. In order to find the optimal parameter combination and reduce the impact of human factors on the accuracy of the algorithm, an improved Reptile Search Algorithm is proposed to optimize the hyperparameter of the ASL-CatBoost algorithm and improve the fault detection speed and detection accuracy of ASL-CatBoost algorithm.
In 2021, Laith Abualigah proposed a meta-heuristic optimizer called the Reptile Search Algorithm (RSA) [28]. The main function of this algorithm is to simulate the hunting behavior of crocodiles. The two main features of crocodile behavior in the algorithm are considered to be ‘rounding up’ and hunting; switching between these two different behaviors is affected by the current number of iterations and the maximum number of iterations. When the current number of iterations is t ≤ T/2, the encirclement strategy is executed; when t > T/2, the hunting phase is performed. The round-up process also includes two steps: high-altitude walking or belly walking. Hunting is achieved through hunting coordination or hunting cooperation. The specific process of the algorithm is as follows:
(1)
Initialization phase
In RSA, the optimization process starts with a set of candidate solutions, and in each iteration, the optimal solution obtained is considered to be close to the optimal value. Among them, X is a randomly generated set of candidate solutions, as shown in Equation (6).
X i , j = r a n d · ( U B L B ) + L B , i = 1 , 2 , 3 , , N   j = 1 , 2 , 3 , 4 , , n
In the equation, X i ,   j represents the position of the i-crocodile individual in the j dimension, N is the number of candidate solutions, n is the dimension of the given problem, r a n d belongs to the random function in the interval [0, 1], and LB and UB represent the given lower and upper bounds of the problem.
(2)
Encirclement stage
When tT/2, the algorithm is in the early stage of its iteration, where the crocodile population searches globally and enters the bounding phase. When tT/4, the crocodile population adopts a high-altitude walking strategy, and when T/4 < tT/2, the crocodile population implements an abdominal walking strategy. The position update equation for the crocodile population during the encirclement exploration phase is shown in Equation (7).
x ( i , j ) ( t + 1 ) = { B e s t j ( t ) η ( i , j ) ( t ) · β R ( i , j ) ( t ) · r a n d , t T 4 B e s t j · x ( r 1 , j ) · E S ( t ) · r a n d ,                               T 4 < t T 2
In the equation, B e s t j ( t ) represents the position of the optimal solution at the current moment, t is the current number of iterations, T is the maximum number of iterations, and η ( i , j )   ( t ) represents the hunting behavior of the i candidate solution in the j dimension’s operator; the calculation is shown in Equation (8). β is a sensitive parameter which controls the exploration accuracy of the encirclement stage during the iterative process, and it is fixed at 0.1. R ( i , j ) ( t ) is a reduction function used to reduce the search area value, and it is calculated using Equation (9). r 1 is a random integer between (1, N), x ( r 1 , j ) , indicating the j dimension position of the r 1 random candidate solution. N is the number of candidate solutions and evolution factor E S ( t ) is a probability ratio. During the entire iteration process, the value randomly decreased between 2 and −2, and it was calculated using Equation (10).
η ( i , j ) = B e s t j ( t ) · P ( i , j )
R ( i , j ) = B e s t j ( t ) x ( r 2 , j ) B e s t j ( t ) + ϵ
E S ( t ) = 2 r 3 ( 1 t T )
In the equation, ϵ is a very small positive number, r 2 is a random integer of [1, N], r 3 represents a random integer between [−1, 1], and P ( i , j ) represents the percentage difference between the optimal solution and the j dimension position of the current solution, calculated as shown in Equation (11).
P ( i , j ) = α + x ( i , j ) M ( x i ) B e s t j ( t ) · ( U B ( j ) L B ( j ) ) + ϵ
M ( x i ) represents the average position of the i candidate solution, and its calculation is shown in Equation (12). U B ( j ) and L B ( j ) represent the upper and lower bounds of the j dimensional position, respectively. α is a sensitive parameter used to control the search accuracy of hunting cooperation during the iteration process (the difference between candidate solutions), which is fixed to 0.1 in this paper.
M ( x i ) = 1 n j = 1 n x ( i , j )
(3)
Hunting stage
When T/2 < t, the population has entered a later stage of iteration, and the crocodile population enters the hunting stage. In this mode, when T/2 < t ≤ 3T/4, crocodiles perform hunting coordination. When 3T/4 < tT, crocodiles perform hunting cooperation. The relevant equation is shown in Equation (13).
x ( i , j ) ( t + 1 ) = { B e s t j ( t ) · P ( i , j ) ( t ) · r a n d , T 2 <   t 3 T 4 B e s t j ( t ) η ( i , j ) ( t ) · ϵ R i , j ( t ) · r a n d , 3 T 4 <   t T

2.4. Improvement Strategy of TtRSA

As mentioned above, the initial positions of RSA crocodile individuals are randomly generated within the search space, and this randomness makes it difficult for the population to obtain a more uniform distribution of initial positions. An uneven distribution of the population may increase the severity of an individual’s blind spot and reduce the population’s diversity. In addition, team cooperation, the search range, and the hunting mechanism of the crocodile population are all updated in terms of the current optimal value, and the individual’s iterative update process lacks mutation mechanisms. If the current optimal individual falls into a local optimum, the population may quickly converge within a short period, resulting in the algorithm being unable to break free from the constraints of the local extreme value. To address the shortcomings of the RSA, this paper considers introducing the Tent chaotic mapping and t-distribution mutation strategy to improve the RSA.

2.4.1. Tent Chaotic Mapping

In response to the problem of uneven population distribution caused by the random initialization of the RSA algorithm, this article introduces Tent chaotic mapping to solve this problem. The Tent chaotic map is a method of implementing chaos control using the tent function as the control function. By introducing the Tent chaotic map to generate pseudo-random numbers to initialize the RSA crocodile population, the traversal of the pseudo-random numbers enables the population to be more evenly distributed throughout the entire search space. This is beneficial for reducing the ‘blind areas’ of crocodile individuals, thus allowing individuals to quickly find better solutions, which improves the convergence speed of the algorithm. Chaotic maps have characteristics such as randomness, traversal, and order, and they can be used to increase the diversity of the population, accelerate the convergence speed of the algorithm in the early stages, and different between chaotic map operators that have different optimization effects. Among them, the Tent chaotic map can produce a uniform chaotic sequence through mapping within the range of (0, 1), and thus, applying the Tent chaotic map to population initialization can increase the diversity of the algorithm population and improve its global optimization ability. The relevant equation is shown in Equation (14).
h n + 1 = f ( h n ) = { h n / α , h n [ 0 , α ) ( 1 + h n ) / ( 1 α ) , h n ( α , 1 ]
In the equation, α ∈ (0,1) is the chaos parameter, h n is a random number within the range of 0 to 1, and n is the chaos variable index. The equation for generating the RSA crocodile population using the Tent chaotic map function is shown in Equation (15).
X i , j = h n · ( U B L B ) + L B
Figure 1a shows the population distribution based on random initialization, and Figure 1b shows the population distribution based on Tent chaotic mapping. It can be observed that in the two-dimensional space, although the population generated by Tent chaotic mapping does not have the same level of randomness as the population generated by the rand function, the individual position distribution is more uniform and there are no overlapping points or small search blind spots; this can improve population diversity and enable the population to quickly find optimal solutions. The frequency histogram of the population distribution is shown in Figure 2.

2.4.2. t-Distribution Mutation Strategy

The t-distribution is a probability distribution commonly used for parameter estimations and hypothesis testing in situations with small sample sizes. It was proposed by British statistician William Gosset in 1908. The shape of the t-distribution is determined by the degrees of the freedom parameter, where t(n = 1) → N (0,1) and t(n→∞) → C (0,1), where N (0,1) is the normal distribution and C (0,1) is the Cauchy distribution, which are two boundary cases of the t-distribution.
With the development of intelligent optimization algorithms, introducing Gaussian and Cauchy mutations has been proven to effectively improve the algorithm’s ability to search the population and escape local optima. In the early stages of algorithm iteration, the degree of the freedom parameter n is set to a small value, and the t-distribution tends towards the Cauchy distribution, which can effectively increase the diversity of the population and improve the algorithm’s global search ability. As the algorithm iterates during later stages, the degree of the freedom parameter n gradually increases, and the t-distribution tends towards the Gaussian distribution, which narrows the population search range and can effectively improve the algorithm’s ability to explore the local space. In the RSA, the expression for the effect of the t-distribution mutation is shown in Equation (16).
X n e w s j = X b e s t j + T D ( n ) · X b e s t j
The equation can be expressed as follows: X n e w s j is the position of the best solution in the j dimension after adaptive t-distribution mutation perturbation, X b e s t j is the position of the best solution in the j dimension before mutation perturbation, and TD(n) represents the t-distribution with a degree of freedom of n.

2.4.3. Summary of TtRSA

In summary of the above, we propose to improve RSA by using the Tent chaos mapping and t-distribution mutation strategy to address the problems of the initialized population of the RSA algorithm; for instance, the population is not uniformly distributed and easily falls into a local optimum during iteration. Based on the above proposed improvements, the TtRSA algorithm was ultimately suggested. Section 3 validated the feasibility of the improvement strategy of the TtRSA algorithm based on 11 benchmark test functions.

3. Improved Intelligent Optimization Algorithm Experiments

3.1. Experimental Design and Test Functions

Experimental Setup: The experiments were conducted on a computer system with the Windows 11 operating system, AMD R7 5800H 3.2GHz processor, and 16 GB of RAM. MATLAB R2022a was used for conducting the experiments. To evaluate the optimization performance of the improved TtRSA algorithm, 11 benchmark test functions were selected, including both unimodal and multimodal functions, that could evaluate the optimization performance of the algorithm for different types of problems. Among them, functions f1–f5 are continuous unimodal functions that are often used to test the optimization accuracy of search algorithms. Functions f9–f13 are multimodal test functions that can evaluate the convergence speed and accuracy of the algorithm and function f15 is a typical fixed-dimension multimodal function, commonly used to test the algorithm’s ability to escape local optima. The relevant information concerning the benchmark test functions is shown in Table 1.

3.2. Improvement Analysis of Optimization Algorithm Performance

In this section, the particle swarm optimization algorithm (PSO) [29], whale optimization algorithm (WOA) [30], chimpanzee optimization algorithm (CHOA) [31], and RSA [32] were used as benchmark algorithms to compare with the optimization performance of the improved TtRSA. The experimental settings included a population size of N = 30, a spatial dimension of D = 30, and a maximum number of iterations of T = 1000. Each algorithm was independently run 30 times with the test functions, and the average result of the 30 runs was taken as the final result.
Based on the comparison results in Table 2, it can be observed that under the same constraints, for the single-peaked test functions f1–f5, the optimization results of the TtRSA were several orders of magnitude (or even several tens of orders of magnitude) higher than those of other improved algorithms. Moreover, f1–f4 were able to converge to the theoretical optimal value of 0. For the complex multi-peaked test functions f9–f13, the optimization results of TtRSA were also better than those of other improved algorithms, and f9 and f11 were able to find the optimal value of 0. For the fixed-dimension multi-peaked test function f15, TtRSA was able to generally converge to the vicinity of the theoretical optimal value. The overall optimization performance of TtRSA was excellent for all 11 benchmark test functions, whether single-peaked test functions or complex multi-peaked test functions. This demonstrates the outstanding stability and robustness of TtRSA, and it proves that the TtRSA algorithm, which integrates multiple strategies, has strong global exploration and local development capabilities.

3.3. Convergence Performance Analysis of the Improved Optimization Algorithm

To visually and intuitively compare the convergence of algorithms in the function optimization process, the convergence curves were analyzed, as shown in Figure 3, where the vertical axis represents the fitness value of the corresponding function, and the horizontal axis represents the number of iterations of the optimization algorithm. Figure 3a–e show the running results of five optimization algorithms on a unimodal function. It is evident that the convergence curve of the TtRSA algorithm decreases faster than the other 4 algorithms, whereas the convergence curves of the remaining four algorithms all exhibit varying degrees of stagnation, indicating a lower optimization accuracy. This suggests that applying the improved Tent chaotic mapping strategy to initialize the population increases the diversity of the crocodile population; this makes the initial solution distribution more uniform, and it indicates that the algorithm can find the optimal solution quickly and more easily. Figure 3f–h show the convergence curves of the five optimization algorithms on a multimodal function. It is evident that in the first stage, the TtRSA algorithm’s convergence speed is significantly faster, further demonstrating the effectiveness of the improved Tent chaotic mapping and t-distribution mutation strategy; this changes the crocodile population’s search step and greatly improves the optimization accuracy and speed of the RSA [33].
Regarding the above experiments, it is evident that the proposed method, based on the Tent chaotic mapping and t-distribution mutation strategy, can effectively solve the effects of uneven population distribution and the difficulties in jumping out of local optima during the initialization of the RSA algorithm. The TtRSA algorithm has significant advantages over the RSA and the other four algorithms in terms of optimization precision, convergence speed, and ability to escape from local optima. Section 4.5 verifies the feasibility of optimizing the hyperparameters of the ASL-CatBoost algorithm based on the TtRSA algorithm to achieve ice fault detection in wind turbines.

4. Ice Fault Detection Experiment for Wind Turbines

Prognostics and health management are crucial for the lifecycle monitoring of equipment, especially complex equipment such as wind turbines that operate in harsh environments. Improving the speed and accuracy of fan fault detection can reduce maintenance costs and optimize work efficiency. This section aims to verify the effectiveness of the proposed ASL-CatBoost fault detection algorithm and TtRSA with regard to the application of wind turbine fault detection.

4.1. Wind Turbine Icing Fault Dataset

The fault detection experiment dataset used in this paper uses the SCADA system data information of two three-bladed wind turbines, F1 and F2, provided by Goldwind, under real operating conditions. There are three state modes in the dataset: icing fault, normal state, and invalid state (wherein it is difficult to determine the type of state). The dataset contains 27 feature dimensions in total, the time span of the F1 wind turbine is two months, and the time span of the F2 wind turbine is one month. The dataset provided the normal operation data and data concerning the specific time period wherein the wind turbine failed due to blade icing. Some dataset examples are shown in Table 3.
The time periods of normal operation and icing faults in the dataset are provided in separate Excel files, as shown in Table 4. Therefore, it is necessary to annotate the dataset based on specific state time periods. The Python ‘append()’ method can be used for annotation, where normal operation data are labeled as 0, and fault operation data are labeled as 1. Data analysis revealed the presence of a few unannotated invalid data points in the dataset, which have unknown operating states. These unannotated data points can negatively impact the accuracy of model training and increase computational overheads. Therefore, they should be removed during the data preprocessing stage.

4.2. Evaluating Indicator

Fault detection is a typical binary classification problem. The Confusion matrix is often used to measure the accuracy of the classifier, as shown in Table 5 below. The icing data are 1, and the normal data are 0. TP is the true example, representing both the diagnostic category and the actual category as icing data. FN is a false negative case, representing the diagnostic category, Normal data, and the actual category is icing data. FP is a false positive example, indicating that the diagnostic category is icing data when the actual category is Normal data. TN is a true negative example, indicating that both the diagnostic category and the actual category are normal data.
In accordance with the Confusion matrix, three evaluation indicators are extended: Precision, Recall, and F1 score. As shown in Equation (17), Precision refers to the proportion of the number of correctly predicted fault samples identified by the algorithm to the total number of predicted fault samples. As shown in Equation (18), the recall rate (Recall) refers to the proportion of the number of correctly predicted fault samples identified by the algorithm to the total number of true fault samples. As shown in Equation (19), the F1 score is the harmonic mean of accuracy and recall. The higher the values of the above three indicators, the better the algorithm performance.
P r e c i s i o n = T P / ( T P + F P )
R e c a l l = T P / ( T P + F N )
F 1 s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
To more intuitively demonstrate the advantages and disadvantages of the algorithm in terms of classification problems, this article also introduces the Receiver Operating Characteristic Curve (ROC) to evaluate the performance of the classification model, as shown in Figure 4. ROC is a curve that visually describes the true positive rate and false positive rate of a classification model based on different thresholds. The horizontal axis of the ROC curve is FPR, which represents the false positive rate, and the vertical axis is TPR, indicating sensitivity. The data points of the ROC curve are calculated using the TPR and FPR values obtained from the classification model at different thresholds. On the ROC curve, it is generally hoped that the curve will be closer to the upper left corner because at this point, the true probability (TPR) is high, whereas the false probability (FPR) is low, indicating that the classification model performs better. Usually, the better the performance of a classifier, the larger the Area Under Curve (AUC) below the ROC. The range of AUC values is 0.5 to 1, where 0.5 represents a completely random classification effect and 1 represents a perfect classifier. Therefore, both ROC and AUC can be used to evaluate the performance of classification models; when ROC is closer to the upper left corner and AUC values are closer to 1, it indicates that the model’s performance is better.

4.3. ASL-CatBoost Experiment

This chapter’s experiment aims to demonstrate the effectiveness of the proposed ASL-CatBoost algorithm. During the training process of the wind turbine fault detection algorithm, based on the ASL-CatBoost algorithm, the F1 wind turbine dataset was used as the training set and validation set. In order to prevent the model from overfitting, a 10-fold cross-validation method was used to improve the generalization ability of the model during the training process, and the optimal model was retained after the training was completed. To test the performance of the fault detection algorithm, the F2 wind turbine data was used as the test dataset, which included 10,638 fault data and 168,930 normal data.
This section’s aim is to verify the effectiveness of the improved ASL-CatBoost algorithm. Using the F1 wind turbine icing dataset as the training set, to prevent overfitting of the model, a 10-fold cross validation method was used during the training process to enhance the model’s generalization ability. To test the performance of the fault detection algorithm, the F2 wind turbine data in the second section were used as the test dataset, which included 10,638 fault data and 168,930 normal data. The ASL-CatBoost algorithm model, as well as classic machine learning algorithm models, such as the GBDT and Deep Learning model (LSTMAE), were used for comparative experiments. The training and testing datasets used for each algorithm were the same, and default parameters were used for hyperparameters. The fault detection performance of different algorithms is shown in Table 6. To verify the effectiveness of the two improvement methods of the loss function in this paper, ablation experiments were conducted. The CatBoost1 algorithm used the asymmetric focusing strategy of Equation (3) to complete the decoupling of the loss weight γ parameters in the focal loss function, and CatBoost2 only refers to Equation (4) for the hard thresholding of negative samples with high confidence. The experimental results show that the two improved strategies exhibit certain improvements, with regard to various evaluation indicators, compared with the initial CatBoost algorithm. Overall, the improved ASL-CatBoost algorithm in this article improved the recall rate by approximately 1% and it improved the accuracy and F1 score by approximately 2%, as compared with the original algorithm. Compared with the LSTMAE model, it improved the accuracy by 9% and the recall rate by 6%. The experimental results validate the feasibility of the improved algorithm.
To more intuitively compare the advantages and disadvantages of each algorithm, Figure 5a–e shows the ROC and AUC values of each algorithm in the wind turbine icing fault detection data validation set. From the figure, it is evident that the improved ASL-CatBoost algorithm exhibits a better fault detection effect than the traditional integrated learning algorithm.

4.4. TtRSA Algorithm Optimization ASL-CatBoost Algorithm Introduction

4.4.1. TtRSA Optimized ASL-CatBoost Algorithm Process

The ASL-CatBoost algorithm is greatly affected by hyperparameters, and the artificially set hyperparameter may not achieve the optimal effect during the algorithm training process. Therefore, this section proposes to use the improved TtRSA optimization algorithm to optimize the ASL-CatBoost algorithm with hyperparameters. The specific steps are as follows, and the process is shown in Figure 6.
Step 1: Data preprocessing. There are issues with missing samples and the incomplete labeling of sample labels in the icing fault data of wind turbines. It is necessary to preprocess the dataset to ensure that it meets the training requirements.
Step 2: Dataset partitioning. Divide the preprocessed dataset and determine the training, testing, and validation sets for the ASL-CatBoost algorithm.
Step 3: Set model parameters. Set the crocodile population size N and the maximum number of iterations T. In accordance with Equation (14), the Tent chaotic map is used to randomly initialize the individual positions of the crocodile population, i = 1, 2,..., N. Let the parameter t of the current number of iterations = 1. Set the value range of the maximum number of iterations of the CatBoost algorithm decision tree to (100, 2000). Set the value range of the learning rate to (0, 0.2). Set the value range of l2_leaf_reg to (0, 10) and the depth range of the tree to (0, 16).
Step 4: Calculate the fitness values of all crocodile individuals and save the current optimal crocodile individual position XBest.
Step 5: Determine whether t ≤ T/2 is true, and if it is true, use Equation (7) to implement the encirclement mechanism. When t ≤ T/4, implement the high-level walking strategy to update the individual crocodile position; when T/4 < t ≤ T/2, implement the abdominal crawling strategy to update the individual crocodile position. If t > T/2, use Equation (13) to implement the hunting mechanism. When T/2 < t ≤ 3T/4, execute the hunting coordination strategy to update the individual position of the crocodile; when 3T/4 < t ≤ T, execute the hunting cooperation mechanism. The policy updates the position of the crocodile individual.
Step 6: Use Equation (16) to perturb the t-distribution mutation strategy on some crocodile individuals and compare the fitness value of the crocodile individual after the updated position with the original individual. Reorder the crocodile individuals in accordance with fitness value and retain the current optimal fitness Degree value XBest.
Step 7: Set t = t + 1 to judge whether the current termination condition is satisfied, that is, whether the maximum number of iterations of the algorithm Itermax has been reached. If the maximum number of iterations of the algorithm has been reached, the currently saved optimal individual fitness value of the crocodile and the best parameter XBest are outputted, and the algorithm ends, otherwise, go to step 4.
Step 8: After the cycle ends, the obtained global optimal result, that is, the optimal hyperparameters of the ASL-CatBoost, may be substituted into the algorithm for model training, and the optimization effect may be tested.

4.4.2. Experiment for Optimizing ASL-CatBoost with TtRSA

In this section, the ASL-CatBoost is trained using the hyperparameter that was optimized by TtRSA. The experimental dataset and experimental equipment are the same as those in Section 4.3. Table 7 introduces the ASL-CatBoost algorithm obtained through optimization and the hyperparameter related to the comparison algorithm.
The experimental results of the optimized fault detection algorithm, based on the TtRSA algorithm, are shown in Table 8. From the table, it is evident that the ASL-CatBoost algorithm, optimized using TtRSA, has a better detection accuracy and recall rate throughout the whole dataset than the ASL-CatBoost algorithm with manually set hyperparameters. Regarding Table 6, it is evident that the accuracy and recall of machine learning algorithms (LightGBM, SVM, etc.) and the Deep Learning algorithm (LSTMAE) are improved after TtRSA optimization. The results demonstrate the effectiveness of the hyperparameter search for fault detection algorithms based on the TtRSA algorithm. Moreover, the training time complexity analysis of each algorithm is given in Table 4, which shows that the improved TtRSA-ASL-CatBoost algorithm has significantly less training time than the other algorithms. This is due to the fact that the algorithm obtains a higher accuracy faster at a lower number of iterations after hyperparameter optimization. For fault data detection, each algorithm can detect in time, therefore, the detection speed is negligible.

4.5. Enhanced Model Robustness

The robustness of the fault detection algorithm in different scenarios can be improved by adjusting the hyperparameters. To improve the robustness of the model using a different number of features, the model can be trained with 8 features, 16 features, and 22 features extracted from the wind turbine icing fault dataset used in this paper. The method can be used to predict the optimal model parameters in a specific scenario and to improve the robustness and generalization of the model under different scenarios. The optimal hyperparameters and accuracy rates for the three feature count cases are shown in Table 9.

5. Conclusions

Icing faults of wind turbines can easily lead to serious economic losses. This paper proposes using the improved ASL-CatBoost algorithm to solve the problem of unbalanced positive and negative samples in the wind turbine fault dataset, and to solve the problem concerning the fault detection algorithm, which is sensitive to the setting of hyperparameters; hence an improved crawler search algorithm is proposed to optimize hyperparameters. The following conclusions can be obtained:
(1)
Replacing the Cross-entropy Loss function of CatBoost algorithm with the asymmetric Loss function can improve the detection accuracy of the algorithm regarding fault data.
(2)
The use of the Tent chaotic mapping and t-distribution mutation strategy can improve the problem of imbalanced population distribution during RSA initialization and the tendency to fall into local optima during the iteration process.
(3)
Optimizing the hyperparameters of the ASL-CatBoost algorithm, based on the TtRSA algorithm, can effectively improve the detection speed and accuracy of the ASL-CatBoost algorithm.
However, this article also has the following limitations; it only details a binary classification problem and it fails to accurately determine which fault is in a multi classification state. In the future, further optimizations should be made to the algorithm to improve the accuracy and efficiency of fault detection. Moreover, future optimizations should enable the algorithm to clearly indicate which category the fault belongs to under multiple fault states.

Author Contributions

L.K.: Conceptualization, Visualization, Methodology, Software, Investigation, Validation, Formal analysis, writing original draft, Writing—review and editing. H.L.: Conceptualization, Methodology, Supervision, Project administration, Funding acquisition. G.L.: Conceptualization, Methodology, Supervision, Project administration, Funding acquisition. S.L.: Investigation, Visualization, Methodology, Software. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China (Grant Nos. 61973180 and 6217072142), Shandong Provincial Production-Education Integration Postgraduate Joint Training Demonstration Base Project (Grant Nos. 2020-19).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are not publicly available as they comprise sensitive information.

Acknowledgments

We appreciate the previous research, as referenced. Many thanks to the reviewers for their valuable suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Guo, X.; Chen, X.; Chen, X.; Sherman, P.; Wen, J.; McElroy, M. Grid integration feasibility and investment planning of offshore wind power under carbon-neutral transition in China. J. Nat. Commun. 2023, 14, 2447. [Google Scholar] [CrossRef] [PubMed]
  2. Beauson, J.; Laurent, A.; Rudolph, D.P.; Jensen, J.P. The complex end-of-life of wind turbine blades: A review of the European context. J. Renew. Sustain. Energy Rev. 2022, 155, 111847. [Google Scholar] [CrossRef]
  3. Kale, A.P.; Wahul, R.M.; Patange, A.D.; Soman, R.; Ostachowicz, W. Development of Deep Belief Network for Tool Faults Recognition. Sensors 2023, 23, 1872. [Google Scholar] [CrossRef] [PubMed]
  4. Patange, A.D.; Jegadeeshwaran, R.; Bajaj, N.S.; Khairnar, A.N.; Gavade, N.A. Application of machine learning for tool condition monitoring in turning. Sound Vib. 2022, 56, 127–145. [Google Scholar] [CrossRef]
  5. Bajaj, N.S.; Patange, A.D.; Jegadeeshwaran, R.; Pardeshi, S.S.; Kulkarni, K.A.; Ghatpande, R.S. Application of metaheuristic optimization based support vector machine for milling cutter health monitoring. Intell. Syst. Appl. 2023, 18, 200196. [Google Scholar] [CrossRef]
  6. Santolamazza, A.; Dadi, D.; Introna, V. A data-mining approach for wind turbine fault detection based on SCADA data analysis using artificial neural networks. Energies 2021, 14, 1845. [Google Scholar] [CrossRef]
  7. Wang, A.; Pei, Y.; Qian, Z.; Zareipour, H.; Jing, B.; An, J. A two-stage anomaly decomposition scheme based on multi-variable correlation extraction for wind turbine fault detection and identification. J. Appl. Energy 2022, 321, 119373. [Google Scholar] [CrossRef]
  8. Jawad, R.S.; Abid, H. Fault Detection in HVDC System with Gray Wolf Optimization Algorithm Based on Artificial Neural Network. Energies 2022, 15, 7775. [Google Scholar] [CrossRef]
  9. Samanta, B.; Al-Balushi, K.R.; Al-Araimi, S.A. Artificial neural networks and support vector machines with genetic algorithm for bearing fault detection. J. Eng. Appl. Artif. Intell. 2003, 16, 657–665. [Google Scholar] [CrossRef]
  10. Zhang, S.; Lang, Z.Q. SCADA-data-based wind turbine fault detection: A dynamic model sensor method. J. Control Eng. Pract. 2020, 102, 104546. [Google Scholar] [CrossRef]
  11. Aziz, U.; Charbonnier, S.; Bérenguer, C.; Lebranchu, A.; Prevost, F. Critical comparison of power-based wind turbine fault-detection methods using a realistic framework for SCADA data simulation. J. Renew. Sustain. Energy Rev. 2021, 144, 110961. [Google Scholar] [CrossRef]
  12. Song, W.; Lin, J.W.; Zhou, F.Z. Wind turbine bearing fault diagnosis method based on an improved denoising AutoEncoder. J. Power Syst. Prot. Control 2022, 50, 61–68. [Google Scholar]
  13. Liu, J.R.; Yang, G.T.; Wang, X.W. A Wind Turbine Fault Diagnosis Method Based on Siamese Deep Neural Network. J. Syst. Simul. 2022, 34, 2348–2358. [Google Scholar]
  14. Liu, J.; Wang, X.; Wu, S.; Xie, F. Wind turbine fault detection based on deep residual networks. J. Expert Syst. Appl. 2023, 213, 119102. [Google Scholar] [CrossRef]
  15. Tang, M.; Cao, C.; Wu, H.; Zhu, H.; Tang, J.; Peng, Z.; Wang, Y. Fault Detection of Wind Turbine Gearboxes Based on IBOA-ERF. Sensors 2022, 22, 6826. [Google Scholar] [CrossRef]
  16. Kanović, Ž.; Rapaić, M.R.; Jeličić, Z.D. Generalized particle swarm optimization algorithm-Theoretical and empirical analysis with application in fault detection. J. Appl. Math. Comput. 2011, 217, 10175–10186. [Google Scholar] [CrossRef]
  17. Lee, D.H.; Ahn, J.H.; Koh, B.H. Fault detection of bearing systems through EEMD and optimization algorithm. Sensors 2017, 17, 2477. [Google Scholar] [CrossRef] [Green Version]
  18. Lei, M.; Lv, Y.; Wei, W.; Ren, Q. Fault Warning of Power Plant Fans based on Long Short-term Memory Neural Network and Bayesian Optimization. J. Eng. Therm. Energy Power 2022, 37, 213–220. [Google Scholar]
  19. Zhang, Y.F.; Pi, Z.Y.; Zhu, R.Q.; Song, J.X.; Shi, J.J. Wind Power Prediction Based on WOA-BiLSTM Neural NetworkElectric. J. Eng. 2022, 28–31. [Google Scholar]
  20. Huang, L.; Wang, Y.; Guo, Y.; Hu, G. An improved reptile search algorithm based on lévy flight and interactive crossover strategy to engineering application. Mathematics 2022, 10, 2329. [Google Scholar] [CrossRef]
  21. Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased boosting with categorical features. J. Adv. Neural Inf. Process. Syst. 2018, 31, 6639–6649. [Google Scholar]
  22. Zhang, W.; Yu, J.; Zhao, A.; Zhou, X. Predictive model of cooling load for ice storage air-conditioning system by using GBDT. J. Energy Rep. 2021, 7, 1588–1597. [Google Scholar] [CrossRef]
  23. Osman, A.; Ahmed, A.; Chow, M.F. Extreme gradient boosting (Xgboost) model to predict the groundwater levels in Selangor Malaysia. J. Ain Shams Eng. J. 2021, 12, 1545–1556. [Google Scholar] [CrossRef]
  24. Ke, G.; Meng, Q.; Finley, T. Lightgbm: A highly efficient gradient boosting decision tree. J. Adv. Neural Inf. Process. Syst. 2017, 30, 3149–3157. [Google Scholar]
  25. Patange, A.D.; Pardeshi, S.S.; Jegadeeshwaran, R.; Zarkar, A.; Verma, K. Augmentation of Decision Tree Model through Hyper-Parameters Tuning for Monitoring of Cutting Tool Faults Based on Vibration Signatures. J. Vib. Eng. Technol. 2022, 1–19. [Google Scholar] [CrossRef]
  26. Hancock, J.T.; Khoshgoftaar, T.M. CatBoost for big data: An interdisciplinary review. J. Big Data 2020, 7, 1–45. [Google Scholar] [CrossRef]
  27. Lin, T.Y.; Goyal, P.; Girshick, R. Focal loss for dense object detection. Proc. IEEE Int. Conf. Comput. Vis. 2017, 2980–2988. [Google Scholar]
  28. Abualigah, L.; Abd, E.M.; Sumari, P. Reptile Search Algorithm (RSA): A nature-inspired meta-heuristic optimizer. J. Expert Syst. Appl. 2022, 191, 116158. [Google Scholar] [CrossRef]
  29. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95—International Conference on Neural Networks, Perth, Australia, 27 November–1 December 1995; Volume 4, pp. 1942–1948. [Google Scholar]
  30. Mirjalili, S.; Lewis, A. The whale optimization algorithm. J. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  31. Khishe, M.; Mosavi, M.R. Chimp optimization algorithm. J. Expert Syst. Appl. 2020, 149, 113338. [Google Scholar] [CrossRef]
  32. Ekinci, S.; Izci, D.; Abu, Z.R.; Alsoud, A.R.; Abualigah, L. Development of Lévy flight-based reptile search algorithm with local search ability for power systems engineering design problems. J. Neural Comput. Appl. 2022, 34, 20263–20283. [Google Scholar] [CrossRef]
  33. Almotairi, K.H.; Abualigah, L. Improved reptile search algorithm with novel mean transition mechanism for constrained industrial engineering problems. J. Neural Comput. Appl. 2022, 34, 17257–17277. [Google Scholar] [CrossRef]
Figure 1. Population initialization distribution diagram. Subgraph (a) uses random method for population initialization; subgraph (b) uses Tent method for population initialization.
Figure 1. Population initialization distribution diagram. Subgraph (a) uses random method for population initialization; subgraph (b) uses Tent method for population initialization.
Sensors 23 06741 g001
Figure 2. Initialized population frequency histogram.
Figure 2. Initialized population frequency histogram.
Sensors 23 06741 g002
Figure 3. Initial function image and algorithm convergence curve. (a) Sphere function convergence curve; (b) Schwefel’ 2.21 function convergence curve; (c) Schwefel’ 1.2 function convergence curve; (d) Schwefel’ 2.21 function convergence curve; (e) Rosenbrock function convergence curve; (f) Rastrigin function convergence curve; (g) Ackley function convergence curve; (h) Criewank function convergence curve.
Figure 3. Initial function image and algorithm convergence curve. (a) Sphere function convergence curve; (b) Schwefel’ 2.21 function convergence curve; (c) Schwefel’ 1.2 function convergence curve; (d) Schwefel’ 2.21 function convergence curve; (e) Rosenbrock function convergence curve; (f) Rastrigin function convergence curve; (g) Ackley function convergence curve; (h) Criewank function convergence curve.
Sensors 23 06741 g003
Figure 4. Receiver operating characteristic curve. Curve a and curve b represent the ROC curves under different accuracy rates, and the situation represented by curve a in the figure is better than curve b.
Figure 4. Receiver operating characteristic curve. Curve a and curve b represent the ROC curves under different accuracy rates, and the situation represented by curve a in the figure is better than curve b.
Sensors 23 06741 g004
Figure 5. ROC of each algorithm during wind turbine icing fault detection. (a) GBDT algorithm Receiver operating characteristics; (b) XGBoost algorithm Receiver operating characteristics; (c) LightGBM algorithm Receiver operating characteristics; (d) CatBoost algorithm Receiver operating characteristics; (e) ASL-CatBoost algorithm Receiver operating characteristics.
Figure 5. ROC of each algorithm during wind turbine icing fault detection. (a) GBDT algorithm Receiver operating characteristics; (b) XGBoost algorithm Receiver operating characteristics; (c) LightGBM algorithm Receiver operating characteristics; (d) CatBoost algorithm Receiver operating characteristics; (e) ASL-CatBoost algorithm Receiver operating characteristics.
Sensors 23 06741 g005
Figure 6. Flowchart of the optimization of the ASL-CatBoost algorithm with TtRSA.
Figure 6. Flowchart of the optimization of the ASL-CatBoost algorithm with TtRSA.
Sensors 23 06741 g006
Table 1. Benchmark functions.
Table 1. Benchmark functions.
NumberTest FunctionRangeOptimaType
f1Sphere[−100, 100]0unimodal
f2Schwefel’ 2.22[−10, 10]0unimodal
f3Schwefel’ 1.2[−100, 100]0unimodal
f4Schwefel’ 2.21[−100, 100]0unimodal
f5Rosenbrock[−30, 30]0unimodal
f9Rastrigin[−5.12, 5.12]0multimodal
f10Ackley[−32, 32]0multimodal
f11Criewank[−600, 600]0multimodal
f12Penalized 1[−50, 50]0multimodal
f13Penalized 2[−50, 50]0multimodal
f15Kowalik[−5, 5]0.0003multimodal
Table 2. Optimization performance analysis of improved optimization algorithms.
Table 2. Optimization performance analysis of improved optimization algorithms.
FunctionPSOWOACHOARSATtRSA
MeanSDMeanSDMeanSDMeanSDMeanSD
f11.40 × 10−302.11 × 10−322.83 × 10−1641.94 × 10−1631.05 × 10−301.30 × 10−200.000.000.000.00
f24.21 × 10−205.54 × 10−231.52 × 10−1108.28 × 10−551.24 × 10−258.55 × 10−216.88 × 10−602.58 × 10−750.000.00
f37.01 × 10−182.21 × 10−216.14 × 10−21.05 × 10−56.05 × 10−191.31 × 10−175.81 × 10−424.55 × 10−730.000.00
f41.08 × 10−403.17 × 10−318.81 × 10−1721.05 × 10−1052.75 × 10−702.85 × 10−274.88 × 10−1758.88 × 10−1650.000.00
f59.67 × 106.01 × 104.39 × 1031.05 × 1053.13 × 1042.57 × 10−141.71 × 101.37 × 102.39 × 10−201.85 × 10−21
f94.67 × 1021.16 × 100.000.001.41 × 10−011.65 × 10−266.68 × 101.16 × 100.000.00
f102.76 × 10−165.09 × 10−219.42 × 10−371.05 × 10−51.96 × 101.79 × 10−78.86 × 10−160.007.62 × 10−785.80 × 10−56
f111.21 × 10−17.74 × 10−31.05 × 10−187.05 × 10−254.79 × 10−025.05 × 10−189.37 × 10−27.50 × 10−10.000.00
f126.92 × 10−71.19 × 10−26.55 × 10−55.06 × 10−73.98 × 10−015.06 × 10−171.24 × 10−63.31 × 10−18.04 × 10−117.50 × 10−11
f136.68 × 10−88.91 × 10−38.78 × 10−31.76 × 10−152.05 × 10−011.76 × 10−151.52 × 10−64.19 × 10−16.95 × 10−408.36 × 10−5
f155.82 × 10−52.21 × 10−42.68 × 10−74.35 × 10−197.36 × 10−024.35 × 10−192.74 × 10−131.15 × 10−32.34 × 10−207.35 × 10−18
Table 3. Example of a dataset section.
Table 3. Example of a dataset section.
TimeWind_SpeedGenerator_SpeedPowerWind_DirectionEnvironment_Tmp
2015/11/1 17:332.671345891.3166610632.571868051−0.7866036930.337770344
2015/11/1 17:343.0585823511.2933944292.537817968−0.9247122350.337770344
2015/11/1 17:343.2798603291.1870326712.551855132−0.9626920840.337770344
2015/11/1 17:343.2319167671.2701277942.54983978−0.8263098990.337770344
2015/11/1 17:343.3646835541.3299562832.557854321−0.8677424610.337770344
2015/11/1 17:343.0106387891.1870326712.54983978−1.1577703990.337770344
2015/11/1 17:343.3609955871.2867468192.565868862−1.2337300970.337770344
2015/12/1 18:591.5575800681.2235945251.6366976461.4611128231.314590648
Table 4. Time periods of normal data and fault data.
Table 4. Time periods of normal data and fault data.
Fault Operating Time PeriodNormal Operating Time Period
Start TimeEnd TimeStart TimeEnd Time
2015/11/4 22:152015/11/4 23:332015/11/1 17:332015/11/4 19:42
2015/11/9 3:212015/11/9 5:142015/11/5 11:062015/11/9 1:23
2015/11/9 21:262015/11/9 23:182015/11/9 12:202015/11/9 19:27
2015/11/13 2:592015/11/13 4:512015/11/10 12:432015/11/13 0:38
2015/11/16 15:312015/11/16 15:572015/11/13 9:102015/11/15 16:35
2015/11/23 20:402015/11/23 22:332015/11/17 12:142015/11/23 18:41
2015/11/24 5:422015/11/24 6:312015/11/24 1:242015/11/24 2:39
2015/11/24 14:582015/11/24 16:512015/11/24 10:492015/11/24 12:12
2015/11/25 20:552015/11/25 22:482015/11/25 18:002015/11/25 18:56
2015/11/26 1:472015/11/26 3:402015/11/26 10:102015/11/28 2:16
2015/11/28 4:152015/11/28 6:082015/11/28 11:522015/11/29 2:30
2015/11/29 4:292015/11/29 6:222015/11/29 11:482015/11/29 14:36
2015/11/29 17:442015/11/30 8:522015/11/30 10:112015/11/30 13:08
Table 5. Confusion matrix.
Table 5. Confusion matrix.
Icing DiagnosisNormal Diagnosis
Actual icingTPFN
Actual normalFPTN
Table 6. Performance comparison of each algorithm model.
Table 6. Performance comparison of each algorithm model.
ModelsPreciousRecallF1-ScoreTrain Times
GBDT0.9227030.9366140.92960612 m 16 s
XGBoost0.9070690.9252870.91608712 m 30 s
LightGBM0.9137420.9365780.92501913 m 15 s
CatBoost0.9261480.9304030.92827112 m 11 s
SVM0.8075920.7357280.77088611 m 20 s
LSTMAE0.8574280.8867410.87183815 m 30 s
CatBoost10.9343160.9416280.93795812 m 30 s
CatBoost20.9357430.9328490.93429412 m 20 s
ASL-CatBoost0.9494270.9432760.94634112 m 35 s
Table 7. Algorithm-related parameters after the optimization of the TtRSA.
Table 7. Algorithm-related parameters after the optimization of the TtRSA.
MethodHyper-Parameters
ASL-CatBoostIterations = 1000, depth = 6, learning_rate = 0.05,
l2_leaf_reg = 0.4
TtRSA-LSTMAEHidden_num = 8, windowsize = 100, stride = 1,
learning_rate = 0.001, epoch = 16
TtRSA-SVMc = 40.001, g = 0.008
TtRSA-ASL-CatBoostIterations = 300, depth = 8, learning_rate = 0.1, l2_leaf_reg = 0.6
TtRSA-XGBoostmax_depth = 5, min_child_weight = 1, subsample = 0.7,
colsample_bytree = 0.8, scale_pos_weight = 1
TtRSA- LightGBMn_estimators = 144, max_depth = 8, learning_rate = 0.1,
random_state = 42, subsample = 0.7, num_leaves = 524
Table 8. Comparison of the fault data detection effects of various algorithms.
Table 8. Comparison of the fault data detection effects of various algorithms.
ModelsPreciousRecallF1-ScoreTrain Times
ASL-CatBoost0.9494270.9432760.94634112 m 30 s
TtRSA-SVM0.8374240.8657920.85137211 m 52 s
TtRSA-LSTMAE0.8824970.9237690.90266118 m 20 s
TtRSA-LightGBM0.9355980.9374730.93653515 m 43 s
TtRSA-XGBoost0.9287360.9185670.92362421 m 17 s
TtRSA-ASL-CatBoost0.9501360.9490260.9495818 m 52 s
Table 9. Comparison of Hyperparameters and Effects of Fault Detection Algorithms under Different Number of Features.
Table 9. Comparison of Hyperparameters and Effects of Fault Detection Algorithms under Different Number of Features.
Number of FeaturesOptimal Case HyperparametersPreciousRecallF1-Score
8Iterations = 1000, depth = 4, learning_rate = 0.02, l2_leaf_reg = 0.60.9035810.6763720.758174
16Iterations = 800, depth = 6, learning_rate = 0.1, l2_leaf_reg = 0.70.9177950.8367420.975396
22Iterations = 500, depth = 7, learning_rate = 0.1, l2_leaf_reg = 0.60.9383310.8970310.917216
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kong, L.; Liang, H.; Liu, G.; Liu, S. Research on Wind Turbine Fault Detection Based on the Fusion of ASL-CatBoost and TtRSA. Sensors 2023, 23, 6741. https://doi.org/10.3390/s23156741

AMA Style

Kong L, Liang H, Liu G, Liu S. Research on Wind Turbine Fault Detection Based on the Fusion of ASL-CatBoost and TtRSA. Sensors. 2023; 23(15):6741. https://doi.org/10.3390/s23156741

Chicago/Turabian Style

Kong, Lingchao, Hongtao Liang, Guozhu Liu, and Shuo Liu. 2023. "Research on Wind Turbine Fault Detection Based on the Fusion of ASL-CatBoost and TtRSA" Sensors 23, no. 15: 6741. https://doi.org/10.3390/s23156741

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop