Next Article in Journal
Two-Level Excitation Current Driver to Reduce the Driving Power of an Electromagnetic Contactor
Next Article in Special Issue
Improving the Performance of Automatic Lip-Reading Using Image Conversion Techniques
Previous Article in Journal
Design of UWB Electrically Small Antenna Based on Distributed Passive Network Loading
Previous Article in Special Issue
Deep Learning of Sensor Data in Cybersecurity of Robotic Systems: Overview and Case Study Results
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Aero-Engine Classification Method Based on Fourier Transform Infrared Spectrometer Spectral Feature Vectors

1
Department of Electronic and Optical Engineering, Space Engineering University, Beijing 101416, China
2
Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100094, China
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(5), 915; https://doi.org/10.3390/electronics13050915
Submission received: 21 December 2023 / Revised: 23 February 2024 / Accepted: 24 February 2024 / Published: 28 February 2024

Abstract

:
Aiming at the classification identification problem of aero-engines, this paper adopts a telemetry Fourier transform infrared spectrometer to collect aero-engine hot jet infrared spectrum data and proposes an aero-engine classification identification method based on spectral feature vectors. First, aero-engine hot jet infrared spectrum data are acquired and measured; meanwhile, the spectral feature vectors based on CO2 are constructed. Subsequently, the feature vectors are combined with the seven mainstream classification algorithms to complete the training and prediction of the classification model. In the experiment, two Fourier transform infrared spectrometers, EM27 developed by Bruker and a self-developed telemetry FT-IR spectrometer, were used to telemeter the hot jet of three aero-engines to obtain infrared spectral data. The training data set and test data set were randomly divided in a ratio of 3:1. The model training of the training data set and the label prediction of the test data set were carried out by combining spectral feature vectors and classification algorithms. The classification evaluation indicators were accuracy, precision, recall, confusion matrix, and F1-score. The classification recognition accuracy of the algorithm was 98%. This paper has considerable significance for the fault diagnosis of aero-engines and classification recognition of aircrafts.

1. Introduction

Infrared spectroscopy technology [1,2,3] is a technique for detecting the molecular structure and chemical composition of substances. This technology utilizes the energy level transition of molecules in substances under infrared radiation to measure the wavelength and intensity of absorbed or emitted light, producing specific spectrogram, which are used to analyze and judge the chemical bonds and structures of substances. This technology has significant research applications in environmental monitoring [4], garbage classification [5], and life chemistry [6]. The Fourier transform infrared spectrometer (FT-IR spectrometer) [7,8,9] is an important means of measuring infrared spectra. It obtains interferogram through an interferometer and restores the interferogram to spectrogram based on Fourier transform. Passive FT-IR is frequently used for the detection of atmospheric pollutants. It has the ability to collect data from any direction, allowing for all-weather, continuous, long-distance, real-time monitoring and rapid analysis of targets.
The classification characteristics of aero-engines are frequently related to fuel type, combustion method, and emission characteristics. Different types of aero-engines produce different gas components and emissions during the combustion process. The vibration and rotation of these molecules form specific infrared absorption and emission spectra. By analyzing the infrared spectra of aero-engine hot jet, the characteristics of the gas components and emissions produced by aero-engine combustion can be obtained. After a considerable number of hot jet infrared spectra are analyzed, a spectral feature library will be established, then the types of aero-engines are going to be determined so that the identification of aero-engines will be realized.
In this paper, an algorithm for the classification and recognition of aero-engines is proposed, which combines the infrared spectrum feature vectors of aero-engine hot jet with the current mainstream classifiers. The classifiers include supervised learning method SVM, integrated learning methods XGBoost, CatBoost, AdaBoost, Random Forest, LightGBM, and a neural network method. Accuracy, precision, recall, F1 value, and confusion matrix are used as classification evaluation criteria. After many experiments, the accuracy of the classification of aero-engines has reached 98%.
The major contributions of this paper are as follows:
  • The infrared spectrum detection method for aero-engine hot jet is used as the basis and data source input of the identification of aero-engines. Aero-engine hot jet is an important infrared radiation characteristic of aero-engine, and the infrared spectrum provides the characteristic information of substances at the molecular level, so it is more scientific to utilize this method for classification.
  • FT-IR is used to measure the infrared spectrum information of aero-engine hot jet. An FT-IR spectrometer has the advantages of fast scanning speed, excellent resolution, wide measurement spectral range, and high measurement accuracy. It can achieve exceptional spectral measurement, which is of great significance for the non-contact classification and recognition of aero-engines.
The architecture of this paper is as follows: Section 1 describes the infrared spectroscopy technology, and this paper uses the classification method, innovation, and article architecture; in Section 2, the spectral components of the hot jet are analyzed, and the construction method of spectral feature vectors is proposed; Section 3 introduces seven mainstream classifier methods; Section 4 describes the experimental content, including the experimental design of aero-engine spectrum acquisition, data set production, spectral feature vectors extraction, and the accuracy evaluation of classification prediction results; Section 5 summarizes the paper and puts forward the idea of the next research direction.

2. Spectral Feature Analysis and Spectral Feature Vector Construction

In this section, the brilliant temperature spectrum (BTS) and components of aero-engine hot jet are analyzed, and a method of constructing spectral feature vectors based on CO2 characteristic peaks is proposed.
When passive FT-IR is used in gas detection and identification studies, the method of calculating the BTS with a constant baseline can be utilized due to the high emissivity of most of the substances in nature that serve as the background. The brilliant temperature [10,11] of a material object is equivalent to the temperature of a blackbody at the same wavelength when the intensity of the spectral radiation of the material object and the blackbody are equal. It is based on equal brightness and is utilized to characterize the material object’s own radiation. The use of BTS does not require pre-measurement of the background spectrum and enables the target gas characteristics to be extracted directly from the BTS analysis.
In order to obtain the BTS from the passive infrared spectrum, it is necessary to first deduct the spectral signal measured by the spectrometer from the bias and response of the instrument to obtain the spectral radiance spectrum entering the spectrometer, and from the radiance spectrum obtained, the equivalent temperature of the radiance spectrum T ( v ) can be calculated according to Planck’s law of radiation by transforming Planck’s formula to obtain the following formula:
T ( v ) = h c v k ln { [ L ( v ) + 2 h c 2 v 3 ] / L ( v ) }
where h is Planck’s constant with a value of 6.62607015 × 10^−34 J·S, c represents the speed of light with a value of 2.998 × 10^8 m/s, v is the wave number in cm−1, k is Boltzmann’s constant with a value of 1.380649 × 10^−23 J/K, and L ( v ) represents the radiance about the wave number.
The experimentally measured infrared spectra of aero-engine hot jets of three turbojet engines are presented in Figure 1, where the horizontal coordinates are the wave numbers and the vertical coordinates are the BTS.
The emission products of an aero-engine typically include oxygen (O2), nitrogen (N2), carbon dioxide (CO2), steam (H2O), carbon monoxide (CO), et al. The combustion products are primarily divided into ① air composition not participating in combustion, including O2 and N2, ② products of combustion reactions, mainly NOx, ③ the end product of an ideal combustion process, including CO2, H2O, and CO [12].
Based on the main emissions of the aero-engines and the measured infrared spectral data of the aero-engines, the peaks of the three most likely products of combustion, CO2, H2O, and CO, were compared and analyzed. It was discovered that in the spectral curve, the characteristic peaks of CO2 at 667 cm−1 and 2349 cm−1 were obvious and stable, the spectrum of CO at the characteristic band of 2000–2222 cm−1 gradually weakened with the increase of rotational speed, and the characteristic peaks of H2O were not obvious and not informative. Therefore, the two characteristic peaks (667 cm−1 and 2350 cm−1) of CO2 in the mid-wave infrared (MWIR) region (400–4000 cm−1) and the two stable and obvious characteristic peaks (719 cm−1 and 2390 cm−1) in the measured spectral curve were selected for the construction of spectral feature vectors. In this paper, the spectral feature vectors are constructed based on the BTS, and the numerical difference of the BTS is related to the exhaust temperature of the aero-engines as well as the concentration and temperature of the gas in the hot jet.
Four characteristic peaks in the MWIR region of the BTS of the measured aero-engine hot jet were selected for the construction of spectral feature vectors, and the corresponding wave numbers of the peaks are 2350 cm−1, 2390 cm−1, 719 cm−1, and 667 cm−1; their locations are shown in Figure 2 below:
Spectral feature vectors a = [ a 1 , a 2 ] are composed of individual spectra by calculating the peak difference between 2390 cm−1 and 2350 cm−1 and the peak difference between 719 cm−1 and 667 cm−1.
a 1 = T v = 2390 c m 1 T v = 2350 c m 1 a 2 = T v = 719 c m 1 T v = 677 c m 1
Affected by the environment, the peak positions of the selected characteristic peaks may be shifted, the four characteristic peaks of 2350 cm−1, 2390 cm−1, 719 cm−1, and 667 cm−1 of the experimentally measured infrared spectral data are extracted from the area range where the maximum and minimum peaks are located for the extraction of spectral feature vectors, and the specific selection of the threshold range is shown in Table 1:

3. Spectral Eigenvector Classification Methods

This section provides a brief description of the mainstream classifiers SVM [13,14,15], XGBoost 1.0 [16,17], CatBoost 1.0.4 [18,19,20], AdaBoost [21], Random Forest [22], LightGBM [23], and neural network [24,25] classifiers.
The current mainstream classification methods are supervised learning methods, unsupervised learning methods, semi-supervised learning methods, reinforcement learning methods, deep learning methods, and ensemble learning methods. Supervised learning methods are methods that use training data with labels to construct a model used to classify new samples, including decision tree, support vector machines (SVM), and logistic regression. Ensemble learning methods include Bagging and Boosting, where Bagging is characterized by the absence of strong dependencies between individual evaluators, and a series of individual learners can be generated in parallel, representing the algorithm Random Forest; Boosting is characterized by the presence of strong dependencies between individual learners, and a series of individual learners basically Boosting is characterized by a strong dependency relationship between individual learners, and a series of individual learners basically need to be generated serially, representing algorithms like AdaBoost, XGBoost, LightGBM, etc.
In view of the characteristics of the aero-engine infrared spectral data and the way of measurement in this paper, it is more appropriate to use the training data set, test data set, and label set for model training to carry out classification and prediction. In this paper, SVM, XGBoost, CatBoost, AdaBoost, Random Forest, LightGBM, and neural network algorithms are used in conjunction with spectral feature vectors to complete the classification task of the aero-engines.
Support vector machine (SVM) classification method
SVM is a kernel function-based classification algorithm machine learning binary classification model for both binary and multi-classification problems. The main task of the SVM model is to find the optimal over-planning for classifying the data points. For binary classification, the SVM algorithm is shown in Figure 3:
SVM classifies the data by hyperplane y , which can be expressed as
y = ω T x + b
Calculate the hyperplane equations, the thresholds on both sides, and the optimization function for the best vector in the hyperplane. Define the margin line as passing through the nearest point in each class to obtain the equation for the boundary line as:
ω T x + b = 0 ω T x + b = 1 ω T x + b = 1
where, ω T x + b = 0 is the hyperplane equation, ω T x + b = 1 is the edge line equation for the positive region, and ω T x + b = 1 is the edge line equation for the region with negative values.
Finding the distance between two edges can be accomplished by the following:
ω T ( x 2 x 1 ) = 2 x 2 x 1 = 2 ω T
Maximize the margin line function to find the optimal threshold. The final SVM model is obtained:
( ω * , b * ) máx 2 w T y i * ( ω T x i + b i ) > = 1
The SVM algorithm is suitable for high-dimensional spatial data processing and has good performance in the field of text classification and image recognition, and at the same time, it can maintain good performance when small amounts of sample data are processed. The SVM algorithm is also one of the most commonly used algorithms in current classification tasks.
XGBoost classification method
XGBoost is an efficient classification and regression algorithm based on gradient boosted decision trees. The XGBoost algorithm generates multiple decision trees in an iterative manner by integrating weak classifiers and training based on the residuals of the previous decision tree and completely integrates multiple decision trees to improve the performance of the model and complete the classification.
In the XGBoost classifier, firstly, the training data set { ( x i y i ) } i = 1 N , the differentiable loss function L ( y , F ( x ) ) , multiple weak learning M, and the learning rate α are defined as the input parameters of the XGBoost model.
Initialization operations are performed on the model manipulating constant values:
f ^ ( 0 ) ( x ) = arg m i n θ i = 1 N L ( y i , θ )
As the model starts from 1 iteration to M, the gradient is first calculated:
g ^ m ( x i ) = [ L ( y i , f ( x i ) ) f ( x i ) ] f ( x ) f ^ ( m 1 ) ( x )
Second, calculate the Hessians matrix:
h ^ m ( x i ) = [ 2 L ( y i , f ( x i ) ) f ( x i ) 2 ] f ( x ) f ^ ( m 1 ) ( x )
Fit the base learner to the training data set { x i , g ^ m ( x i ) h ^ m ( x i ) } i = 1 N evolutionary optimization of the formula is obtained:
φ ^ m = arg m i n φ Φ i = 1 N 1 2 h m ( x i ) [ φ ( x i ) g ^ m ( x i ) h m ^ ( x i ) ] 2
f ^ m ( x ) = α ϕ ^ m ( x )
Finally, the model is updated:
f ^ ( m ) ( x ) = f ^ ( m 1 ) ( x ) + f ^ m ( x )
After the iteration is completed, the output of the final model equation is:
f ^ ( x ) = f ^ ( M ) ( x ) = m = 0 M f ^ m ( x )
The XGBoost algorithm shows higher performance and efficiency in classification tasks on large-scale data sets. The following are some benefits of the XGBoost algorithm: firstly, it is extraordinarily efficient in handling large-scale data; it supports L1 and L2 regularization, which helps to prevent over-fitting; it is effective to automatically select the salient features, which reduces the work of feature engineering; it supports a variety of loss functions, which can handle multiple tasks, such as regression, classification, sorting, etc.; it is capable of utilizing multiple core processors in parallel, sorting, and many other tasks; it is able to perform parallel computation using multi-core processors.
AdaBoost classification method
The AdaBoost algorithm is an adaptive enhancement algorithm for ensemble learning; the method is used to add and train new weak decision makers serially and weigh the combination of decision makers so that the loss function continues to decrease until the addition of decision makers is ineffective and finally all the decision makers are integrated into one whole for decision making. The AdaBoost algorithm is illustrated in Figure 4, and the arrows in the figure indicate the data flow direction:
First, the AdaBoost algorithm defines the training data set as T = { ( x 1 , y 1 ) , ( x 2 , y 2 ) , , ( x n , y n ) } , where y { 1 , + 1 } , the learner is defined as G m ( x ) , the training session is set as M, and the initial weight distribution is set as w i ( 1 ) = 1 N , where i = 1,2 , 3 , , N .
During the training iterations, the base learner G m ( x ) is first obtained by learning using a training data set with a distribution of power values:
G m ( x ) = arg m i n G ( x ) i = 1 N w i ( m ) I ( y i G ( x i ) )
Based on G m ( x ) , calculate the error rate of the learner G m ( x ) on the training data set:
ϵ m = i = 1 N w i ( m ) I ( y i G m ( x i ) ) i = 1 N w i ( m )
Calculate the coefficient α m of G m ( x ) :
α m = 1 2 l n 1 ϵ m ϵ m
Update the sample weight distribution w i ( m + 1 ) :
w i ( m + 1 ) = w i ( m ) e y i α m ( ξ m ( x i ) Z ( m ) , i = 1,2 , 3 N
where Z ( m ) is the normalization factor, Z ( m ) = i = 1 N w ( m ) i e y i α m G m ( x i ) , which ensures that all w i ( m + 1 ) constitute a distribution.
The final output of the model G ( x ) :
G ( x ) = s i g n [ m = 1 M G m ( x ) ]
The AdaBoost algorithm can adapt to the respective training error rates of weak learners and is suitable for a variety of classification problems that do not frequently require tuning. The benefits of the AdaBoost algorithm are as follows: it is simple to implement and adjust and does not require too much parameter tuning; becoming overly fit is difficult, and by iteratively lowering the weight of the wrong samples, the risk of over-fitting can be reduced; it can be utilized in conjunction with a variety of basic classifiers of weak learners such as decision trees, neural networks, etc.; it is applicable with unbalanced data sets and deals with unbalanced data through the adjustment of weights.
Light gradient boosting machine (LightGBM) classification method
Developed by the Microsoft team, LightGBM is a fast classification and regression algorithm based on a gradient boosting decision tree, which represents an optimized improvement of the XGBoost algorithm. The improvement of the LightGBM algorithm lies in the use of the Histogram algorithm to process the data, leaf-wise growth strategy to construct the tree, and the optimal splitting point by optimizing the objective function to select the optimal splitting point. It can be comprehended that the LightGBM algorithm is a combination of XGBoost, Histogram, GOSS, and EFB algorithms.
The construction of the LightGBM algorithm model is explained in Figure 5:
The LightGBM algorithm has a fast, efficient, distributed structure and high-performance characteristics that can be used in sorting, classification, regression, and many other machine learning tasks. The LightGBM algorithm’s advantages are the following: the introduction of the Histogram algorithm, which reduces the time complexity consumed by traversal; during the training process, the one-sided gradient algorithm can filter the samples with small gradient to reduce the amount of computation; the leaf-wise growth strategy to construct the tree is also introduced to reduce the computational expense; the optimized feature-parallel and data-parallel methods are used to accelerate the computation, and the ticket-parallel strategy can be adopted when the data volume is large and the cache is also optimized.
CatBoost classification method
The CatBoost algorithm is an open-source gradient boosting classification algorithm that uses symmetric binary tree structure for training and introduces a new loss function and optimization method. A fully symmetric binary tree, based on symmetric decision trees (oblivious trees), is used by the CatBoost algorithm’s base learner. This algorithm provides fewer parameters and supports categorical variables and an extremely accurate GBDT framework, which can efficiently and reasonably process categorical features. It also proposes methods for dealing with gradient bias and prediction shift problems to improve the accuracy and generalization ability of the algorithm.
First, the CatBoost algorithm defines the training data set as | D | = { ( x 1 , y 1 ) , ( x 2 , y 2 ) , , ( x n , y n ) } , and then the prediction set is as:
y ^ i = ϕ ( x i ) = k = 1 K f k ( x i )
where f k represent the regression trees and K is the number of regression trees. The formula proves that after performing an input x i , the output K regression tree adds up to the predicted values.
Define the objective function, the loss function, and the regular term to obtain the optimized objective function:
L ( ϕ ) = i l ( y ^ i , y i ) + k Ω ( f k )
where Ω ( f ) = Υ T + 1 2 λ w 2 .
The CatBoost algorithm automatically handles category features and excels in both performance and effectiveness. The CatBoost algorithm has several benefits, including its remarkable capacity to manage category characteristics, resilience, exceptional performance, support for GPU acceleration, automated feature selection, and compatibility on sparse data.
Random Forest (RF) classification method
RF is a supervised classification algorithm based on decision trees, which predicts classification results by assembling multiple decision trees. In 2001, Bremen combined classification trees into a Random Forest, i.e., randomized the use of variables (columns) and the use of data (rows) to generate many classification trees and then aggregated the results of the classification trees. RF improves the prediction accuracy without significant increase in arithmetic.
RF is an extension of Bagging, the model input is defined as the training data set D = { ( x 1 , y 1 ) , ( x 2 , y 2 ) , , ( x m , y m ) } , the base learning algorithm I , and the number of training rounds   T .
Conduct the T iteration of the learning algorithm, and the base learning algorithm I is updated in the iteration:
h t = I ( D , D b x )
The output model H ( x ) was obtained:
H ( x ) = arg m a x y T t = 1 T ( h t ( x ) = y )
RF is an extended variant of Bagging that further introduces the selection of random attributes in the training process of decision trees based on the decision tree as the base learner to build the Bagging integration. RF is shown in Figure 6, and the yellow in the figure represents the flow of data:
RF increases the differences between classification models by constructing different training data sets, thereby improving the extrapolated prediction ability of the combined classification model. Through training, a sequence of classification models is obtained, and then they are utilized to form a multi-classification model system with the final classification decision as H ( x ) , as in Equation (22).
RF supports parallel processing and does not require normalization of features or processing of missing values; the model is stable, generalizes well, and can output the importance of features; it uses Out of Bag and does not need to divide the test set separately. However, it takes a long time to construct the tree, and the algorithm occupies a large amount of memory.
Neural network (NN) classification method
Neural networks utilize neural networks for classification by modeling the way the human nervous system works.
A neuron can be understood as a multi-dimensional linear function or a unit that achieves a linear combination. In the figure, { x } is the input to the neuron, H ( θ ) is the threshold function, and f represents the output.ω is the weights in the linear combination or the slope of the line. To facilitate the representation and computation of a considerable number of weights, they are typically represented as vectors or matrices.
The NN approach is shown in Figure 7:
NN is the current mainstream algorithm used for image classification, which supports automatic learning of features and patterns in the data and has good adaptability to non-linear relationships, Its computational units support highly paralleled computation, which speeds up the training speed, its distributed storage and processing improves the fault tolerance of the system, and it has a good generalization ability after sufficient training and is able to perform accurate classification on unseen data. When dealing with large-scale data and complex tasks, NN requires longer training time and more massive computational resources and can be improved in terms of reasonable choice of network structure, adjustment of hyper parameters, and avoidance of over-fitting to address the problem.
The categories, application directions, advantages, and disadvantages of the seven algorithms are compared in Table 2, which is shown underneath:

4. Experiments and the Results

This section describes the specific experimental process and methodology of the aero-engine classification experiment, which consists of three parts: aero-engines, spectral acquisition experimental design, data set production and spectral feature vector extraction, and classification prediction result accuracy assessment. Among them, the part of aero-engines spectral measurement experiment design describes the field arrangement of the aero-engine hot jet spectral measurement experiment, the part of data set fabrication and spectral feature vector extraction describes the training data set, test data set, label set production, and spectral feature vector extraction adopted in the classification experiment, and the classification prediction result accuracy assessment section evaluates the prediction of the classification method used in this paper on the real data set. Ultimately, the experimental result graphs and evaluation index tables are provided.

4.1. Experimental Design of Aero-Engine Spectral Measurement

In the first place, the infrared spectral data of three different aero-engines’ types were collected by field measurement. The FT-IR spectrometers used in the experiment are the EM27 and the telemetry FT-IR spectrometer developed by the Aerospace Information Research Institute. The specific parameters of the two devices are shown in Table 3:
The experimental preparation stage requires the experimental devices to be set up according to the experimental conditions in the external field.
Initially, according to the spectrometers’ field of view angle and hot jet information, the measurement distance was adjusted with a telescope to ensure the hot jet fills the field of view. The EM27 and telemetry FT-IR spectrometer were mounted on two tripods, the laser and scopes were used to assist the aiming, and the height and angle of the tripod were adjusted so that the optical axis of the equipment was aligned with the center of the aero-engine tail nozzle.
Next, after fixing the position, increase the tripod counterweight to improve stability and fix the thermos-hygrometer near the measurement position. Determine the position of the infrared thermal camera to clearly photograph the hot jet to be measured.
Finally, the workstation display time and the control room control system time were strictly aligned. After the cooling process is finished, launch the two computer programs listed above in that order. In the EM27 program, adjust the measurement mode, spectral resolution, and number of superimposed cycles. Then, work with the displayed ADJUST value to adjust the tripod angle until the highest value is recorded. Following the completion of the settings, background data were collected from the spectrum during the times when the aero-engine was not running.
The layout of the experimental site is presented in Figure 8:
During the experiment, real-time communication was conducted with the person in charge of the aero-engine through walkie-talkies, requesting real-time prompts when the rotational speed was changed, and requesting that each test rotational speed be stabilized for 1 min as much as possible (since 100% rotational speed is difficult to be maintained for 1 min in the actual measurement process, the amount of spectral data collected in this part is small). The ambient temperature and humidity were recorded at each adjustment of the rotational speed. The environmental factors of the experiment were recorded as shown in Table 4:
Table 4 provides the conditions under which the experiment was carried out. The temperature and humidity in the environment cause an absorbing and attenuating effect on the spectrum, which responds to atmospheric radiative transfer. When solar radiation and surface thermal radiation are transmitted in the atmosphere, they are affected by the absorption and scattering of atmospheric molecules like H2O, mixed gases (CO2, CO, N2O, CH4, O2), O3, N2, etc., on the one hand, and by the scattering or absorption of aerosolized particulate matter on the other hand, which results in the attenuation of the solar radiation and surface thermal radiation. At a distance from the target, the atmosphere will have a non-negligible effect on the collected spectrum. As of currently, we utilize the method of ground testing, the experimental distance is relatively close, and the difference between aero-engines’ hot jets and the background is very large, so the attenuation can be ignored.

4.2. Data Set Production and Spectral Feature Vectors Extraction

Based on the actual measurements of the aero-engines in the field, the controllable rotational speed ratios differed for each engine. Therefore, the infrared spectral data of 70%, 80%, 90%, and 100% of the maximum rotational speed ratios common to the three aero-engines were selected as the data source. There are a total of 211 spectral data in the data source, and after removing 2 erroneous data, 209 reliable data remain.
The experiment was conducted based on the aero-engine models, so the original data were divided by the aero-engines’ types. In the first place, 209 pieces of data were assigned random numbers from 1 to 209 to ensure the numbers were unrepeated, and the labels corresponding to the data were recorded (the labels corresponded to the engine model), 3/4 of them and their labels were selected at random as the training data set, and the rest of the data with the labels were used as the test data set.
The spectral features of the experimentally measured three aero-engines at 70%, 80%, 90%, and 100% of the maximum rotational speed were counted, and the statistical results were plotted as a two-dimensional feature map (as shown in the figure below), where the horizontal coordinates are   a 1 and the vertical coordinates are a 2 , as shown in Figure 9.
The aero-engines of the three aircraft are relatively sufficient for fuel combustion, and the exhaust gas spectrum at different speeds is close, so the gas composition characteristics of the gas discharged at the cruise speed of the three aircraft aero-engines are compared. Due to the various environment, temperature, and other conditions of the three field experiments, the spectrograms were compared after deducting the background. From the two-dimensional feature map, it can be regarded that the two-dimensional feature vectors of the three types of engines have been distributed in different regions of the feature space, and the overlap region between each other is relatively small, so the constructed spectral feature vectors initially have the ability to classify.
The spectral feature vectors and the classifier are combined to train and predict the spectral data. The flow chart of the classification algorithm used in the experiment is shown in Figure 10:

4.3. Assessment of the Accuracy of Classification Prediction Results

The experimentally constructed infrared spectra training data set and test data set are tested for classification of spectral feature vectors with seven classifiers, SVM, XGBoost, CatBoost, AdaBoost, Random Forest, LightGBM, and neural networks, and the classification results are shown in Figure 11:
Where, red, green, and blue represent the feature vectors of the three correctly classified aero-engine hot jet infrared spectra, while black represents the misclassified feature vectors.
At the same time, the parameters of the seven classification algorithms are given in Table 5, which is shown below:
The evaluation criteria for aero-engine spectral classification consist of accuracy, precision, recall, F1 value (F1-score), and confusion matrix [26]. It is assumed that if the instance is a positive class and is predicted to be positive, i.e., true class, it is denoted as TP (true positive), and if it is predicted to be negative, i.e., false negative, it is denoted as FN (false negative); on the contrary, if the instance is a negative class and it is predicted to be positive, i.e., false positive, it is denoted as FP (false positive), and if it is predicted to be negative, i.e., true negative, it is denoted as TN (true negative). Based on the above assumptions, the accuracy, precision, recall, F1 value, and confusion matrix of the evaluation criteria are defined separately:
① Accuracy: proportion of correctly categorized samples to total samples.
Accuracy   = T P + T N T P + T N + F P + F N
② Precision: the ratio of the number of samples correctly predicted to be positive to the number of all samples predicted to be positive.
Precision   = T P T P + F P
③ Recall: the ratio of the number of samples correctly predicted to be in the positive category to the number of samples in the true positive category.
Recall   = T P T P + F N
④ F1-score: the F1 value combines the harmonic mean of precision and recall and is used to measure the overall performance of the model.
F 1 - score   = 2 * P * R P + R
Among them, P represents Precision, R represents Recall.
⑤ Confusion matrix: The confusion matrix shows how well the classifier categorized the different categories, including true examples, false positive examples, true negative examples, and false negative examples. It proves the difference between the actual and predicted values, and the values on the diagonal of the confusion matrix indicate the number of correct predictions made by the classifier for that category. The confusion matrix is shown in Table 6:
According to the above five evaluation criteria for the algorithm of combining spectral feature vectors and classifiers used in this paper to predict the labels of the prediction set, the prediction results are shown in Table 7:
The evaluation criteria for combining the aero-engine hot jet infrared spectral feature vectors with the seven classifier combination algorithms are analyzed according to the preceding table. Since the experimental data set is generated by random numbering based on the experimental measurement spectra, the probability of predicting the data possesses a certain degree of chance. Therefore, several experiments were conducted to evaluate the general accuracy, and the correctness is shown in Figure 12, where the data of CatBoost (shown yellow) basically overlaps with AdaBoost.
The data from the 30 experiments conducted were counted to obtain the classifier prediction probability statistics as shown in Table 8:
According to Table 8, in terms of accuracy, CatBoost, AdaBoost, Random Forest, and LightGBM in repeated experiments maintain good accuracy and can achieve relatively accurate prediction in multiple experiments. The prediction accuracy of SVM is average, while the performance of neural networks is unsatisfactory. In terms of time measures, the seven methods have a similar running time, while CatBoost is slightly slower. Substantially, the mainstream classifiers have achieved a relatively ideal classification accuracy.

5. Conclusions

In this paper, for the aero-engine classification problem, two Fourier transform infrared spectrometers, Bruker’s EM27 and self-developed telemetry FT-IR spectrometer, were used to telemetry the infrared spectra of the hot jet of three aero-engine engines in different states, the training data set and test data set were randomly divided in the ratio of 3:1, and the spectral feature vectors were used to combine with the classification algorithm for the training of the training data set and the labeling of the test set. The classification evaluation indexes are accuracy, precision, recall, confusion matrix, and F1-score, and the classification accuracy of the algorithm is about 98%.
The spectral feature vectors proposed in this paper remain a preliminary concept for the aero-engine classification and identification problem, and in the subsequent stage, we will further study the infrared radiation model of the aero-engines’ hot jet and statistically analyze the more stable feature peaks in the infrared spectrum of the hot jet to find out the most stable feature peaks to build more robust spectral feature vectors, which can be used for more accurate classification of the aero-engines; furthermore, the spectral feature vectors proposed in this paper can be used for more accurate classification of the aero-engines. Expand the hot jet infrared spectral data of aero-engines and try using the measured spectral data to expand the aero-engine hot jet infrared spectral library, so as to develop a foundation for the recognition of aero-engines; under the premise of insufficient data, the method of deep migration learning can be introduced to expand the amount of training samples, so as to improve the training degree of the model.

Author Contributions

Formal analysis, Y.L.; Investigation, S.D. and Z.L.; Software, W.H.; Validation, Z.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study is available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Razeghi, M.; Nguyen, B.M. Advances in mid-infrared detection and imaging: A key issues review. Rep. Prog. Phys. 2014, 77, 082401. [Google Scholar] [CrossRef]
  2. Chikkaraddy, R.; Arul, R.; Jakob, L.A.; Baumberg, J.J. Single-molecule mid-IR detection through vibration ally-assisted luminescence. arXiv 2022, arXiv:2205.07792. [Google Scholar]
  3. Knez, D.; Toulson, B.W.; Chen, A.; Ettenberg, M.H.; Nguyen, H.; Potma, E.O.; Fishman, D.A. Spectral imaging at high definition and high speed in the mid-infrared. Sci. Adv. 2022, 8, eade4247. [Google Scholar] [CrossRef]
  4. Zhang, J.; Gong, Y. Automated identification of infrared spectra of hazardous clouds by passive FTIR remote sensing. In Multispectral and Hyperspectral Image Acquisition and Processing; SPIE: Bellingham, DC, USA, 2001; Volume 4548, pp. 356–362. [Google Scholar]
  5. Roh, S.B.; Oh, S.K. Identification of Plastic Wastes by Using Fuzzy Radial Basis Function Neural Networks Classifier with Conditional Fuzzy C-Means Clustering. J. Electr. Eng. Technol. 2016, 11, 103–116. [Google Scholar] [CrossRef]
  6. Kumar, V.; Kashyap, M.; Gautam, S.; Shukla, P.; Joshi, K.B.; Vinayak, V. Fast Fourier infrared spectroscopy to characterize the biochemical composition in diatoms. J. Biosci. 2018, 43, 717–729. [Google Scholar] [CrossRef]
  7. Han, X.; Li, X.; Gao, M.; Tong, J.; Wei, X.; Li, S.; Ye, S.; Li, Y. Emissions of Airport Monitoring with Solar Occultation Flux-Fourier Transform Infrared Spectrometer. J. Spectrosc. 2018, 2018, 1069612. [Google Scholar] [CrossRef]
  8. Cięszczyk, S. Passive Open-Path FTIR Measurements and Spectral Interpretations for in situ Gas Monitoring and Process Diagnostics. Acta Phys. Pol. A 2014, 126, 673–678. [Google Scholar] [CrossRef]
  9. Schütze, C.; Lau, S.; Reiche, N.; Sauer, U.; Borsdorf, H.; Dietrich, P. Ground-based remote sensing with open-path Fourier-transform infrared (OP-FTIR) spec-troscopy for large-scale monitoring of greenhouse gases. Energy Procedia 2013, 37, 4276–4282. [Google Scholar] [CrossRef]
  10. Doubenskaia, M.; Pavlov, M.; Grigoriev, S.; Smurov, I. Definition of brightness temperature and restoration of true temperature in laser cladding using infrared camera. Surf. Coat. Technol. 2013, 220, 244–247. [Google Scholar] [CrossRef]
  11. Homan, D.C.; Cohen, M.H.; Hovatta, T.; Kellermann, K.I.; Kovalev, Y.Y.; Lister, M.L.; Popkov, A.V.; Pushkarev, A.B.; Ros, E.; Savolainen, T. MOJAVE. XIX. Brightness Temperatures and Intrinsic Properties of Blazar Jets. Astrophys. J. 2021, 923, 67. [Google Scholar] [CrossRef]
  12. Schumann, U. On the effect of emissions from aircraft engines on the state of the atmosphere. Ann. Geophys. 2005, 12, 365–384. [Google Scholar] [CrossRef]
  13. Boser, B.E.; Guyon, I.M.; Vapnik, V.N. A training algorithm for optimal margin classifiers. In Proceedings of the Fifth Annual Workshop on Computational Learning Theory, Pittsburgh, PA, USA, 27–29 July 1992; pp. 144–152. [Google Scholar]
  14. Zhang, Y.; Li, T. Three different SVM classification models in Tea Oil FTIR Application Research in Adulteration Detection. J. Phys. Conf. Ser. 2021, 1748, 022037. [Google Scholar] [CrossRef]
  15. Menezes, M.V.; Torres, L.C.; Braga, A.P. Width optimization of RBF kernels for binary classification of support vector machines: A density estimation-based approach. Pattern Recognit. Lett. 2019, 128, 1–7. [Google Scholar] [CrossRef]
  16. Chen, T.; Guestrin, C. XGBoost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  17. Nalluri, M.; Pentela, M.; Eluri, N.R. A Scalable Tree Boosting System: XGBoost. Int. J. Res. Stud. Sci. Eng. Technol. 2020, 7, 36–51. [Google Scholar]
  18. Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased boosting with categorical features. arXiv 2018, arXiv:1706.09516. [Google Scholar]
  19. Dorogush, A.V.; Ershov, V.; Gulin, A. CatBoost: Gradient boosting with categorical features support. arXiv 2018, arXiv:1810.11363. [Google Scholar]
  20. Dorogush, A.V.; Gulin, A.; Gusev, G.; Kazeev, N.; Prokhorenkova, L.O.; Vorobev, A. Fighting biases with dynamic boosting. arXiv 2017, arXiv:1706.09516. [Google Scholar]
  21. Freund, Y.; Schapire, R.; Abe, N. A short introduction to boosting. J.-Jpn. Soc. Artif. Intell. 1999, 14, 771–780. [Google Scholar]
  22. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  23. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.Y. Lightgbm: A highly efficient gradient boosting decision tree. Adv. Neural Inf. Proc. Syst. 2017, 30, 3149–3157. [Google Scholar]
  24. Zeng, P. Artificial Neural Networks Principle for Finite Element Method. Z. Angew. Math. Mech. 1996, 76, 565–566. [Google Scholar]
  25. ArulRaj, K.; Karthikeyan, M.; Narmatha, D. A View of Artificial Neural Network Models in Different Application Areas. E3S Web Conf. 2021, 287, 03001. [Google Scholar] [CrossRef]
  26. Sokolova, M.; Lapalme, G. A systematic analysis of performance measures for classification tasks. Inf. Proc. Manag. 2009, 45, 427–437. [Google Scholar] [CrossRef]
Figure 1. Experimentally measured hot jets’ infrared spectra of three turbojet engines.
Figure 1. Experimentally measured hot jets’ infrared spectra of three turbojet engines.
Electronics 13 00915 g001
Figure 2. Schematic representation of the positions of the four characteristic peaks of the three aero-engines as measured in practice.
Figure 2. Schematic representation of the positions of the four characteristic peaks of the three aero-engines as measured in practice.
Electronics 13 00915 g002
Figure 3. Schematic representation of dichotomy SVM data classification.
Figure 3. Schematic representation of dichotomy SVM data classification.
Electronics 13 00915 g003
Figure 4. Schematic representation of the AdaBoost algorithm.
Figure 4. Schematic representation of the AdaBoost algorithm.
Electronics 13 00915 g004
Figure 5. Schematic representation of the LightGBM algorithm.
Figure 5. Schematic representation of the LightGBM algorithm.
Electronics 13 00915 g005
Figure 6. Schematic representation of the RF algorithm.
Figure 6. Schematic representation of the RF algorithm.
Electronics 13 00915 g006
Figure 7. Schematic representation of the NN algorithm.
Figure 7. Schematic representation of the NN algorithm.
Electronics 13 00915 g007
Figure 8. Schematic representation of the aero-engine hot jet infrared spectrum measurement experiment site.
Figure 8. Schematic representation of the aero-engine hot jet infrared spectrum measurement experiment site.
Electronics 13 00915 g008
Figure 9. Characteristic statistical results of three different turbojet aero-engines at 70%, 80%, 90%, and 100% maximum speed.
Figure 9. Characteristic statistical results of three different turbojet aero-engines at 70%, 80%, 90%, and 100% maximum speed.
Electronics 13 00915 g009
Figure 10. Flowchart of the spectral feature vector aero-engine classification algorithm.
Figure 10. Flowchart of the spectral feature vector aero-engine classification algorithm.
Electronics 13 00915 g010
Figure 11. Aero-engine hot jet infrared spectral feature vector classification effect map.
Figure 11. Aero-engine hot jet infrared spectral feature vector classification effect map.
Electronics 13 00915 g011aElectronics 13 00915 g011bElectronics 13 00915 g011c
Figure 12. Distribution of correct rates for each algorithm under 30 experiments.
Figure 12. Distribution of correct rates for each algorithm under 30 experiments.
Electronics 13 00915 g012
Table 1. Characteristic peak threshold takes the value range.
Table 1. Characteristic peak threshold takes the value range.
Characteristic Peak TypeEmission Peak (cm−1)Absorption Peak (cm−1)
Peak standard features23502390719667
Characteristic peak range values2350.5–23482377–2392722–718666.7–670.5
Table 2. Comparison table of classification algorithms.
Table 2. Comparison table of classification algorithms.
MethodsCategoriesApplication ScenariosAdvantagesDisadvantages
SVMSupervised learning methodClassification and regression, text classification, image recognitionAdaptability of high-dimensional spatial data processing and small sample data processingLess efficient processing of large data sets
XGBoostEnsemble learningBiclassification, multiclassification, and regression problemsHigh efficiency, flexibility, automatic selection of important features, prevention of overfitting, parallel computationLess efficient processing of large data sets
CatboostEnsemble learningBiclassification, multiclassification, and regression problemsStrong ability of handling categorical features, robustness, high performance, GPU acceleration support, automatic feature selection, and friendly treatment of sparse dataPoor robustness of large-scale data and non-linear relationship processing
AdaboostEnsemble learningBiclassification, multiclassification, and regression problemsEasy to implement and adjust, not easy to overfit, combination with various basic classifiersNoise sensitive
Random ForestEnsemble learningImage recognition, data predictionParallel processing, stable model, good generalization abilityNoise sensitive, long tree construction time, large memory consumption
LightGBMEnsemble learningMachine learning and data mining areasEfficient, distributed structure, and high performanceNoise sensitive
Neural NetworkMachine learningImage classification, computer vision, natural language processingHigh classification accuracy, strong parallel distributed processing, noise robustnessExtensive parameter adjustment, non-intuitive learning process, and long learning time
Table 3. Parameters of the Fourier transform infrared spectrometers used for the experiment.
Table 3. Parameters of the Fourier transform infrared spectrometers used for the experiment.
NameManufacturerMeasurement PatternSpectral Resolution (cm−1)Spectral Measurement Range (μm)Full Field of View Angle
EM27BrukerActive/PassiveActive: 0.5/1 Passive: 0.5/1/42.5~1230 mrad (no telescope) (1.7°)
Telemetry Fourier Transform Infrared SpectrometerAerospace Information Research InstitutePassive12.5~121.5°
Table 4. Table of experimental aero-engines and environmental factors.
Table 4. Table of experimental aero-engines and environmental factors.
Aero-Engine Serial NumberEnvironmental TemperatureEnvironmental HumidityDetection Distance
130 °C43.5% Rh11.8 m
220 °C71.5% Rh5 m
319 °C73.5% Rh10 m
Table 5. Parameter list of classification algorithms.
Table 5. Parameter list of classification algorithms.
MethodsParameter Settings
SVMdecision_function_shape = ‘ovr’, kernel = ‘rbf’
XGBoostobjective = ‘multi:softmax’, num_classes = num_classes
CatBoostloss_function = ‘MultiClass’
Adaboostn_estimators = 200
Random Forestn_estimators = 300
LightGBMobjective’: ‘multiclass’,
‘num_class’: num_classes
Neural Networkhidden_layer_sizes = (100), activation = ‘relu’, solver = ‘adam’, max_iter = 200
Table 6. Confusion matrix.
Table 6. Confusion matrix.
Forecast Results
Positive samplesNegative samples
Real resultsPositive samplesTPTN
Negative samplesFPFN
Table 7. Table of classification and evaluation indexes of aero-engine hot jet infrared spectrum feature vectors and seven classifier algorithms.
Table 7. Table of classification and evaluation indexes of aero-engine hot jet infrared spectrum feature vectors and seven classifier algorithms.
Evaluation CriterionAccuracyPrecision ScoreRecallF1Confusion
Matrix
Running Time/s
Classification Methods
feature vectors + SVM98.04%98.77%97.78%98.22% 26 0 0 0 10 0 1 0 14 2.48
Feature vectors + XGBoost98.04%98.77%97.78%98.22% 26 0 0 0 10 0 1 0 14 2.62
Feature vectors + CatBoost98.04%98.77%97.78%98.22% 26 0 0 0 10 0 1 0 14 5.27
Feature vectors + AdaBoost98.04%98.77%97.78%98.22% 26 0 0 0 10 0 1 0 14 2.91
Feature vectors + Random Forest98.04%98.77%97.78%98.22% 26 0 0 0 10 0 1 0 14 3.09
Feature vectors + LightGBM96.08%96.38%96.38%96.38% 26 0 1 0 10 0 1 0 13 2.63
Feature vectors + Neural Networks80.39%76.19%90.99%76.27% 27 0 10 0 10 0 0 0 4 2.41
Table 8. Statistics of prediction probability of classifiers.
Table 8. Statistics of prediction probability of classifiers.
Method
Order
SVMXGBoostCatBoostAdaBoostRandom ForestLightGBMNeural Networks
Average value97.17%97.74%98.13%98.00%98.32%98.07%74.52%
Variance0.06%0.04%0.03%0.04%0.03%0.02%1.84%
Standard deviation2.41%1.96%1.71%1.92%1.73%1.52%13.56%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Du, S.; Han, W.; Shi, Z.; Liao, Y.; Li, Z. An Aero-Engine Classification Method Based on Fourier Transform Infrared Spectrometer Spectral Feature Vectors. Electronics 2024, 13, 915. https://doi.org/10.3390/electronics13050915

AMA Style

Du S, Han W, Shi Z, Liao Y, Li Z. An Aero-Engine Classification Method Based on Fourier Transform Infrared Spectrometer Spectral Feature Vectors. Electronics. 2024; 13(5):915. https://doi.org/10.3390/electronics13050915

Chicago/Turabian Style

Du, Shuhan, Wei Han, Zhengyang Shi, Yurong Liao, and Zhaoming Li. 2024. "An Aero-Engine Classification Method Based on Fourier Transform Infrared Spectrometer Spectral Feature Vectors" Electronics 13, no. 5: 915. https://doi.org/10.3390/electronics13050915

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop