Next Article in Journal
Integrated Earthquake Catalog III: Gakkel Ridge, Knipovich Ridge, and Svalbard Archipelago
Next Article in Special Issue
Design and Testing of a Measurement Device for High-Speed Bearing Evaluation
Previous Article in Journal
Assessment of Insulation against Contact Heat and Radiant Heat of Composites with TiO2-ZrO2-Al and Parylene C Coatings Intended for Protective Gloves Supported by Computational Fluid Dynamics
Previous Article in Special Issue
Anomalistic Symptom Judgment Algorithm for Predictive Maintenance of Ship Propulsion Engine Using Machine Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research Method for Ship Engine Fault Diagnosis Based on Multi-Head Graph Attention Feature Fusion

1
Marine Engineering College, Dalian Maritime University, Dalian 116026, China
2
Dalian Maritime University Smart Ship Limited·Company, Dalian 116026, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(22), 12421; https://doi.org/10.3390/app132212421
Submission received: 13 October 2023 / Revised: 8 November 2023 / Accepted: 14 November 2023 / Published: 16 November 2023
(This article belongs to the Special Issue Technical Diagnostics and Predictive Maintenance)

Abstract

:
At present, there are problems such as low fault data, insufficient labeling information, and poor fault diagnosis in the field of ship engine diagnosis. To address the above problems, this paper proposes a fault diagnosis method based on probabilistic similarity and rank-order similarity of multi-head graph attention neural networks (MPGANN) models. Firstly, the ship engine dataset is used to explore the similarity between the data using the probabilistic similarity of T_SNE and the rank order similarity of Spearman’s correlation coefficient to define the neighbor relationship between the samples, and then the appropriate weights are selected for the early fusion of the two graph structures to fuse the feature information of the two scales. Finally, the graph attention neural networks (GANN) incorporating the multi-head attention mechanism are utilized to complete the fault diagnosis. In this paper, comparative experiments such as graph construction and algorithm performance are carried out based on the simulated ship engine dataset, and the experimental results show that the MPGANN outperforms the comparative methods in terms of accuracy, F1 score, and total elapsed time, with an accuracy rate of 97.58%. The experimental results show that the model proposed in this paper can still fulfill the ship engine fault diagnosis task well under unfavorable conditions such as small samples and insufficient label information, which is of practical significance in the field of intelligent ship cabins and fault diagnosis.

1. Introduction

As the main power source for ship navigation and power generation, ship engines occupy an important position in the field of ships. Due to its importance, if the engine operation failure cannot be found in time, it will cause huge economic losses or even casualties. Therefore, timely monitoring and diagnosis of the engine’s working condition can effectively improve the reliability, economy, and safety of normal ship operation.
With the wave of the era of ship intelligence and the continuous development and improvement of artificial intelligence and algorithmic arithmetic power, more and more scholars apply artificial intelligence to the field of condition detection and fault diagnosis, and the fault diagnosis methods of ship engines also progress and develop. Fault diagnosis of ship engines faces challenges such as data scarcity, multiple fault types, data noise interference, and insufficient labeling information. In order to provide safety for normal ship navigation, Zhong et al. [1] introduced deep learning into ship diesel engine fault diagnosis and proposed a fault diagnosis method based on correlation distribution and deep confidence networks, which achieved good results. However, this type of method does not take into account the problem of imbalance in the ship fault data categories, resulting in some defects in the training and application of the model. Aiming at the above problems Hou et al. [2] first used principal component analysis (PCA) to reduce the dimensionality of the data. Then, the sample size optimization algorithm (SSO) was used to solve the problem of data imbalance, and the three-dimensional Arnold mapping to the particle swarm optimization algorithm was introduced to improve the holding vector machine (SVM), which effectively improves the generalization ability of the classification model. Finally, the model is validated by example using the fault data of the fuel supply system, and the experimental results show that the method is able to alleviate the impact of unbalanced data. Zhong et al. [3] proposed a semi-supervised principal component analysis (SSPCA) to improve the diagnostic accuracy of the model by fusing labeled and unlabeled samples as an alternative to unsupervised learning, with better robustness to false alarms.
The above model has achieved good results in fault monitoring, but it can only determine whether the fault occurs or not and cannot diagnose the specific cause of the fault, which makes it difficult to play a guiding role for the subsequent operation and maintenance decision-making work. In this regard, Ren et al. [4] established a fault tree, transformed the fault tree into a Bayesian network, constructed a Bayesian network diagnostic model of the lubrication system, and, in the example validation, the model was able to accurately diagnose the cause of the failure. Han et al. [5] introduced a convolutional neural network to solve the problem of propulsion fault detection and classification of time series. Xu et al. [6] fused an artificial neural network (ANN) model, a belief rule-based reasoning (BRB) model, and an ER rule model and used a genetic algorithm to optimize the importance weights of each model to improve the overall performance of the fusion system, and finally, through the three models, joint decision-making to realize the fault diagnosis of the ship system. In recent years, algorithmic improvements have achieved significant results in the field of industrial fault diagnosis [7,8,9,10,11]. Samet [12] proposed a new kNN-based classifier (PFS-kNN) to find the k-nearest neighbors using Minkowski’s metric of the image fuzzy soft matrix, which was validated on the UCI medical dataset, and the PFS-kNN outperformed the most state-of-the-art kNN-based algorithms. Yang et al. [13] proposed a diagnostic method based on fault sign simulation to solve the problem of a too-small sample size of faults in the fault diagnosis of a particular engine. Agrawal et al. [14] proposed a data-driven model for predictive health maintenance to monitor the health status of the equipment and take action before a fault occurs. The model will alert for any major defects in the system and can be effective in avoiding work interruptions in the production process.
A graph neural network (GNN) is a neural model that captures graph dependencies through message passing between graph nodes. Zhou et al. [15] divided GNN applications into six directions: natural language processing, computational technology, natural science research, knowledge graph, combinatorial optimization, and graph generation, and made a more comprehensive and detailed summary of GNN applications.
In recent years, variants of GNNs such as Graph Convolutional Networks (GCNs) [16] and GANN [17] have achieved breakthroughs in many deep learning tasks. Kipf et al. [18] applied GCNs to fast and scalable semi-supervised classification of nodes in graphs, which demonstrated significant superiority over a large number of datasets. Wang et al. [19] proposed a multi-graph-based GCN fault diagnosis method that fuses two feature embeddings into one combined embedding by introducing a self-attention mechanism to improve classification accuracy and stability in unbalanced datasets. Li et al. [20] proposed a domain adversarial graph convolutional network (DAGCN), which incorporates three types of information—class label, domain label, and data structure—into a deep network for modeling, thus realizing unsupervised domain adaptive (UDA) fault diagnosis. Zheng et al. [21] proposed a new quantum graph convolutional neural network (QGCN) model by drawing on quantum neural networks and graph convolutional neural networks, which utilize quantum parametric circuits and the high computational power of quantum systems to accomplish the traditional task of graph classification in machine learning. In all of the above studies, GCN has achieved good results in the field of graph structure classification due to its advantage of being able to handle non-Euclidean distance data well. However, GCN’s mean-aggregation operation in the metric matrix makes it insufficiently attentive to important nodes. Therefore, Veličković et al. [22] constructed GANN by adding a self-attention layer to traditional GCNs, which are able to set different weights for different nodes in the neighborhood, thus solving the drawback that GCNs cannot consider the importance of neighbors. On this basis, Yang et al. [23] proposed a full graph attention neural network (FGANN), which also considers the influence of nodes other than neighboring nodes and can handle the graph classification task well. Zhang et al. [24] noticed that in sentiment classification tasks, traditional models are not sensitive to syntactic structural information due to the complexity of syntactic analysis relationships, and the models lack the utilization of external sentiment knowledge. In this regard, a new graph attention neural network is proposed that realizes attention to multiple levels of syntax, semantics, and knowledge by introducing the mechanism of multi-head attention and acquires knowledge about sentiment by considering information such as syntax.
The relationship between numerous thermal parameters of ship engines is not just a simple linear relationship but is often very complex and agnostic. In this paper, to address this problem, graph learning theory is introduced to explore the relationship of geometric structure among data, and a MPGANN ship engine fault diagnosis model is proposed. Based on the ship engine failure dataset, the model uses both T_SNE and Spearman to construct the probabilistic graph structure and ordinal graph structure, selects appropriate weights for early fusion of the two graph structures, and finally adopts GANN to transfer and extract the combined features. The core of the model is that it can mine the similarity between the data from two scales and more comprehensively describe the relationship between the data and the graph structure to provide more useful information for the model. The main contributions of this paper can be summarized as follows:
(1)
Transform the ship engine dataset into two graph structures from different scales and make the two graph structures contain similarity relationships from multiple perspectives by extracting the neighbor relationships between samples so as to achieve complementation and extension of the model input information.
(2)
Introducing fusion weights, the two graph structures are structurally fused according to appropriate weights to obtain a fused graph structure that contains deeper information.
(3)
Input the obtained fusion graph structure into the GANN of fused multi-head attention for multi-channel feature extraction, and finally connect the Softmax layer to realize fault diagnosis.
(4)
The ship engine thermal parameter dataset is used to verify that the MPGANN proposed in this paper outperforms other classical algorithms and achieves higher accuracy.
The remaining sections of this paper are organized as follows: In Section 2, the theory related to the MPGANN model is introduced in detail. Section 3 presents experiments and comparisons of the construction of the model proposed in this paper, the setting of hyperparameters, and the performance of the algorithm through case studies. Section 4 then summarizes and outlooks the paper.

2. Related Work and Theory

2.1. Data Topology Construction

Due to the complexity of the work of the ship engine, there is a more complex connection between the various data. In order to fully explore the deep connection between the data, this paper will use the conditional probability and the correlation coefficient between the features of the different nodes to calculate the adjacency relationship between two different scales, as a way to construct the probability graph structure and the similarity graph structure, and the fusion of the two graph structures to get the new fusion graph structure as the input of the network for fault diagnosis.

2.1.1. Probabilistic Graph Structure

Inspired by the T-SNE method [25], in this paper, we use the conditional probability calculated by transforming the distance between the data to represent the correlation between the data, and the correlation C o r ( x i , x j ) serves as the degree of correlation between the data samples x i and x j , and the larger C o r ( x i , x j ) is, the higher the likelihood that x i will be a neighbor of x j .
For a dataset X = { x 1 , x 2 , x 3 x n } , the conditional probability (denoted as p j | i ) between any pair of sample points, x j and x j , is expressed by Equation (1):
p j | i = exp ( x i x j 2 / 2 σ i 2 ) k i ( x i x k 2 / 2 σ i 2 ) ,
In this context, when both p j | i and q j | i are set to 0, σ i represents the Gaussian variance centered on point x i . It is worth noting that because data density can vary, there is no universally optimal value for σ i across all datasets. The entropy of this distribution increases as σ i grows larger, so a binary search approach is employed to determine the appropriate value of σ i . Perplexity (denoted as P e r P i ) is defined as the metric for computing σ i and typically falls within the range of [5, 50]. Its definition is as follows:
P e r ( P i ) = 2 H ( P i ) ,
By specifying a perplexity range as described earlier, the information entropy H P i in the equation above is calculated using the following definition:
H ( P i ) = j p j | i log 2 p j | i ,
In the low-dimensional space, the symbol q j | i is employed to denote the conditional probability between samples and to symbolize the data correlation C o r ( x i , x j ) , as shown in Equation (4):
C o r ( x i , x j ) = q j | i = exp ( y i y j 2 ) k i ( y i y k 2 ) ,
We utilize the KL divergence to quantify the level of concurrence between q j | i and p j | i . When the KL divergence is 0, it signifies that q j | i =   p j | i , signifying that y i and y j in the low-dimensional space precisely mirror the correct degree of correlation between data points x i and x j in the high-dimensional space. The KL divergence is defined as Equation (5):
C = K L ( P i Q i ) = i j p j | i log p j | i q j | i ,
The KL divergence is minimized through the process of gradient descent:
δ C δ y i = 2 j ( p j | i q j | i + p i | j q i | j ) ( y i y j ) ,
After the above formula, the degree of correlation between the two samples C o r ( x i , x j ) is calculated, and the larger C o r ( x i , x j ) is, the stronger the correlation between x i and x j . The probabilistic correlation C o r ( x i , x j ) value between each node constitutes the probabilistic feature matrix A p p:
A p = [ p 11 p 12 p 1 N p 21 p 22 p 2 N p N 1 p N 2 p N N ] ,

2.1.2. Rank-Order Graph Structure

In this paper, the Spearman correlation coefficient [26] is utilized to derive the similarity graph structure among nodes. The samples in the original data are sorted in ascending order. Let x i and x j represent the positions of x i and x j after this sorting; these positions are referred to as the rank order of samples x i and x j . The Spearman correlation coefficient between two samples is calculated as Equation (8):
S p e ( x i , x j ) = r s = 1 6 i = 1 n d i 2 n ( n 2 1 ) ,
where d is the difference between x i and x j and n represents the number of samples. The Spearman correlation coefficient values S p e x i , x j between the nodes form the correlation characterization matrix A o :
A O = [ o 11 o 12 o 1 N o 21 o 22 o 2 N o N 1 o N 2 o N N ] ,

2.1.3. Feature Graph Fusion

Multi-scale feature fusion is mainly used to fuse the probabilistic graph structure and rank-order graph structure constructed earlier and fully excavate the intrinsic feature connection between each graph structure. The more widely used feature fusion methods are early fusion [27], late fusion, and attention feature fusion. Attention feature fusion is able to effectively deal with multimodal data, and it has strong adaptability [28,29]. Late fusion can capture the temporal correlation and interaction effects between data sources well and has better performance in time series tasks [30]. However, the above two methods have the shortcomings of a complex model structure, a high demand for computational resources, and a high cost of training time when the number of samples is small [31]. Therefore, for the discrete small-sample ship engine failure sample set, this paper chooses early fusion [32] as the method of fusion of the multi-graph structure, sets the feature weight W i and neighbor threshold T of each graph to construct the fusion graph structure, and its formula is as follows:
A F = i = 1 K W i N o l ( A i ) ,
N o l ( x ) = x min ( x ) max ( x ) min ( x ) ,
A T = [ a 11 a 12 a 1 N a 21 a 22 a 2 N a N 1 a N 2 a N N ] ,
a i j = [ 1 , A F i j T 0 , A F i j < T ,
where A F is the fusion feature matrix; N o l A i represents the feature matrix of each graph after normalization; A i is the original feature matrix; A T signifies the adjacency matrix after multi-graph fusion; and a i j denotes the elements of the adjacency matrix in the ith row and jth column of A F after applying the adjacency threshold T .
The larger the value of the element A F i j in A F , the stronger the similarity between node i and node j. How to set the threshold T to define the adjacency matrix A T is the key. In this paper, we use quartiles to determine the threshold. The calculated A F i j values are arranged from small to large and divided into four equal parts. The three values in the numerical split point position are defined as Q1, Q2, and Q3, and Q3 is set to be the critical value with T . Compare A F i j with T . When A F i j is greater than T , it means that node i and node j are connected, and a i j = 1 . When A F i j is smaller than T , it means that i-node and j-node are not connected, then a i j = 0 .

2.2. Graphical Attention Neural Network

Both GANN and GCN studied in this paper are variants of GNN. Different from the Laplacian smoothing of all directly adjacent nodes at each node of GCN [33], GANN uses the attention mechanism to calculate the edge weight coefficient. The attention mechanism should be to the operation of aggregating neighbors of GCN. According to the three elements of the attention mechanism: query, source, and attention value [34], we will determine to set query, which is the feature vector of the current central node, source to the feature vector of all neighbors, and attention value to the new feature vector of the center node after the aggregation operation.
Consider the feature vector corresponding to node x i in layer 1 of the graph as h i h i R d l , where d l denotes the feature length of the node. After applying an aggregation operation with an added attention mechanism, the output is a new feature vector h i h i R d l + 1 for each node, with d l + 1 denoting the length of the output feature vector. We refer to this aggregation operation as the Graph Attention Layer (GAL).
Define the central node as x i , then the correlation from neighboring nodes x j to x i is as follows:
e i j = a ( W h i , W h j ) ,
The weight parameter of the feature transformation of the nodes in the layer is denoted as W R d l + 1 × d l . The function a · calculates the correlation between two nodes, and in this paper, a single fully connected layer is chosen:
e i j = L e a k y R e L U ( a T [ W h i W h j ] ) ,
The weight vector a R 2 d l + 1 is used along with the activation function LeakyReLU with a negative slope α = 0.2 . Here, “ ” denotes a vector splicing operation. To better assign the weights, the computed correlation is normalized using Softmax:
α i j = exp ( e i j ) k N i exp ( e i j ) ,
α is the weight coefficient, which is calculated in the equation above to ensure that the aggregated weight coefficients of all neighboring nodes of the central node sum to one. Figure 1 illustrates the computation process. W h i and W h j are first parameterized by a , then activated by LeakyReLU, and finally normalized by Softmax to obtain the weight parameters. Here’s the complete formula for calculating the weight coefficients:
α i j = exp ( L e a k y R e L U ( a T [ W h i W h j ] ) ) k N i exp ( L e a k y R e L U ( a T [ W h i W h k ] ) ) ,
The update formula for the feature vector of the node x i is calculated as follows based on the weight coefficients obtained from the above formula:
h ( l + 1 ) i = σ ( k N i α i j W h ( l ) j ) ,
where h i l + 1 is the updated feature vector of node xi in layer l + 1; α represents the activation function (commonly used is LeakyReLU); N i denotes the neighbors of node x i ; α i j is the weight coefficient calculated using the previously mentioned weight coefficient formula; and W is the weight matrix parameter. This formula calculates the updated feature vector for each node by aggregating information from its neighbors, with the aggregation weights determined by the attention mechanism.
To enhance the expressive capacity of the attention layer, this paper employs the multi-head attention mechanism, similar to the approach utilized by Vaswani et al. [35], to enhance the model’s expressive ability. The typical method for combining the outputs involves concatenating the outputs from k sets of mutually independent attention mechanisms in the equation provided above:
h ( l + 1 ) i = k = 1 K σ ( k N i α i j ( k ) W ( k ) h ( l ) j ) ,
This equation represents the concatenation operation. Each α i j ( k ) is the weight coefficient computed by the k th set of attention mechanisms, and W ( k ) corresponds to the associated learnable parameter. This approach combines the outputs from different attention heads, each with its own set of weights and parameters, allowing the model to capture diverse patterns and relationships in the data.

2.3. Algorithmic Process

Figure 2 shows a schematic diagram of the structure of the MPGANN model proposed in this paper. Figure 3 is a schematic diagram of the algorithm flow of the MPGANN model in this paper. Firstly, the values of the detection indexes of the engine in different states are collected by sensors, and the data are normalized. Furthermore, each sample is regarded as a node, and by constructing a probabilistic graph structure and an ordinal graph structure, the two structural graphs are fused as inputs to the multi-head GANN model by the multi-graph accumulation method, so that the model can learn the sample features at multiple scales. The features of neighboring nodes are aggregated in the model and computed iteratively to maximize the classification accuracy between the extracted feature values and the labeled values.

3. Case Study

3.1. Data Description

The subject of this experiment is a Wärtsilä 9L34DF marine engine. The real ship failure simulation is costly, dangerous, and destructive; therefore, the engine failure simulation performed by AVL-BOOST (R2019.2) software is used in this paper. The simulation model of the marine engine is shown in Figure 4, where SB is the system boundary; CL is the air filter; TC is the supercharger; CO is the air cooler; PL is the intake and exhaust manifold; C is the cylinder; MP is the arrangement of the detection points; and the black line is the corresponding connecting pipe.
Before the fault, simulation needs to verify the effectiveness of the model. The text for the parameters under different working loads is verified for the comparison of measured values and simulation values. Error analysis and comparison results are shown in Table 1.
As can be seen from Table 1, the error between the measured and real values of each main performance parameter under different loads is less than 4%. Especially under 100% load, the error of each main performance parameter is within 2%, and the error is within the acceptable range of requirements, so this paper chooses the model under 100% working conditions for fault simulation and fault sample collection and simulates five states, namely, injection timing advance (F1), injection timing delay (F2), supercharger efficiency decline (F3), air cooler efficiency decline (F4), and normal operation (F5). Five states are simulated, and the fault simulation scheme is shown in Table 2.
After the above fault simulation scheme, in order to fully explore and analyze the relationship between internal engine parameters and faults and realize the diagnosis of engine faults, a total of 400 sets of samples of 23-dimensional detection index parameters were selected, of which 80 sets of samples for each state were used to carry out the fault diagnosis study. The dataset is as specific as Table 3, and the collected detection indexes are shown in Table 4.

3.2. Comparison of Graph Structures

In this section, by comparing and analyzing the commonly used correlation coefficients, including Pearson correlation coefficient, Kendall correlation coefficient, and Spearman correlation coefficient, as well as their constructed correlation graph structure and probability graph structure, respectively, we compare the model accuracy and loss and select the graph structure construction method that is suitable for the proposed model in this paper. The activation function selected for the model is ReLU, the learning rate is 0.01, the number of iterations is 1000, and the number of multi-attention heads is 16. The ratio of the model training set, validation set, and test set is 7:1:2, which is selected randomly. Each graph structure construction method was subjected to 20 experiments, and the average accuracy of the 20 ornaments was used as the basis for judgment.
As can be seen in Figure 5, the average accuracy of the Spearman correlation coefficient in the correlation graph structure is better than that of the Pearson correlation coefficient and Kendall correlation coefficient, and the average accuracy of the fusion graph obtained after the fusion of the graph structures is improved as a model input. This is due to the fact that different graph construction methods extract information from different aspects of the data, and the fused graph structure contains information from both aspects of the data, which can be input into the neural network with more information than a single graph structure. Especially when the probabilistic graph structure and the correlation graph structure constructed by the Spearman correlation coefficient are fused, the average accuracy of the model is greatly improved, reaching 96.36%. It can be seen that the model performance is optimal when the correlation graph structure is constructed with the Spearman correlation coefficient, which is chosen to construct the graphs in all subsequent experiments.
Figure 6 shows the fused graph structure after fusion using the T_SNE probabilistic graph structure and the Spearman rank order graph structure. The edges between the nodes indicate that the nodes are neighbors to each other, as can be seen from the figure that the node relationships are only roughly delineated in a single graph structure, and after fusion of the graph structures, the neighbor relationships between the nodes have been clearly described.
After constructing the neighbor relationship, the labels of some nodes are removed; nodes with color in Figure 7 indicate that the node has label information, while nodes without color indicate that the node has no label information. During the graph learning process, these unlabeled nodes do not provide labeling information, but they can aggregate the neighbor’s node information to be used as auxiliary data in order to provide the information used to explore the similarity structure between the nodes. The information of the nodes is used to aggregate the neighboring information through multi-head attention, and after the graph learning, the updated new nodes are classified, and all the samples are tagged with labels. The above figure shows the classification results, and it can be seen that although the model is not given all the labeling information, it can still correctly label the unlabeled nodes.

3.3. Network Parameter Setting

In this subsection, we will delve into the critical hyperparameters of the neural network model for fault diagnosis. These include the activation function, the number of attention heads, and the model learning rate. We maintain the ratio of the training set, validation set, and test set at 7:1:2, conducting a total of 20 experiments. The results from these experiments will be averaged to determine the average accuracy rate, serving as the basis for the final assessment and guiding the selection of the most suitable network structure.

3.3.1. Activation Function

The activation function is a very important part of the neural network structure, which can nonlinearly transform the input data. In this paper, five typical activation functions, Sigmoid, ReLU, ELU, ReLU6, and Tanh, as well as three novel activation functions, SupEx [36], αSechSig, and αTanhSig [37], are selected for comparison, and the number of iterations is set to be 1000 and the learning rate is set to be 0.01. Figure 8 shows the total time consumed and the average accuracy in the test set for the models with different activation functions selected. of the total elapsed time and the average accuracy of the test set.
In Figure 8, it can be seen that Sigmoid has the lowest accuracy when it is used as an activation function, and the total time consumed by αSechSig and ELU is significantly higher than the other activation functions, and the total time consumed by the rest of the activation functions, even though the total time consumed by SupEx and αTanhSig is less, the total time consumed is not long in the data of the small samples in the first place, so the accuracy is used as the main criterion and the total time consumed is used as the auxiliary criterion, and the model’s classification accuracy for the test set is over 0.97 when Tanh and ReLU6 are used as activation functions, the classification accuracy of the model on the test set exceeds 0.97, reaching 0.9758 and 0.9727, of which Tanh not only has high accuracy but also the total time consumed is less than ReLU6.
In Figure 9, we can intuitively see that Tanh, as the activation function, has the highest average accuracy; the fluctuation range of the experimental results is also the smallest; and most of the experimental results are in the 0.975 or so range. The synthesis of the above two results can be seen in this paper’s model Tanh, as the activation function is the optimal effect.

3.3.2. Number of Attention Heads

Multi-head attention allows the model to focus on different channels of the input at the same time, thus improving the expressiveness and performance of the model. It should be noted that while multi-head attention increases the complexity of the model, it also increases the computational overhead of training and inference, so the optimal number of heads needs to be determined experimentally to achieve the best performance of the model. The activation function selected in the experiment is Tanh, the value range of H is set to [1, 50], and other settings are the same as in the previous experiment. The final comparison of the average accuracy of the test set under different numbers of heads and the experimental results is shown in the following figure.
Based on Figure 10, it is evident that the model achieves higher accuracy within the range of head counts H = 16, H = 22, and H = 30. The highest accuracy is attained at a learning rate of Lr = 0.01 with H = 16. However, it is worth noting that excessively high head counts lead to longer running times and increased computational costs. Therefore, the optimal configuration of the model is Lr = 0.01 and H = 16.

3.3.3. Algorithm Performance Comparison

In order to verify the validity of the proposed model in this paper, GCN, CNN, SVM, and BPNN algorithms are selected in the experiments in this subsection to compare the algorithm performance with the MPGANN model proposed in this paper. The parameters of GCN, such as the neighbor graph construction method, the number of hidden layers, the number of neurons in the hidden layer, and the dropout rate, are all the same as those of MPGANN. CNN has a total of four convolutional layers. The size of the convolution kernel is 5-3-2-2, the stride is 1-1-1-1, and each convolution layer is connected to an average pooling layer after each convolution layer. The window size is 2-2-2-2, padding is 0, and stride is 1. The kernel function of SVM is the Gaussian kernel function; set the penalty factor c = 1.0 and set the gamma to auto. In the BPNN hidden layer, the number of neurons is set to 16, the number of iterations for all algorithms is set to 1000, and the learning rate is set to 0.01.
Accuracy is an overall performance indicator; if it is used as a performance evaluation index for diagnostic algorithms, it will likely mask the specific classification of diagnostic algorithms, so in order to solve this type of problem, this paper introduces precision, recall, and the F1 score, calculated as:
R = T P T P + F N ,
P = T P T P + F P ,
F = 2 1 P + 1 R = 2 R P R + P ,
where R is the recall rate; P is the precision rate; F is the F1 score; T P is the number of samples that classify positive samples as positive samples; F N is the number of positive samples classified as negative samples, and F P is the number of negative samples classified as positive samples.
The precision and recall of each algorithm used in this paper are shown in Table 5. From the table, it can be seen that each algorithm can basically correctly classify the normal working conditions, but the complexity of the engine fault data leads to a relatively poor classification effect for the traditional diagnostic algorithms. In this experiment, the precision rate and recall rate of BPNN are lower than other diagnostic algorithms, and although the precision rate and recall rate of GCN and CNN diagnosis for each condition are better than those of SVM and BPNN, the diagnostic effect is still unsatisfactory in F1 and F2 conditions. The MPGANN proposed in this paper achieves 100% accuracy and recall in all working conditions except for 0.89 accuracy and 0.88 recall in F2 and F1 working conditions, which is an obvious advantage over other diagnostic algorithms and is more in line with practical industrial needs.
From Figure 11, it can be seen that fault F1 is misclassified in all diagnostic algorithms, while after the feature mining of probabilistic graph relationships and rank order graph structure, MPGANN and GCN only have two F1 fault samples misclassified, and the MPGANN proposed in this paper only has two samples misclassified, which shows that the MPGANN model can mine more information between the data to improve the diagnostic algorithm’s classification accuracy and stability and is better than other models.
Figure 12 shows the average accuracy, average precision, average recall, and F1 score of each diagnostic algorithm. The performance of each diagnostic algorithm can be seen more clearly by comparing the figures. CNN, as a network model for deep learning, is widely used in the field of mechanical fault diagnosis and has achieved excellent results. However, due to the fact that there is no large amount of data for model training in this experiment, the training effect of CNN is not good, and it cannot make correct classification judgments. GCN is limited by its own information propagation rules, and even though it adopts the same topology structure for fault diagnosis, the diagnostic effect is still not good. However, the MPGANN proposed in this paper uses the multi-head attention mechanism in the aggregation operation, which makes the feature aggregation more reasonable and increases the learning efficiency of the model, and the average precision, average accuracy, average recall, and F1 score reach 0.978, 0.976, 0.975, and 0.976.

4. Conclusions

In this paper, a new MPGANN model is proposed, and in the experiments on the ship engine dataset, firstly, different graph structure construction methods are compared, and the fused graph structure enhances the expressive ability, especially the fusion of probabilistic and ordinal graph structures, which dramatically improves the diagnostic accuracy of the model. Secondly, the accuracy of the model under different activation functions, head counts, and learning rates is greatly different, which shows that the influence of hyperparameters on the model is also not negligible. Finally, when MPGANN is compared with other algorithms, it outperforms the various methods compared in diagnosis in terms of accuracy, precision, recall, and F1 score. Therefore, the MPGANN proposed in this paper has a certain practical significance in the field of intelligent ship fault diagnosis. In summary, the research for this paper is summarized as follows:
(1)
The two graph structure construction methods in the MPGANN model can effectively obtain probabilistic similarity and ordinal similarity between data samples and transform the data into probabilistic graph structure and ordinal graph structure.
(2)
Early fusion is employed to combine the probability map structure and rank-order map structure with the incorporation of feature weights. This integration process effectively amalgamates information from samples at various scales.
(3)
The multi-head attention mechanism is applied to conduct multi-channel feature screening on the fusion graph structure, extracting feature information with higher relevance to enhance the diagnostic performance of the model.
(4)
The model effect was validated using the ship engine fault dataset, and compared with other models in terms of accuracy, precision, recall, and F1 score, MPGANN was the most effective, with a diagnostic accuracy as high as 97.58% and an F1 score of 97.6%.
The proposed diagnostic model in this paper fuses features from different aspects of the data, mining the information of the dataset at a deeper level, but it also has some limitations, such as: the fused features contain only probabilistic similarity and rank-order similarity; the fusion weights need to be set in advance; and so on.
There are three goals for future work: (1) Physical connections in real space will also imply information, and the actual structural connections of the engine will be transformed into graph structures to mine the correlation information between the samples. (2) Choose more graph fusion methods or assign meanings to edge features so that the model has more information for training to improve its diagnostic accuracy. (3) Introduce the attention mechanism into the fusion process to achieve adaptive allocation of fusion weights.

Author Contributions

Conceptualization, Z.A. and H.C.; methodology, Z.A.; validation, Z.A., H.C. and J.W.; formal analysis, H.C.; investigation, Z.C. and J.W.; resources, H.C.; data curation, Z.A., L.W. and K.J.; writing—original draft preparation, Z.A.; writing—review and editing, Z.A., H.C. and L.W.; visualization, Z.A. and Z.C.; supervision, H.C.; project administration, H.C.; funding acquisition, H.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the project: Development of Ship Operation Condition Monitoring and Simulation Platform (funding number: 1638882993269) and the research and application of the Smart Ship Digital Twin Information Platform (fundingt number: 2022JH1/10800097).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

Author Hui Cao was employed by Dalian Maritime University Smart Ship Limited Company. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Zhong, G.Q.; Wang, H.Y.; Zhang, K.Y.; Jia, B.Z. Fault diagnosis of Marine diesel engine based on deep belief network. In Proceedings of the 2019 Chinese Automation Congress (CAC), Hangzhou, China, 22–24 November 2019; pp. 3415–3419. [Google Scholar] [CrossRef]
  2. Hou, L.; Zhang, J.; Du, B. A Fault Diagnosis Model of Marine Diesel Engine Fuel Oil Supply System Using PCA and Optimized SVM. J. Phys. Conf. Ser. 2020, 1576, 012045. [Google Scholar] [CrossRef]
  3. Zhong, K.; Li, J.B.; Wang, J.; Han, M. Fault Detection for Marine Diesel Engine Using Semi-supervised Principal Component Analysis. In Proceedings of the 9th International Conference on Information Science and Technology (ICIST), Hulunbuir, China, 2–5 August 2019; pp. 146–151. [Google Scholar]
  4. Ren, D.P.; Zeng, H.; Wang, X.L.; Pang, S.; Wang, J.D. Fault Diagnosis of Diesel Engine Lubrication System Based on Bayesian Network. In Proceedings of the 6th International Conference on Control, Automation and Robotics (ICCAR), Electr Network, Singapore, 20–23 April 2020; pp. 423–429. [Google Scholar]
  5. Han, P.H.; Li, G.Y.; Skulstad, R.; Skjong, S.; Zhang, H.X. A Deep Learning Approach to Detect and Isolate Thruster Failures for Dynamically Positioned Vessels Using Motion Data. IEEE Trans. Instrum. Meas. 2021, 70, 1–11. [Google Scholar] [CrossRef]
  6. Xu, X.J.; Zhao, Z.Z.; Xu, X.B.; Yang, J.B.; Chang, L.L.; Yan, X.P.; Wang, G.D. Machine learning-based wear fault diagnosis for marine diesel engine by fusing multiple data-driven models. Knowl. Based Syst. 2020, 190, 105324. [Google Scholar] [CrossRef]
  7. Cao, H.; Zhang, J.D.; Cao, X.; Li, R.; Wang, Y.R. Optimized SVM-Driven Multi-Class Approach by Improved ABC to Estimating Ship Systems State. IEEE Access 2020, 8, 206719–206733. [Google Scholar] [CrossRef]
  8. Wang, R.H.; Chen, H.; Guan, C. Random convolutional neural network structure: An intelligent health monitoring scheme for diesel engines. Measurement 2021, 171, 108786. [Google Scholar] [CrossRef]
  9. Memis, S.; Arslan, B.; Aydin, T.; Enginoglu, S.; Camci, C. Distance and Similarity Measures of Intuitionistic Fuzzy Parameterized Intuitionistic Fuzzy Soft Matrices and Their Applications to Data Classification in Supervised Learning. Axioms 2023, 12, 463. [Google Scholar] [CrossRef]
  10. Choo, K.B.; Cho, H.; Park, J.H.; Huang, J.F.; Jung, D.; Lee, J.; Jeong, S.K.; Yoon, J.; Choo, J.; Choi, H.S. A Research on Fault Diagnosis of a USV Thruster Based on PCA and Entropy. Appl. Sci. 2023, 13, 3344. [Google Scholar] [CrossRef]
  11. Zhao, N.Y.; Mao, Z.W.; Wei, D.H.; Zhao, H.P.; Zhang, J.J.; Jiang, Z.N. Fault Diagnosis of Diesel Engine Valve Clearance Based on Variational Mode Decomposition and Random Forest. Appl. Sci. 2020, 10, 1124. [Google Scholar] [CrossRef]
  12. Memis, S. Picture Fuzzy Soft Matrices and Application of Their Distance Measures to Supervised Learning Picture Fuzzy Soft k-Nearest Neighbor (PFS-kNN). Electronics 2023, 12, 4129. [Google Scholar] [CrossRef]
  13. Yang, K.; Fan, H.Y. Research on Fault Diagnosis Method of Diesel Engine Thermal Power Conversion Process. In Proceedings of the 3rd International Forum on Energy, Environment Science and Materials (IFEESM), Shenzhen, China, 25–26 November 2017; pp. 1445–1450. [Google Scholar]
  14. Agrawal, S.K.; Banerjee, S.; Sinha, A.; Das, D. SafeEngine: Fault Detection with Severity Prediction for Diesel Engine. In Proceedings of the 2022 IEEE 10th Region 10 Humanitarian Technology Conference (R10-HTC), Hyderabad, India, 16–18 September 2022; pp. 216–220. [Google Scholar]
  15. Zhou, J.; Cui, G.; Hu, S.; Zhang, Z.; Yang, C.; Liu, Z.; Wang, L.; Li, C.; Sun, M. Graph neural networks: A review of methods and applications. AI Open 2020, 1, 57–81. [Google Scholar] [CrossRef]
  16. Nie, J.; Xu, Y.; Huang, Y.; Li, J. The review of image processing based on graph neural network. In Proceedings of the International Conference on Intelligent Robotics and Applications, Yantai, China, 22–25 October 2021; pp. 534–544. [Google Scholar]
  17. Zhou, A.Z.; Li, Y.F. Structural attention network for graph. Appl. Intell. 2021, 51, 6255–6264. [Google Scholar] [CrossRef]
  18. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  19. Wang, R.; Chen, H.; Guan, C. DPGCN Model: A Novel Fault Diagnosis Method for Marine Diesel Engines Based on Imbalanced Datasets. IEEE Trans. Instrum. Meas. 2022, 72, 1–11. [Google Scholar] [CrossRef]
  20. Li, T.; Zhao, Z.; Sun, C.; Yan, R.; Chen, X. Domain adversarial graph convolutional network for fault diagnosis under variable working conditions. IEEE Trans. Instrum. Meas. 2021, 70, 1–10. [Google Scholar] [CrossRef]
  21. Zheng, J.; Gao, Q.; Lü, Y. Quantum graph convolutional neural networks. In Proceedings of the 2021 40th Chinese Control Conference (CCC), Shanghai, China, 26–28 July 2021; pp. 6335–6340. [Google Scholar]
  22. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
  23. Yang, F.; Zhang, H.Y.; Tao, S.M. Semi-supervised classification via full-graph attention neural networks. Neurocomputing 2022, 476, 63–74. [Google Scholar] [CrossRef]
  24. Zhang, S.; Gong, H.; She, L. An aspect sentiment classification model for graph attention networks incorporating syntactic, semantic, and knowledge. Knowl. Based Syst. 2023, 275, 110662. [Google Scholar] [CrossRef]
  25. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  26. Zhang, W.-Y.; Wei, Z.-W.; Wang, B.-H.; Han, X.-P. Measuring mixing patterns in complex networks by Spearman rank correlation coefficient. Phys. A Stat. Mech. Appl. 2016, 451, 440–450. [Google Scholar] [CrossRef]
  27. Wang, Y.; Qin, Y.; Guo, J.; Cao, Z.; Jia, L. Multi-point short-term prediction of station passenger flow based on temporal multi-graph convolutional network. Phys. A Stat. Mech. Appl. 2022, 604, 127959. [Google Scholar] [CrossRef]
  28. Wang, L.; Hu, Z.; Kong, Q.; Qi, Q.; Liao, Q. Infrared and Visible Image Fusion via Attention-Based Adaptive Feature Fusion. Entropy 2023, 25, 407. [Google Scholar] [CrossRef] [PubMed]
  29. Luo, Z.; Li, J.; Zhu, Y. A deep feature fusion network based on multiple attention mechanisms for joint iris-periocular biometric recognition. IEEE Signal Process. Lett. 2021, 28, 1060–1064. [Google Scholar] [CrossRef]
  30. Yang, J.; Yang, J.-Y.; Zhang, D.; Lu, J.-F. Feature fusion: Parallel strategy vs. serial strategy. Pattern Recognit. 2003, 36, 1369–1381. [Google Scholar] [CrossRef]
  31. Wang, Y.; Xu, X.; Yu, W.; Xu, R.; Cao, Z.; Shen, H.T. Comprehensive Framework of Early and Late Fusion for Image–Sentence Retrieval. IEEE MultiMedia 2022, 29, 38–47. [Google Scholar] [CrossRef]
  32. Boulahia, S.Y.; Amamra, A.; Madi, M.R.; Daikh, S. Early, intermediate and late fusion strategies for robust deep learning-based multimodal action recognition. Mach. Vis. Appl. 2021, 32, 121. [Google Scholar] [CrossRef]
  33. Ruiz, L.; Gama, F.; Ribeiro, A. Graph neural networks: Architectures, stability, and transferability. Proc. IEEE 2021, 109, 660–682. [Google Scholar] [CrossRef]
  34. Brauwers, G.; Frasincar, F. A general survey on attention mechanisms in deep learning. IEEE Trans. Knowl. Data Eng. 2021, 35, 3279–3298. [Google Scholar] [CrossRef]
  35. Weston, J.; Chopra, S.; Bordes, A. Memory networks. arXiv 2014, arXiv:1410.3916. [Google Scholar]
  36. Kiliçarslan, S.; Közkurt, C.; Bas, S.; Elen, A. Detection and classification of pneumonia using novel Superior Exponential (SupEx) activation function in convolutional neural networks. Expert Syst. Appl. 2023, 217, 119503. [Google Scholar] [CrossRef]
  37. Közkurt, C.; Kiliçarslan, S.; Bas, S.; Elen, A. αSechSig and αTanhSig: Two novel non-monotonic activation functions. Soft Comput. 2023, 27, 18451–18467. [Google Scholar] [CrossRef]
Figure 1. Weighting factors of the calculation process.
Figure 1. Weighting factors of the calculation process.
Applsci 13 12421 g001
Figure 2. Structure of the MPGANN model.
Figure 2. Structure of the MPGANN model.
Applsci 13 12421 g002
Figure 3. Algorithm flow chart.
Figure 3. Algorithm flow chart.
Applsci 13 12421 g003
Figure 4. Simulation model of the ship’s engine.
Figure 4. Simulation model of the ship’s engine.
Applsci 13 12421 g004
Figure 5. Test accuracy for each graph structure and total elapsed time.
Figure 5. Test accuracy for each graph structure and total elapsed time.
Applsci 13 12421 g005
Figure 6. Fusion of probabilistic graph structures with rank-order graph structures.
Figure 6. Fusion of probabilistic graph structures with rank-order graph structures.
Applsci 13 12421 g006
Figure 7. Graph learning process.
Figure 7. Graph learning process.
Applsci 13 12421 g007
Figure 8. Average accuracy and total elapsed time for different activation functions.
Figure 8. Average accuracy and total elapsed time for different activation functions.
Applsci 13 12421 g008
Figure 9. Boxplot of the accuracy of different activation functions.
Figure 9. Boxplot of the accuracy of different activation functions.
Applsci 13 12421 g009
Figure 10. Heat map of model accuracy at different learning rates and head counts.
Figure 10. Heat map of model accuracy at different learning rates and head counts.
Applsci 13 12421 g010
Figure 11. Categorization of test set failures: (a) MPGAT, (b) GCN, (c) CNN, (d) SVM, and (e) BPNN.
Figure 11. Categorization of test set failures: (a) MPGAT, (b) GCN, (c) CNN, (d) SVM, and (e) BPNN.
Applsci 13 12421 g011aApplsci 13 12421 g011b
Figure 12. Algorithm performance comparison.
Figure 12. Algorithm performance comparison.
Applsci 13 12421 g012
Table 1. Comparison results between model simulation data and measured data.
Table 1. Comparison results between model simulation data and measured data.
Working ConditionFormPower/kWBooster Outlet Pressure/BarAir Cooler Outlet Temperature/°CFuel Consumption/
g/Cycle
100% loadsimulation value4104.994.673246.64.14
measured value40784.528464.12
inaccuracies/%0.66181.18151.30430.4854
75% loadsimulation value3393.193.206742.93.17
measured value33753.74423.18
inaccuracies/%0.5390−2.14712.143−0.3145
50% loadsimulation value2295.032.301836.72.23
measured value22502.57382.24
inaccuracies/%2.001−3.4319−3.4211−0.4464
25% loadsimulation value1164.881.27728.41.21
measured value11251.42291.26
inaccuracies/%3.5449−3.0282−2.0689−3.9683
Working ConditionFormInlet Flow/g/sAverage Effective Pressure/barExhaust Temperature/°C
100% loadsimulation value8370.982320.0947455.3
measured value830019.96460
inaccuracies/%0.85520.6748−0.1022
75% loadsimulation value6477.096617.4753438.75
measured value680016.52445
inaccuracies/%−3.27803.9667−1.4045
50% loadsimulation value4249.762111.4935437.8
measured value460011.01430
inaccuracies/%−3.2663.45181.8140
25% loadsimulation value2371.39595.7910427.7
measured value26005.51410
inaccuracies/%−3.98483.59573.5593
Table 2. Engine failure simulation scheme.
Table 2. Engine failure simulation scheme.
Malfunction CodeParameterization
Failure CharacteristicsRegular ValueDegree of Failure
LV1LV2LV3LV4
F1Injection timing advance−13.1 deg−14.1−15.1−16.1−17.1
F2Delayed injection timing−13.1 deg−12.1−11.1−10.1−9.1
F3Decline in supercharger efficiency100%95%90%85%80%
F4Reduced air cooler efficiency100%95%90%85%80%
Table 3. Engine failure dataset.
Table 3. Engine failure dataset.
Working ConditionDimensionSample Size
Injection timing advance (F1)2380
Delayed injection timing (F2)2380
Decline in supercharger efficiency (F3)2380
Reduced air cooler efficiency (F4)2380
Normal operation (F5)2380
Table 4. Detection indicator indexes.
Table 4. Detection indicator indexes.
NotationMonitoring IndicatorsUnitNotationMonitoring IndicatorsUnit
Pmp3Booster outlet pressurebarTmp23Exhaust manifold temperature°C
Vmp3Supercharger outlet flowm/sPmp24Turbine outlet pressurebar
Tmp3Supercharger inlet temperature°CVmp24Turbine Outlet Flowm/s
Pmp4Air cooler inlet pressurebarTmp24Turbine outlet temperature°C
Vmp4Air cooler outlet flowm/sFAir intakekg/s
Tmp4Air cooler outlet temperature°CgFuel consumption rateg/kW·h
Pmp5Cylinder inlet pressurebarPzMaximum burst pressurebar
Vmp5Cylinder inlet flowm/sλMaximum voltage risebar/deg
Tmp5Cylinder inlet temperature°CNePower (output)kW
Tmp14Cylinder exhaust temperature°CPIIMEPbar
Pmp23Exhaust manifold pressurebarPBBMEPbar
Vmp23Exhaust manifold flowm/s///
Table 5. Precision and recall for each algorithm.
Table 5. Precision and recall for each algorithm.
AlgorithmsPrecisionRecall
F1F2F3F4F5F1F2F3F4F5
MPGANN10.891110.881111
GCN0.880.8840.94110.880.9410.940.94
CNN0.80.781110.750.88110.94
SVM0.710.761110.750.810.9410.94
BPNN0.650.720.880.8810.690.810.880.880.81
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ai, Z.; Cao, H.; Wang, J.; Cui, Z.; Wang, L.; Jiang, K. Research Method for Ship Engine Fault Diagnosis Based on Multi-Head Graph Attention Feature Fusion. Appl. Sci. 2023, 13, 12421. https://doi.org/10.3390/app132212421

AMA Style

Ai Z, Cao H, Wang J, Cui Z, Wang L, Jiang K. Research Method for Ship Engine Fault Diagnosis Based on Multi-Head Graph Attention Feature Fusion. Applied Sciences. 2023; 13(22):12421. https://doi.org/10.3390/app132212421

Chicago/Turabian Style

Ai, Zeren, Hui Cao, Jihui Wang, Zhichao Cui, Longde Wang, and Kuo Jiang. 2023. "Research Method for Ship Engine Fault Diagnosis Based on Multi-Head Graph Attention Feature Fusion" Applied Sciences 13, no. 22: 12421. https://doi.org/10.3390/app132212421

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop