Next Article in Journal
Rejection of Synchronous Vibrations of AMB System Using Nonlinear Adaptive Control Algorithm with a Novel Frequency Estimator
Previous Article in Journal
An LWPR-Based Method for Intelligent Lower-Limb Prosthesis Control by Learning the Dynamic Model in Real Time
Previous Article in Special Issue
Linear Quadratic Optimal Control with the Finite State for Suspension System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sound-Based Intelligent Detection of FOD in the Final Assembly of Rocket Tanks

1
Key Laboratory of Education Ministry for Modern Design & Rotor-Bearing System, Xi’an Jiaotong University, Xi’an 710049, China
2
Shanghai Aerospace Equipments Manufacturer Co., Ltd., Shanghai 200245, China
*
Author to whom correspondence should be addressed.
Machines 2023, 11(2), 187; https://doi.org/10.3390/machines11020187
Submission received: 13 December 2022 / Revised: 13 January 2023 / Accepted: 17 January 2023 / Published: 31 January 2023

Abstract

:
The traditional method of relying on human hearing to detect foreign object debris (FOD) events during rocket tank assembly processes has the limitation of strong reliance on humans and difficulty in establishing objective detection records. This can lead to undetected FOD entering the engine with the fuel and causing major launch accidents. In this study, we developed an automatic, intelligent FOD detection system for rocket tanks based on sound signals to overcome the drawbacks of manual detection, enabling us to take action to prevent accidents in advance. First, we used log-Mel transformation to reduce the high sampling rate of the sound signal. Furthermore, we proposed a multiscale convolution and temporal convolutional network (MS-CTCN) to overcome the challenges of multi-scale temporal feature extraction to detect suspicious FOD events. Finally, we used the proposed post-processing strategies of label smoothing and threshold discrimination to refine the results of FOD event detection and ultimately determine the presence of FOD. The proposed method was validated through FOD experiments. The results showed that the method had an accuracy rate of 99.16% in detecting FOD and had a better potential to prevent accidents compared to the baseline method.

1. Introduction

The tank is a crucial component of a rocket, serving as a storage unit for cryogenic propellants and providing structural support and protection for the rocket [1]. It is typically composed of multiple tanks that are joined together through a docking process, which includes drilling, riveting, and bolting. However, this process may inadvertently result in the presence of Foreign Object Debris (FOD) such as small bolts, nuts, screws, titanium wire, and other metal remnants inside the tank. FOD is a major contributor to launch accidents of space products, as they can enter the engine along with the propellant during launch if not detected and removed in a timely manner, leading to serious accidents. Thus, the development of FOD detection technology for rocket tanks is vital for ensuring the reliability of aerospace engineering.
The commonly used Particle Impact Noise Detection (PIND) method [2,3,4,5,6,7,8] in the field of FOD in the aerospace industry requires the manufacture of a large excitation platform to be effective in detecting FOD in large rocket tanks, which is not practical. The methods based on machine vision [9,10], on the other hand, require cameras to enter the complex internal structures of the tanks and are susceptible to being affected by factors such as occlusion and lighting, which can also cause secondary damage to the tanks. Therefore, the most common method used for detecting FOD in rocket tanks is the rotation listening method, which involves rotating the tank slowly while it is being inspected and listening for the sound of metal FOD sliding against or colliding with the tank’s walls. The presence of FOD is determined by human experience, but this method has some drawbacks, including low detection efficiency, high workload, and high dependence on human judgment. Additionally, it is difficult to form data records of detection and quantitative evaluation results. Because of these drawbacks, there is a need for a more efficient, accurate, and automatic method for detecting FOD in rocket tanks.
Nowadays, acoustic-based condition monitoring methods using microphones have gained popularity in special applications due to their non-destructive and non-contact advantages [11], such as in trackside acoustic diagnostic systems [12], arc magnet internal defect detection [13], and milling process monitoring [14]. These systems offer a new approach for detecting FOD in rocket tanks by using microphone sensors to acquire acoustic signals and applying suitable signal processing and FOD identification algorithms for intelligent, quantitative assessment of FOD status. However, in the field of acoustic-based diagnosis, it is common to intercept a small segment of the signal for processing, which is suitable for high-speed rotating machinery such as gears [15,16], bearings [17,18] and motors [19,20]. This approach may not be effective for FOD detection in tanks, as FOD can cause random scratching or collision events with the slowly rotating (1-2 r/min) inner walls of the tank, and the response time for FOD events can range from a few tens of milliseconds to a few seconds. Therefore, sound-based FOD detection must be able to detect suspicious FOD sound events of varying lengths within the entire input signal and determine the presence of an FOD event through post-processing criteria.
For the task of long-signal sound event detection, there are two main types of methods: traditional methods and intelligent methods. Traditional methods, such as those based on Gaussian Mixture Models–Hidden Markov Models (GMM-HMM) [21] or Non-negative Matrix Factorization (NMF) [22], rely on manual feature design and are dependent on user expertise. In contrast, data-driven deep learning models, such as those based on Convolutional Recurrent Neural Networks (CRNN) [23,24,25], are increasingly popular due to their ability to perform automatic feature extraction. These models utilize recurrent neural network structures, such as Long Short-Term Memory (LSTM) [26] and Gate Recurrent Units (GRUs) [27], to model the time sequence characteristics of sound events. However, some studies have demonstrated that recurrent neural network structures have limitations in modeling long-term dependencies [28]. In contrast, Temporal Convolutional Networks (TCNs) are able to extract features from both short-term and long-term time sequences by expanding the receptive field through multiple layers of dilated convolutions with different dilation rates and have been shown to be effective in tasks such as sound event detection [29] and localization [30].
Unfortunately, TCN-based methods still have the following limitations for FOD detection tasks. First, locally, the frictional sound generated by the FOD in the tank consists of many short-weak shocks with random intervals. The duration of these shocks and the time interval between shocks can be long or short and need to be analyzed from different scales. However, traditional TCNs use only one size of convolution kernel in dilated convolution, and a single size of convolution kernel does not capture the local multi-scale features [15]. Secondly, the TCN-based event detection method only makes a simple threshold judgment on the prediction result of each frame, and when the prediction probability is greater than a set value, a specific event is considered to have occurred in that frame. This will lead to unstable detection results for the FOD events consisting of intermittent short-weak shocks.
To solve the above problems, the MS-CTCN method is proposed in this paper for sound-based FOD detection. The main contributions of this paper are as follows:
(1) A new method for detecting FOD in rocket tanks using sound signals and deep neural networks is proposed, addressing the deficiencies of traditional methods.
(2) The Multi-Kernel Size (MKS) convolution is introduced in TCN-based temporal feature extraction for local feature fusion and solving FOD multi-scale short weak shock extraction.
(3) A post-processing strategy with label smoothing and decision-making based on FOD response frames is proposed for stable detection.
(4) A simulation tank for FOD detection experiments is constructed to test the proposed method’s effectiveness and performance using various FOD sound signals.
The remainder of this paper is organized as follows: in Section 2, we review relevant previous research. The fundamental theories utilized in this study are outlined in Section 3. The proposed method is described in detail in Section 4. The experimental setup is presented in Section 5. The results and discussion are provided in Section 6. Finally, we provide our conclusions in Section 7.

2. Literature Review

2.1. FOD Detection in Rocket Tanks

PIND [2,3,4,5,6,7,8] is considered the most well-established method in the field of FOD. This method utilizes a specialized excitation platform to generate mechanical impacts that cause FOD to collide with the inner walls of the object being inspected. The collision signals are then captured using acceleration sensors or acoustic emission sensors and analyzed using signal processing techniques to identify the presence of FOD [7,8]. Additionally, researchers have developed FOD detection methods using machine vision technology [9,10], which use specialized cameras to scan the interior of the object and compare images to determine the presence of FOD. However, these methods are not suitable for detecting FOD in rocket tanks due to their complex structures. The PIND method is limited in its application to larger objects, and the excitation process can potentially cause damage to the tanks. Machine-vision-based methods, on the other hand, have the risk of introducing additional FOD and are affected by factors such as lighting and obstruction. As a result, the current method for detecting FOD in rocket tanks during final assembly involves slowly rotating the tank under the motor’s drive and listening for the acoustic signals produced by the FOD colliding or slipping against the walls. However, this method is labor-intensive, relies heavily on human experience, and produces limited data records and quantitative results.
As hardware and software technologies have advanced, microphone-based acoustic state detection technology for machinery has been widely studied [12,13,14,31]. It has the advantage of non-destructive and non-contact [11]. Therefore, conducting research on FOD detection in rocket tanks based on sound signals has promising potentials. However, these methods for rotating machinery deal with periodic signals with fast rotational speeds and are not applicable to rocket storage tank sound signals with slow rotational speeds and random non-stationary FOD events.

2.2. Sound Event Detection

Sound event detection refers to the process of detecting and identifying specific sound events within a given recording or live audio stream. Currently, deep learning-based [32,33] sound event detection techniques have the advantage of automatically extracting features and stronger fitting capabilities compared to traditional machine learning methods based on GMM-HMM [21] or NMF [22], which makes them widely used. Convolutional neural networks (CNNs) are particularly effective in extracting local features due to their local connections and weight sharing, which is important for many sound event detection tasks [34,35]. However, CNNs are not suitable for capturing long-term temporal dependencies in signals, which leads to poor performance when detecting events with long time spans. To address this limitation, some methods combine CNN with other types of models, such as RNN [26,27] or Transformers [36], to capture both local and long-term temporal features. CRNNs [23,24,25,37] use CNNs as the feature extractors and feed the extracted features into RNNs, which are able to model temporal dependencies in the input signal. The CNN–Transformer uses a self-attention mechanism to model the temporal relationship among features [38]. The performance of the CNN–Transformer is comparable with CRNN, and it has the advantage of being more computationally efficient due to its parallel computation nature [39]. However, Transformer models have been shown to have a weaker inductive bias when compared to some other models such as CNNs [40]. This means that they may perform poorly when there is are a limited number of data. Another variant of CNN, known as TCN [41], is another option for capturing short- and long-term temporal dependencies in the input signals while maintaining a relatively lower computational complexity. TCN uses multiple layers of convolutional layers, each with a different dilation factor, to expand the receptive field of the network. By using dilated convolutions, TCN can capture long-term dependencies in the input signal without the need for recurrent connections or self-attention mechanisms, which makes it computationally more efficient than RNNs or Transformers while still able to effectively extract temporal features in the input signal. It is considered to be a simple yet powerful alternative to complex architectures, and its performance on sound event detection [29,42] and localization [30] has been proven.
However, the utilization of TCN for FOD detection is associated with certain limitations. The FOD response is composed of a series of brief and weak shocks with unpredictable intervals and duration events, which can vary in length. As a result, utilizing a single kernel-sized TCN may not effectively capture the multi-scale properties [15] in the temporal sequence. Furthermore, the current event detection methods rely solely on simple thresholding to make final decisions, which may not be suitable for detecting weaker FOD events.

3. Theoretical Background

3.1. Logarithmic Mel Transform

The sound signal of the rotating tank is characterized by (a) time-varying non-stationary and (b) high sampling rate. To address these characteristics, this study incorporates the logarithmic Mel transform [30,43], which is a method that is based on the Short-Time Fourier Transform (STFT) [44] and can capture the time-frequency information of time-varying signals. The Mel transform also reduces the dimensionality of the STFT features of high sampling rate signals while preserving useful information [45]. Additionally, the Mel transform simulates the varying sensitivity of the human ear [45] to different frequencies of sound, thereby introducing further a priori knowledge. The process of logarithmic Mel time-frequency transformation is as follows.
STFT operation. A window function of static length is used to intercept a very short part (frame) of the total time-varying signal, and the discrete Fourier transform is used to obtain the local spectrum of each frame, as expressed in Equation (1).
S ( t , k ) = n = 1 N f x t [ n ] · w [ n ] · e j 2 π k n N f
where S ( t , k ) R T × N f represents the k-th discrete frequency component of the short-time spectrum at time t; t = 1 , 2 , , T and k = 1 , 2 , , N f denote the frame number and discrete frequency number, respectively; and x t [ n ] and w [ n ] refer to the t-th framed signal and window function, respectively.
Mel frequency filter banks construction. The main parameters for constructing Mel filter banks include (a) number of Mel filters, F, (b) minimum frequency in Hz, f Hz m i n , and (c) maximum frequency in Hz, f Hz m a x . For the most general case, f Hz m i n is equal to zero and f Hz m a x is equal to half the sampling frequency. Then, Mel-scale minimum frequency and maximum frequency f Mel m a x and f Mel m i n can be computed using Equation (2). After that, Mel frequency filters can be constructed as expressed in Equations (3)–(5).
f Mel = 2595 · log 10 ( 1 + f Hz / 700 )
H ( k , m ) = 0 , f Hz ( k ) < f Hz c ( m 1 ) f Hz ( k ) f Hz c ( m 1 ) f Hz c ( m ) f Hz c ( m 1 ) , f Hz c ( m 1 ) f Hz ( k ) f Hz c ( m ) f Hz ( k ) f Hz c ( m + 1 ) f Hz c ( m ) f Hz c ( m + 1 ) , f Hz c ( m ) f Hz ( k ) f Hz c ( m + 1 ) 0 , f Hz ( k ) f Hz c ( m + 1 )
f Hz c ( m ) = 700 ( 10 f Mel c ( m ) / 2595 1 )
f Mel c ( m ) = m · ( f Mel m a x f Mel m i n ) / F
where H ( k , m ) R N f × F denotes the gain at the k-th Hz-scale frequency f Hz ( k ) in the m-th Mel filter; f Mel c and f Hz c denotes the center frequency of the filter in the Hz-scale and Mel-scale, respectively; and m = 1 , 2 , , F denotes the Mel filter number.
Mel filter bank features extraction. The constructed Mel filter banks H are used to extract features from the original short-time spectrum S . Eventually, the two-dimensional MFB features X ( t , m ) are obtained according to Equation (6), which includes the subsequent absolute and logarithmic operations.
X ( t , m ) = 20 × log 10 ( | S ( t , k ) | · H ( k , m ) ) R T × F

3.2. Convolutional Neural Network (CNN)

CNNs are particularly adept at extracting features from two-dimensional (2D) data, such as images. Compared to traditional Artificial Neural Networks (ANNs), CNNs have a smaller number of trainable parameters and have been widely and successfully applied in various fields [46,47], including the field of sound-based intelligent equipment maintenance [48,49], where it has demonstrated impressive results in various studies. The fundamental operations in a convolutional layer include convolution, activation, batch normalization, and max pooling.
The convolution operation with the most used Rectified Linear Unit (ReLU) activation function is described in Equations (7) and (8).
y a = Re LU ( W cn x cn + b cn )
ReLU ( x ) = max { x , 0 }
where W cn , b cn , and ⊗ denote weight, bias, and convolution operation, respectively; x cn R W × H refers to input feature map; and y a denotes feature map after convolution and activation.
Batch Normalizing (BN) transform is described in Equations (9) and (10). The purpose of using BN is to reduce internal covariate shift to better train the network [50].
y a ^ = y a μ σ 2 + ε
y bn = γ y a ^ + β
where μ and σ are the expectation and the variance, y bn represents the output features over a mini-batch, γ and β are two parameters to be learned, and ϵ is a constant close to zero.
Maximum pooling is a sample-based discretization process. The objective is to down-sample an input representation (image, hidden-layer output matrix, etc.), reducing its dimensionality and enabling features that are more robust after convolution [51]. The mathematical description is given as follows:
y mp ( i ) = max { y bn ( i : i + r 1 ) }
where y mp ( i ) denotes the maximum value in the i-th corresponding pooling region and r is the pool size.

3.3. Temporal Convolutional Network (TCN)

Unlike 2D-CNN, the processing object of TCN is 1D time sequences. The basic TCN consists of 1D dilated convolutions with different dilation rates. As the dilation rate increases, the receptive field of the convolution kernel increases accordingly [25]. By concatenating multiple layers of dilated convolutions with gradually increasing dilated rates, the network can learn the time sequence correlation of sound signals that expand from local to global. As illustrated in Figure 1, a three-layer structure with a kernel size of 3 and a dilated convolution with dilation rates of 1, 2, and 4 is given, and the receptive field changes from an initial 3 to 9. The output in the i-th layer at moment t with an odd kernel size is described in Equation (12).
x i + 1 ( t ) = u = ξ ξ w i ( u ) · x i ( t + u · d i ) + b i ( u )
where ξ = int ( k i / 2 ) , w i ( u ) and b i ( u ) are the u-th element’s weight and bias of convolution kernel in the i-th layer.
For the i-th layer of dilated convolution, the output length T i + 1 is shown in Equation (13).
T i + 1 = T i + 2 p i d i ( k i 1 ) 1 s i + 1
where p i , d i , k i , and s i denote padding size, dilated rate, kernel size, and stride of dilated convolution. To make the input and output sequence equal in length, specifically, T i + 1 = T i = T , the boundary padding p i = 1 / 2 [ s i ( T 1 ) T + d i ( k i 1 ) + 1 ] .

4. Proposed Method

During FOD detection process of rocket tanks, the tank is rotated by a motor, causing any FOD inside the tank to be lifted along the inner wall. When it reaches a certain height, the FOD falls off the wall due to gravity and generates an abnormal sound from impacts and friction. These mechanical sound waves are transmitted to the outside of the tank and captured by microphones placed externally. The subsequent detection system processes the sound signals acquired by the microphones and automatically determines the presence of FOD. By gradually moving the position of the microphone along the axial direction of the rocket tank, the FOD detection operation can be completed for the entire tank. The proposed method, the MS-CTCN-based FOD detection system, is the core of this process. As illustrated in Figure 2, it includes five sequential steps: (1) preprocessing, (2) CNN block, (3) MS-TCN block, (4) classification block, and (5) post-processing.

4.1. Preprocessing

The utilization of high sampling rate sound signals as input directly necessitates deep and extensive network models, which incurs significant computational costs. Therefore, a signal-processing method with a preliminary feature extraction function and dimensionality reduction capability is required. Additionally, the time-varying characteristics of the FOD signal must be considered. As outlined in Section 3.1, the problem of time-varying feature conversion of FOD signals is addressed by first applying the STFT to the raw time-domain high sampling rate signals, followed by extracting the MFB features, which effectively reduces the dimensionality while preserving relevant information.
For convenience, the continuous sound signal is divided into time frames with a duration of 2048 sampling points at a 48k sampling rate (42.7 ms), and the frame overlap rate is 50%. The log-Mel transform with the feature dimension F = 128 is applied to each frame to extract the MFB features x t F 0 , which leads to the input sequence X 1 = [ x 1 F 1 , x 2 F 1 , , x T F 1 ] R T × F 1 for the subsequent network, where T and F 1 present the length of the time sequence and the dimension of the MFB features, respectively.

4.2. CNN Block

The MFB feature is actually a 2D image with the location of each pixel point in the length and height directions in time and frequency, respectively. The lack of a feature extractor to further generalize the MFB into a feature form that the network can appreciate will make it difficult to develop prediction results. In image recognition, 2D CNNs, with less parameters and better feature extraction capability than fully connected networks, are frequently used to automatically comprehend 2D image information and further abstract the local translation-invariant [52] features in images. The 2D convolution operation described in Section 3.2 is also introduced before the TCN in the detection system. The purpose of the CNN block is to extract frequency invariant features from the time sequence X 1 obtained in the previous step.
However, in pooling operation, unlike the traditional 2D max-pooling, which has the same pooling size in both dimensional directions, the max-pooling is only applied in the frequency direction [53], specifically, the height direction of the feature map. Its purpose is to ensure that the time resolution remains unchanged when using TCN to extract temporal features.
After a series of convolution blocks, the obtained multi-channel feature maps are stacked in the frequency direction. One can thereby obtain the single-channel feature map X 2 = [ x 1 F 2 , x 2 F 2 , . . . , x T F 2 ] R T × F 2 with unchanged sequence length T, where F 2 represents the feature dimension after stacking. If the number of channels of the last layer of convolution operation is C c n , the height of the output feature map is H cn , and F 2 = C cn · H cn .

4.3. MS-TCN Block

The signal of FOD slipping from the tank is characterized by short and weak shocks, and the interval between these short and weak shocks can be long or short. The duration of the FOD event consisting of a series of short and weak shocks can also be long or short. To prevent missed detection, the local response of the FOD signal should be extracted and the long temporal features of the FOD need to be concerned. Therefore, a TCN with dilated convolution is developed to increase the receptive field by increasing the cavity rate layer-by-layer for obtaining a wider range of long-time features. Meanwhile, MKS is joined to the dilated convolution operation to accommodate the short and weak shock features with different time intervals. Finally, the features extracted from different layers are fused to achieve the expected multi-scale feature fusion and enhance the feature expression level.
In the proposed multi-scale temporal convolution block as shown in Figure 3, three dilated convolution layers with different dilation rates are used to obtain global receptive fields at different scales. In each layer, an MKS dilation convolution operation is used for considering local features at different scales, and multi-scale feature fusion is performed on this basis.
The MKS dilated convolution operation in layer performs convolution and feature fusion as described in Equations (14) and (15).
x out ( t ) = j = 1 3 u = ξ j ξ j w i j ( u ) · x in ( t + u · d i ) + b i j ( u )
ξ j = int ( k i j / 2 )
where k i j , w i j ( u ) , and b i j ( u ) denote the kernel size the u-th weight, and the bias of j-th kernel in MKS dilated convolution, respectively, and x in ( t ) and x out ( t ) denote input and output feature at the time, respectively. In the proposed method, we set k i 1 = 3 , k i 2 = 5 and k i 3 = 7 in the MKS-dilated convolution, as presented in Figure 3.
The activation function f a ( x ) after dilated convolution and feature fusion is combined with the tanh function and the sigmoid function, as expressed in Equations (16)–(18). This activation mode is better than the ReLU function in the time sequence processing of the sound signal [54].
tanh ( x ) = e x e x e x + e x
σ ( x ) = 1 1 + e x
f a ( x ) = tanh ( x ) σ ( x )
where ⊙ denotes element-wise multiplication.
The data flow for the multi-scale temporal convolution block can be expressed as follows. The input sequence X i 1 = X 2 enters Layer 1, and MKS dilated convolution, activation, and a 1D convolution are performed. The output of Layer 1, X o 1 , flows in two directions. One is be added to X i 1 to form the residual input of Layer 2 to enhance the model’s trainable ability [55], and another one is fused with the output features X o 2 and X o 3 of Layer 2 and Layer 3. This multi-scale fusion is completed by feature stacking. Similarly, the same operation is performed on Layer 2 and Layer 3; the difference, however, is that the input to layers 2 and 3 are different from layer 1, the input to Layer 2 is X i 2 = X i 1 + X o 1 , and the input to Layer 3 is X i 3 = X i 2 + X o 2 . Then, the optimal output of the multi-scale temporal convolution block is obtained as described in Equation (19).
X 3 = X o 1 X o 2 X o 3 = [ x 1 F 3 , x 2 F 3 , . . . , x T F 3 ] R T × ( 3 × F ¯ )
where F ¯ denotes the size in the frequency direction of the output feature maps of each layer, and F 3 = 3 × F ¯ .

4.4. Classification Block

We expect fine detection of each frame of the signal to improve the interpretability and persuasiveness of the detection. Therefore, the state of each frame needs to be output. This can be regarded as an FOD state classification for each frame.
The classification block includes several fully connected layers and the sigmoid activation function of the last layer, which outputs the probability of each frame belonging to each sound event. If the number of sound event categories is C, then P y R T × C , and 0 P y ( t , c ) 1 . In this article, the sound event category only includes two types, with FOD (abnormal) and without FOD (normal), i.e., C = 2 . In other words, we do not distinguish between the types of FOD. Thus, the network’s final output is shown in Equation (20).
Y = [ y 1 , , y T ] = arg max ( P y ) R T × 1
where y i { 0 , 1 } ( i = 1 , , T ) represents the detection result of each frame and 0 and 1 represent normal and abnormal, respectively.

4.5. Post-Processing

The output of the above network represents only the state of each frame of the signal, while the final output required by the detection system is whether this segment of the signal contains FOD. In addition, the predicted state of each frame is not guaranteed to be 100% correct. Therefore, a post-processing strategy is needed to improve the robustness of the final decision. It includes (1) predicted label smoothing and (2) final discriminant criteria as described in Figure 4.
(1) Predicted label smoothing: From experimental results, it was found that the FOD sound events always last for some time (more than the length of each frame) in each long signal segment. Since the duration of each frame is chosen to be small when the sound signal is split, this makes the FOD sound events, once they occur, often last for several frames. In addition, the continuous response of each FOD event is locally composed of short and weak shock and transition intervals. Most of these transition segments are the same or close to the normal waveform, which may cause the system to output incorrect results. Therefore, it is necessary to adopt a smoothing strategy to correct these unstable prediction results. As shown in Figure 4a, the method proposed in this paper ignores frames predicted to be labeled "FOD" if the number of consecutive frames for which FOD is detected is less than the consecutive threshold α 1 . Only when the number of consecutive abnormal frames is greater than or equal to α 1 , are the segments of these frames considered to be an FOD event. In addition, two FOD events with the frame interval less than the interval threshold α 2 are considered as the same event. By the above strategies, the elimination of instability of event detection results is achieved.
(2) Final discriminatory criteria: the continuous sound signal of the rocket tank is processed by pre-processing, network model prediction, and a frame smoothing strategy to obtain all suspected FOD events within the detection period of the rocket tank. As shown in Figure 4b, if the sum of the frames of all suspected FOD events exceeds the allowed threshold α 3 , the FOD is believed to be presented in this part of the tank.
The specific algorithm pseudo-code for post-processing is shown in Algorithm 1.
Algorithm 1: Post-processing
   Input  :prediction result Y = [ y 1 , , y T ] R T × 1 , continuous threshold value α 1 ,
       interval threshold α 2 , allowable frame length threshold α 3
   Output: final detection result R e s { 0 , 1 } , where 0 and 1 denote normal and
       abnormal respectively.
Machines 11 00187 i001

5. Experimental Setup

5.1. Signal and Data Description

In the experimental part of this research, the rocket tank with a complex internal structure was simulated, and the rocket tank model and related detection devices as shown in Figure 5. were manufactured. The rocket tank model has a three-stage structure, including fuel tank Section 1, a connecting section, and fuel tank Section 2. Each section has a diameter of 800 mm and a length of 600 mm. A motor drives the two rollers on the left side to rotate the tank with a speed of 2 r/min. At the same time, 16 microphones arranged on the carbon fiber bracket collect sound signals in different positions. The abnormal state with FOD is simulated by adding metal FOD to the inner wall of the tank. The types of FOD configured in the experiment are shown in the right of Figure 5, including titanium alloy wires, nuts, bolts, and rivets.
The signals of different FOD types are presented in Figure 6. The amplitude of the original time domain signals is close and dominated by the background noise of the drive motor. Thus, it is difficult to distinguish them from each other. From the view of the short time spectrum, the high-frequency portion is dark, so the energy of the signal is mainly concentrated in the low frequency. This indicates that if STFT is used as the input, it will bring many meaningless high-frequency features and increase the computational effort in vain. In contrast, in the Mel spectrum, the central frequency distribution of the Mel filter is densely decayed from low to high frequencies. Therefore, the Mel spectrum of the FOD signal increases the resolution of the low-frequency band compared to the short-time spectrum, and the useful information is further amplified.
In the process of the production of the data set as shown in Figure 7, each 60 s signal is divided into many small segments of 5.46 s (256 frames), and the overlap between segments is 2.73 s (128 frames) for data augmentation. Then they are converted to MFB features with the length of 256 and feature dimension of 128, and labels for each frame are added manually. The final data set is shown in Table 1. Abnormal samples contain both normal frames and abnormal frames, while normal samples have only normal frames. Thus, only abnormal samples are used for model training, while both normal and abnormal samples are used for testing to evaluate the performance of the method.

5.2. Metrics

In this paper, the metrics [56] of precision P, recall R, and F-score F are used to evaluate the effectiveness of the proposed sound-based FOD detection method, and the definition of each indicator is shown in Equations (21)–(23). For the FOD detection task, precision can reflect the level of missed detection and recall can reflect the level of false alarm, while F-score is a comprehensive consideration of the above two factors.
P = T P T P + F P
R = T P T P + F N
F = 2 P R P + R
where T P , F P , and F N denote true positive, false positive, and false negative, respectively. They are intermediate metrics from the confusion matrix.
In addition, the performance of the method is evaluated by calculating the overall accuracy of the samples with and without FOD, as described in Equation (24).
P total = T P + T N N
where T N denotes true negative and N denotes the total number of test samples.

5.3. Comparison Methods and Parameter Setting

The network model used in this paper consists of a three-layer 2D-CNN, a three-layer TCN, and a two-layer fully connected layer. The detailed network structure parameters are shown in Table 2. Since the process of FOD detection can be considered essentially as a frame-by-frame multi-label detection task, the network uses a binary cross-entropy (BCE) loss function. At the same time, a sigmoid function is used to output the probability of the accuracy of the signal detection results for each frame.
To verify the effectiveness of the MS-CTCN-based method for FOD detection, the method in this paper is compared with a CRNN that considers the sound signal’s time sequence characteristics and a CNN that only contains convolution layers and fully connected layers. In both CNN and CRNN, the same three-layer 2D-CNN structure and the resulting post-processing method as the proposed method in this paper are adopted, while the recurrent structure in CRNN adopts the bidirectional gated recurrent unit (Bi-GRU).
The factors that affect the model performance during the training of all models are parameter initialization strategy, batch size, learning rate, and epoch number. For the sake of fairness, the same settings are adopted in all model training processes. All models are built based on Pytorch and use its default parameter initialization method. The batch size, learning rate, and number of epochs are 64, 0.001, and 80, respectively. In addition, the most classical Adam optimizer is used to train the network parameters.
When post-processing the prediction results, three thresholds, α 1 , α 2 , and α 3 , should be determined. α 1 and α 2 are determined to have the most appropriate values of 2 and 1 through a grid search; that is, the isolated abnormal frames are discarded and two abnormal events with an interval of one frame are combined into the same event. As shown in Figure 8, the total number of abnormal frames for most samples in the dataset is between 25 and 150, and there are almost no abnormal samples with fewer than 10 frames. The same grid search method is used and α 3 is set to 10.

6. Results and Discussion

6.1. Overall Performance Analysis of MS-CTCN

The final detection performance of the proposed method is presented in Table 3. As can be observed from the results, the CNN method, which does not take into account temporal characteristics, incorrectly classifies all samples as normal states, rendering it unsuitable for this system. In comparison, the MS-CTCN method, which incorporates temporal characteristics, demonstrates an improvement in precision, recall, F-score, and overall accuracy of 0.29%, 5.07%, 3.50%, and 2.70%, respectively, when compared to the CRNN method. These results indicate that the performance of the MS-CTCN method is superior.
Due to the extremely poor performance of the CNN frame-by-frame method, only CRNN and MS-CTCN are further compared. The overall two-category confusion matrix is shown in Figure 9. For the case of misreporting the normal condition as abnormal, there are 10 cases of CRNN and 7 cases of MS-CTCN. The performance of the two methods is very close in terms of false alarm rate. However, in terms of the miss detection rate, MS-CTCN has only 13 cases, while CRNN has 75 cases. Thus, MS-CTCN performs better for FOD detection.
In addition, the detection performance of each kind of FOD is analyzed, as shown in Table 4. For nuts, large titanium alloy wires, and bolts with large masses, the detection accuracy can reach 100%. However, for rivets and small titanium alloy wires with a mass of less than 0.5 g, the detection accuracy of the simple CRNN method is not as good as that of the MS-CTCN method. Especially for small titanium alloy wires with a mass of only 0.33 g, the detection accuracy of the MS-CTCN method is 20.42% higher than that of the CRNN. Therefore, the method proposed in this paper has superior detection performance for smaller FOD.

6.2. Frame-Wise Event Detection Performance Analysis

To evaluate the performance of the proposed method for frame-wise event detection, a segment of bolt signal was further analyzed by different methods, as shown in Figure 10. The four red signal segments in Figure 10a are the four continuous bolt events marked manually. Due to the subjective nature of the manual marking, it is not possible to mark the exact start and end times of the events correctly, but an approximate range can be determined. Figure 10b shows the results using CNN. The CNN has poor feature extraction ability for signals with faint FOD because the characteristics of the time sequences are not considered. Only a few frames predict the FOD response, but they do not agree with the true labels. In addition, FOD events occur less frequently than normal events. In most cases, most of the data frames input to the system are normal frames without FOD, so the training of CNN belongs to a typical CNN training problem under imbalanced sample conditions, where the CNN tends to predict all frames as FOD-free events due to the influence of majority class samples. The performance of CRNN is shown in Figure 10c, where it isolates only two FOD events, one that has its start time grossly misestimated. In contrast, the MS-CTCN method in Figure 10d successfully detects four FOD events, and the duration of each event includes the range of manually labeled events. In addition, the post-processed smoothed labels successfully eliminated the instability of the FOD event start and end positions. Therefore, the method in this paper is more suitable for frame-wise detection of FOD events.

6.3. Performance of MS-TCN with Different Kernel Sizes

Experiments were conducted in MS-TCN blocks using convolution kernel sizes of 3, 5, and 7 to verify the excellent performance of convolution using MKS in this paper. The results are shown in Table 5. It is inferred that the possible reason is that the global time sequence features extracted at different scales are best suited for FOD sound event detection after the convolution kernel of size 3 expands the receptive field with dilation rates of 1, 2, and 4. In MKS, the addition of the temporal convolution with kernel sizes 5 and 7 corresponds to the enrichment of local multiscale time sequence features that are beneficial for FOD detection; i.e., local multiscale time sequence feature extraction is achieved.
Compared to using a single-size convolutional kernel, the F-score with MKS increased by at least 0.94% and the recall increased by at least 2.25%, but the precision decreased slightly, though not more than 0.42%. Corresponding to Equations (21) and (22), the decrease in precision reflects an increase in the system’s false reporting of normal samples in the test set as abnormal, and the increase in recall indicates a decrease in the number of missed detections by the system. Compared to misreporting, missed detections are more important to avoid in the actual testing task, and the reduction in missed detections in MKS is 5.36 times greater than the increase in misreporting. Therefore, MKS is more beneficial to improve the overall performance of the system. On the other hand, it also shows that the additional use of MKS with convolutional kernels of the size 5 and 7 is more beneficial for the FOD detection task.
To explore the event detection performance of the models with different convolutional kernel sizes, we performed frame-wise abnormal detection of the signal in Figure 10a with the corresponding model, and the results are shown in Figure 11. Before smoothing, MKS predicts the start and end positions of FOD events more stably compared to KS-3, KS-5, and KS-7. Figure 11d shows the start and end positions of the four FOD events. The oscillations of MKS are significantly smaller than Figure 11a–c. In addition, for the signals in the figure, KS-5 performs the worst, a result that is also consistent with its worst F-score in Table 5. In conclusion, the size of the convolution kernel has an important effect on the FOD detection in the dilated convolution where TCN has been used. For convolution kernel sizes of 3, 5, and 7, size 3 works best, followed by 7, and then 5, while the method using MKS is better than that using the three aforementioned convolution kernels alone.

7. Conclusions

In this paper, an intelligent FOD-detection model is proposed based on MS-CTCN. The results show that the method has superior FOD detection accuracy to the traditional CNN and CRNN models. The impacts of different sizes of convolution kernels on the FOD detection performance in MS-CTCN are investigated, and the optimal network parameters applicable to FOD detection are given.
The aim of this study is to identify the presence of FOD. In future work, the research on sound-based FOD detection will be devoted to achieving the identification of FOD types and the localization of FOD through multi-channel sound signals. As a starting point, researchers can focus on developing methods for identifying different types of FOD based on their unique sound signatures. This could involve using advanced signal-processing techniques, such as frequency-domain analysis or machine learning algorithms [57], to extract relevant features from the sound signals. Another area of focus can be on developing methods for localizing FOD based on multi-channel sound signals. This could involve using techniques such as beamforming [58], which uses information from multiple microphone channels to pinpoint the location of a sound source, or using techniques from source separation to separate the sounds from the FOD from the background noise. Incorporating these new techniques into the detection model could enable the model to identify not just the presence of FOD but also the type of FOD and its location, providing more efficient and reliable support for quality assurance in the rocket-production process.

Author Contributions

Conceptualization, T.L., Y.Z. and Z.R.; methodology, T.L. and S.H.; software, T.L.; validation, T.L., Y.Z. and Z.R.; formal analysis, T.L.; investigation, K.H. and X.Z.; resources, K.Y.; data curation, K.H. and X.Z.; writing—original draft preparation, T.L. and Z.R.; writing—review and editing, Y.Z. and K.Y.; visualization, K.H.; supervision, Y.Z.; project administration, S.H.; funding acquisition, Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Defense Industrial Technology Development Program grant number JCKY2017203A007.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, J.; Liang, G. Simulation of mass and heat transfer in liquid hydrogen tanks during pressurizing. Chin. J. Aeronaut. 2019, 32, 2068–2084. [Google Scholar] [CrossRef]
  2. Porto, J.F.D.; Loescher, D.H.; Olson, H.C.; Plunkett, P.V. SEM/EDAX Analysis of PIND Test Failures. In Proceedings of the 19th International Reliability Physics Symposium, Las Vegas, NV, USA, 7–9 April 1981; pp. 163–166. [Google Scholar] [CrossRef]
  3. Scaglione, L. Neural network application to particle impact noise detection. In Proceedings of the 1994 IEEE International Conference on Neural Networks (ICNN’94), Orlando, FL, USA, 28 June–2 July 1994; Volume 5, pp. 3415–3419. [Google Scholar] [CrossRef]
  4. Wang, S.; Chen, R.; Zhang, L.; Wang, S. Detection and material identification of loose particles inside the aerospace power supply via stochastic resonance and LVQ network. Trans. Inst. Meas. Control. 2012, 34, 947–955. [Google Scholar] [CrossRef]
  5. Zhai, G.; Chen, J.; Li, C.; Wang, G. Pattern recognition approach to identify loose particle material based on modified MFCC and HMMs. Neurocomputing 2015, 155, 135–145. [Google Scholar] [CrossRef]
  6. Wang, G.T.; Liang, X.W.; Xue, Y.Y.; Li, C.; Ding, Q. Algorithm Used to Detect Weak Signals Covered by Noise in PIND. Int. J. Aerosp. Eng. 2019, 2019, e1637953. [Google Scholar] [CrossRef] [Green Version]
  7. Gao, H.l.; Zhang, H.; Wang, S.j. Research on Auto-detection for Remainder Particles of Aerospace Relay Based on Wavelet Analysis. Chin. J. Aeronaut. 2007, 20, 75–80. [Google Scholar] [CrossRef] [Green Version]
  8. Wang, S.j.; Gao, H.l.; Zhai, G.f. Research on Feature Extraction of Remnant Particles of Aerospace Relays. Chin. J. Aeronaut. 2007, 20, 253–259. [Google Scholar] [CrossRef] [Green Version]
  9. Liu, Y.; Li, S.; Wang, J.; Zeng, H.; Lu, J. A computer vision-based assistant system for the assembly of narrow cabin products. Int. J. Adv. Manuf. Technol. 2015, 76, 281–293. [Google Scholar] [CrossRef]
  10. Xu, H.; Han, Z.; Feng, S.; Zhou, H.; Fang, Y. Foreign object debris material recognition based on convolutional neural networks. EURASIP J. Image Video Process. 2018, 2018, 21. [Google Scholar] [CrossRef]
  11. Wang, Y.S.; Liu, N.N.; Guo, H.; Wang, X.L. An engine-fault-diagnosis system based on sound intensity analysis and wavelet packet pre-processing neural network. Eng. Appl. Artif. Intell. 2020, 94, 103765. [Google Scholar] [CrossRef]
  12. Zhang, S.; He, Q.; Ouyang, K.; Xiong, W. Multi-bearing weak defect detection for wayside acoustic diagnosis based on a time-varying spatial filtering rearrangement. Mech. Syst. Signal Process. 2018, 100, 224–241. [Google Scholar] [CrossRef]
  13. Huang, Q.; Xie, L.; Yin, G.; Ran, M.; Liu, X.; Zheng, J. Acoustic signal analysis for detecting defects inside an arc magnet using a combination of variational mode decomposition and beetle antennae search. ISA Trans. 2020, 102, 347–364. [Google Scholar] [CrossRef] [PubMed]
  14. Tran, M.Q.; Liu, M.K.; Elsisi, M. Effective multi-sensor data fusion for chatter detection in milling process. ISA Trans. 2022, 125, 514–527. [Google Scholar] [CrossRef]
  15. Yao, Y.; Zhang, S.; Yang, S.; Gui, G. Learning Attention Representation with a Multi-Scale CNN for Gear Fault Diagnosis under Different Working Conditions. Sensors 2020, 20, 1233. [Google Scholar] [CrossRef] [Green Version]
  16. Parey, A.; Singh, A. Gearbox fault diagnosis using acoustic signals, continuous wavelet transform and adaptive neuro-fuzzy inference system. Appl. Acoust. 2019, 147, 133–140. [Google Scholar] [CrossRef]
  17. Zhang, G.; Wang, J.; Han, B.; Jia, S.; Wang, X.; He, J. A Novel Deep Sparse Filtering Method for Intelligent Fault Diagnosis by Acoustic Signal Processing. Shock Vib. 2020, 2020, e8837047. [Google Scholar] [CrossRef]
  18. Li, X.; Wan, S.; Liu, S.; Zhang, Y.; Hong, J.; Wang, D. Bearing fault diagnosis method based on attention mechanism and multilayer fusion network. ISA Trans. 2022, 128, 550–564. [Google Scholar] [CrossRef]
  19. Glowacz, A.; Glowacz, W.; Glowacz, Z.; Kozik, J. Early fault diagnosis of bearing and stator faults of the single-phase induction motor using acoustic signals. Measurement 2018, 113, 1–9. [Google Scholar] [CrossRef]
  20. Nakamura, H.; Asano, K.; Usuda, S.; Mizuno, Y. A Diagnosis Method of Bearing and Stator Fault in Motor Using Rotating Sound Based on Deep Learning. Energies 2021, 14, 1319. [Google Scholar] [CrossRef]
  21. Heittola, T.; Mesaros, A.; Eronen, A.; Virtanen, T. Context-dependent sound event detection. EURASIP J. Audio Speech Music. Process. 2013, 2013, 1. [Google Scholar] [CrossRef] [Green Version]
  22. Mesaros, A.; Heittola, T.; Dikmen, O.; Virtanen, T. Sound event detection in real life recordings using coupled matrix factorization of spectral representations and class activity annotations. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), South Brisbane, QLD, Australia, 19–24 April 2015; pp. 151–155. [Google Scholar] [CrossRef]
  23. Huang, G.; Heittola, T.; Virtanen, T. Using Sequential Information in Polyphonic Sound Event Detection. In Proceedings of the 2018 16th International Workshop on Acoustic Signal Enhancement (IWAENC), Tokyo, Japan, 17–20 September 2018; pp. 291–295. [Google Scholar] [CrossRef]
  24. Adavanne, S.; Politis, A.; Virtanen, T. Multichannel Sound Event Detection Using 3D Convolutional Neural Networks for Learning Inter-channel Features. In Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–7. [Google Scholar] [CrossRef] [Green Version]
  25. Li, Y.; Liu, M.; Drossos, K.; Virtanen, T. Sound Event Detection Via Dilated Convolutional Recurrent Neural Networks. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 286–290. [Google Scholar] [CrossRef] [Green Version]
  26. Yu, Y.; Si, X.; Hu, C.; Zhang, J. A Review of Recurrent Neural Networks: LSTM Cells and Network Architectures. Neural Comput. 2019, 31, 1235–1270. [Google Scholar] [CrossRef]
  27. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar] [CrossRef]
  28. Wang, Y.; Zhao, G.; Xiong, K.; Shi, G.; Zhang, Y. Multi-Scale and Single-Scale Fully Convolutional Networks for Sound Event Detection. Neurocomputing 2021, 421, 51–65. [Google Scholar] [CrossRef]
  29. Guirguis, K.; Schorn, C.; Guntoro, A.; Abdulatif, S.; Yang, B. SELD-TCN: Sound Event Localization & Detection via Temporal Convolutional Networks. In Proceedings of the 2020 28th European Signal Processing Conference (EUSIPCO), Amsterdam, The Netherlands, 18–21 January 2021; pp. 16–20. [Google Scholar] [CrossRef]
  30. Frigieri, E.P.; Brito, T.G.; Ynoguti, C.A.; Paiva, A.P.; Ferreira, J.R.; Balestrassi, P.P. Pattern recognition in audible sound energy emissions of AISI 52100 hardened steel turning: A MFCC-based approach. Int. J. Adv. Manuf. Technol. 2017, 88, 1383–1392. [Google Scholar] [CrossRef]
  31. Warren Liao, T. Feature extraction and selection from acoustic emission signals with an application in grinding wheel condition monitoring. Eng. Appl. Artif. Intell. 2010, 23, 74–84. [Google Scholar] [CrossRef]
  32. Brahim, J.; Loubna, R.; Noureddine, F. Rnn-And Cnn-Based Weed Detection For Crop Improvement: An Overview. Foods Raw Mater. 2021, 9, 387–396. [Google Scholar]
  33. Jabir, B.; Falih, N.; Rahmani, K. Accuracy and Efficiency Comparison of Object Detection Open-Source Models. Int. J. Online Biomed. Eng. (iJOE) 2021, 17, 165. [Google Scholar] [CrossRef]
  34. Zinemanas, P.; Cancela, P.; Rocamora, M. End-to-end Convolutional Neural Networks for Sound Event Detection in Urban Environments. In Proceedings of the 2019 24th Conference of Open Innovations Association (FRUCT), Moscow, Russia, 8–12 April 2019; pp. 533–539. [Google Scholar] [CrossRef]
  35. Wang, C.Y.; Tai, T.C.; Wang, J.C.; Santoso, A.; Mathulaprangsan, S.; Chiang, C.C.; Wu, C.H. Sound Events Recognition and Retrieval Using Multi-Convolutional-Channel Sparse Coding Convolutional Neural Networks. IEEE/ACM Trans. Audio Speech Lang. Process. 2020, 28, 1875–1887. [Google Scholar] [CrossRef]
  36. Khan, S.; Naseer, M.; Hayat, M.; Zamir, S.W.; Khan, F.S.; Shah, M. Transformers in Vision: A Survey. ACM Comput. Surv. 2022, 54, 200:1–200:41. [Google Scholar] [CrossRef]
  37. Imoto, K.; Kyochi, S. Sound Event Detection Utilizing Graph Laplacian Regularization with Event Co-Occurrence. Ieice Trans. Inf. Syst. 2020, E103.D, 1971–1977. [Google Scholar] [CrossRef]
  38. Kim, S.J.; Chung, Y.J. Multi-Scale Features for Transformer Model to Improve the Performance of Sound Event Detection. Appl. Sci. 2022, 12, 2626. [Google Scholar] [CrossRef]
  39. Kong, Q.; Xu, Y.; Wang, W.; Plumbley, M.D. Sound Event Detection of Weakly Labelled Data With CNN-Transformer and Automatic Threshold Optimization. IEEE/ACM Trans. Audio Speech Lang. Process. 2020, 28, 2450–2460. [Google Scholar] [CrossRef]
  40. D’Ascoli, S.; Touvron, H.; Leavitt, M.L.; Morcos, A.S.; Biroli, G.; Sagun, L. ConViT: Improving Vision Transformers with Soft Convolutional Inductive Biases. In Proceedings of the 38th International Conference on Machine Learning, PMLR, Virtual, 18–24 July 2021; pp. 2286–2296. [Google Scholar]
  41. Wang, J.J.; Wang, C.; Fan, J.S.; Mo, Y.L. A deep learning framework for constitutive modeling based on temporal convolutional network. J. Comput. Phys. 2022, 449, 110784. [Google Scholar] [CrossRef]
  42. Wang, Y.; Zhao, G.; Xiong, K.; Shi, G. MSFF-Net: Multi-scale feature fusing networks with dilated mixed convolution and cascaded parallel framework for sound event detection. Digit. Signal Process. 2022, 122, 103319. [Google Scholar] [CrossRef]
  43. Kopparapu, S.K.; Laxminarayana, M. Choice of Mel filter bank in computing MFCC of a resampled speech. In Proceedings of the 10th International Conference on Information Science, Signal Processing and their Applications (ISSPA 2010), Kuala Lumpur, Malaysia, 10–13 May 2010; pp. 121–124. [Google Scholar] [CrossRef] [Green Version]
  44. Allen, J.; Rabiner, L. A unified approach to short-time Fourier analysis and synthesis. Proc. IEEE 1977, 65, 1558–1564. [Google Scholar] [CrossRef]
  45. Tak, R.N.; Agrawal, D.M.; Patil, H.A. Novel Phase Encoded Mel Filterbank Energies for Environmental Sound Classification. In Proceedings of the Pattern Recognition and Machine Intelligence; Shankar, B.U., Ghosh, K., Mandal, D.P., Ray, S.S., Zhang, D., Pal, S.K., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2017; pp. 317–325. [Google Scholar] [CrossRef]
  46. Li, Y.; Du, X.; Wan, F.; Wang, X.; Yu, H. Rotating machinery fault diagnosis based on convolutional neural network and infrared thermal imaging. Chin. J. Aeronaut. 2020, 33, 427–438. [Google Scholar] [CrossRef]
  47. Vafeiadis, A.; Votis, K.; Giakoumis, D.; Tzovaras, D.; Chen, L.; Hamzaoui, R. Audio content analysis for unobtrusive event detection in smart homes. Eng. Appl. Artif. Intell. 2020, 89, 103226. [Google Scholar] [CrossRef] [Green Version]
  48. Hasan, M.J.; Islam, M.M.M.; Kim, J.M. Acoustic spectral imaging and transfer learning for reliable bearing fault diagnosis under variable speed conditions. Measurement 2019, 138, 620–631. [Google Scholar] [CrossRef]
  49. Hasan, M.J.; Islam, M.M.M.; Kim, J.M. Multi-sensor fusion-based time-frequency imaging and transfer learning for spherical tank crack diagnosis under variable pressure conditions. Measurement 2021, 168, 108478. [Google Scholar] [CrossRef]
  50. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. arXiv 2015. [Google Scholar] [CrossRef]
  51. Jiao, J.; Zhao, M.; Lin, J.; Liang, K. A comprehensive review on convolutional neural network in machine fault diagnosis. Neurocomputing 2020, 417, 36–63. [Google Scholar] [CrossRef]
  52. Biscione, V.; Bowers, J. Learning Translation Invariance in CNNs. arXiv 2020. [Google Scholar] [CrossRef]
  53. Sainath, T.N.; Vinyals, O.; Senior, A.; Sak, H. Convolutional, Long Short-Term Memory, fully connected Deep Neural Networks. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), South Brisbane, QLD, Australia, 19–24 April 2015; pp. 4580–4584. [Google Scholar] [CrossRef]
  54. Oord, A.v.d.; Dieleman, S.; Zen, H.; Simonyan, K.; Vinyals, O.; Graves, A.; Kalchbrenner, N.; Senior, A.; Kavukcuoglu, K. WaveNet: A Generative Model for Raw Audio. arXiv 2016. [Google Scholar] [CrossRef]
  55. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  56. Mesaros, A.; Heittola, T.; Virtanen, T. Metrics for Polyphonic Sound Event Detection. Appl. Sci. 2016, 6, 162. [Google Scholar] [CrossRef] [Green Version]
  57. Gao, D.W.; Zhu, Y.S.; Yan, K.; Fu, H.; Ren, Z.J.; Kang, W.; Guedes Soares, C. Joint learning system based on semi–pseudo–label reliability assessment for weak–fault diagnosis with few labels. Mech. Syst. Signal Process. 2023, 189, 110089. [Google Scholar] [CrossRef]
  58. Shaw, A.; Smith, J.; Hassanien, A. MVDR Beamformer Design by Imposing Unit Circle Roots Constraints for Uniform Linear Arrays. IEEE Trans. Signal Process. 2021, 69, 6116–6130. [Google Scholar] [CrossRef]
Figure 1. 1D dilated convolution with convolution kernel size 3.
Figure 1. 1D dilated convolution with convolution kernel size 3.
Machines 11 00187 g001
Figure 2. The MS-CTCN-based FOD detection system.
Figure 2. The MS-CTCN-based FOD detection system.
Machines 11 00187 g002
Figure 3. Multi-scale temporal convolution block.
Figure 3. Multi-scale temporal convolution block.
Machines 11 00187 g003
Figure 4. Post-processing: (a) predicted label smoothing, (b) final discriminant criteria.
Figure 4. Post-processing: (a) predicted label smoothing, (b) final discriminant criteria.
Machines 11 00187 g004
Figure 5. The test bench of sound-based detection for FOD.
Figure 5. The test bench of sound-based detection for FOD.
Machines 11 00187 g005
Figure 6. Signals of different FOD: (a) Nut, (b) Ti-wire (3.93 g), (c) Ti-wire (0.33 g), (d) Rivet, (e) Bolt, (f) Without FOD. In the spectrum and Mel spectrum, a brighter color means higher energy.
Figure 6. Signals of different FOD: (a) Nut, (b) Ti-wire (3.93 g), (c) Ti-wire (0.33 g), (d) Rivet, (e) Bolt, (f) Without FOD. In the spectrum and Mel spectrum, a brighter color means higher energy.
Machines 11 00187 g006
Figure 7. Manual labeling and data set expansion.
Figure 7. Manual labeling and data set expansion.
Machines 11 00187 g007
Figure 8. The statistical distribution of the number of frames of the samples with FOD.
Figure 8. The statistical distribution of the number of frames of the samples with FOD.
Machines 11 00187 g008
Figure 9. Confusion matrix: (a) Confusion matrix of CRNN, (b) Confusion matrix of MS-CTCN.
Figure 9. Confusion matrix: (a) Confusion matrix of CRNN, (b) Confusion matrix of MS-CTCN.
Machines 11 00187 g009
Figure 10. Frame-by-frame prediction and smoothing results for different methods: (a) FOD signal and true label, (b) CNN’s prediction results and smoothed results, (c) CRNN’s prediction results and smoothed results, (d) MS-CTCN’s prediction results and smoothed results.
Figure 10. Frame-by-frame prediction and smoothing results for different methods: (a) FOD signal and true label, (b) CNN’s prediction results and smoothed results, (c) CRNN’s prediction results and smoothed results, (d) MS-CTCN’s prediction results and smoothed results.
Machines 11 00187 g010
Figure 11. The frame-by-frame prediction results of the model when using different convolution kernel sizes for dilated convolution for the signal in Figure 10: (a) KS-3’s prediction label, (b) KS-5’s prediction label, (c) KS-7’s prediction label, (d) MKS’s prediction label.
Figure 11. The frame-by-frame prediction results of the model when using different convolution kernel sizes for dilated convolution for the signal in Figure 10: (a) KS-3’s prediction label, (b) KS-5’s prediction label, (c) KS-7’s prediction label, (d) MKS’s prediction label.
Machines 11 00187 g011
Table 1. Data set information.
Table 1. Data set information.
Acoustic Condition of the TankNumber of the Training SamplesNumber of the Testing Samples
With FOD
(abnormal)
Nut960240
Ti-wire (3.39 g)960240
Ti-wire (0.33 g)960240
Rivet960240
Bolt960240
Without FOD
(normal)
--1200
Total-48002400
Table 2. The main parameters of the proposed method.
Table 2. The main parameters of the proposed method.
OperationOutput SizeConfiguration
Preprocessing256 × 128-
Network input1 × 256 × 128-
CNN1(32, 5, 5)32 × 256 × 128Max-pooling (1, 4)
CNN2(32, 5, 5)32 × 256 × 32Max-pooling (1, 4)
CNN3(32, 3, 3)32 × 256 × 4Max-pooling (1, 2)
Reshape256 × 128-
TCN1256 × 128Dilation rate = 1
TCN2256 × 128Dilation rate = 2
TCN3256 × 128Dilation rate = 4
Connection256 × 384-
FC1256 × 128-
FC2256 × 2-
Sigmoid256 × 2-
Post-processing1 α 1 = 2 , α 2 = 1 , α 3 = 10
Table 3. Performance of different models.
Table 3. Performance of different models.
ModelP(%)R(%)F(%)Ptotal(%)
CNN1000.080.1750.04
CRNN99.1293.7596.3696.46
MS-CTCN99.4198.8299.1799.16
Table 4. The accuracy of CRNN and MS-CTCN for different FOD.
Table 4. The accuracy of CRNN and MS-CTCN for different FOD.
Type of FODCRNN(%)MS-CTCN(%)
Nut100100
Ti-wire (3.93)100100
Ti-wire (0.33)77.5097.92
Rivet93.7597.50
Bolt100100
Table 5. Model performance when using dilated convolution with different convolution kernel sizes.
Table 5. Model performance when using dilated convolution with different convolution kernel sizes.
ModelP(%)R(%)F(%)
KS-399.8299.6798.22
KS-599.7496.0097.83
KS-799.8296.0897.91
MKS99.4098.9299.16
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lin, T.; Zhu, Y.; Ren, Z.; Huang, K.; Zhang, X.; Yan, K.; Huang, S. Sound-Based Intelligent Detection of FOD in the Final Assembly of Rocket Tanks. Machines 2023, 11, 187. https://doi.org/10.3390/machines11020187

AMA Style

Lin T, Zhu Y, Ren Z, Huang K, Zhang X, Yan K, Huang S. Sound-Based Intelligent Detection of FOD in the Final Assembly of Rocket Tanks. Machines. 2023; 11(2):187. https://doi.org/10.3390/machines11020187

Chicago/Turabian Style

Lin, Tantao, Yongsheng Zhu, Zhijun Ren, Kai Huang, Xinzhuo Zhang, Ke Yan, and Shunzhou Huang. 2023. "Sound-Based Intelligent Detection of FOD in the Final Assembly of Rocket Tanks" Machines 11, no. 2: 187. https://doi.org/10.3390/machines11020187

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop