Next Article in Journal
A Decision Support System for Hospital Configurations in Construction Projects
Next Article in Special Issue
Experimental Research on the Compression Property of Geopolymer Concrete with Molybdenum Tailings as a Building Material
Previous Article in Journal
Short Review on the Application of Recycled Powder in Cement-Based Materials: Preparation, Performance and Activity Excitation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Big Data-Based Performance Analysis of Tunnel Boring Machine Tunneling Using Deep Learning

1
State Key Laboratory of Eco-hydraulics in Northwest Arid Region, Xi’an University of Technology, Xi’an 710048, China
2
Smart Construction Laboratory, Department of Building and Real Estate, the Hong Kong Polytechnic University, Hong Kong 999077, China
3
Tianjin Port Engineering Institute Ltd. of CCCC First Harbor Engineering Company Ltd., Tianjin 300222, China
*
Author to whom correspondence should be addressed.
Buildings 2022, 12(10), 1567; https://doi.org/10.3390/buildings12101567
Submission received: 13 August 2022 / Revised: 21 September 2022 / Accepted: 23 September 2022 / Published: 29 September 2022

Abstract

:
In tunnel boring machine (TBM) construction, the advance rate is a crucial parameter that affects the TBM driving efficiency, project schedule, and construction cost. During the operation process, various types of indicators that are monitored in real-time can help to control the advance rate of TBM. Although some studies have already been carried out in advance rate prediction, the research is almost all based on statistical methods and shallow machine learning algorithms, thereby having difficulties in dealing with a very large amount of monitored data and in modeling the time-dependent characteristics of the parameters. To solve this problem, a deep learning model is proposed based on the CNN architecture, bidirectional Long Short-Term Memory module, and the attention mechanism, which is called the CNN-Bi-LSTM-Attention model. In the first step, the monitored data is processed, and the CNN architecture is adopted to extract features from the data sequence. Then the Bi-LSTM module is adopted to obtain the time-dependent indicators. The significant features can be addressed by the added attention mechanism. In the model training process, the rotation speed of the cutter head (N), thrust (F), torque (T), penetration rate (P), and chamber earth pressure (Soil_P) are adopted to predict the advance rate. The influence of the training periods on the model performance is also discussed. The result shows that not only the data amount, but also the data periods have an influence on the prediction. The long-term data may lead to a failure of the advance rate of TBM. The model evaluation result on the test data shows that the proposed model cannot predict the monitored data in the starting stage, which denotes that the working state of TBM in the starting stage is not stable. Especially when the TBM starts to work, the prediction error is big. The proposed model is also compared with several traditional machine methods, and the result shows the excellent performance of the proposed model.

1. Introduction

With the development of city construction, tunnel boring machines (TBMs) have been widely applied in civil engineering. The TBM is a type of mechanization equipment with the characteristics of high efficiency, environment friendly, and safety [1], which are widely used in mine tunneling engineering, hydraulic engineering, subway construction, etc. In the process of TBM operation, the divers control the cutter head and the advance speed by inputting the related parameters. However, it is hard to tune the parameters in a timely manner and making good decisions is based on experience, which may lead to low construction efficiency and cutter wear. The TBM sensors should monitor different kinds of operational parameters, such as rotation speed of cutter, thrust of cutter head, torque of cutter head, advance velocity, and penetration index. These parameters reflect the real-time interaction state of the tunnel face and the cutter head. Based on the recorded data, the TBM performance and driving parameters can be reasonably predicted. It is beneficial to develop an appropriate tunneling plan, minimize the frequency of common risks, and improve the drilling system efficiency in construction [2,3].
Many achievements in TBM performance prediction have been gained from previous studies, and some theoretical and experience-based models have been proposed [4,5,6] to reveal the inner mechanism of TBM operations and establish the relationship between different variables. Sapigni et al. [7] found that the rate of penetration (ROP) is related to the rock mass rating (RMR) based on the data from three tunnels. The relationship of the two parameters can be expressed as a quadratic polynomial regression equation. Kahraman et al. [8] established a regression model for ROP based on the monitored data. The result showed that ROP is determined by the rock properties. Hassanpour et al. [9] built an empirical formula according to multi geo-parameters for ROP based on the data from the NO. 2 Tunnel of Nowsood. It is thought that ROP is closely related to rock mass properties, especially for the field penetration index (FPI). The research also showed that the common indexes applied in rock mass evaluation, such as uniaxial compressive strength (UCS) and rock quality designation (RQD), are convenient for building the prediction model. However, the experience-based model is established for a specific engineering. Hence, the model’s application is limited.
Some researchers also applied numerical simulation in the TBM operation process. Kasper and Meschke [10] analyzed the influence of TBM operational parameters on the foundation using a 3D finite element simulation model. Gong and Zhao [11] analyzed the interaction of the tunnel face and the cutter head using a numerical simulation method, and the result shows that ROP advanced with the growth of the degree of rock brittleness. Li et al. [12] applied the discrete element method (DEM) to analyze the relationship between ROP and wedge angle and confinement stress, and the result showed that large wedge angle lead to tensile cracks and lowers the ROP. Although the numerical models are inspiring for TBM performance prediction, there are also weaknesses in TBM operation analysis. The boundary condition of the model is usually simplified in most cases of numerical simulation calculation. The results of these methods are not accurate and cannot show the minor differences in TBM operational parameter changes.
With the advancement of data science, the intelligent analysis of complex data in engineering projects can be realized. The machine learning methods applied in engineering can improve efficiency and lower cost [13]. Based on the monitoring of data in real projects, intelligent techniques are also applied in TBM operating parameters prediction. Sun et al. [14] combined geological data with the operational data and applied random forest to predict TBM load. Random forest [15] and support vector regression (SVR) [16] with parameters optimization were also applied to predict TBM advance rate and penetration rate, respectively. Noori et al. [17] discussed the factors influencing TBM performance prediction using neural networks and particle swarm optimization (ANN-PSO). Afradi et al. [18] applied fuzzy logic to predict TBM penetration rate, with the optimization algorithm of harmony search and particle swarm. However, it is hard for the shallow networks, such as SVR and random forest, to deal with a very large amount of monitored data in TBM construction. Meanwhile, the time series characteristics of TBM operation data are not considered in establishing the prediction model.
In recent years, deep learning theory has been developed in unstructured data analysis [19,20,21], especially for images and language. It is a subset of artificial neural networks of machine learning. Deep learning models usually have more layers and are deeper than shallow networks. The specific applications of deep learning models include image interpretation [22,23], sound classification [24,25], and text analysis [26,27]. These complex models applied in research are essentially extensions of the simple deep models, such as recurrent neural networks (RNNs) [28], convolutional neural networks (CNNs) [29], deep belief networks [30], and auto-encoder networks [31]. Deep learning models are widely used in engineering, such as dam displacement prediction [32,33], dredging construction operation [34], and mechanical responses prediction of caisson foundation [35]. Some of the models were also applied in TBM performance prediction [36,37,38]. Among these models, RNNs is a special type of network that can output the targets by considering both the previous and current information, thus having advantages in analyzing time series data. However, traditional RNNs suffer from the effects of gradients vanishing or exploding when dealing with long sequences. Therefore, the long short-term memory (LSTM) networks [39] were proposed to improve the RNNs performance. The three types of gates in LSTM, i.e., the input gate, the forget gate, and the output gate, control the data flow in the LSTM unit. The LSTM-based models are also applied in TBM performance prediction. Liu et al. [40] applied LSTM with a global attention mechanism to predict tunnel lithology based on big TBM data. Gao et al. [41] established a LSTM-based model to predict TBM operatingchange parameters and verified the model with a large amount of data collected from a practical project. Fu and Zhang [42] considered spatio-temporal feature fusion in a LSTM model, and conducted a global sensitivity analysis for TBM performance prediction. Li et al. [43] combined CNNs and LSTM to establish the prediction model and analyzed the TBM cutter head speed and penetration rate. However, it is difficult to deal with complex high-dimensional tunneling parameters using the maximum likelihood and classic artificial intelligence (AI) techniques, and the prediction accuracy needs to be improved.
Inspired by the previous achievements, a novel deep learning model was established by combining CNNs and LSTM for the prediction task of TBM performance. The one-dimensional CNNs and the bidirectional LSTM (Bi-LSTM) were adopted to build the whole model. The local and global information is considered in the CNNs, and the time series information is involved in the Bi-LSTM. The significant features are addressed using the attention mechanism. Meanwhile, how much data should be used to build a TBM advance rate is another challenging issue, as deep learning is a totally data-driven approach. In the research, five models were built to address the problem. Moreover, the traditional machine learning methods and the proposed algorithm were also compared. The results showed that the LSTM-based model outperforms the traditional machine learning models. The terms of monitored data are also discussed, and the results show that long-term data may lead to a prediction failure of the TBM advance rate.

2. Methodology

TBM performance prediction is a significant problem in engineering. It is a dynamic process that needs operators to make quick and accurate operational parameter decisions. In the research, the CNN-Bi-LSTM-Attention model was established to predict the advance rate of TBM. The monitored data, including the rotation speed of the cutter head (N), thrust (F), torque (T), penetration rate (P), and chamber earth pressure (Soil_P), were adopted to build the model. The model architecture is shown in Figure 1. The flow chart of the research is as follows.
  • In step 1, the monitored data was collected using different types of sensors installed on the TBM. The construction data in the strata of medium-weathered sandstone was adopted to train the model. The monitored data sequence for free rotation was also cut in the working process. The error in the monitored data was also detected, and the errors in the data sequence were processed. Normalization was also adopted to scale the data into the range of (0,1).
  • In step 2, the CNN-Bi-LSTM-Attention model was established using the CNN architecture, bidirectional Long Short-Term Memory module, and the attention mechanism. The CNN architecture was adopted to extract global features from the data sequence. The Bi-LSTM module was adopted to get the time-dependent features. The attention mechanism was used to address the local features, which was significant for the TBM advance rate prediction. Root mean square error (RMSE), Mean absolute error (MAE), and R2 were used to evaluate the models. The comparison of the predicted result and the monitored data are also shown.
  • In step 3, the monitored data from Day 1 to Day 27 was applied. The monitored data of Day 27 were taken as the test data. The training data periods were considered in model training. The data of Day 1-Day 26, Day 20-Day 26, Day 24-Day 26, Day 25-Day 26, and Day 26 were used to train the model. The different schemes of model training were used to evaluate the effectiveness of the data amount and periods. The comparison between the monitored advance rate and the predicted values shows the performance of the different models and where the errors appear in the corking cycle of the TBM.
Finally, the performance of the traditional machine learning methods and the proposed model were compared. The results showed that the proposed model outperforms the traditional models. It indicates that the monitored data of TBM working cycles is time-dependent. The CNN-Bi-LSTM-Attention model can utilize multiple types of features in advance rate prediction.

2.1. CNN

Convolutional neural networks (CNN) have been widely used in image classification, identification, and segmentation tasks, which perform better than traditional image processing methods. The low and high-level features are extracted and integrated to be new features for image analysis. There are three significant ideas for CNN architecture.
  • Local connection: The kernels of the CNN layer are only connected to specific ones of the previous layer, which can obtain the effective features of the sequence.
  • Weight sharing: The feature map of the input sequence is processed by the same kernel using sliding windows. Therefore, all the neurons of the same kernel share the same parameters, which reduces the time of the training process.
  • Pooling layers: Pooling denotes some or all features based on the values. It is implemented in the low- and high-level features integration process.
In this research, the one-dimensional CNN is applied to process the monitoring data. The sequence features are extracted from the sequence through a CNN structure which contains two convolutional layers.

2.2. Bi-LSTM-Attention

2.2.1. Long Short-Term Memory (LSTM)

LSTM is a kind of RNN architecture with feedback connections. The memory unit in a LSTM network makes it available for information learning over a long period. A common LSTM block includes the input gate, the output gate, and the forget gate in a single unit, as shown in Figure 2. The information transmission is controlled by the three gates in the LSTM block. The three gates in the LSTM block make it possible to solve gradient return problems, including gradient vanishing and exploding. At time t, every gate produces a variable, it, ft, and ot for the memory ct. The output of the block is ht. The hidden state ht−1 determines the amount of information from xt to ct. The variable ft determines the amount of stored information in cell state ct−1. The variable ot controls the final output, ht. The specific formulae of LSTM are shown in Equations (1)–(5).
f t = σ W f h t 1 , x t + b f
i t = σ W i h t 1 , x t + b i
o t = σ W o h t 1 , x t + b o
C t = f t C t 1 + i t tanh W c h t 1 , x t + b c
h t = O t tanh C t

2.2.2. Bi-LSTM-Attention

In general data processing of the traditional LSTM networks, the model only processes the previous data and ignores future information. The bidirectional LSTM (Bi-LSTM) includes two LSTM modules with a forward direction and a backward direction. Both previous and later information are stored for the time series data processing at the current time. Bi-LSTM can overcome the limitations of LSTM cells that only process the previous information as opposed to using both the previous and the later ones.
The attention mechanism is proposed, based on the fact that human visual attention should be attracted to a part of the scene and should ignore unnecessary information, which is different from the idea of global attention. The attention mechanism addresses focusing on the significant region of the scene rather than on global computation. As a result, the critical part of the scene should be assigned a greater weight to address the important information. Therefore, the Bi-LSTM model can be optimized using an attention mechanism.
In this research, an attention block, which can choose the critical part of the data sequence from the global information, was adopted for Bi-LSTM. Based on the influence on the TBM advance rate prediction result, the features generated by the Bi-LSTM were assigned the corresponding weights, i.e., the attention values of the features. Therefore, the effective generated features by the Bi-LSTM layer were addressed. The attention block could create the corresponding weights automatically, and the output of the attention block is shown in Equation (6).
ω i = j = 1 n α i j l j
The variable lj denotes the global features, αij denotes the weights of the features generated by the attention block, ωi is the output. Attention mechanism includes soft attention and hard attention. Hard attention means selecting the concerned area as input data to decrease the training cost. However, hard attention is not suitable for time series-related problems. Moreover, hard attention makes Bi-LSTM training non-differentiable, which leads to the difficulty of training increasing sharply. The input features generated by soft attention weights focus on specific areas. The training process using soft attention is differentiable, and the end-to-end network structure can be adopted. As a result, soft attention was employed. The schematic representation of Bi-LSTM is shown in Figure 3.

2.3. Normalization and Model Evaluation Metrics

Before training, the data normalization was adopted to scale the values of the data into [0, 1], as given by:
x n = x x m i n x m a x x m i n
where x is raw data and xn is the normalized data; xmax and xmin are the maximum and minimum data of the sequence, respectively.
The model was established using the training data and evaluated using the test data. Reverse normalization was also applied to process the prediction result. The reverse normalization process is given by:
y = y n y m a x y m i n + y m i n
where y is the real prediction result and yn is the result with normalization; ymax and ymin are the maximum and minimum prediction values of the sequence, respectively.
In this research, three model evaluation metrics, i.e., Root Mean Square Error (RMSE), Mean Absolute Error (MAE), and R-square (R2), were adopted to evaluate the model. The three model metrics are shown in Equations (9)–(11).
RMSE = 1 n i = 1 n y i y i 2
MAE = 1 n i = 1 n y i y i
R 2 = 1 i = 1 n y i y i 2 i = 1 n y i y ¯ i 2
where yi is the prediction result, y i is the ground truth, y ¯ i is the mean value of output data. According to these equations, when RMSE and MAE are lower, and R2 is higher, the prediction accuracy is higher.

3. Modeling and Results

3.1. Engineering Project Review

The TBM application engineering project was a subway construction project in Chengdu, China. The length of the subway tunnel was 43.186 km. There were 12 stations on the subway line. The maximum and minimum distances between the stations were 5.84 km and 1.46 km, respectively. The mean distance between adjacent stations was 3.58 km. The studied construction section was 4 km, and TBM was adopted for the whole section. The strata of the construction site were mainly comprised of medium-weathered sandstone and mudstone. The profile of the geological section is shown in Figure 4, indicating that the geological conditions are not complex and are easy and convenient for TMB construction.
The TBM operation data is recorded per second. The monitoring data contain 527 TBM operational parameters besides the ID and date. The TBM operational parameters include advance rate, cutter head torque, cutter head thrust, penetration, rotational speed, etc. For the study section, the monitored data of TBM operation is more than 20 GB, which is more than enough for training deep learning models.

3.2. Data Profile and Preprocessing

Limited by the length of the clamp cylinder either side of the TBM, the TBM continues working through the interaction of the bracket boots, gripper, and thrusting clamp cylinder. Meanwhile, the TBM should stop working to adjust working conditions or for repairs after a working cycle. The period may last for several hours or days. The advancing-regripping state will repeat in the whole TBM working process. The monitored data should include multiple working cycles and intervals in the data sequence, as shown in Figure 5. Therefore, the working periods between the TBM working cycles should be cut off when the TBM performance analysis is carried out. Only the monitored data during the working cycles are effective for TBM performance prediction.
In a working cycle of TBM construction, there are four stages for the whole process, i.e., (I) free rotation, (II) starting stage, (III) stable operation stage, (IV) stopping stage. The specific definitions are described in [32]. The starting stage and stable operation stage are the two stages in which TMB is working. The free rotation stages are short and should be ignored in TBM performance prediction because the stages are not stable for the working state.
In this research, the monitored construction data from Day 1 to Day 27, with a total of 176,959 records, were applied. The construction section was between 98 + 901 and 99 + 045, and the strata were mainly medium-weathered sandstone, as shown in Figure 6. The data distribution of date and position is shown in Table 1.
Some cases, such as operational mistakes, machine vibration, equipment error, and complex geological conditions, should negatively influence the data monitoring of TBM operation. Data monitoring errors may occur in the data recording process. As a result, it is necessary to remove the error from the monitoring data. In the process of TBM construction data recording, the common error is that the parameter values change sharply in a very short time. The errors often appear at the end of the working cycle and should be removed or replaced by the value of the previous recorded data. In a practical working cycle, it is impossible for the mechanical parameters to change in one or two seconds. However, the unstable working state leads to data recording errors. For example, Figure 7 shows the thrust (F) data in a working cycle of the TBM construction. At the end of the working cycle, the thrust changes to 0 from about 20,000 KN. The recorded data in red circle is regarded as an error, which is not beneficial for TBM performance analysis.
The advance rate is an important index in the TBM working process. It should be set first before the working cycle begins. The high-dimension non-linear relationship between the monitored data and the advance rate should be established. The monitored data includes the rotation speed of the cutter head (N), thrust (F), torque (T), penetration rate (P), and chamber earth pressure (Soil_P). The data structure is shown in Table 2. Table 3 and Figure 8 shows the specific variable information and distributions during the monitoring period. Table 3 shows the maximum, minimum, and medium values of the monitored data. It also shows the unit of the different types of parameters. Figure 6 shows the distributions of the different parameters. The six parameters are approximately normally distributed. It also shows the mean values and the standard deviations. The manual control of the advance rate is determined by the parameters. As a result, the advance rate should commensurate with other parameters during operation.

3.3. Model Establishment and Evaluation

In the research, the dataset was split into training data and test data. The TBM construction monitoring data of Day 27 was set as the test data. The number of test data points was 14,732, which was the last 14,732 number in the data sequence. The applied data is shown in Figure 9. All data sequences were split into training and test data.
The model structure included 64 LSTM units. It was composed of an input layer, a convolutional layer, a Bi-LSTM layer, two dropout layers, two dense layers, and two attention layers. The convolutional layer contained 448 parameters, the Bi-LSTM layer contained 66,048 parameters, and the two dense layers included 16,512 and 1025 parameters, respectively. The total number of training parameters was 84,033. Adam optimizer was adopted in model training. The MSE was selected as the loss function.
This research focused on the performance of the deep model regarding the massive amount of monitored data. Generally, more data means higher performance for deep learning models. However, because the LSTM-related models are time-dependent, it is crucial to determine the suitable data size for advance rate prediction. The data from Day 1 to Day 27 is applied, which was responding to the construction section of 98 + 901 to 99 + 045. The excavation distance was 144 m in total. To discuss the influences of the data sequence length on the model’s performance, different sizes of monitored data were applied to train the model. The five schemes are shown in Table 4.
The model was established using the Tensorflow and Keras framework. The CPU and GPU were Intel® Core™ i9-10900K and NVIDIA Geofroce 3060 with 12G memory, respectively. The number of units was 64. In the training process of the five models, the epoch was set to 100. In one epoch, all the training data were trained once. The model loss value should decrease with the training steps. Figure 10 shows the loss change in the training process for the five models.
In Figure 10, it can be seen that the loss values converge to a small value in the training process. In Scheme 1, the loss value converges to about 3.50. In Schemes 2–4, the loss values converge to about 7.00–7.50. It shows that the five models are stable. However, the five models should be tested on the monitored data of Dec 15th.
The five models were evaluated using both the training data and the test data. The RMSE, MAE, and R2 were adopted to evaluate the five models. The monitored TBM advance rate and the predicted values with time are shown in Figure 10. Because the data are massive, only the comparison of the test data and the predicted values are shown. The comparison of the model’s performances on the training and test data are shown in Table 5. The comparison of the monitored advance rate and the predicted values are shown in Figure 11.
In Figure 11, it can be seen that the models of Schemes 3–5 perform well. It denotes that unrelated data cannot promote the model performance. Meanwhile, for the deep learning models, the data size and the time should both be considered in the model establishment. In Scheme 1, although the data size is much bigger than that of other schemes, the evaluation results are worse than the others. The data number from Day 9 to Day 12 is 300 to 600. It denotes that the TBM is operated for 5–10 min. During this period, the TBM may be in a repair state. The data of the period should be ignored. In Schemes 3–5, the model of Scheme 4 outperforms the others. The construction data from Day 26 to Day 27 are set as the training data. The training data number is 26,039. In Scheme 2, the training data number is 46,433, while the model evaluation performance is worse than that of the models in Schemes 3–5. It denotes that the influence of time should be considered in data selection. Moreover, Figure 11 shows that the big prediction errors appear at the beginning of the working cycles. The result denotes that the working state of the TBM in the starting stage is still not stable and the errors increase when the TBM working state is not stable.
The predicted and monitored advance rate in the five schemes have been compared in Figure 12. ypredicted denotes the predicted advance rate; ymonitored denotes the monitored advance rate. It shows the R2 distribution of the five models, indicating that the model of Scheme 4 is the best.
The R2 values of the TBM advance rate are 0.443, 0.845, 0.920, 0.955 and 0.945, respectively. The different errors are given in different colors. Red means that the error is larger than 40. Yellow means that the error is in the range of (20, 40). Light blue means that the error is in the range of (10, 20). Blue means that the error is in the range of (0, 10). In Figure 11, it can be seen that the advance rate of the TBM is more than 30 mm/min. In most of the time, the advance rate is 40–50 mm/min in the working cycles. In other words, the advance rate values of 0–30 mm/min are in the starting stage. In the starting period, the TBM working state is not stable, and it negatively influences the prediction of the advance rate. As a result, the error increases when the advance rate is in the range of (0, 30). Especially when the advance rate is lower than 5 mm/min, the error is in the range of (40, 60). These results also denote that the unstable working state of the TBM negatively influences the prediction of advance rate.
A comparison of the proposed method and the traditional machine learning methods was made, including the support vector machine (SVM), random forest (RF), linear regression (LR), multilayer perceptron (MLP), and K-nearest neighbors (KNN). The scikit-learn was applied to build the models, and the parameters were chosen by the grid searching method. The results are shown in Table 6, and prove that the proposed method is effective for TBM advance rate prediction.
According to Table 6, the proposed model, i.e., CNN-Bi-LSTM-Attention, has the best performance of all models. KNN is an unsupervised algorithm, where only the Euler distances between samples are utilized; thus, it is difficult to achieve a good performance in such a complex task. It is also hard to catch the non-linear information of the data with LR, a classical linear model. Therefore, the R2 values of KNN and LR are less than 0.90. The R2 values of SVM, RF, and MLP are larger than 0.90, while they are still worse than the proposed model. The result shows that the proposed model can obtain effective information and realize an accurate prediction. Some pieces of research also support the result. Wang et al. [44] compared LSTM, RF, SVM, and DFN, and the result supports the above conclusion. Mokhtari and Mooney [45] established an SVM model with R2 from 0.88 to 0.95. Fu et al. [46] employed a neural network optimized by genetic algorithm to predict advance rate, with an R2 of 0.92. The result denotes that there are limitations of the shallow models in big data learning. Other researchers also applied RNN models in TBM parameter analysis. Gao et al. [36] applied RNN, LSTM, and GRU to analyze the monitored parameters, and the models achieved a high performance for torque, advance rate, and thrust prediction. Liu et al. [40] adopted LSTM and TBM real-time data to realize the tunnel lithology prediction with an accuracy of 96%. Gao et al. [41] applied an LSTM-based model to train the model using the machine parameters, rock mass parameters, and geological survey data. The result showed that the model is better than the RNN-based models as well as the autoregressive integrated moving average with explanation variables (ARIMAX) model. The data monitoring frequency is 1 Hz in TBM construction. LSTM-based models are suitable for the very large amount of monitored TBM data according to the references. The simulation result based on the subway construction data proves the conclusion. In future research, the robustness should be verified by testing the proposed model using further TBM construction data and other traditional methods.

4. Discussion

In this research, CNN-Bi-LSTM-Attention is proposed to predict the TBM advance rate. The five data sequences with different sizes were applied to build the TBM advance rate prediction model. The five models were compared, and the best training data sequence size was determined. The proposed model training time was about 400–700 s. The R2 of the proposed model could reach 0.955. The result denotes that the model could be used to predict the TBM advance rate. Meanwhile, training time is available in engineering.
The limitations of the proposed model lie in a type of preference during prediction. Figure 11 shows that advance rate data between 0 and 30 mm/min was much less than that of the advance rate between 30 and 70 mm/min. As a result, the performance of the two data sequences was different. Even for Schemes 4 and 5, it denotes that the fitting of performance data with the advance rate in [0, 30) was worse than that with the advance rate in [30, 70). The result denotes that big data is necessary for the proposed method. Meanwhile, according to the result of Schemes 1, 2, and 3, the data in an unstable construction state negatively affects the model performance. In other words, it is necessary to apply clean and big data to train the proposed model.
Moreover, the proposed model could be utilized in other research studies. However, the data number should be sufficient for model training. In the research, the model in Scheme 4 with 26,039 data was the best one. The performance of the model in Scheme 5 with 13,753 data decreased. The result highlights the importance of data amount. If the proposed method is applied to another construction, clean and enough data are required.

5. Conclusions

In tunnel construction using TBM, big data is always applied in parameter prediction. However, how long the data sequence must be to be regarded as big data, and what method to choose, are the two problems in engineering projects. In this research, the effectiveness of data length is discussed, and a deep learning model is proposed for TBM monitored data analysis. LSTM-based models have been applied widely in engineering data analysis recently, and it proves that the model is effective for prediction. As a result, the deep learning model based on the CNN architecture, bidirectional Long Short-Term Memory module, and attention mechanism can be adopted to predict the TBM advance rate. RMSE, MAE, and R2 were adopted to evaluate the model. The different sizes of the data sequence were applied to establish the models to verify that the data amount and data periods were both effective for the TBM advance rate. Through the comparison of the predicted values and the monitored values of the advance rate, it can be seen that the data in the starting stage cannot reflect the working state of the TBM, which may lead to prediction failure for the advance rate.
The results show that data quality is important in TBM advance rate prediction. The data in an unstable operation state is detrimental to the prediction performance. In Figure 12, it is clear that the predictions of the advance rate near 0 are very poor. The data period relates to the starting or stopping stage of TBM construction. The LSTM-based model cannot capture the information from the unstable construction state. This is because these states are too noisy, and they significantly affect the model’s performance. The results also showed that data amount is significant for TBM advance rate prediction. In Figure 12, it is clear that the prediction performance of the advance rate [0, 30) is worse than that of the advance rate [30, 70). The reason is that deep learning is a data-hungry technique. Because the data of [0, 30) are much less than the data of [30, 70), it is difficult for the model to reach a very high accuracy in this range. The proposed model outperforms all the models in the comparison with traditional machine learning methods. The results denote that it is suitable for the time-dependent data sequence because of the LSTM units. They are capable of extracting the time features to make predictions on advance rate with high accuracy.
In summary, this study showed the advantages of deep learning in the automatic analysis of the performance of TBM, and the proposed method is expected to be an auxiliary method for TBM operation, by providing an intelligent method for controlling advance rate. In the research, the model was only verified on the data from the subway construction. The rock mass is mainly composited of sandstone, which belongs to soft rock. In the future, the model should be tested on more data, especially using data from hard rock operations.

Author Contributions

Conceptualization, B.L.; methodology, Y.Z.; software, J.C.; validation, J.C.; formal analysis, S.H.; data curation, J.C.; writing—original draft, Y.Z.; writing—review, editing, supervision, Y.Z. and S.H.; funding acquisition, Y.Z. and S.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China (NSFC) (Grant No. 52009109), the PhD Research Startup Foundation of Xi’an University of Technology (Grant No. 104-451120005), and the Start-up Fund for RAPs under the Strategic Hiring Scheme of the Hong Kong Polytechnic University (Grant No. P0042478).

Data Availability Statement

Data generated and analyzed during this study are available from the corresponding author by request.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Yang, H.; Liu, X.; Song, K. A novel gradient boosting regression tree technique optimized by improved sparrow search algorithm for predicting TBM penetration rate. Arab. J. Geosci. 2022, 15, 1–15. [Google Scholar] [CrossRef]
  2. Yang, H.; Wang, H.; Zhou, X. Analysis on the rock–cutter interaction mechanism during the TBM tunneling process. Rock Mech. Rock Eng. 2016, 49, 1073–1090. [Google Scholar] [CrossRef]
  3. Zeng, J.; Roy, B.; Kumar, D.; Mohammed, A.S.; Armaghani, D.J.; Zhou, J.; Mohamad, E.T. Proposing several hybrid PSO-extreme learning machine techniques to predict TBM performance. Eng. Comput. 2021, 1–17. [Google Scholar] [CrossRef]
  4. Barton, N.R. TBM Tunnelling in Jointed and Faulted Rock; CRC Press: Boca Raton, FL, USA, 2000. [Google Scholar]
  5. Blindheim, O.T. Boreability Predictions for Tunneling; The Norwegian Institute of Technology: Trondheim, Norway, 1979. [Google Scholar]
  6. Bruland, A. Hard Rock Tunnel Boring; Fakultet for Ingeniørvitenskap og Teknologi: Trondheim, Norway, 2000. [Google Scholar]
  7. Sapigni, M.; Berti, M.; Bethaz, E.; Busillo, A.; Cardonee, G. TBM performance estimation using rock mass classifications. Int. J. Rock Mech. Min. Sci. 2002, 39, 771–788. [Google Scholar] [CrossRef]
  8. Kahraman, S.; Bilgin, N.; Feridunoglu, C. Dominant rock properties affecting the penetration rate of percussive drills. Int. J. Rock Mech. Min. Sci. 2003, 40, 711–723. [Google Scholar] [CrossRef]
  9. Hassanpour, J.; Rostami, J.; Khamehchiyan, M.; Bruland, A. Developing new equations for TBM performance prediction in carbonate-argillaceous rocks: A case history of Nowsood water conveyance tunnel. Geomech. Geoengin. Int. J. 2009, 4, 287–297. [Google Scholar] [CrossRef]
  10. Kasper, T.; Meschke, G. On the influence of face pressure, grouting pressure and TBM design in soft ground tunnelling. Tunn. Undergr. Space Technol. 2006, 21, 160–171. [Google Scholar] [CrossRef]
  11. Gong, Q.M.; Zhao, J. Development of a rock mass characteristics model for TBM penetration rate prediction. Int. J. Rock Mech. Min. Sci. 2009, 46, 8–18. [Google Scholar] [CrossRef]
  12. Li, X.; Li, H.; Liu, Y.; Zhou, Q.; Xia, X. Numerical simulation of rock fragmentation mechanisms subject to wedge penetration for TBMs. Tunn. Undergr. Space Technol. 2016, 53, 96–108. [Google Scholar] [CrossRef]
  13. Zhong, G.; Wang, L.-N.; Ling, X.; Dong, J. An overview on data representation learning: From traditional feature learning to recent deep learning. J. Financ. Data Sci. 2016, 2, 265–278. [Google Scholar] [CrossRef]
  14. Sun, W.; Shi, M.; Zhang, C.; Zhao, J.; Song, X. Dynamic load prediction of tunnel boring machine (TBM) based on heterogeneous in-situ data. Autom. Constr. 2018, 92, 23–34. [Google Scholar] [CrossRef]
  15. Mahdevari, S.; Shahriar, K.; Yagiz, S.; Shirazi, M.A. A support vector regression model for predicting tunnel boring machine penetration rates. Int. J. Rock Mech. Min. Sci. 2014, 72, 214–229. [Google Scholar] [CrossRef]
  16. Zhou, J.; Qiu, Y.; Zhu, S.; Armaghani, D.J.; Li, C.; Nguyen, H.; Yagiz, S. Optimization of support vector machine through the use of metaheuristic algorithms in forecasting TBM advance rate. Eng. Appl. Artif. Intell. 2021, 97, 104015. [Google Scholar] [CrossRef]
  17. Noori, A.M.; Mikaeil, R.; Mokhtarian, M.; Haghshenas, S.S.; Foroughi, M. Feasibility of intelligent models for prediction of utilization factor of TBM. Geotech. Geol. Eng. 2020, 38, 3125–3143. [Google Scholar] [CrossRef]
  18. Afradi, A.; Ebrahimabadi, A.; Hallajian, T. Prediction of TBM penetration rate using fuzzy logic, particle swarm optimization and harmony search algorithm. Geotech. Geol. Eng. 2022, 40, 1513–1536. [Google Scholar] [CrossRef]
  19. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  20. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT press: Cambridge, MA, USA, 2016. [Google Scholar]
  21. Bengio, Y.; Goodfellow, I.; Courville, A. Deep Learning; MIT press: Cambridge, MA, USA, 2017. [Google Scholar]
  22. Tian, C.; Fei, L.; Zheng, W.; Xu, Y.; Zuo, W.; Lin, C.W. Deep learning on image denoising: An overview. Neural Netw. 2020, 131, 251–275. [Google Scholar] [CrossRef]
  23. Wang, Z.; Chen, J.; Hoi, S.C.H. Deep learning for image super-resolution: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 3365–3387. [Google Scholar] [CrossRef]
  24. Salamon, J.; Bello, J.P. Deep convolutional neural networks and data augmentation for environmental sound classification. IEEE Signal Processing Lett. 2017, 24, 279–283. [Google Scholar] [CrossRef]
  25. Kim, J.; Min, K.; Jung, M.; Chi, S. Occupant behavior monitoring and emergency event detection in single-person households using deep learning-based sound recognition. Build. Environ. 2020, 181, 107092. [Google Scholar] [CrossRef]
  26. Majumder, N.; Poria, S.; Gelbukh, A.; Erik, C. Deep learning-based document modeling for personality detection from text. IEEE Intell. Syst. 2017, 32, 74–79. [Google Scholar] [CrossRef]
  27. Chatterjee, A.; Gupta, U.; Chinnakotla, M.K.; Srikanth, R.; Galley, M.; Agrawal, P. Understanding emotions in text using deep learning and big data. Comput. Hum. Behav. 2017, 93, 309–317. [Google Scholar] [CrossRef]
  28. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  29. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  30. Hinton, G.E.; Osindero, S.; Teh, Y.W. A fast learning algorithm for deep belief nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef]
  31. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef]
  32. Ren, Q.B.; Li, M.C.; Li, H.; Shen, Y. A novel deep learning prediction model for concrete dam displacements using interpretable mixed attention mechanism. Adv. Eng. Inform. 2021, 50, 101407. [Google Scholar] [CrossRef]
  33. Li, M.C.; Li, M.H.; Ren, Q.B.; Li, H.; Song, L.G. DRLSTM: A dual-stage deep learning approach driven by raw monitoring data for dam displacement prediction. Adv. Eng. Inform. 2022, 51, 101510. [Google Scholar] [CrossRef]
  34. Bai, S.; Li, M.C.; Lu, Q.R.; Tian, H.J.; Qin, L. Global Time Optimization Method for Dredging Construction Cycles of Trailing Suction Hopper Dredger Based on Grey System Model. J. Constr. Eng. Manag. 2022, 148, 04021198. [Google Scholar] [CrossRef]
  35. Zhang, P.; Yin, Z.Y.; Zheng, Y.Y.; Gao, F.P. A LSTM surrogate modelling approach for caisson foundations. Ocean. Eng. 2020, 204, 107263. [Google Scholar] [CrossRef]
  36. Gao, X.; Shi, M.; Song, X.; Zhang, C.; Zhang, H. Recurrent neural networks for real-time prediction of TBM operating parameters. Autom. Constr. 2019, 98, 225–235. [Google Scholar] [CrossRef]
  37. Feng, S.; Chen, Z.; Luo, H.; Wang, S.; Zhao, Y.; Liu, L.; Ling, D.; Jing, L. Tunnel boring machines (TBM) performance prediction: A case study using big data and deep learning. Tunn. Undergr. Space Technol. 2021, 110, 103636. [Google Scholar] [CrossRef]
  38. Chen, Z.; Zhang, Y.; Li, J.; Li, X.; Jing, L. Diagnosing tunnel collapse sections based on TBM tunneling big data and deep learning: A case study on the Yinsong Project, China. Tunn. Undergr. Space Technol. 2021, 108, 103700. [Google Scholar] [CrossRef]
  39. Graves, A. Long Short-Term Memory. Supervised Sequence Labelling with Recurrent Neural Networks; Springer: Berlin/Heidelberg, Germany, 2012; pp. 37–45. [Google Scholar] [CrossRef]
  40. Liu, Z.; Li, L.; Fang, X.; Qi, W.; Shen, J.; Zhou, H.; Zhang, Y. Hard-rock tunnel lithology prediction with TBM construction big data using a global-attention-mechanism-based LSTM network. Autom. Constr. 2021, 125, 103647. [Google Scholar] [CrossRef]
  41. Gao, B.; Wang, R.; Lin, C.; Guo, X.; Liu, B.; Zhang, W. TBM penetration rate prediction based on the long short-term memory neural network. Undergr. Space 2021, 6, 718–731. [Google Scholar] [CrossRef]
  42. Fu, X.; Zhang, L. Spatio-temporal feature fusion for real-time prediction of TBM operating parameters: A deep learning approach. Autom. Constr. 2021, 132, 103937. [Google Scholar] [CrossRef]
  43. Li, L.; Liu, Z.; Zhou, H.; Zhang, J.; Shen, W.; Shao, J. Prediction of TBM cutterhead speed and penetration rate for high-efficiency excavation of hard rock tunnel using CNN-LSTM model with construction big data. Arab. J. Geosci. 2022, 15, 1–17. [Google Scholar] [CrossRef]
  44. Wang, Q.; Xie, X.; Shahrour, I. Deep learning model for shield tunneling advance rate prediction in mixed ground condition considering past operations. IEEE Access 2020, 8, 215310–215326. [Google Scholar] [CrossRef]
  45. Mokhtari, S.; Mooney, M.A. Predicting EPBM advance rate performance using support vector regression modeling. Tunn. Undergr. Space Technol. 2020, 104, 103520. [Google Scholar] [CrossRef]
  46. Fu, X.; Gong, Q.; Wu, Y.; Zhao, Y.; Li, H. Prediction of EPB Shield Tunneling Advance Rate in Mixed Ground Condition Using Optimized BPNN Model. Appl. Sci. 2022, 12, 5485. [Google Scholar] [CrossRef]
Figure 1. Overall framework of CNN-Bi-LSTM-Attention model.
Figure 1. Overall framework of CNN-Bi-LSTM-Attention model.
Buildings 12 01567 g001
Figure 2. The structure of an LSTM block.
Figure 2. The structure of an LSTM block.
Buildings 12 01567 g002
Figure 3. Schematic representation of Bi-LSTM.
Figure 3. Schematic representation of Bi-LSTM.
Buildings 12 01567 g003
Figure 4. Geological conditions of the TBM construction section.
Figure 4. Geological conditions of the TBM construction section.
Buildings 12 01567 g004
Figure 5. Working cycles and intervals in TBM construction monitored data.
Figure 5. Working cycles and intervals in TBM construction monitored data.
Buildings 12 01567 g005
Figure 6. The different stages of TBM working cycle.
Figure 6. The different stages of TBM working cycle.
Buildings 12 01567 g006
Figure 7. Error recording in TBM working cycle.
Figure 7. Error recording in TBM working cycle.
Buildings 12 01567 g007
Figure 8. Distribution of the TBM construction parameters. (a) V (b) N (c) P (d) F (e) T (f) Soil_P.
Figure 8. Distribution of the TBM construction parameters. (a) V (b) N (c) P (d) F (e) T (f) Soil_P.
Buildings 12 01567 g008
Figure 9. The training and test monitored data of the TBM construction.
Figure 9. The training and test monitored data of the TBM construction.
Buildings 12 01567 g009aBuildings 12 01567 g009b
Figure 10. Loss changes in the training processes of all the training schemes.
Figure 10. Loss changes in the training processes of all the training schemes.
Buildings 12 01567 g010
Figure 11. The evaluation of the models in the five schemes.
Figure 11. The evaluation of the models in the five schemes.
Buildings 12 01567 g011aBuildings 12 01567 g011b
Figure 12. Comparison of the advance rate monitored and predicted.
Figure 12. Comparison of the advance rate monitored and predicted.
Buildings 12 01567 g012aBuildings 12 01567 g012b
Table 1. TBM monitored data distribution for date and chainage.
Table 1. TBM monitored data distribution for date and chainage.
DateChainageData Number
Day 198 + 901–98 + 9053985
Day 298 + 905–98 + 9084466
Day 398 + 908–98 + 91811,095
Day 998 + 918–98 + 918553
Day 1098 + 918–98 + 918353
Day 1198 + 918–98 + 918305
Day 1298 + 918–98 + 919345
Day 1798 + 919–98 + 9236014
Day 1898 + 923–98 + 93212,135
Day 1998 + 923–98 + 94011,487
Day 2098 + 940–98 + 95213,845
Day 2198 + 952–98 + 96414,169
Day 2298 + 964–98 + 97917,936
Day 2398 + 979–98 + 99519,106
Day 2498 + 995–99 + 0028744
Day 2599 + 002–99 + 01211,650
Day 2699 + 012–99 + 02212,286
Day 2799 + 022–99 + 03313,753
Day 2899 + 033–99 + 04514,732
Total-176,959
Table 2. Data structure.
Table 2. Data structure.
Advance Rate (V)NFTPSoil_P
0.1464842.0375442,829.4551,279.0060.0718930.646701
2.2703252.0308162,895.9041,147.5401.1179370.641276
2.2703252.0349392,943.1621,080.6931.1156720.666884
2.1973572.03320325,555.9903,005.4471,040.5851.080737
2.1973572.0366753,047.9971,029.4441.0788940.710067
1.6845702.0323353,088.0111,009.3900.8288840.735460
Table 3. Data application profile.
Table 3. Data application profile.
Operation ParametersMaxMinMediumUnit
Advance rate (V)112.3530.07346.451mm/min
Rotation speed of cutter head (N)2.2450.5101.800rpm
Penetration rate (P)120.9770.03725.889mm/r
Thrust (F)
Torque (T)
25,555.991,041.28716,995.810kN
8,834.944311.9535,275.055kN·m
Chamber earth pressure (Soil_P)2.1550.0300.772bar
Table 4. The data profile of the five models.
Table 4. The data profile of the five models.
Model SchemesDateSectionTraining Data NumberTraining Time (s)
1Day 1–Day 2798 + 901–99 + 033162,2274360
2Day 20–Day 2798 + 940–99 + 033111,4892994
3Day 24–Day 2798 + 995–99 + 03346,4331248
4Day 26–Day 2799 + 012–99 + 03326,039642
5Day 2799 + 022–99 + 03313,753396
Table 5. Performance of the five models.
Table 5. Performance of the five models.
ModelsData typeRMSEMAER2
Scheme 1 Training7.1202.3260.569
Test6.8292.3640.443
Scheme 2 Training3.8701.5630.845
Test3.8111.5880.826
Scheme 3 Training1.6061.1960.926
Test1.6651.1300.920
Scheme 4 Training1.4631.0280.960
Test1.6571.0440.955
Scheme 5 Training1.5341.1280.948
Test1.7641.1030.945
Table 6. Comparison with the traditional machine learning models.
Table 6. Comparison with the traditional machine learning models.
Data TypeRMSEMAER2
SVM3.3813.1110.901
RF1.9201.8230.951
LR4.1963.9490.845
MLP2.2521.7990.939
KNN6.5135.1540.492
CNN-Bi-LSTM-Attention1.6571.0440.955
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, Y.; Chen, J.; Han, S.; Li, B. Big Data-Based Performance Analysis of Tunnel Boring Machine Tunneling Using Deep Learning. Buildings 2022, 12, 1567. https://doi.org/10.3390/buildings12101567

AMA Style

Zhang Y, Chen J, Han S, Li B. Big Data-Based Performance Analysis of Tunnel Boring Machine Tunneling Using Deep Learning. Buildings. 2022; 12(10):1567. https://doi.org/10.3390/buildings12101567

Chicago/Turabian Style

Zhang, Ye, Jinqiao Chen, Shuai Han, and Bin Li. 2022. "Big Data-Based Performance Analysis of Tunnel Boring Machine Tunneling Using Deep Learning" Buildings 12, no. 10: 1567. https://doi.org/10.3390/buildings12101567

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop