Next Article in Journal
Extremal Solutions for Surface Energy Minimization: Bicubically Blended Coons Patches
Previous Article in Journal
Data-Driven RBFNN-Enhanced Model-Free Adaptive Traffic Symmetrical Signal Control for a Multi-Phase Intersection with Fast-Changing Traffic Flow
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Temperature-Induced Deflection Data Missing Recovery of a Cable-Stayed Bridge Based on Bayesian Robust Tensor Learning

1
YunJi Intelligent Engineering Co., Ltd., Shenzhen 518000, China
2
Key Laboratory of Concrete and Pre-Stressed Concrete Structures of the Ministry of Education, Southeast University, Nanjing 210096, China
3
Hong Kong-Zhuhai-Macao Bridge Authority, Zhuhai 519060, China
*
Authors to whom correspondence should be addressed.
Symmetry 2023, 15(6), 1234; https://doi.org/10.3390/sym15061234
Submission received: 23 May 2023 / Revised: 5 June 2023 / Accepted: 6 June 2023 / Published: 9 June 2023
(This article belongs to the Section Engineering and Materials)

Abstract

:
Changes in the deflection of cable-stayed bridges due to thermal effects may adversely affect the bridge structure and reflect the degradation of bridge performance. Therefore, complete deflection field data are important for bridge health monitoring. A strong linear correlation has been found between temperature-induced deflections in different positions of the same span of a cable-stayed bridge in many studies, which make the deflection data matrix/tensor have a low-rank structure. Therefore, it is appropriate to use a low-rank matrix/tensor learning to model the temperature–deflection field of a cable-stayed bridge. Moreover, to avoid disturbing the recovery results via abnormal data (e.g., baseline shift and outliers), a Bayesian robust tensor learning method is proposed to extract the spatio-temporal characteristics of the bridge temperature–deflection field. The missing data recovery and abnormal data cleaning are achieved simultaneously in the process of reconstructing the temperature-induced field via tensor learning. The performance of the method is verified with actual continuous monitoring data from a cable-stayed bridge. The experimental results show that low-order tensor (i.e., matrix) learning has a good recovery and cleaning performance. The extension to higher-order tensor learning is proposed to extract the spatial symmetry of the sensor locations, which is experimentally proven to have better missing recovery and abnormal data cleaning performance.

1. Introduction

Infrastructure construction around the world, especially large bridges, are constantly growing, as is the economy. The form of cable-stayed bridge is favored in the construction of long-span bridges due to their long span and flexibility [1]. Many new bridges built since the 1970s have been cable-stayed bridges [2]. However, damage to the main girder and cables of cable-stayed bridges can cause serious damage. For instance, a stay cable of the Maracaibo Lake Bridge in Venezuela suddenly broke due to severe corrosion of the anchor box, causing the bridge to partially collapse [3]. Therefore, the health monitoring of cable-stayed bridges is extremely important. The evolution rule of temperature-induced deflection in the main girder is an important indicator of the condition assessment of cable-stayed bridges [4], which directly reflects the deterioration of the bridge structural or the nonstructural components [5].
Structural health monitoring systems (SHMS) have been widely installed on large span bridges for 40 years [6]. Based on the monitoring data, many studies can be conducted on bridge temperature-induced deflections. Yang et al. [5] found a significant linear correlation existing between temperature-induced deflections at different positions of the same span of a cable-stayed bridge, which is an important feature of the deflection field of cable-stayed bridges. In addition, a significant correlation was found between the temperature-induced deflections and temperature of bridges [7,8]. Many scholars have also conducted research on structural condition assessments based on monitoring data and machine learning [9,10,11,12].
Although the SHMS can collect a large amount of data, there are inevitably missing data in the monitoring data due to the harsh environment where the SHMS operates, which can hinder the condition assessments of bridges [13]. Therefore, high-quality bridge temperature-induced deflection data are the key to using the evolution rule of temperature-induced deflections for bridge condition assessments. There are many machine learning methods currently being used for missing data recovery. Chen et al. [14] used functional data analysis techniques to analyze and model the inter-sensor relationship between two strain sensors located at different locations. The correlations were modeled in explicit ways and were used to recover the missing data. However, the model selection is time-consuming work in this method. Zhang et al. [15] recovered missing stress data by analyzing spatial correlations of different strain sensors. Deep learning is a new branch of machine learning, which has also been widely used for missing data recovery. Liu et al. [16] used LSTM networks to recover the missing structural temperature data from bridge SHM systems. Nevertheless, deep learning performs better in handling highly nonlinear data [17], which may not be the most suitable way to process existing significant linear correlations of the deflection dataset. Deep learning methods require high-quality data as a training dataset, which is not easily available in practice. However, a simple linear regression model cannot meet the accuracy requirements of many application scenarios (e.g., missing recovery). Thus, a more suitable and accurate method is needed to process temperature-induced deflections.
Tensor learning is a branch of machine learning used to process data in the format of a tensor, which has contributed to image processing, recommender systems and traffic data analysis [18,19,20,21,22]. Temperature-induced deflection data collected from sensors at different locations are a kind of spatio-temporal data, which can be organized in the format of a matrix (e.g., sensor × time) or a tensor. Moreover, the low-rank structure of the matrix/tensor of temperature-induced deflection data is easy to implement due to the strong linear correlation between the temperature-induced deflections at different locations. Therefore, low-rank tensor learning is an appropriate method to handle temperature-induced deflection field data. In addition, the actual monitoring data inevitably contain abnormal data (e.g., outliers and baseline shifts) except missing data, which can obscure the true information contained in the data and lead to false alerts [23]. The abnormal data contained in the data may interfere with the results of missing recovery. However, most current studies on missing data recovery do not simultaneously consider the presence of abnormal data. Therefore, there is an urgent need for a more robust data recovery method, namely recovering the missing data of a dataset containing abnormal data. To this end, this research inherits the modeling idea for Bayesian robust matrix factorization (BRMF) of Naiyan Wang and Dit-Yan Yeung [24]. The recovery of missing data and the cleaning of abnormal data can be achieved simultaneously via BRMF. Because of the full Bayesian framework, there are no problems of overfitting and manual parameter tuning existing in the optimization-based methods. However, BRMF cannot handle data with more than two modes of variation because of the low dimensionality of the matrix. To solve this problem, the BRMF model is extended to a high-order tensor so that the symmetry of the deflection field data due to the symmetry of the bridge structure and sensor locations can be captured to increase recovery accuracy.

2. Methodology

2.1. Problem Description

The bridge temperature-induced deflection data measured with multiple sensors can be organized in the format of a matrix or tensor [25]. For instance, the matrix Y R M × T can be used to represent temperature-induced deflection data containing missing and abnormal data, which are collected from M sensors on T time stamps. Data can be denoted by a tensor Y R M × N × T , which is a more flexible organized form that is able to represent “sensor types × sensors × time stamps”, “sensors in the same span×spans×time stamps” or ‘‘sensors × days × time stamps per day” [26]. The process of recovering missing data involves factorizing the original matrix/tensor under a Bayesian framework to obtain its low-rank approximation that does not contain missing or abnormal data [27].

2.2. Bayesian Robust Matrix Factorization

2.2.1. Full Bayesian Model

A partially observed spatio-temporal matrix Y R M × T can be decomposed into a spatial factor matrix U R K × M and a temporal factor matrix X R K × T , viz.,
Y U T X ,
where K denotes the rank of matrix Y . The column vector of U and X is u i R K and x t R K , respectively. Therefore, Equation (1) can be written at the elemental level as
y i , t u i T x t ,
An indicator matrix Ω = { ( i , t ) | y i , t   i s   o b s e r v e d } is defined to locate the position of the missing data. The conceptual diagram of the basic matrix factorization model is shown in Figure 1.
The standard matrix factorization model is a good method to recover missing data. However, when the data matrix is contaminated with abnormal data (e.g., outliers), the estimations may not match the ground truth significantly due to the sensitivity of the l 2 norm to the abnormal data [28]. A common solution is to use the l 1 norm instead of the l 2 norm [29,30]. However, the model with the l 1 loss function may still perform not robustly enough when dealing with massive abnormal data. To this end, a Laplace mixture with the generalized inverse Gaussian (GIG) distribution is considered to model the noise to enhance model robustness. Following the above ideas, the graphical model for the BRMF framework is built as shown in Figure 2. The model is entirely built on observed data in Ω , which means it can be trained on datasets with missing and abnormal values. Each component in this probabilistic graphic model is introduced below.
To perform the matrix factorization under the l 1 loss function, a Laplace distribution is assumed as the prior distribution of entries of the matrix Y , i.e.,
y i , t ~ L ( u i T x t , η i , t ) , ( i , t ) Ω ,
where η i , t is a spatially varying and temporally varying precision parameter modeling the noise of y i , t . To make the model more robust, the noise is modeled by a Laplace mixture with the generalized inverse Gaussian (GIG) distribution, which has been demonstrated in [31] and performs better than the l 1 norm when it is used to define a regularizer for variable selection. We expect it to be better when it is used as a loss function as well. η i , t is assumed to follow a generalized inverse Gaussian distribution, i.e.,
η i , t ~ G I G ( p , a , b ) ,
where G I G ( · ) represents the generalized inverse Gaussian distribution, and p , a , b are hyperparameters. However, there is a computational efficiency problem because the Laplace distribution is not easy to sample. An important property of Laplace distributions is used to solve this problem. When a random variable ε follows the Laplace distribution L ( ε | μ , λ 2 ) , the probability density function of ε is
p ( ε | μ , λ 2 ) = λ 2 2 exp ( λ 2 | ε μ | ) .
Then, the Laplace distribution can be equivalently written as an infinite Gaussian mixture with the exponential distribution, i.e.,
L ( ε | μ , λ 2 ) = 0 L ( ε | μ , τ ) E ( τ | λ 2 ) d τ ,
where E ( τ | β 2 ) denotes an exponential distribution. An important property is called the hierarchical view of the Laplace distribution. A matrix T R M × T whose entry τ i , t is the precision parameter of y i , t is defined to apply the hierarchical view of the Laplace distribution. The exponential prior is placed on τ i , t .
τ i , t ~ E ( η i , t 2 ) .
Thus, the prior distribution of y i , t should be changed from a Laplace distribution in Equation (3) to the following Gaussian distribution:
y i , t ~ 𝒩 ( u i T x t , τ i , t ) .
Due to the introduction of a hierarchical view, sampling from the Laplace distribution is converted to sampling from the Gaussian and exponential distributions, where they are much easier to sample than from the Laplace distribution. The computational efficiency is greatly improved by the hierarchical view of the Laplace distribution.
Then, the spatial factors are modeled. Each column vector u i of spatial factor U follows an i.i.d. multivariate Gaussian distribution:
u i ~ 𝒩 ( μ u , Λ u 1 ) ,
where μ u R K is a mean vector and Λ u R K × K is the precision matrix. We further place a conjugate Gaussian–Wishart prior on the parameters of u i , viz.,
μ u ~ 𝒩 ( μ 0 , ( β 0 Λ u ) 1 ) , Λ u ~ 𝒲 ( W 0 , v 0 ) ,
where μ 0 , β 0 , W 0 , v 0 are hyperparameters, 𝒲 ( · ) denotes the Wishart distribution with degrees of freedom v 0 and scale matrix W 0 R K × K .
The modeling process of temporal features is like that of spatial features. A multivariate Gaussian distribution prior is placed on the temporal feature vector x t :
x t ~ 𝒩 ( μ x , Λ x 1 ) .
Similarly, the conjugate Gaussian–Wishart distribution is placed on the parameters of the temporal feature, namely
μ x ~ 𝒩 ( μ 0 , ( β 0 Λ x ) 1 ) , Λ x ~ 𝒲 ( W 0 , v 0 ) .
where the mean vector μ x R K and the precision vector Λ x R K × K .
Note that the column vectors of both U and X can be sampled in parallel due to their independence from each other. This potential for parallelization is important for large-scale applications and computational efficiency.

2.2.2. Model Inference

Monte Carlo–Markov chain (MCMC) sampling is adopted by deriving the full conditional distributions for all parameters and hyperparameters to perform Bayesian inference. Due to the use of conjugate priors in Section 2.2.1, especially the corresponding posterior distribution of the mixture of exponential power distributions with a generalized inverse Gaussian also following a generalized inverse Gaussian distribution [30], the analytical solutions of all conditional distributions can be obtained. The detailed process of model inference is given below.
Sample spatial feature parameters  μ u   and   Λ u : Based on the property of conjugate priors, the joint posterior distribution of μ u   and   Λ u is the Gaussian–Wishart distribution as well:
p ( μ u , Λ u | U , W 0 , μ 0 , v 0 , β 0 ) = 𝒩 ( μ u | μ 0 , ( β 0 Λ u ) 1 ) 𝒲 ( Λ u | W 0 , v 0 ) ,
where
μ 0 = β 0 μ 0 + M u ¯ β 0 + M , β 0 = β 0 + M ,   v 0 = v 0 + M ,   ( W 0 ) 1 = W 0 1 + Σ ¯ u + β 0 M β 0 + M ( u ¯ μ 0 ) ( u ¯ μ 0 ) T .
x ¯ and Σ ¯ in Equation (14) are two parameters:
u ¯ = 1 M i = 1 M u i , Σ ¯ u = i = 1 M ( u i u ¯ ) ( u i u ¯ ) T .
The sampling process for μ x and Λ x is similar.
Sample spatial feature  u i : According to Bayes’ rule, the conditional distribution of u i related to the temporal factor X, the partially observed Y, the accuracy parameter τ i , t and all relevant hyperparameters is given as follows:
p ( u i | Y , X , μ u , Λ u , T ) = t = 1 T 𝒩 ( y i , t | u i T x t , τ i , t ) 𝒩 ( u i | μ u , ( Λ u ) 1 ) ,
where
Λ u = Λ u + t = 1 T x t x t T τ i , t ,   μ u = ( Λ u ) 1 ( t = 1 T y i , t x t τ i , t + Λ u μ u ) ,   ( i , t ) ϵ Ω .
The sampling process for X t is similar.
Sample  τ i , t : Based on Bayes’ rule, the posterior distribution of τ i , t is
p ( 1 τ i , t | Y , U , X , η i , t ) = I G ( 1 τ i , t | η i , t | r i , t | , η i , t ) i = 1 M t = 1 T 𝒩 ( y i , t | u i T x t , τ i , t ) E ( τ i , t | η i , t 2 ) ,
where
r i , t = y i , t u i T x t ,   ( i , t ) ϵ Ω .
Sample  η i , t : Based on the corresponding posterior distribution of the mixture of exponential power distributions with a generalized inverse Gaussian also follows a generalized inverse Gaussian distribution:
p ( η i , t | τ i , t , p , a , b ) = G I G ( η i , t | p , a , b ) E ( τ i , t | η i , t 2 ) G I G ( η i , t | p , a , b ) ,
where
p = p + 1 ,   a = τ i , t + a ,   b = b .
However, it is not convenient to sample directly from the generalized inverse Gaussian distribution. Thus, a form of the generalized inverse Gaussian distribution was chosen to sample. Specifically, we let p = −1/2, and the posterior distribution was converted to the inverse Gaussian distribution, which was more efficient to sample:
1 η i , t ~ I G ( τ i , t + a b , τ i , t + a ) ,

2.2.3. High-Order Extension to Bayesian Tensor Learning

Limited by the low dimensionality of the matrix, there is only one form of data organization. However, it has been shown that data often have more than two modes of variation [31,32,33]. As a result, the best form of data organization is tensor. In practice, the locations of sensor networks always have a certain spatial feature, such as symmetry. To reflect this spatial feature in the form of data organization, we extended BRMF to Bayesian robust tensor factorization (BRTF) by organizing the data as a a high-order tensor Y R M × N × T . The complete expansion process to BRTF is described below.
The Tucker decomposition and CP decomposition both can be used for tensor factorization. Because the result of Tucker decomposition may not be unique when the rank is fixed, the CP decomposition is adopted here, which is one of the most commonly used tensor factorizations that captures multi-linear structures [33]. CP decomposition represents a high-order tensor as a sum of several rank 1 tensors. As shown in Figure 3, the partially observed data Y Σ p = 1 p u p v p x p , where u p R M , v p R N , x p R T , are the rank 1 tensors of the pth row factor matrices U , V and X , respectively. P is the rank of the tensor. The entries in Y can be expressed as
y i , j , t ~ 𝒩 ( p = 1 P u i , p v j , p x t , p , τ i , j , t ) , ( i , j , t ) ϵ Ω ,
where Ω = { ( i , j , t ) | y i , j , t   i s   o b s e r v e d } is a third-order indicator tensor. Consistent with BRMF, the column vectors u i ,   x t and v j of three factor matrices in BRTF follow a multivariate Gaussian distribution, i.e.,
u i ~ 𝒩 ( μ u , Λ u 1 ) ,   v j ~ 𝒩 ( μ v , Λ v 1 ) ,   x t ~ 𝒩 ( μ x , Λ x 1 ) .
Regarding the parameters of u i , x t and v j , the conjugate Gaussian–Wishart priors are placed on them. Because there is only an addition factor matrix V in BRTF compared to BRMF, only the prior of the parameters of V is written here.
μ v ~ 𝒩 ( μ 0 , ( β 0 Λ v ) 1 ) , Λ v ~ 𝒲 ( W 0 , v 0 ) .
The posterior distribution of μ v and Λ v is
p ( μ v , Λ v | V , μ v , Λ v , T ) = n = 1 N 𝒩 ( v n | μ v , Λ v 1 ) 𝒩 ( μ v | μ 0 , ( β 0 Λ v ) 1 ) 𝒲 ( Λ v | W 0 , v 0 ) ,
where
μ 0 = β 0 μ 0 + N v ¯ β 0 + N , β 0 = β 0 + N ,   v 0 = v 0 + N ,   ( W 0 ) 1 = W 0 1 + Σ ¯ v + β 0 N β 0 + N ( v ¯ μ 0 ) ( v ¯ μ 0 ) T .
Similarly, v ¯ and Σ ¯ v are defined as follows:
v ¯ = 1 N j = 1 N v j , Σ ¯ v = j = 1 N ( v j v ¯ ) ( v j v ¯ ) T .
The priors of the parameters of u i and x t are the same as those in Equations (10) and (12), respectively. In addition, the posterior distribution of the parameters of u i is the same as those in Equations (13) and (14), respectively.
Because the partially observed data Y are organized as a third-order tensor rather than a matrix in BRMF, the posterior distributions of u i , v j and x t in BRTF are totally different. The posterior distribution of u i is given by u i ~ 𝒩 ( μ u , ( Λ u ) 1 ) , where
Λ u = Λ u + j , t W j , t W j , t T τ i , j , t ,   μ u = ( Λ u ) 1 ( j , t y i , j , t W j , t τ i , j , t + Λ u μ u ) , W j , t = v j x t , ( i , j , t ) ϵ Ω ,
⨀ denotes the Hadamard product. The posterior distributions over v j and x t have the same form as that in Equation (29). The posterior distribution of η i , j , t and τ i , j , t can be obtained by simply replacing the two-dimensional subscripts (i, t) with three-dimensional subscripts (i, j, t) in Equations (18), (19) and (22), i.e.,
1 η i , j , t ~ I G ( τ i , j , t + a b , τ i , j , t + a ) ,
1 τ i , j , t ~ I G ( η i , j , t | r i , j , t | , η i , j , t ) ,
where
r i , j , t = y i , j , t p = 1 p u i , p v j , p x t , p , ( i , j , t ) ϵ Ω .

2.3. Bridge and SHM System Description

The Tongling highway–railway bridge is a cable-stayed bridge that connects the two sides of the Yangtze River in Anhui Province. As shown in Figure 4, the bridge has a symmetric span configuration with a midspan of 630 m, a side span of 240 m, and an auxiliary span of 91.8 m. The main truss is an N-shaped truss structure. The cross section of the bridge is shown in Figure 5. The three main trusses are 17.1 m apart at the center with a 34.2 m width and 15.5 m height. The steel orthogonal anisotropic plate is adopted in the highway bridge; the railroad bridge deck is an orthogonal anisotropic integral deck with a ballast track structure.
The locations of the deflection sensors are shown in Figure 4. There are 12 deflection sensors placed in 10 sections of the bridge. As shown in Figure 5, Sections 1–5 and 7–11 are each installed with one deflection sensor; Section 6 is installed with two deflection sensors. The sampling frequency of the deflection sensors is 5 Hz. The raw data are shown in Figure 6a. The bridge girder deflection caused by vehicle loads is a dynamic response, whereas the temperature-induced deflection of the bridge girder is a quasi-static response. Therefore, the vehicle loads should be removed from the raw data. The raw data can be resampled at a rate of 5 min intervals to obtain the temperature-induced deflection, as shown in Figure 6b. The SHM system recorded monitoring data from March 2019 to August 2019. The temperature-induced deflection data measured by each sensor are shown in Figure 7. To prevent confusion, the temperature-induced deflections of the main span and side spans are plotted separately, as shown in Figure 7a,b.

3. Results

3.1. The Correlation between Temperature-Induced Deflections of Different Sensors

In this section, the correlation between temperature-induced deflections of different sensors is studied. The sensors can be divided into main span sensors (i.e., ND3~10) and side span sensors (i.e., ND1~2 and ND11~12). The correlation coefficients of the different deflections are shown in Figure 8 to quantitatively demonstrate the correlation between the deflections of different position. As shown in Figure 8, all the correlation coefficients of the deflections of the same span approximate to 1, which shows that there is a significant linear correlation between the deflections of the same span. In contrast, all the correlation coefficients of the deflections of different spans are significantly less than 1, which indicates that the correlation between the temperature-induced deflection of the main span and the side spans is weak.
Although there is a significant linear correlation between the temperature-induced deflections of different positions of the same span, the accuracy of simple linear regression is still not enough for some scenarios (e.g., missing recovery). For example, the temperature-induced deflections of ND3 are used to linearly fit the temperature-induced deflection of ND4. As shown in Figure 9a, the linear regression equation is given by y = 2.0950 x + 3.6987 . The fitted temperature-induced deflection data from ND4 can be given by using the linear regression equation. As shown in Figure 9b, the fitted temperature–deflection data from ND4 deviate significantly from the ground truth. To calculate the accuracy of the fitted results, the recovery loss rate ρ r is defined as the mean absolute error (MAE) between the estimations and ground truth, normalized by the absolute mean of the target value:
ρ r = 1 n i = 1 n | y i y i | 1 n i = 1 n | y i | ,
where y i and y i represent the ground truth and estimated values at the same position I, respectively, and n is the total number of missing values. The ρ r of linear regression’s fitted results for ND4 temperature–deflection data is as high as 19.39%, which indicates that the accuracy of linear regression cannot meet the requirements of missing recovery.

3.2. Experimental Verification

To present the details in pictures, the main span deflection data of July 2019 are used to test the performance of the proposed method. The deflection data of July 2019 are shown in Figure 10, where there exists an outlier in the data of ND6.

3.2.1. Abnormal Data Types of Cable-Stayed Bridge Deflection Field Data

As shown in Figure 11, the common types of abnormal data in structural health monitoring data are outliers, baseline shifts and noise. Outliers are points that deviate significantly from the overall trend from the perspective of statistical probability, as shown in Figure 11a. The specific characteristics of noise are shown in Figure 11b, which is a section of noise in the deflection data with the original sampling frequency of 5 Hz. As shown in Figure 11c, a baseline shift is a phenomenon in which part of the monitoring data significantly deviates from the overall trend in the time domain. However, the abnormally deviated part of the data retains a normal component in terms of pattern and trend.
The above three types of abnormal data are common in raw data. However, abnormal data types in temperature-induced deflection data are slightly different. Resampling is used to smooth raw data to obtain temperature-induced deflection data, which results in outliers and noise in the raw data being cleaned. However, the small segment baseline shift is transformed into an outlier by resampling. The data with the sampling frequency of 5 Hz in Figure 11 is resampled at a rate of 5 min intervals. The corresponding results are shown in Figure 12. A large segment (e.g., one day or a few days) baseline shift is still a baseline shift after resampling, which is easy to understand. In summary, there are only two common types of abnormal data in bridge temperature-induced deflection data, namely outliers and baseline shifts. Therefore, only outliers and baseline shifts are added to the data in subsequent experiments. Finally, the abnormal rate γ a is introduced into the experiments, which is the ratio of the amount of abnormal data to the total amount of data.
The cleaning loss rate ρ c is calculated using the mean absolute error (MAE) and is normalized by the absolute mean of the target values, as follows:
ρ c = 1 m j = 1 m | y j y j | 1 m j = 1 m | y j | ,
where y i and y i represent estimated values and the ground truth at position i, respectively.

3.2.2. Missing Recovery Based on Low-Order Tensor Learning

In the framework of low-order tensor (i.e., matrix) learning, deflection data containing missing and abnormal data can be organized as a matrix Y R M × T , where M is the number of the deflection sensors and T is the number of timestamps during a continuous monitoring period. In our experiments, the main span deflection data of July 2019 are organized as the matrix Y R 8 × 8928 In our experiments, the hyperparameters are initialized as follows: μ 0 = 0 , W 0 = I ,   β 0 = 2 , v 0 = K , a = 10 4 , b = 10 .
The missing data should be set in temperature-induced deflection data to test the performance of missing recovery. The missing rate γ m is introduced in the missing setting, which means the ratio of the amount of the missing data to the total number of the entries of the matrix/tensor. In the experimental design of [34], three types of missing data were set up, namely random missing, structural missing and mixed missing data. Random missing is a scenario in which each entry in the data matrix is set as missing randomly. Structural missing data are more common in actual monitoring data, where data are missing continuously due to sensor failure or other reasons. Mixed missing data consist of random missing and structural missing data. Considering that it is more common and challenging that large proportions of structural missing data occur to multiple sensors at the same time in real situations, only scenarios of large proportions of structural missing data are set in experiments. Specifically, multiple sensors are chosen to add large proportions of structural missing data at the same time.
To demonstrate that the proposed method can recover missing data if the data contain abnormal data, abnormal data should be set in the dataset. According to the analysis in Section 3.2.1, only outliers and baseline shifts are set in the bridge temperature–deflection data. Abnormal data are randomly set at any entry (except for missing entries) of the matrix. The specific scenarios are shown in Table 1. To leverage the low-rank structure of the deflection data matrix’s low-rank structure, rank K is set as 5 in the experiments. The results of ND4 are selected to present missing recovery results in each experimental scenario. Figure 13 shows the results of missing recovery in different scenarios, where the shaded area presents missing data.
As shown in Figure 13, the estimations match well with the ground truth in all scenarios despite the observed data containing a lot of abnormal data. It is worth mentioning that a large percentage of missing data can greatly decrease the number of observations, which makes the actual abnormal rate larger than the set abnormal rate, namely γ a . The loss rate in the four scenarios is 3.03%, 3.07%, 3.16% and 3.35%, respectively, which indicates that the high accuracy of the proposed method does not decrease significantly as the missing rate and the abnormal rate increase. In addition, the convergence of the model is very fast. The model converges after about 50 iterations in settled scenarios.

3.2.3. Missing Recovery Based on High-Order Tensor Learning

Effectively modeling and exploiting complex correlations across spatial and temporal dimensions is key to recovering missing data [35]. However, the spatial correlation may not be leveraged when the deflection data of different sections is organized in the format of a matrix. Note that the structure of the Tongling Bridge and the deflection sensor locations are symmetrical, which leads to symmetry between the deflection data measured by the symmetrical sensors. First, the symmetric sensor pair is introduced in the experiments, which refers to a pair of sensors arranged in a symmetrical position (e.g., ND3 and ND10). Then, a three-order tensor Y is used to characterize this spatial symmetry, which represents the “sensors in the same span × spans × time stamps” with a size of 8 × 2 × 8928. The horizontal slices of tensor Y are shown in Figure 14a, which are two matrices where white boxes represent missing data and red boxes represent abnormal data. Each vertical slice of tensor Y is a matrix formed by the data measured by a symmetric sensor pair, as shown in Figure 14b. To test the performance of tensor learning, we compare the ρ r and ρ c of the tensor learning results with those of the matrix learning results in the same missing scenarios. In the experiments, data from one or several symmetric sensor pairs are set as missing, and the effect of missing data in symmetric sensor pairs on the results of tensor learning and matrix learning is compared. The deflection data of the main span from July 2019 is used as the test dataset in this section. The rank of the matrix is set as 5, and the rank of the tensor is set as 10. The specific test scenarios are shown in Table 2. The relative recovery accuracy rate λ r and relative cleaning accuracy rate λ c are introduced to quantitatively compare the performance of tensor learning and matrix learning, which is defined by the following equation:
λ r = ρ r , m ρ r , t ρ r , m × 100 % ,
λ c = ρ c , m ρ c , t ρ c , m × 100 % ,
where ρ r , m is the recovery loss rate of matrix learning, ρ r , t is the recovery loss rate of tensor learning, ρ c , m is the cleaning loss rate of matrix learning and ρ c , t is the cleaning loss rate of tensor learning.
The results of tensor learning and matrix learning are shown in Figure 15. As shown in Figure 15a, it can be seen that the recovery accuracy of tensor learning is higher than that of matrix learning in all missing scenarios. However, the improvement of the accuracy of tensor learning (i.e., λ r ) decreases as the number of symmetric sensor pairs containing the missing data increases, as shown in Figure 15b. This is because the latent spatial symmetry in the data is broken severely when the number of symmetric sensor pairs containing missing data is too large. In conclusion, the experimental results show that tensor learning has significant advantages when the symmetric structure of the data is not badly broken. On the other hand, the cleaning accuracy of tensor learning is much higher than that of matrix learning, as shown in Figure 15c. As shown in Figure 15d, λ c does not decrease as the number of symmetric sensor pairs containing missing data increases like λ r , which shows the great advantage of high-order tensor learning.

4. Discussion and Conclusions

The Bayesian matrix/tensor learning for the missing recovery of temperature–deflection field data of cable-stayed bridge girders is presented in this paper. First, the correlation between temperature-induced deflections at different locations is studied. The results show that there is a strong linear correlation between temperature-induced deflections at different locations of the same span of a cable-stayed bridge, which makes the low-rank matrix/tensor learning method ideal for the missing recovery of temperature–deflection data. A Laplace mixture with a generalized inverse Gaussian distribution is used to enhance the robustness of the model to recover missing data while cleaning the abnormal data inevitably existing in real monitoring data. The experimental results show that the proposed method has satisfactory missing recovery and abnormal data cleaning performance. The extension of high-order tensor learning is proposed to characterize spatial symmetry in bridge structure and sensor arrangements. The experimental results show that tensor learning outperforms matrix learning for missing recovery in many scenarios.
The highlights of this paper are described below. First, low-rank tensor learning is proposed to model the temperature–deflection data of cable-stayed bridges with a high linear correlation, which is proven to be a very effective method. Second, a Laplace mixture with a generalized inverse Gaussian distribution is used to improve model robustness, which allows the model to recover missing data while cleaning abnormal data. Third, the Bayesian framework of the model avoids overfitting and manual tuning of the parameters. Last, the high-order extension to Bayesian tensor learning is proposed to characterize the symmetry of deflection data, which shows higher missing recovery accuracy and abnormal data cleaning accuracy than those of low-order tensor learning.
This study indicates that tensor learning has great potential for SHM applications. Some future research directions are proposed: first, the application of tensor learning to other bridge types can be investigated; second, the variational inference method with computational efficiency should not be ignored; and finally, applications of tensor learning in other areas of SHM, such as anomaly detection, should also be investigated.

Author Contributions

Methodology, S.S.; Software, Z.W.; Validation, Z.X. and Z.Y.; Data curation, Y.D.; Writing—original draft preparation, L.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially supported by the National Key Research and Development Program (2019YFB1600702), Shenzhen Technology Research Project (CJGJZD20210408092601005), and Program of the National Natural Science Foundation of China (51978154).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All the data, models or code that support the findings of this paper are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, L.X.; Qiu, G.Y.; Chen, Z.S. Structural health monitoring methods of cables in cable-stayed bridge: A review. Measurement 2021, 168, 108343. [Google Scholar] [CrossRef]
  2. Qin, S.Q.; Gao, Z.Y. Developments and Prospects of Long-Span High-Speed Railway Bridge Technologies in China. Engineering 2017, 3, 787–794. [Google Scholar] [CrossRef]
  3. De Rincon, O.; Contreras, D.; Sanchez, M.; Padilla, J.; Lasarte, C.; Sarcos, A. Inspection and repair of Lake Maracaibo Bridge suspension cables. Mater. Perform. 2001, 40, 34–38. [Google Scholar]
  4. Yue, Z.-X.; Ding, Y.-L.; Zhao, H.-W. Deep Learning-Based Minute-Scale Digital Prediction Model of Temperature-Induced Deflection of a Cable-Stayed Bridge: Case Study. J. Bridge Eng. 2021, 26, 05021004. [Google Scholar] [CrossRef]
  5. Yang, D.-H.; Yi, T.-H.; Li, H.-N.; Zhang, Y.-F. Correlation-Based Estimation Method for Cable-Stayed Bridge Girder Deflection Variability under Thermal Action. J. Perform. Constr. Facil. 2018, 32, 04018070. [Google Scholar] [CrossRef]
  6. Sun, L.; Shang, Z.; Xia, Y.; Bhowmick, S.; Nagarajaiah, S. Review of Bridge Structural Health Monitoring Aided by Big Data and Artificial Intelligence: From Condition Assessment to Damage Detection. J. Struct. Eng. 2020, 146, 04020073. [Google Scholar] [CrossRef]
  7. Chen, C.-C.; Wu, W.-H.; Liu, C.-Y.; Lai, G. Elimination of environmental temperature effect from the variation of stay cable force based on simple temperature measurements. Smart Struct. Syst. 2017, 19, 137–149. [Google Scholar] [CrossRef]
  8. Cao, Y.; Wang, M.L. Cable Stress Monitoring for a Cable Stayed Bridge. In Proceedings of the 5th European Workshop on Structural Health Monitoring, Naples, Italy, 29 June–4 July 2010; pp. 1196–1201. [Google Scholar]
  9. Barkhordari, M.S.; Armaghani, D.J.; Asteris, P.G. Structural damage identification using ensemble deep convolutional neural network models. Comput. Model. Eng. Sci. 2022, 134, 835–855. [Google Scholar] [CrossRef]
  10. Barkhordari, M.S.; Barkhordari, M.M.; Armaghani, D.J.; Rashid, A.S.A.; Ulrikh, D.V. Hybrid Wavelet Scattering Network-Based Model for Failure Identification of Reinforced Concrete Members. Sustainability 2022, 14, 12041. [Google Scholar] [CrossRef]
  11. Barkhordari, M.S.; Tehranizadeh, M. Data-driven Dynamic-classifiers-based Seismic Failure Mode Detection of Deep Steel W-shape Columns. Period. Polytech. Civ. Eng. 2023. [Google Scholar] [CrossRef]
  12. Barkhordari, M.S.; Jawdhari, A. Machine learning based prediction model for plastic hinge length calculation of reinforced concrete structural walls. Adv. Struct. Eng. 2023. [Google Scholar] [CrossRef]
  13. Jiang, H.; Wan, C.; Yang, K.; Ding, Y.; Xue, S. Continuous missing data imputation with incomplete dataset by generative adversarial networks-based unsupervised learning for long-term bridge health monitoring. Struct. Health Monit. Int. J. 2021, 21, 1093–1109. [Google Scholar] [CrossRef]
  14. Chen, Z.; Li, H.; Bao, Y. Analyzing and modeling inter-sensor relationships for strain monitoring data and missing data imputation: A copula and functional data-analytic approach. Struct. Health Monit. 2018, 18, 1168–1188. [Google Scholar] [CrossRef]
  15. Zhang, Z.Y.; Luo, Y.Z. Restoring method for missing data of spatial structural stress monitoring based on correlation. Mech. Syst. Signal Process. 2017, 91, 266–277. [Google Scholar] [CrossRef]
  16. Liu, H.; Ding, Y.-L.; Zhao, H.-W.; Wang, M.-Y.; Geng, F.-F. Deep learning-based recovery method for missing structural temperature data using LSTM network. Struct. Monit. Maint. 2020, 7, 109–124. [Google Scholar]
  17. Ye, X.W.; Jin, T.; Yun, C.B. A review on deep learning-based structural health monitoring of civil infrastructures. Smart Struct. Syst. 2019, 24, 567–585. [Google Scholar]
  18. Ji, Y.W.; Wang, Q.; Li, X.; Liu, J. A Survey on Tensor Techniques and Applications in Machine Learning. IEEE Access 2019, 7, 162950–162990. [Google Scholar] [CrossRef]
  19. Nie, X.; Yin, Y.; Sun, J.; Liu, J.; Cui, C. Comprehensive Feature-Based Robust Video Fingerprinting Using Tensor Model. IEEE Trans. Multimed. 2017, 19, 785–796. [Google Scholar] [CrossRef]
  20. Hu, W.R.; Tao, D.C.; Zhang, W.S.; Xie, Y.; Yang, Y.H. The Twist Tensor Nuclear Norm for Video Completion. IEEE Trans. Neural Netw. Learn. Syst. 2017, 28, 2961–2973. [Google Scholar] [CrossRef]
  21. Xie, K.; Wang, L.; Wang, X.; Xie, G.; Wen, J.; Zhang, G.; Cao, J.; Zhang, D. Accurate Recovery of Internet Traffic Data: A Sequential Tensor Completion Approach. IEEE/Acm Trans. Netw. 2018, 26, 793–806. [Google Scholar] [CrossRef]
  22. Zheng, C.; Song, H.E.M.; Song, J. CMPTF: Contextual Modeling Probabilistic Tensor Factorization for recommender systems. Neurocomputing 2016, 205, 141–151. [Google Scholar] [CrossRef]
  23. Bao, Y.Q.; Chen, Z.C.; Wei, S.Y.; Xu, Y.; Tang, Z.Y.; Li, H. The State of the Art of Data Science and Engineering in Structural Health Monitoring. Engineering 2019, 5, 234–242. [Google Scholar] [CrossRef]
  24. Wang, N.; Yeung, D.-Y. Bayesian Robust Matrix Factorization for Image and Video Processing. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013. [Google Scholar]
  25. Chen, X.; Sun, L. Bayesian Temporal Factorization for Multidimensional Time Series Prediction. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 4659–4673. [Google Scholar] [CrossRef]
  26. Ren, P.; Chen, X.; Sun, L.; Sun, H. Incremental Bayesian matrix/tensor learning for structural monitoring data imputation and response forecasting. Mech. Syst. Signal Process. 2021, 158, 107734. [Google Scholar] [CrossRef]
  27. Zhao, Q.B.; Zhang, L.Q.; Cichocki, A. Bayesian CP Factorization of Incomplete Tensors with Automatic Rank Determination. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1751–1763. [Google Scholar] [CrossRef] [Green Version]
  28. Peng, Y.G.; Ganesh, A.; Wright, J.; Xu, W.L.; Ma, Y. RASL: Robust Alignment by Sparse and Low-Rank Decomposition for Linearly Correlated Images. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2233–2246. [Google Scholar] [CrossRef] [PubMed]
  29. Zhang, Z.D.; Liang, X.; Ganesh, A.; Ma, Y. TILT: Transform Invariant Low-Rank Textures. In Proceedings of the 10th Asian Conference on Computer Vision (ACCV), Queenstown, New Zealand, 8–12 November 2010; Volume 6494, pp. 314–337. [Google Scholar]
  30. Zhang, Z.; Wang, S.; Liu, D.; Jordan, M.I. EP-GIG Priors and Applications in Bayesian Sparse Learning. J. Mach. Learn. Res. 2012, 13, 2031–2061. [Google Scholar]
  31. Kolda, T.G.; Bader, B.W. Tensor Decompositions and Applications. SIAM Rev. 2009, 51, 455–500. [Google Scholar] [CrossRef]
  32. Acar, E.; Yener, B. Unsupervised Multiway Data Analysis: A Literature Survey. IEEE Trans. Knowl. Data Eng. 2009, 21, 6–20. [Google Scholar] [CrossRef]
  33. Acar, E.; Dunlavy, D.M.; Kolda, T.G.; Mørup, M. Scalable tensor factorizations for incomplete data. Chemom. Intell. Lab. Syst. 2011, 106, 41–56. [Google Scholar] [CrossRef] [Green Version]
  34. Chen, X.; He, Z.; Wang, J. Spatial-temporal traffic speed patterns discovery and incomplete data recovery via SVD-combined tensor decomposition. Transp. Res. Part C Emerg. Technol. 2018, 86, 59–77. [Google Scholar] [CrossRef]
  35. Bahadori, M.T.; Yu, Q.; Liu, Y. Fast Multivariate Spatio-temporal Analysis via Low Rank Tensor Learning. In Proceedings of the 28th Conference on Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 8–13 December 2014; Volume 27. [Google Scholar]
Figure 1. Conceptual diagram of the basic matrix decomposition model. White boxes represent missing data, gray boxes denote normal observed data, and red boxes denote abnormal data.
Figure 1. Conceptual diagram of the basic matrix decomposition model. White boxes represent missing data, gray boxes denote normal observed data, and red boxes denote abnormal data.
Symmetry 15 01234 g001
Figure 2. Probabilistic graphic model for Bayesian robust matrix factorization.
Figure 2. Probabilistic graphic model for Bayesian robust matrix factorization.
Symmetry 15 01234 g002
Figure 3. Schematic diagram of the third-order tensor CP decomposition.
Figure 3. Schematic diagram of the third-order tensor CP decomposition.
Symmetry 15 01234 g003
Figure 4. Elevation of the bridge, and deflection sensor locations.
Figure 4. Elevation of the bridge, and deflection sensor locations.
Symmetry 15 01234 g004
Figure 5. Deflection sensor locatiosn in cross sections: (a) Sections 1–5, 7–11; (b) Section 6.
Figure 5. Deflection sensor locatiosn in cross sections: (a) Sections 1–5, 7–11; (b) Section 6.
Symmetry 15 01234 g005
Figure 6. Girder deflection–time curves: (a) Raw data; (b) 5 min average.
Figure 6. Girder deflection–time curves: (a) Raw data; (b) 5 min average.
Symmetry 15 01234 g006
Figure 7. Temperature-induced deflection measured by each sensor: (a) Temperature-induced deflections of main span; (b) Temperature-induced deflection of side spans.
Figure 7. Temperature-induced deflection measured by each sensor: (a) Temperature-induced deflections of main span; (b) Temperature-induced deflection of side spans.
Symmetry 15 01234 g007
Figure 8. Correlation coefficients between deflections of different positions.
Figure 8. Correlation coefficients between deflections of different positions.
Symmetry 15 01234 g008
Figure 9. Linear regression of main span deflections between sensor ND3 and ND4: (a) Linear regression of deflection data from ND3 and ND4; (b) Comparison of fitted temperature-induced deflection data and ground truth of ND4.
Figure 9. Linear regression of main span deflections between sensor ND3 and ND4: (a) Linear regression of deflection data from ND3 and ND4; (b) Comparison of fitted temperature-induced deflection data and ground truth of ND4.
Symmetry 15 01234 g009
Figure 10. Deflection data of July 2019, where there exists an outlier in the data of ND6.
Figure 10. Deflection data of July 2019, where there exists an outlier in the data of ND6.
Symmetry 15 01234 g010
Figure 11. Three common types of abnormal data in health monitoring data: (a) Outliers; (b) Noise (c) Baseline shifts.
Figure 11. Three common types of abnormal data in health monitoring data: (a) Outliers; (b) Noise (c) Baseline shifts.
Symmetry 15 01234 g011
Figure 12. Three common types of abnormal data after resampling: (a) Outliers after resampling; (b) Noise after resampling; (c) Baseline shift after resampling.
Figure 12. Three common types of abnormal data after resampling: (a) Outliers after resampling; (b) Noise after resampling; (c) Baseline shift after resampling.
Symmetry 15 01234 g012
Figure 13. Recovery results of main span sensors: (a) Scenario 1; (b) Scenario 2; (c) Scenario 3; (d) Scenario 4.
Figure 13. Recovery results of main span sensors: (a) Scenario 1; (b) Scenario 2; (c) Scenario 3; (d) Scenario 4.
Symmetry 15 01234 g013
Figure 14. Horizontal and vertical slices of tensor Y : (a) Horizontal slices of tensor Y ; (b) Vertical slices of tensor Y . White boxes represent missing data, gray boxes denote normal observed data, and red boxes denote abnormal data.
Figure 14. Horizontal and vertical slices of tensor Y : (a) Horizontal slices of tensor Y ; (b) Vertical slices of tensor Y . White boxes represent missing data, gray boxes denote normal observed data, and red boxes denote abnormal data.
Symmetry 15 01234 g014
Figure 15. Comparison of the results of tensor learning and matrix learning: (a) Recovery loss rate; (b) Relative recovery accuracy rate; (c) Cleaning loss rate; (d) Relative cleaning accuracy rate.
Figure 15. Comparison of the results of tensor learning and matrix learning: (a) Recovery loss rate; (b) Relative recovery accuracy rate; (c) Cleaning loss rate; (d) Relative cleaning accuracy rate.
Symmetry 15 01234 g015
Table 1. Scenarios of large proportions of structural missing data.
Table 1. Scenarios of large proportions of structural missing data.
Scenario γ a γ m Sensors Containing Missing Data
110%50%ND4, ND5, ND7, ND9
210%50%ND4, ND6, ND8, ND9
310%50%ND4 ND5, ND7, ND9, ND10
415%50%ND4 ND5, ND7, ND9, ND10
Table 2. Missing scenarios for the comparison experiments of tensor learning and matrix learning.
Table 2. Missing scenarios for the comparison experiments of tensor learning and matrix learning.
Scenario γ a γ m Sensors Containing Missing DataNumber of Symmetric Sensor Pairs Containing Missing Data
110%50%ND4, ND5, ND6, ND71
210%50%ND6, ND7, ND8, ND91
310%50%ND4, ND6, ND8, ND91
410%50%ND4, ND6, ND8, ND9, ND101
510%50%ND4, ND5, ND7, ND91
610%50%ND4, ND5, ND7, ND9, ND101
710%50%ND4, ND5, ND8, ND92
810%50%ND4, ND5, ND6, ND8, ND92
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sun, S.; Wang, Z.; Xia, Z.; Yi, L.; Yue, Z.; Ding, Y. The Temperature-Induced Deflection Data Missing Recovery of a Cable-Stayed Bridge Based on Bayesian Robust Tensor Learning. Symmetry 2023, 15, 1234. https://doi.org/10.3390/sym15061234

AMA Style

Sun S, Wang Z, Xia Z, Yi L, Yue Z, Ding Y. The Temperature-Induced Deflection Data Missing Recovery of a Cable-Stayed Bridge Based on Bayesian Robust Tensor Learning. Symmetry. 2023; 15(6):1234. https://doi.org/10.3390/sym15061234

Chicago/Turabian Style

Sun, Shouwang, Zhiwen Wang, Zili Xia, Letian Yi, Zixiang Yue, and Youliang Ding. 2023. "The Temperature-Induced Deflection Data Missing Recovery of a Cable-Stayed Bridge Based on Bayesian Robust Tensor Learning" Symmetry 15, no. 6: 1234. https://doi.org/10.3390/sym15061234

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop