Next Article in Journal
Research on the Soil-Plugging Effect on Small-Diameter Jacked Piles through In Situ Testing and DEM Simulation
Next Article in Special Issue
Response Characteristics of Pre-Stressed Strand Cables Subjected to Low-Velocity Impact: Experiment Test
Previous Article in Journal
Post-Earthquake Assessment and Strengthening of a Cultural-Heritage Residential Masonry Building after the 2020 Zagreb Earthquake
Previous Article in Special Issue
Rapid Reconstruction of 3D Structural Model Based on Interactive Graph Cuts
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dynamic Characteristic Monitoring of Wind Turbine Structure Using Smartphone and Optical Flow Method

1
Institution of Earthquake Protection and Disaster Mitigation, Lanzhou University of Technology, Lanzhou 733050, China
2
International Research Base on Seismic Mitigation and Isolation of GANSU Province, Lanzhou University of Technology, Lanzhou 733050, China
3
Disaster Prevention and Mitigation Engineering Research Center of Western Civil Engineering, Lanzhou University of Technology, Lanzhou 733050, China
*
Author to whom correspondence should be addressed.
Buildings 2022, 12(11), 2021; https://doi.org/10.3390/buildings12112021
Submission received: 18 October 2022 / Revised: 31 October 2022 / Accepted: 16 November 2022 / Published: 18 November 2022
(This article belongs to the Special Issue Damage Detection Based on Smartphones in Buildings)

Abstract

:
The dynamic characteristics of existing wind turbine structures are usually monitored using contact sensors, which is not only expensive but also time-consuming and laborious to install. Recently, computer vision technology has developed rapidly, and monitoring methods based on cameras and UAVs (unmanned aerial vehicles) have been widely used. However, the high cost of UAVs and cameras make it difficult to widely use them. To address this problem, a target-free dynamic characteristic monitoring method for wind turbine structures using portable smartphone and optical flow method is proposed by combining optical flow method with robust corner feature extraction in ROI (region of interest). Firstly, the ROI region clipping technology is introduced after the structural vibration video shooting, and the threshold value is set in the ROI to obtain robust corner features. The sub-pixel displacement monitoring is realized by combining the optical flow method. Secondly, through three common smartphone shooting state to monitor the structural displacement, the method of high pass filtering combined with adaptive scaling factor is used to effectively eliminate the displacement drift caused by the two shooting states of standing and slightly walking, which can meet the requirements of structural dynamic characteristics monitoring. After that, the structural displacement is monitored by assembling the telephoto lens on the smartphone. The accuracy of displacement monitored by assembling the telephoto lens on the smartphone is investigated. Finally, the proposed monitoring method is verified by the shaking table test of the wind turbine structure. The results show that the optical flow method, combined with smartphones, can accurately identify the dynamic characteristics of the wind turbine structure, and the smartphone equipped with a telephoto lens is more conducive to achieving low-cost wind turbine structure dynamic characteristics monitoring. This research can provide a reference for evaluating the condition of wind turbine structures.

Graphical Abstract

1. Introduction

As wind energy stands out among many renewable energy sources, people are paying more and more attention to the safety performance of wind turbine structures [1,2]. The amount of wind energy obtained is related to wind turbine blades. In order to obtain more wind energy and generate more power, the size of wind turbine blades has grown exponentially in recent years [3,4]. Wind turbine blades may be defective or damaged due to production defects, turbulent winds, lightning, irregular loads, etc. [5], which may lead to surface changes that affect the aerodynamic efficiency of the blades [6,7], thus causing serious safety problems. As the main component of wind turbine to capture wind energy, the power generation efficiency and safety of wind turbine structure mainly depends on the health status of the blades. To ensure that the wind turbine structure will not be damaged, thus effectively reducing economic losses, it is of great significance to monitor the vibration response of the wind turbine structure.
The traditional wind turbine detection is mainly manual, which is not only costly and unsafe, but also requires the experience of engineers, so it is not suitable for efficient detection of wind turbine structures. In addition, ultrasonic, thermal imaging, telescope, and other equipment are often used for wind turbine structure monitoring [8,9]. However, due to its high cost and the high professional level required for the use of instruments, it has not been widely used in the detection of wind turbine structures. Damage detection based on vibration analysis is mainly carried out by changing dynamic characteristic parameters [10,11]. Therefore, it is meaningful to identify dynamic characteristic parameters of wind turbine structure through vibration data. Civil engineering monitoring mainly uses contact sensors to monitor dynamic characteristics [12]. However, the disadvantage of using sensors is that they are fixed to the surface of the structure. This not only causes damage to the structure, but also substantially modifies the structure mode due to the mass load effect [13]. These wind turbines are huge structures, so conventional methods cannot be used for monitoring. Some wind turbines operate in complex natural environments, including wind, salt fog, rain, etc. Thus, the sensor-based wind turbine structure monitoring faces different types of challenges, such as sensor damage, extensive wiring, and labor intensity. Data acquisition is still challenging despite the large number of non-contact sensors emerging in structural health monitoring. Therefore, monitoring of the structural dynamic characteristics has not seen its popularity [14].
Computer vision-based methods for structural health monitoring have been proposed and applied in practice by many scholars [15,16,17,18]. Engineers favor computer vision technology with application advantages, such as non-contact, long distance, fast, low cost, and low labor to routine operations. Feng et al. [19] performed a vision-based measurement of dense full-field displacement with simply supported beams using the template-matching algorithm. It was verified practically when the trains were passing though the bridge. Dong et al. [20] used feature point extraction and optical flow tracking algorithm to identify dynamic characteristics of the stadium stand compared with contact sensors. This overcame the problem of small sensor monitoring area to achieve multi-point measurement. Khadka et al. [21] used digital image correlation (DIC) method to perform target tracking on wind turbine blades and used marker points to identify dynamic characteristics of wind turbine blades. Song et al. [22] perfectly segmented the background through the depth learning full convolution network (FCN) and conditional random field (CRF), and then used DIC to measure the displacement. The feasibility of this method was verified by experiments under different illumination conditions. Through the research of a large number of scholars, computer vision is feasible as a supplement to traditional structural health monitoring.
A large number of structural dynamic characteristics identification methods based on computer vision use cameras or unmanned aerial vehicles. However, because of the high cost of cameras or UAVs, low-cost equipment is needed to monitor the structural dynamic characteristics through the same imaging difference with the camera. In recent years, smartphones have developed on an unprecedented scale and can now be used as an effective measurement tool in structural health monitoring [23]. Most researchers use the internal accelerometers of smartphones in combination with the actual structure for structural health monitoring [24,25,26]. However, such methods require tying bind smartphones to structures and the smartphone can only monitor data from one point, so they are greatly limited in structural dynamic characteristics. Zhao et al. [27] developed an APP: D-viewer that can monitor bridge displacements by using color-matching algorithm and smartphones, and conducted static and dynamic tests on the bridge. However, the article does not consider the complex backgrounds, and the monitoring is not effective under lighting conditions, which is not conducive to long-term monitoring. Zhao et al. [28] proposed a new visual cable force measurement method based on smartphone cameras, and preliminarily verified its feasibility and practicability through cable model tests. Li et al. [29] provided the possibility of crowd perception of all buildings in urban areas after the earthquake with smartphone-based monitoring technology. Ozer et al. [30] introduced the concept of smartphone structural health monitoring and constructed a hybrid structure vibration response measurement framework by using multi-sensor smartphone functions. A novel hybrid motion sensing platform has been successfully implemented through the integration of various sensor types and devices. Wang et al. [31] developed an automatic damage detection system based on smartphones which can realize real-time damage detection of masonry buildings through experimental verification. The research on the monitoring of structural dynamic characteristics by cheap and commonly used smartphones was initially minimal. It is still limited to the monitoring of artificial markers and single environments at present. Therefore, it is necessary to explore a smartphone-based method for monitoring the dynamic characteristics of wind turbine structures which is suitable for low-cost, target-free structural dynamic characteristics monitoring.
This paper focuses on the research of monitoring the dynamic characteristics of wind turbine blades based on the combination of computer vision and smartphones. In Section 2, a smartphone-based target-free wind turbine structure dynamic characteristics monitoring method was proposed using the optical flow method and ROI clipping. After shooting the structural vibration videos, ROI was selected to set the threshold value to obtain robust corner features and combined with the optical flow method to achieve sub-pixel displacement monitoring. In Section 3, camera calibration and vibration tests were conducted for different models of smartphones to verify the feasibility of smartphone monitoring. Through the analysis of three common smartphone shooting states to monitor the structural displacement, the method of high pass filtering combined with adaptive scaling factor was used to effectively eliminate the displacement drift caused by the two shooting states of standing and slightly walking to monitor the structural displacement. After that, the structural displacement was monitored by assembling the telephoto lens on the smartphone, and the accuracy of displacement monitored by assembling the telephoto lens on the smartphone was investigated. In Section 4, the proposed monitoring method was verified by using a small shaking table test. The results show that computer vision combined with smartphones can accurately identify the natural frequency of the wind turbine structure, and that a smartphone equipped with a telephoto lens is more conducive to achieving low-cost wind turbine structure dynamic characteristics monitoring. Finally, the modal shape of wind turbine structure is obtained. Section 5 summarizes the research content of this paper.

2. Dynamic Displacement Monitoring Based on Computer Vision

Computer vision detects, extracts, recognizes, and tracks moving objects in image sequences to obtain moving object parameters. The dynamic characteristics monitoring of wind turbine structures using smartphones and visual algorithms consists of four parts: camera calibration, feature recognition, target tracking, and jitter processing and displacement calculation.

2.1. Camera Calibration

In recent years, the lenses of modern consumer cameras have been significantly improved. However, cameras tend to be smaller and more convenient. Generally, these cameras use wide-angle lenses. Smartphone cameras increase the field of vision by introducing significant radial distortion. In order to eliminate this distortion and obtain accurate displacement measurements with the consumer camera, the camera must be corrected.
Generally, the calibration process is divided into two steps. The first step is to convert the world coordinate system to the camera coordinate system. This step is to convert the three-dimensional points to three-dimensional points, including the relevant parameters of external camera parameters such as R, t. The second step is to convert the camera coordinate system to the image coordinate system. This step is to convert the 3D points to 2D points, including the relevant parameters of the K isoperimetric camera internal parameters. The camera calibration steps are shown in Figure 1. The synchronous calibration of internal and external parameters generally includes two strategies: optical calibration, that is, using known geometric information (such as fixed length checkerboard) to achieve parameter solution. Another strategy is called self-calibration, that is, using the structural motion in static scenes to estimate the parameters.
Camera calibration is to take the calibration plate by the camera, determining the internal and external parameters of the camera with the intrinsic value of the calibration plate’s feature points. Following that, the image coordinates are converted to physical coordinates through the scale factor. The following equation is used to convert image coordinates to physical coordinates:
( x y 1 ) = [ f x γ c x 0 f y c y 0 0 1 ] [ r 11 r 12 r 13 t 1 r 21 r 22 r 23 t 2 r 31 r 32 r 33 t 3 ] [ X Y Z 1 ]
The simplified expression is:
s x = K [ R | t ] X
where s is the scale factor; ( x , y , z , 1 ) T is the image coordinate; K is the camera internal parameter; ( X , Y , Z , 1 ) T is the world coordinates; f x and f y are the focal lengths of the camera in the horizontal and vertical directions; c x and c y are the offsets of the optical axis; γ is the tilt factor; R and t are the camera the external parameters; r i j and t i are elements of R and t, respectively.
The camera’s internal parameters, tangential distortion, and radial distortion are used to calibrate the video, which can effectively eliminate lens distortion and image distortion, thus obtaining more accurate displacement measurements. In this paper, the smartphone lens uses Zhang’s calibration method to calibrate the camera lens, followed by carrying out the video calibration [32].

2.2. Target Tracking Principle Based on Optical Flow Method

Optical flow is an assumption based on the image brightness motion information. The optical flow calculation is based on two assumptions about the optical characteristics of object motion: constant brightness assumption and small motion assumption. Assume that a pixel ( x , y ) on the image has a corresponding brightness of I ( x , y , t ) at time t, and a brightness of I ( x + d x , y + d y , t + d t ) at time t. From the optical flow consistency assumption:
I ( x , y , t ) = I ( x + d x , y + d y , t + d t )
The basic equation of optical flow can be obtained by using Taylor expansion Equation (3):
I ( x , y , t ) x u + I ( x , y , t ) y v + I ( x , y , t ) t = 0
where u = d x / d t and v = d y / d t are the instantaneous velocities of pixels in the image in x and y directions at time t.
Assuming I x = ( x , y , t ) / x , I y = ( x , y , t ) / y and I t = ( x , y , t ) / t , it can be converted into the optical flow constraint equation:
I x u + I y v + I t = 0
Since one of the above optical flow constraint equations cannot solve the two unknowns ( x , y ) , it is necessary to establish a new constraint equation to solve it. In 1981, the Lucas Kanade (LK) optical flow method [33], proposed by Lucas and Kanade, put forward the assumption of spatial consistency to solve the basic equation of optical flow.
The LK optical flow method assumes that the motion vector of the neighborhood Ω in a space is constant. In a neighborhood Ω of n pixels, each pixel satisfies the following equation:
I x i u + I y i v + I t i = 0    i = 1 , 2 , , n
At this time, the constraint equation of optical flow can be changed into:
E c ( u , v ) = ( I x u + I y v + I t ) 2 d x d y
In neighborhood Ω , the error of LK optical flow is:
E L K ( u , v ) = W 2 ( x , y ) ( I x u + I y v + I t ) 2 d x d y
where W ( x , y ) = { w i | i = 1 , 2 , n } is the weight value of each point in the field, and the farther away from the center point, the smaller the corresponding weight value.
Discretization of Equation (8) leads to:
[ i n w i 2 I x i 2 i n w i 2 I x i I y i i n w i 2 I x i I y i i n w i 2 I y i 2 ] [ u v ] = [ i n w i 2 I x i I t i i n w i 2 I y i I t i ]
where I x i , I y i , and I t i are the gradient values corresponding to the x, y, and z directions of the first pixel in the neighborhood, respectively; w i is the weight of the ith pixel. Assumptions:
A = [ I x 1 , , I x n I y 1 , , I y n ] T
W = d i a g ( w x 1 , w x 2 , , w x n )
b = [ I t 1 , I t 2 , , I t n ] T
Equation (7) can be expressed as:
A T W 2 A [ u v ] = A T W 2 b
Then the vector optical flow can be calculated:
[ u v ] = ( A T W 2 A ) 1 A T W 2 b
Finally, the corresponding pixel position ( x , y ) of each target point is calculated from the vector optical flow u = d x / d t and v = d y / d t .

2.3. Target-Free Feature Extraction Based on Optical Flow Method

The optical flow method can track the target in a low-resolution image. When the vibration is small, the error is small. Moreover, the optical flow method occupies less computer memory to achieve fast calculation. Therefore, the optical flow method can be used to monitor the displacement response of wind turbine structures. Although the optical flow method can effectively track the structural feature points, the optical flow method requires that the feature points must be prominent to be recognized effectively. In the actual monitoring of wind turbine structure, it is impossible to place artificial markers on the structures. To solve this problem, this paper proposes a robust corner feature extraction method based on region of interest (ROI) to realize dynamic characteristics monitoring of target-free structures.
In the optical flow method, the detector is usually used to extract the feature point as Harris corner [34], which has been widely used in engineering practice [35,36]. Since Harris corners use a Gaussian filtering, and the motion speed is relatively slow, there is a risk of corner information loss and information migration. Therefore, tracking errors will occur in optical flow tracking. J. Shi improved Harris corner algorithm in 1994 and proposed an improved Shi–Thomasi corner detection operator [37]. The Shi–Tomasi corner operator solves the problem of feature point aggregation by limiting the shortest distance between two interest points, and only allows points beyond a certain distance to become interest points from the strongest point of Harris corner. It solves the defect of Harris corner and is more suitable for optical flow tracking. Based on this, this paper conducts ROI selection before structural vibration video processing, and then sets the quality factor and threshold value for the number of Shi–Thomasi corners. Pause at the first frame when the video starts playing, and the operator manually selects the ROI. The selected ROI should not only include the range of structural vibration, but also avoid excessive image selection to avoid excessive calculation. Finally, corners with better robustness appear on the structure in the ROI area for target tracking, and sub-pixel coordinate extraction is achieved by capturing image corners and combining them with the optical flow equation. The basic flow of target tracking based on ROI target-free robust corner feature extraction is shown in Figure 2.
Step 1: Establish the region of interest (ROI). After taking the vibration videos, the smartphone selects the ROI in the first frame, which aims to reduce corner recognition and improve the recognition accuracy of the required corner area.
Step 2: Shi–Tomasi corner detection. Based on the first frame of the video, all Shi–Tomasi corners within the ROI are detected.
Step 3: Select the corner. Not all the corners detected in step 2 are required due to engineering requirements. Select the corner with the strongest feature in the corresponding area or multiple corners containing the structure itself for target tracking through such constraints as corner distance, number of corners, quality factor, etc.
Step 4: Build an image pyramid. In view of the small motion assumption among the three assumptions of the optical flow method, the image is sampled down in the form of an upper pyramid to compensate for excessive motion. In this paper, considering the actual working conditions and smartphone lens frame rate, a 4-layer pyramid was built for each frame image.
Step 5: Update the parameters. After each image pyramid is built to meet the optical flow conditions, the strongest corner selected in the first frame will be used for feedback to the next frame. In each frame of image, the corners to be tracked are the corners of the previous frame, so it is necessary to detect the corners of each frame, and then compare them with the corners of the first frame to update the corner parameters.
Step 6: Target tracking. After updating the parameters, optical flow correlation calculation can be carried out according to the corners to obtain the corresponding corner coordinates of each frame. The coordinates extracted from the optical flow equation are sub-pixel coordinates.
In the target tracking based on optical flow method, except for the first step where the ROI area needs to be manually selected, the other steps are automatically carried out, laying a foundation for displacement monitoring in structural dynamic characteristics monitoring.

2.4. Smartphone Jitter Processing

Although smartphones can be fixed with tripods for visual monitoring, considering the convenience of using smartphones, measures should be taken counteract the effects of common smartphone shooting methods on the monitoring of structural dynamic characteristics. When the hand-held smartphone is used for shooting, it will shake, which is similar to the situation when the UAV is monitoring. Li et al. [38] used the method of in-plane high pass filtering and out-of-plane adaptive scaling factor to deal with the jitter of UAV hovering monitoring when using UAV to monitor the wind turbine blades. This paper uses similar methods to deal with the jitter of smartphones.
The high pass filter filters the low frequency noise by passing the measurement signal across the high pass filter and can restore the projection of the displacement on the still image plane to scaling and perspective factors. Therefore, it is expressed in the matrix form as follows:
[ x i h p y i h p ] = [ s x p y x p x y s y ] [ X w o r l d Y w o r l d ]
where x i h p and y i h p respectively represent the results of high pass measurement signals.
The out-of-plane smartphone jitter processing adopts the method of adaptive scaling factor, and its equation is as follows:
S i = L l = L p i 1 p i 2
where S i is the scale factor at the time of the i th frame, L is the actual distance between two structural points, and p i 1 and p i 2 are the image coordinates of p 1 and p 2 on the image at the time of the i th frame.

2.5. Displacement Calculation

The corresponding coordinate P i ( x i , y i ) of the target in each frame of the time series and the coordinate P 0 ( x 0 , y 0 ) of the target in the first frame are obtained by the optical flow method of target tracking. Following that, the absolute displacement of the structure was obtained by converting the image coordinate to the physical coordinate through the adaptive scale factor. Considering the elevation problem when using smartphones to capture videos, the displacement d S of the structure was calculated by angle correction using the following equation:
d S = S i × P i P 0 cos 2 θ
where θ is the angle between the smartphone optical axis and the measured target.
Through the above target tracking and displacement calculation, the absolute displacement of the structure can be calculated. Following that, the displacement responses of multiple monitoring points of the structure are obtained through monitoring. Finally, the modal shape of the wind turbine structure can be calculated by responses.

3. Smartphone Performance Test

3.1. Smartphone Lens Distortion Test

This paper uses iPhone 12 and Honor X10 smartphones to investigate the performance of the proposed method. The camera parameters of the two phones are shown in Table 1.
Due to the process error of the camera optical system during manufacturing, there is geometric distortion between the actual imaging and the ideal imaging. Distortion is mainly divided into radial distortion and tangential distortion, as shown in Figure 3.
The radial distortions (pillow and barrel) are mainly caused by the lens quality and the fact that the light is more bent away from the center of the lens than near the center. Generally, the radial distortion can be corrected by the following equation:
x c o r r = x d i s ( 1 + k 1 r 2 + k 2 r 4 + k 3 r 6 )
y c o r r = y d i s ( 1 + k 1 r 2 + k 2 r 4 + k 3 r 6 )
Tangential distortion (thin lens distortion and centrifugal distortion) is caused by defects in lens manufacturing that make the lens itself not parallel to the image plane. Tangential distortion can be corrected by the following equation:
x c o r r = x d i s + [ 2 p 1 x y + p 2 ( r 2 + 2 x 2 ) ]
y c o r r = y d i s + [ p 1 ( r 2 + 2 y 2 ) + 2 p 2 x y ]
where x c o r r and y c o r r represent the coordinates of the image plane in x and y directions after repair, respectively, x d i s and y d i s represent the coordinates of the image plane in x and y directions with distortion, respectively; k 1 , k 2 , and k 3 are radial distortion parameters; p 1 and p 2 represent tangential distortion parameters.
It can be seen that the image distortion has k 1 , k 2 , k 3 , p 1 , and p 2 , totaling five parameters. For a camera with good quality, the tangential distortion is small, and the radial distortion coefficient can be ignored. Only two parameters need to be calculated. For a camera with good quality, the tangential distortion is very small and can be ignored. The radial distortion coefficient k 3 can also be ignored. Only k 1 and k 2 , need to be calculated.
Zhang’s calibration method can be carried out in the integrated package of MATLAB. Theoretically, only two calibration photos are needed to calibrate the camera parameters. In order to verify the accuracy of the camera parameters identified by the calibration photos taken by smartphones, 33 marking photos taken by a smartphone were used for calibration. The calibration plate was a 25 mm × 25 mm black and white checkerboard. The process of using two smartphones to identify radial distortion parameters is shown in Figure 4.
It can be seen from Figure 4 that the accuracy of parameter calibration tends to be stable with the increase of the number of calibrations during the process of camera calibration parameter identification for the lens of smartphone using the calibration plate. The parameter calibration accuracy of smartphones is worse than that of cameras, but the calibration requirements can be met by using about 20 photos. Finally, through the camera calibration program, the radial distortion parameters k 1 and k 2 of the Honor X10 smartphone are 0.0431 and −0.0102, respectively. The radial distortion parameters k 1 and k 2 of the iPhone 12 smartphone are 0.0611 and −0.0835, respectively.
The smartphone lens was adjusted to a wide angle to form the distortion, and then distortion parameters were used to correct the image, as shown in Figure 5. The distorted and distorted edges of the black and white squares of the image completed by distortion parameter correction have been well-corrected to straight edges. Therefore, the camera calibration method can be used to correct the image distortion in smartphone image acquisition to obtain more accurate monitoring.

3.2. Smartphone Displacement Monitoring

In the case of simulated shutdown, smartphones were used for vibration monitoring of scaled wind turbine models, and an LDS (laser displacement sensor) was used to evaluate the accuracy of visual data. Canon R6 camera and 24–105 mm zoom lens were also used in the test. The test apparatus is shown in Figure 6.
In practical wind turbine structure monitoring state, the background is relatively complex, and the structural health monitoring usually using the optical flow method will be affected by the background [39]. In order to verify that the ROI-based corner feature extraction method of unmarked robustness realizes the dynamic characteristics monitoring of unmarked structures, the test has verified the feature-matching between simple backgrounds and complex backgrounds, as shown in Figure 7.
It can be seen from Figure 7 that, no matter in a simple background or in a complex background, corner-matching using the feature-matching method after selecting ROI is very effective, and there is no error-matching. Therefore, it is feasible to use optical flow method to select ROI for target-free dynamic characteristic monitoring. In the actual monitoring process, multi-point monitoring is realized by selecting multiple ROIs for monitoring at the same time, which solves the limitation that one sensor can only monitor one point in traditional monitoring.
To verify the accuracy of the smartphone to monitor the displacement, the wind turbine blade was released to free vibration after given initial displacement. The LDS is used as the data reference, and the camera is used for verification. The smartphone adopts 1080 × 1920 resolution with a frame rate of 60 fps. The frequency of LDS is set to 50 Hz. The resolution of 1080 × 1920 is adopted with a frame rate of 50 fps. The camera and mobile phone are fixed with a tripod. Through 100 s of data acquisition, the displacement time history is shown in Figure 8.
It can be seen from Figure 8a that the displacement time history through vision is generally consistent with that of LDS. Figure 8b shows that the camera and LDS are basically consistent in the 10~20 s displacement time history details. However, the phase is inconsistent smartphone monitoring during operation. Therefore, such problems must be addressed to enable smartphones to accurately monitor displacement. The frame rate is usually used for time conversion in visual monitoring. However, due to the instability of the sensor when the smartphone captures video, the captured video will not be consistent with the original set frame rate. For example, if 60 fps is set, the final captured video frame rate is 59.58 fps or 60.56 fps. This is also the reason why the phase of the visual displacement curve and the standard displacement curve is not consistent in displacement monitoring. To solve this problem, the video output is read at the real-time frame rate by traversing all frames, and finally time conversion is performed at the real-time read frame rate. The displacement time history curve after frame rate correction is shown in Figure 9.
It can be seen from the time domain information shown in Figure 9 that the problem of phase difference during smartphone monitoring can be effectively solved by frame rate correction, which is consistent with the displacement curve monitored by the camera. Through the free vibration test of wind turbine blades, it was proved that the method of selecting target-free displacement monitoring by smartphones combined with ROI is effective.

3.3. Performance Test of Smartphones in Different States

As one of the indispensable tools in human life, smartphone monitoring has the advantages of high efficiency and low cost. However, people usually do not carry tripods when they travel, so if they want to use smartphones more conveniently to complete structural monitoring tasks, they must explore monitoring methods for convenient photography. When using a smartphone to take pictures of objects without a tripod, people tend to stand still or walk slightly. Therefore, to monitor the dynamic characteristics of the structure based on smartphones without a tripod, the test uses three states to explore: smartphones on the tripod, holding smartphones when standing still, and using smartphones when walking slightly. The test used the same smartphone (Honor X10) to shoot fixed points in three shooting states: tripod, standing handheld smartphone, and slightly walking handheld smartphone. The monitoring statue under the three shooting states was inversely deduced from this stationary fixed point. In order to better distinguish the three shooting states, the direction coordinates centered on the smartphone are specified in this paper. The schematic diagram of the tester’s shooting and the smartphone direction are shown in Figure 10.
The tester stood at the fixed point 2 m away and photographed the fixed point for 4 min. The displacement time histories in three directions under three shooting conditions are shown in Figure 11, and the maximum displacement is shown in Table 2.
It can be seen from Figure 11 that only the tripod is stable in three directions under three conditions, and the last section also proves the feasibility of displacement monitoring with support. Standing hand-held photography and light walking photography have large displacement in three directions—the displacement deviation of light walking in Z-direction is especially large. In the test, the moving distance is controlled within 0.5 m when shooting with slightly walking, so the displacement deviation of this kind of shooting will be larger in actual situations. Table 2 shows the three shooting states using smartphones and the displacement peaks in three directions, from which it can be seen that slightly walking has the largest displacement in three directions. The test data processing adopts the pixel as the unit, and the farther the distance is, the greater the actual error is in the actual monitoring. Therefore, if the smartphone is used to monitor the dynamic characteristics of the structure, the displacement offset under the common states must be processed.
The frequency domain information obtained by processing the displacement time history data captured in Figure 11 through Fast Fourier Transform (FFT) is shown in Figure 12. The X-direction data are used in the tripod equipped and hand-held shooting states, and the Z-direction data are used in the slightly moving shooting state.
It can be seen from the frequency domain information taken in the three states shown in Figure 12 that the frequency domain of the shooting method with a bracket is basically 0, and the frequency domain information of the shooting state of standing, holding, and walking slightly is confused before 0.6 Hz. Therefore, high pass filtering can be used to eliminate the displacement error caused by different shooting states of smartphones.

3.4. Structural Displacement Monitoring Using Smartphone in Different States

Since the above tests have verified the reliability of smartphone and LDS displacement monitoring, the test will hammer the blade to make it vibrate freely, and use three shooting states: tripod, standing hand-held smartphone, and slightly walking hand-held smartphone to monitor the displacement.
Taking the displacement time history obtained by the shooting method equipped with a tripod as a benchmark, the displacement time history monitored by the two shooting states of standing hold smartphone and slightly walking hand-held smartphone is shown in Figure 13 and Figure 14.
Figure 13 and Figure 14 show the displacement time history comparison between the standing hold and slightly walking shooting and the tripod equipped shooting states. Although the displacement time histories of the two states are scattered, the two trajectories inversely calculated from the fixed points can match the structural vibration displacement time histories. When the trajectories shown in Figure 13 and Figure 14 are consistent with the displacement drift path, high pass filtering can be used for noise processing. In this paper, a high pass filter with a cut-off frequency of 0.6 Hz was used to denoise smartphones. Combined with the adaptive scaling factor method, the final processing results are shown in Figure 15.
It can be seen from Figure 15a that the displacement time history of the monitoring method with support can be basically consistent with that of the method using high pass filtering combined with adaptive scaling factor. The influence of displacement is not completely eliminated in the area with small displacement. Since the displacement is relatively small, it often presents low-frequency movement, and the displacement time history is consistent on the whole, which does not affect the actual monitoring. The 10 ~ 20 s displacement time history details show that the displacement is eliminated better by the above methods when the handheld smartphone is shooting, and the amplitude is still inconsistent in some areas when the walking camera is shooting. The displacement drift in the walking process was not completely filtered out due to the influence of the tester’s walking during the processing.
To quantify the errors of the two common ways of holding smartphones to monitor the dynamic characteristics of structures, this paper uses Root Mean Square Error (RMSE), correlation coefficient ( ρ ), and determination coefficient ( R 2 ) for error analysis. The equations are:
RMSE = i ( x v ( i ) x s ( i ) ) 2 / n
ρ = | i ( x s ( i ) μ s ) × ( x v ( i ) μ v ) | i ( x s ( i ) μ s ) 2 i ( x v ( i ) μ v ) 2
R 2 = 1 i ( x v ( i ) x s ( i ) ) 2 i ( x s ( i ) μ s )
RMSE is calculated using Equation (22), where n is the total number of monitoring, and x v and x s are displacement data from vision monitoring and laser displacement sensors, respectively. ρ calculated using Equation (23), where μ v and μ s are the average values of the two displacement trajectories. The calculation equation of R 2 is Equation (24), which is used to determine the matching degree of the two recorded tracks [40]. The error comparison of vibration monitoring data of two common smartphone shooting methods is shown in Table 3, and the error distribution is shown in Figure 16, in which the displacement time history data with tripod is taken as the benchmark.
It can be seen from Table 3 that the errors of standing shoot and slightly walking shooting are within the acceptable range. The maximum error shown in Figure 16 does not exceed 2 mm, and most of the errors are concentrated within 0.5 mm, which is acceptable in time domain monitoring. The displacement time history data monitored by the three shooting methods are transformed by FFT, as shown in Figure 17.
It can be seen from the comparison of monitoring frequency domains under the three common shooting states shown in Figure 17 that, although there is a certain error between the hold smartphone and the slightly walking shooting state in the time domain, the natural frequency identified in the frequency domain is consistent. Therefore, it is feasible to monitor the displacement of structures by three common shooting methods in the frequency domain. Through the method of high pass filtering and adaptive scaling factor, the structural dynamic characteristics can be accurately monitored by removing the noise from displacement monitoring of hand-held smartphones and shooting methods of slightly walking.

3.5. Structural Displacement Monitoring of Smartphone Assembled with Long Focus Lens

The actual wind turbine structure is enormous, and the use of ordinary smartphones cannot obtain clear images due to lens limitations, so accurate monitoring cannot be achieved. The lens is an important part of the camera, but with the development of technology and the improvement of technology, lenses are not limited to cameras. Many manufacturers produce telephoto lenses that can be matched on smartphones. By assembling a telephoto lens in a smartphone, the defect that the smartphone cannot take pictures of distant objects can be effectively solved. Figure 18 shows the contrast of the experimenter using a camera and a smartphone equipped with a telephoto lens to shoot the wind turbine at a distance of 5 m from the wind turbine. The shooting area of the camera in Figure 18a is wider than that of the telephoto lens equipped with the smartphone in Figure 18b, but the longer objects can be shot through the smartphone. Figure 18b is just a picture taken in a smartphone with one pixel and one distance of the telephoto lens. The telephoto lens can be up to 32 times as long as possible and can shoot objects thousands of meters away. The connection between the smartphone and the telephoto lens can be easily fixed by a clamp. Therefore, the displacement of the actual wind turbine structure can be monitored by assembling a telephoto lens on a smartphone.
Through the hammer test on the wind turbine, the smartphone was equipped with a telephoto lens to monitor the blade tip. Limited by the test site, the smartphone was 10 m away from the wind turbine, and the monitoring structure is shown in Figure 19.
It can be seen from Figure 19a that the displacement monitored by the telephoto lens is consistent with that monitored by the camera. The displacement time history monitored by the smartphone equipped with the telephoto lens is consistent with the LDS at both the phase and peak. Therefore, smartphones can be used to assemble telephoto lenses for structural dynamic characteristics monitoring, which enriches the way smartphones monitor dynamic characteristics.

4. Dynamic Characteristic Monitoring of Wind Turbine Structure

4.1. Experimental Equipment

To monitor the dynamic characteristics of the wind turbine structure, the test uses the Honor X10 smartphone, with a lens resolution of up to 4K, a maximum frame rate of 60 fps, a video resolution of 1080P, and a 32 times telephoto lens. At the same time, in order to verify the reliability of smartphones, the Canon R6 camera was also used in the test. The frame rate is 50 fps, and a 24 mm~105 mm zoom lens was used for test verification. Zhang’s camera calibration method was used for video shooting to correct lens distortion. The vibration test was conducted on the scaled wind turbine model under simulated shutdown. The arrangement of test device and measuring points is shown in Figure 20. In order to verify the accuracy of the visual test, 1 LDS and 5 accelerometers were used to monitor the structural responses, where the sampling frequency is set to 50 Hz and the traditional vibration test digital system is used for verification.

4.2. Natural Frequency Identification of Wind Turbine Structure

The vibration of the wind turbine structure has two directions: edgewise and flap-wise. Since the flap-wise direction vibration belongs to out-of-plane vibration, the flap-wise direction is more likely to cause damage to the wind turbine structure. Monitoring the dynamic characteristics of wind turbine structure flap-wise direction is the basis for ensuring the normal operation of wind turbine structure.
The blade tip of the scaled wind turbine model is hammered in the flap-wise direction to make it vibrate freely. The acceleration data during the vibration were monitored by the accelerometers, and the camera and LDS were used as the displacement reference for verification. The acceleration time history curve monitored by the accelerometer is shown in Figure 21. The domain information, monitored by smartphones, cameras, and LDS, is shown in Figure 22.
It can be seen from Figure 21 and Figure 22 that the acceleration time history monitored by the accelerometer after hammering the blade tip attenuates rapidly, while the displacement monitored by LDS and vision attenuates slowly. The main reason is that the accelerometer is attached to the structure and the monitoring frequency is high, so the data monitored by the acceleration sensor tends to zero after the blade vibration slow down. The displacement time histories show, in Figure 22, that the displacement time histories monitored by camera and smartphone are consistent with LDS as a benchmark. Therefore, the use of smartphones can achieve low-cost dynamic characteristics monitoring of wind turbine structures. The displacement time history monitored by accelerometer, LDS, and smartphone is converted into PSD frequency domain information as shown in Figure 23. In order to make the images more intuitive, the horizontal axis of the coordinates is expressed in logarithmic coordinates.
Figure 23 shows that the natural frequencies obtained by monitoring the wind turbine structure with three types of equipment are consistent. The first three natural frequencies of the wind turbine structure under simulated shutdown in the flap-wise direction are 2.5181 Hz, 6.4041 Hz, and 8.6580 Hz, respectively. In Figure 23, the amplitude of the first three natural frequencies recognized by the accelerometer increases sequentially, while the third natural frequencies recognized by LDS and smartphones are fuzzy. Through frequency domain comparison, the monitoring data of structural dynamic characteristics represented by smartphones are consistent with LDS. The smartphone can identify the natural frequency of the structure stably through its camera, especially for low order frequencies. The structure of wind turbines is mainly affected by low order frequency. Therefore, smartphones with low cost, non-contact, and remote shooting, combined with vision technology, can be used to replace the traditional contact sensors to monitor the dynamic characteristics of wind turbine structures.

4.3. Shaking Table Test of Wind Turbine Structure

To verify the monitoring effect of smartphones combined with visual technology on the dynamic characteristics of wind turbine structures, the test used accelerometers, camera, and smartphone to conduct vibration monitoring by inputting seismic waves into the vibration table to excite the wind turbine structures. The visual measuring points were divided into 10 monitoring points P0~P9 from top to bottom by the wind turbine structure (P0~P4 is on the blade and P5~P9 is on the tower), corresponding to P0, P2, P4, P7, and P9, which are verified with acceleration sensors. Figure 24 and Figure 25 show the response data monitored by the acceleration sensor and smartphone in the flap-wise direction. Since LDS is only used for verification at P0 in the test, to make the article more concise, the full data part of displacement time history shown in Figure 25 was monitored by smartphones. Since the error of smartphones and LDS in time domain can be within 0.1mm, smartphones can be used to monitor the displacement time history of wind turbine structures in the whole field. Table 4 shows the peak values of each measuring point in the two monitoring methods.
From the response monitored by the accelerometers shown in Figure 24, the overall structural responses are in accordance with the law of “from big to small”. According to the data obtained from P4 measuring point in Table 4, the response of the acceleration sensor at P4 is relatively small. The reason for this is that P4 is located at the top of the tower, and the blade-waving amplitude is relatively large. Therefore, it is reasonable to state that P4 had the smallest acceleration response. Figure 25 shows that the displacement time history monitored by smartphone is relatively stable. It can be seen from Table 4 that smartphones can be used in combination with visual technology for multi-point monitoring. The displacement time histories of each monitoring point are obtained through smartphones, and then the structural modal shapes of wind turbines were calculated by a stochastic subspace identification (SSI) method. The position of the sensor is important when calculating the modal shape [41]. In this paper, five accelerometers and visual sensors are compared, and the final modal shape of the wind turbine structure is shown in Figure 26.
It can be seen from Figure 26 that the vibration modes calculated by vision and accelerometers are basically the same. Since there are many measuring points monitored by smartphones combined with vision technology, the vibration modes obtained are smoother. However, the frame rate of the smartphone is 60 fps, and the natural frequency below 30 fps can be monitored at most. Therefore, only the first two modes are obtained. If higher order modal shapes are desired, lenses with higher frame rates should be considered. As the vibration of wind turbine structure in the flap-wise direction only needs the first two steps to be sufficient, the dynamic characteristics of wind turbine structure can be monitored through smartphones combined with visual technology.

5. Conclusions

In this study, a target-free dynamic characteristic monitoring method for wind turbine structures using a portable smartphone and optical flow method was proposed. Firstly, the characteristics of smartphones in the monitoring environment were studied to verify the robustness of the proposed algorithm. After that, smartphones in different shooting states were used to monitor the displacement, and high pass filtering combined with adaptive scaling factor was used to process the displacement drift of common smartphone shooting states. Then, the displacement monitoring of smartphone assembling telephoto lens was studied. Finally, the following conclusions can be drawn from the wind turbine structure test verification and result analysis:
(1)
The proposed method based on optical flow method for monitoring the target-free dynamic characteristics of wind turbine structures can better identify targets by simulating simple and complex background projects. In addition, the use of smartphones combined with visual algorithms can simultaneously monitor the spatial displacement of the entire blade through ROI clipping.
(2)
The method of high pass filtering combined with adaptive scaling factor was adopted to effectively eliminate the displacement drift caused by the two shooting states of standing and slightly walking. The error analysis shows that the final error is less than 2 mm, which can meet the requirements of structural dynamic characteristics monitoring. The smartphone is equipped with a telephoto lens to monitor the displacement of the structure, which effectively expands the method of smartphone to monitor the dynamic characteristics of the structure.
(3)
The proposed method for monitoring the dynamic characteristics of wind turbine structures performs well in cooperation with smartphones. Combined with the shaking table test, the results show that using smartphones to monitor the dynamic characteristics of fan structures has higher accuracy in time and frequency domains.

Author Contributions

Conceptualization, W.Z. and W.L.; methodology, W.Z. and W.L.; software, W.Z. and B.F.; validation, W.Z., W.L. and B.F.; formal analysis, W.Z.; investigation, W.L.; resources, W.L. and Y.D.; data curation, W.Z.; writing—original draft preparation, W.Z.; writing—review and editing, Y.D.; visualization, W.Z.; supervision, W.L.; project administration, W.L.; funding acquisition, W.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was jointly funded by the National Natural Science Foundation of China (Nos. 52068049, 51908266), the Science Fund for Distinguished Young Scholars of Gansu Province (No. 21JR7RA267), and Hongliu Outstanding Young Talents Program of Lanzhou University of Technology.

Data Availability Statement

Some or all data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Hernandez-Estrada, E.; Lastres-Danguillecourt, O.; Robles-Ocampo, J.B.; Lopez-Lopez, A.; Sevilla-Camacho, P.Y.; Perez-Sariñana, B.Y.; Dorrego-Portela, J.R. Considerations for the structural analysis and design of wind turbine towers: A review. Renew. Sustain. Energy Rev. 2021, 137, 110447. [Google Scholar]
  2. Beganovic, N.; Söffker, D. Structural health management utilization for lifetime prognosis and advanced control strategy deployment of wind turbines: An overview and outlook concerning actual methods, tools, and obtained results. Renew. Sustain. Energy Rev. 2016, 64, 68–83. [Google Scholar] [CrossRef]
  3. Lin, Z.; Cevasco, D.; Collu, M. A methodology to develop reduced-order models to support the operation and maintenance of offshore wind turbines. Appl. Energy 2020, 259, 114228. [Google Scholar] [CrossRef]
  4. Xu, D.; Liu, P.F.; Chen, Z.P. Damage mode identification and singular signal detection of composite wind turbine blade using acoustic emission. Compos. Struct. 2021, 255, 112954. [Google Scholar] [CrossRef]
  5. Lau, B.C.P.; Ma, E.W.M.; Pech, M. Review of offshore wind turbine failures and fault prognostic methods. In Proceedings of the IEEE 2012 Prognostics and System Health Management Conference, Beijing, China, 23–25 May 2013. [Google Scholar]
  6. Mishnaevsky, L., Jr.; Branner, K.; Petersen, H.K.; Beauson, J.; McGugan, M.; Sørensen, N.F. Materials for wind turbine blades: An overview. Materials 2017, 10, 1285. [Google Scholar]
  7. Keegan, M.H.; Nash, D.H.; Stack, M.M. On erosion issues associated with the leading edge of wind turbine blades. J. Appl. Phys. 2013, 46, 383001. [Google Scholar] [CrossRef] [Green Version]
  8. Du, Y.; Zhou, S.; Jing, X.; Ping, Y.; Wu, H.; Kwok, N. Damage detection techniques for wind turbine blades: A review. Mech. Syst. Signal Process. 2020, 141, 106445. [Google Scholar] [CrossRef]
  9. Jasinien, E.; Raiutis, R.; Voleiis, A.; Vladiauskas, A.; Mitchard, D.; Amos, M. NDT of wind turbine blades using adapted ultrasonic and radiographic techniques. Insight 2009, 51, 477–483. [Google Scholar] [CrossRef] [Green Version]
  10. Behtani, A.; Tiachacht, S.; Khatir, T.; Khatir, S.; Wahab, M.A.; Benaissa, B. Residual Force Method for damage identification in a laminated composite plate with different boundary conditions. Frat. Integrità Strutt. 2022, 16, 35–48. [Google Scholar] [CrossRef]
  11. Khatir, S.; Abdel Wahab, M.; Tiachacht, S.; Le Thanh, C.; Capozucca, R.; Magagnini, E.; Benaissa, B. Damage identification in steel plate using FRF and inverse analysis. Frat. Integrità Strutt. 2021, 58, 416–433. [Google Scholar] [CrossRef]
  12. Hsu, M.H.; Zhuang, Z.Y. An intelligent detection logic for fan-blade damage to wind turbines based on mounted-accelerometer data. Buildings 2022, 12, 1588. [Google Scholar] [CrossRef]
  13. Amenabar, I.; Mendikute, A.; López-Arraiza, A.; Lizaranzu, M.; Aurrekoetxea, J. Comparison and analysis of non-destructive testing techniques suitable for delamination inspection in wind turbine blades. Compos. Part B Eng. 2011, 42, 1298–1305. [Google Scholar]
  14. Felipe-Sesé, L.; Díaz, F.A. Damage methodology approach on a composite panel based on a combination of Fringe Projection and 2D Digital Image Correlation. Mech. Syst. Signal Process. 2018, 101, 467–479. [Google Scholar]
  15. Li, Y.; Peng, L.; Wu, C.; Zhang, J. Street View Imagery (SVI) in the built environment: A theoretical and systematic review. Buildings 2022, 12, 1167. [Google Scholar] [CrossRef]
  16. Spencer, B.F., Jr.; Hoskere, V.; Narazaki, Y. Advances in computer vision-based civil infrastructure inspection and monitoring. Engineering 2019, 5, 199–222. [Google Scholar] [CrossRef]
  17. Wang, N.; Zhao, Q.; Li, S.; Zhao, X.; Zhao, P. Damage classification for masonry historic structures using convolutional neural networks based on still images. Comput.-Aided Civ. Infrastruct. Eng. 2018, 33, 1073–1089. [Google Scholar] [CrossRef]
  18. Sangirardi, M.; Altomare, V.; De Santis, S.; de Felice, G. Detecting damage evolution of masonry structures through computer-vision-based monitoring methods. Buildings 2022, 12, 831. [Google Scholar] [CrossRef]
  19. Feng, D.; Feng, M. Experimental validation of cost-effective vision-based structural health monitoring. Mech. Syst. Signal Process. 2017, 88, 199–211. [Google Scholar] [CrossRef]
  20. Dong, C.Z.; Ye, X.W.; Jin, T. Identification of structural dynamic characteristics based on machine vision technology. Measurement 2018, 126, 405–416. [Google Scholar] [CrossRef]
  21. Khadka, A.; Fick, B.; Afshar, A.; Tavakoli, M.; Baqersad, J. Non-contact vibration monitoring of rotating wind turbines using a semi-autonomous UAV. Mech. Syst. Signal Process. 2020, 138, 106446. [Google Scholar] [CrossRef]
  22. Song, Q.; Wu, J.; Wang, H.; An, Y.; Tang, G. Computer vision-based illumination-robust and multi-point simultaneous structural displacement measuring method. Mech. Syst. Signal Process. 2022, 170, 108822. [Google Scholar] [CrossRef]
  23. Han, R.; Zhao, X.; Yu, Y.; Guan, Q.; Hu, W.; Li, M. A cyber-physical system for girder hoisting monitoring based on smartphones. Sensors 2016, 16, 1048. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Grossi, M. A sensor-centric survey on the development of smartphone measurement and sensing systems. Measurement 2019, 135, 572–592. [Google Scholar] [CrossRef]
  25. Ozer, E.; Feng, M.Q. Biomechanically influenced mobile and participatory pedestrian data for bridge monitoring. Int. J. Distrib. Sens. Netw. 2017, 13, 1550147717705240. [Google Scholar] [CrossRef]
  26. Han, R.; Zhao, X. Shaking table tests and validation of multi-modal sensing and damage detection using smartphones. Buildings 2011, 11, 477. [Google Scholar] [CrossRef]
  27. Zhao, X.; Liu, H.; Yu, Y.; Zhu, Q.; Hu, W.; Li, M.; Ou, J. Displacement monitoring technique using a smartphone based on the laser projection-sensing method. Sens. Actuators A Phys. 2016, 246, 35–47. [Google Scholar] [CrossRef]
  28. Zhao, X.; Ri, K.; Han, R.; Yu, Y.; Li, M.; Ou, J. Experimental research on quick structural health monitoring technique for bridges using smartphone. Adv. Mater. Sci. Eng. 2016, 2016, 1871230. [Google Scholar] [CrossRef] [Green Version]
  29. Li, H.; Chen, X.; Chen, H.; Wang, B.; Li, W.; Liu, S.; Li, P.; Qi, Z.; He, Z.; Zhao, X. Simulation of smartphone-based public participation in earthquake structural response emergency monitoring using a virtual experiment and AI. Buildings 2022, 12, 492. [Google Scholar] [CrossRef]
  30. Ozer, E.; Feng, D.; Feng, M.Q. Hybrid motion sensing and experimental modal analysis using collocated smartphone camera and accelerometers. Meas. Sci. Technol. 2017, 28, 105903. [Google Scholar] [CrossRef]
  31. Wang, N.; Zhao, X.; Zhao, P.; Zhang, Y.; Zou, Z.; Ou, J. Automatic damage detection of historic masonry buildings based on mobile deep learning. Autom. Constr. 2019, 103, 53–66. [Google Scholar] [CrossRef]
  32. Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef] [Green Version]
  33. Lucas, B.D.; Kanade, T. An Iterative Image Registration Technique with an Application to Stereo Vision; Computer Science Department, Carnegie Mellon University: Pittsburgh, PA, USA, 1981; pp. 674–679. [Google Scholar]
  34. Harris, C.; Stephens, M. A combined corner and edge detector. In Proceedings of the Alvey Vision Conference, Manchester, UK, 31 August–2 September 1988. [Google Scholar]
  35. Dong, C.Z.; Celik, O.; Catbas, F.N. Marker-free monitoring of the grandstand structures and modal identification using computer vision methods. Struct. Health Monit. 2019, 18, 1491–1509. [Google Scholar] [CrossRef]
  36. Yan, W.J.; Feng, Z.Q.; Yang, W.; Yuen, K.V. Bayesian inference for the dynamic properties of long-span bridges under vortex-induced vibration with Scanlan’s model and dense optical flow scheme. Mech. Syst. Signal Process. 2022, 174, 109078. [Google Scholar] [CrossRef]
  37. Shi, J. Good features to track. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 21–23 June 1994. [Google Scholar]
  38. Li, W.; Zhao, W.; Gu, J.; Fan, B. Dynamic characteristics monitoring of large wind turbine blades based on target-free DSST vision algorithm and UAV. Remote Sens. 2022, 14, 3113. [Google Scholar] [CrossRef]
  39. Yoon, Y.; Elanwar, H.; Choi, H.; Golparvar-Fard, M.; Spencer, B.F., Jr. Target-free approach for vision-based structural system identification using consumer-grade cameras. Struct. Control Health Monit. 2016, 23, 1405–1416. [Google Scholar] [CrossRef]
  40. Khuc, T.; Catbas, F.N. Completely contactless structural health monitoring of real-life structures using cameras and computer vision. Struct. Control Health Monit. 2017, 24, e1852. [Google Scholar] [CrossRef]
  41. Slimani, M.; Khatir, T.; Tiachacht, S.; Boutchicha, D.; Benaissa, B. Experimental sensitivity analysis of sensor placement based on virtual springs and damage quantification in CFRP composite. J. Mater. Eng. Struct. 2022, 9, 207–220. [Google Scholar]
Figure 1. General steps for camera calibration.
Figure 1. General steps for camera calibration.
Buildings 12 02021 g001
Figure 2. Target tracking based on ROI for target-free robust corner feature extraction.
Figure 2. Target tracking based on ROI for target-free robust corner feature extraction.
Buildings 12 02021 g002
Figure 3. Normal image and distorted image.
Figure 3. Normal image and distorted image.
Buildings 12 02021 g003
Figure 4. Recognition trend of radial distortion coefficient: (a) Parameter recognition of Honor X10 smartphone; (b) iPhone 12 smartphone parameter recognition.
Figure 4. Recognition trend of radial distortion coefficient: (a) Parameter recognition of Honor X10 smartphone; (b) iPhone 12 smartphone parameter recognition.
Buildings 12 02021 g004
Figure 5. Smartphone image correction: (a) Before correction; (b) After correction.
Figure 5. Smartphone image correction: (a) Before correction; (b) After correction.
Buildings 12 02021 g005
Figure 6. Smartphone displacement monitoring test equipment.
Figure 6. Smartphone displacement monitoring test equipment.
Buildings 12 02021 g006
Figure 7. ROI based target-free feature-matching: (a) Simple background; (b) Complex background.
Figure 7. ROI based target-free feature-matching: (a) Simple background; (b) Complex background.
Buildings 12 02021 g007
Figure 8. Time history comparison of vibration displacement monitored by smartphones: (a) Whole displacement time history; (b) 10~20 s displacement time history.
Figure 8. Time history comparison of vibration displacement monitored by smartphones: (a) Whole displacement time history; (b) 10~20 s displacement time history.
Buildings 12 02021 g008
Figure 9. Time history comparison of smartphone displacement monitoring through frame rate correction: (a) Whole displacement time history; (b) 10~20 s displacement time history.
Figure 9. Time history comparison of smartphone displacement monitoring through frame rate correction: (a) Whole displacement time history; (b) 10~20 s displacement time history.
Buildings 12 02021 g009
Figure 10. Directions for testers to take pictures and smartphones.
Figure 10. Directions for testers to take pictures and smartphones.
Buildings 12 02021 g010
Figure 11. Displacement time histories comparison in three directions of three shooting states using a smartphone.
Figure 11. Displacement time histories comparison in three directions of three shooting states using a smartphone.
Buildings 12 02021 g011
Figure 12. Frequency domain comparison of three shooting states using smartphone.
Figure 12. Frequency domain comparison of three shooting states using smartphone.
Buildings 12 02021 g012
Figure 13. Displacement time history comparison in the state of standing hold smartphone shooting.
Figure 13. Displacement time history comparison in the state of standing hold smartphone shooting.
Buildings 12 02021 g013
Figure 14. Displacement time history comparison under the shooting state of slightly walking hand-held smartphone.
Figure 14. Displacement time history comparison under the shooting state of slightly walking hand-held smartphone.
Buildings 12 02021 g014
Figure 15. Comparison of position shift time history de-noising using smartphone in different states: (a) Whole displacement time history; (b) 10~20 s displacement time history.
Figure 15. Comparison of position shift time history de-noising using smartphone in different states: (a) Whole displacement time history; (b) 10~20 s displacement time history.
Buildings 12 02021 g015
Figure 16. The error distribution of displacement monitoring under the common smartphone shooting mode: (a) Standing shoot; (b) Shoot when walking slightly.
Figure 16. The error distribution of displacement monitoring under the common smartphone shooting mode: (a) Standing shoot; (b) Shoot when walking slightly.
Buildings 12 02021 g016
Figure 17. Frequency domain comparison of monitoring under three common smartphone shooting states.
Figure 17. Frequency domain comparison of monitoring under three common smartphone shooting states.
Buildings 12 02021 g017
Figure 18. Comparison between camera shooting and smartphone equipped telephoto lens shooting: (a) Camera; (b) Smartphone equipped with telephoto lens.
Figure 18. Comparison between camera shooting and smartphone equipped telephoto lens shooting: (a) Camera; (b) Smartphone equipped with telephoto lens.
Buildings 12 02021 g018
Figure 19. Comparison of displacement time history of telephoto lens for smartphone assembly: (a) Whole displacement time history; (b) 10~20 s displacement time history.
Figure 19. Comparison of displacement time history of telephoto lens for smartphone assembly: (a) Whole displacement time history; (b) 10~20 s displacement time history.
Buildings 12 02021 g019
Figure 20. Experimental equipment.
Figure 20. Experimental equipment.
Buildings 12 02021 g020
Figure 21. Acceleration time history in flap-wise direction.
Figure 21. Acceleration time history in flap-wise direction.
Buildings 12 02021 g021
Figure 22. Time history comparison of vibration displacement in flap-wise direction: (a) Whole displacement time history; (b) 10~20 s displacement time history.
Figure 22. Time history comparison of vibration displacement in flap-wise direction: (a) Whole displacement time history; (b) 10~20 s displacement time history.
Buildings 12 02021 g022
Figure 23. PSD comparison of different equipment.
Figure 23. PSD comparison of different equipment.
Buildings 12 02021 g023
Figure 24. Accelerometers monitoring wind turbine structural responses.
Figure 24. Accelerometers monitoring wind turbine structural responses.
Buildings 12 02021 g024
Figure 25. Smartphone monitoring wind turbine structure responses.
Figure 25. Smartphone monitoring wind turbine structure responses.
Buildings 12 02021 g025
Figure 26. Structural modal shape of wind turbine.
Figure 26. Structural modal shape of wind turbine.
Buildings 12 02021 g026
Table 1. Parameters of two smartphone cameras.
Table 1. Parameters of two smartphone cameras.
Smartphone
Category
Smartphone PhotoFrame RateMax PixelPixel
Density
Zoom MultipleAperture
iPhone 12Buildings 12 02021 i00130/60 fps12 million460 ppi5f/2.4
Honor X10Buildings 12 02021 i00230/60 fps40 million397 ppi10f/1.8
Table 2. Displacement peak value under different conditions when using a smartphone.
Table 2. Displacement peak value under different conditions when using a smartphone.
Motion StateX-Direction/PixelY-Direction/PixelZ-Direction/Pixel
Equipped with tripod0.0150.0340.018
Standing shooting95.02572.1836.254
Walk slightly150.641282.94461.239
Table 3. Displacement monitoring error of common smartphone shooting methods.
Table 3. Displacement monitoring error of common smartphone shooting methods.
Shooting MethodRMSE ρ R2
Standing shooting0.62190.82540.8763
Walk slightly0.73420.75130.7925
Table 4. Response peak value obtained by two monitoring methods.
Table 4. Response peak value obtained by two monitoring methods.
Measuring PointsAccelerometer (mm/s2)Smartphone (mm)Measuring PointsAccelerometers (mm/s2)Smartphone (mm)
P05.6341 × 10−420.6585P5 8.3780
P1 18.1829P6 6.8293
P22.3902 × 10−415.0244P71.6911 × 10−45.4341
P3 12.0122P8 4.2276
P46.5732 × 10−510.1504P91.0098 × 10−43.4472
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhao, W.; Li, W.; Fan, B.; Du, Y. Dynamic Characteristic Monitoring of Wind Turbine Structure Using Smartphone and Optical Flow Method. Buildings 2022, 12, 2021. https://doi.org/10.3390/buildings12112021

AMA Style

Zhao W, Li W, Fan B, Du Y. Dynamic Characteristic Monitoring of Wind Turbine Structure Using Smartphone and Optical Flow Method. Buildings. 2022; 12(11):2021. https://doi.org/10.3390/buildings12112021

Chicago/Turabian Style

Zhao, Wenhai, Wanrun Li, Boyuan Fan, and Yongfeng Du. 2022. "Dynamic Characteristic Monitoring of Wind Turbine Structure Using Smartphone and Optical Flow Method" Buildings 12, no. 11: 2021. https://doi.org/10.3390/buildings12112021

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop