Next Article in Journal
Effect of Irradiation Process on Physical and Chemical Properties and Mildew Resistance of Bamboo
Previous Article in Journal
Uncertainties of Climate Change on Suitable Habitat of Calligonum mongolicum Turcz. (Polygonaceae) in Desert Areas of Central Asia
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MISF: A Method for Measurement of Standing Tree Size via Multi-Vision Image Segmentation and Coordinate Fusion

1
College of Mathematics and Computer Science, Zhejiang A&F University, Hangzhou 311300, China
2
Information and Education Technology Center, Zhejiang A&F University, Hangzhou 311300, China
3
School of Information Engineering, Huzhou University, Huzhou 313000, China
*
Author to whom correspondence should be addressed.
Forests 2023, 14(5), 1054; https://doi.org/10.3390/f14051054
Submission received: 2 April 2023 / Revised: 6 May 2023 / Accepted: 13 May 2023 / Published: 20 May 2023
(This article belongs to the Section Forest Inventory, Modeling and Remote Sensing)

Abstract

:
With the development of computer vision technology, its applications in forestry are steadily becoming wider. To address the problems of inconvenience in transporting unmanned aerial vehicles (UAVs), as well as the complex operation of large instruments for measurement, a new method based on multi-vision image segmentation and coordinate fusion (MISF) is proposed in this paper for measuring the size of standing trees. In MISF, after images of a standing tree are captured using a camera from multiple angles, a semantic segmentation method based on deep learning is used to segment the main body of the standing tree and automatically detect the edge feature points. Next, the effects of visual field splicing and fusion are analyzed collaboratively using the correlations among images, so as to restore the three-dimensional spatial information of the feature points of the tree to be measured. Lastly, the size attributes of the standing tree, such as height, diameter at breast height (DBH), and crown width, are automatically measured. The urban environment measurement experiment showed that the relative errors of tree height, DBH, and crown width measured using the proposed method, i.e., MISF, were 1.89%, 2.42%, and 3.15%, respectively, representing a significant enhancement compared with binocular measurement. On the one hand, the experimental results exhibited a high degree of measurement accuracy; therefore, MISF can be used for the management inventory of typical forests. On the other hand, MISF cannot be used if a tree’s images cannot be acquired due to environmental or other reasons.

1. Introduction

In forestry, collecting tree-related information is essential for analyzing forestry resources [1,2]. This information consists of the height, diameter at breast height (DBH), and other data of standing trees. Traditional methods measure standing trees using rulers or other measuring tools [3], which is time-consuming and labor-intensive. On the other hand, precise equipment, such as electronic theodolites and total stations, is inconvenient to use because of its relatively complex operation [4,5].
At present, most measurements of standing trees are conducted using hardware, such as laser unmanned aerial vehicles (UAVs), terrestrial laser scanners, and other height-equipment measurement methods [6,7]. For example, Li et al. [8] proposed a new tree diameter measurement instrument developed using a self-reset displacement sensor, which was equipped with a personal computer to enable the measurement, transmission, storage, and analysis of tree diameters. Olofsson et al. [9] proposed a method to automatically detect trees in a TLS-scanned field plot, classify the laser return as either into stem or canopy for each of the detected trees, and extract forest variables from the classified data. Yuan et al. [10] developed a measuring device for tree DBH and tree height using a high-precision laser ranging sensor. The combination of the accurate identification and segmentation of standing trees using multi-vision measurement is less well-studied. The traditional fixed binocular measurement model still has certain limitations and flexibility, whereas multi-vision measurement can make up for these shortcomings, improving the stability and accuracy [11,12].
With the continuous development of science and technology mirroring the progress of computer vision, more and more measurement methods have been proposed. Adilson et al. [13] presented an automatic technique for the mapping and measurement of individual tree stems using vertical terrestrial images collected with a fisheye camera. Krause et al. [14] used UAV photogrammetry to semiautomatically obtain the height of standing trees. Ramli et al. [15] proposed the use of an UAV to capture images of oil palm trees, followed by the generation of digital models and the processing of orthophoto images using Agisoft software; then, they calculated the crown size of oil palm trees by implementing the seed generation tool of the quantum geographic information system (QGIS) and the system for automatic geological analysis (SAGA) segmentation methods. Wu et al. [16] proposed a method based on monocular vision to measure the DBH of standing trees by photographing a single image using a camera.
Addressing the shortcomings identified from the literature review, a flexible method for the measurement of standing trees using multi-vision image segmentation and coordinate fusion (MISF) is proposed in this paper.
The main contributions of this paper include three aspects:
(1) A new multi-view measurement model (MISF) for standing trees is proposed, combining the advantages of image processing to realize the noncontact automatic measurement of standing trees. A camera is used to capture standing trees from three angles, thereby obtaining a wider field of view and improving the accuracy of the matching position of feature points.
(2) The image matching algorithm SURF (speeded up robust features) is improved to enable the matching of feature points in standing tree images. The original SURF algorithm has good stability and low computing complexity. However, the SURF algorithm transforms color images into grayscale, thus partially losing the color information. In view of this problem, we improved the algorithm SURF by adding a 48-dimensional RGB color information descriptor. Accordingly, the color information could be well-retained compared with the original SURF algorithm, thus increasing the matching performance and the robustness of the algorithm.
(3) Extensive experiments were conducted so as to evaluate the performance of the proposed method (MISF). In the experiments, some tree size attributes such as height, DBH, and crown width were measured using MISF to verify its applicability. Its performance was also compared with traditional binocular measurement and other measurement methods in the literature. The experiments revealed that the performance of MISF could meet the requirements of forest management inventory.
The method proposed in this paper achieves a significant improvement in measurement accuracy over the traditional binocular measurement method, improving the practical application. The measurement errors of tree height, diameter at breast height, and crown width were comparable with those of other measurement methods, in addition to meeting the accuracy requirements for the continuous inventory of forest resources.

2. Materials and Methods

2.1. Main Ideas

In the theory of close-up photogrammetry, the utilization of a single camera to capture an image in order to acquire stereo information regarding the target is referred to as monocular measurement. This method is low cost but limited to obtaining only two-dimensional information about the target, and cannot ascertain the distance of the measured target. On the other hand, binocular measurement involves using two cameras to capture images from different perspectives, thereby enabling the acquisition of three-dimensional coordinate information of the target. However, this method is constrained by a certain range of field of view and requires the use of two parallel cameras, which may introduce some degree of error due to human factors.
On this basis, the present paper proposes an innovation that incorporates the intersection measurement method [17]. This approach eliminates the need for parallel optical axes between cameras and instead involves capturing multiple images from varying angles of a standing tree to reconstruct its three-dimensional information. During the image capture process, it is only necessary to ensure that the main body of the tree being measured is within the common field of view among the three shooting perspectives. This condition ensures that the common field of view is sufficiently large to enhance measurement accuracy without requiring strict location specifications.
The main concept behind the proposed method for measuring the size of standing trees, known as MISF, is based on multi-vision image segmentation and coordinate fusion, as illustrated in Figure 1. To facilitate measurement, three binocular subsystems were established using cameras to capture images of standing trees from varying perspectives. In the following step of the method, the segmentation of standing trees in the images is performed to accurately identify their positions. Next, three sets of binocular sub-models are utilized to obtain the group of largest measurements via comparison. The final standing-tree factor measurement is achieved through the fusion of the results obtained from the segmentation of standing trees and size measurement.
When compared to binocular stereo vision measurement methods and height equipment measurement techniques, such as UAV and scanner 3D reconstruction, the MISF method exhibits distinct differences and notable advantages in several key areas, including:
(1) The MISF method employs a deep learning-based image segmentation technique to extract the standing tree factor, which enables automatic contour extraction and segmentation results without requiring manual intervention. As a result, the MISF method is characterized by enhanced efficiency and automation, significantly reducing manpower and time costs.
(2) Compared to a conventional binocular camera, the MISF method can acquire additional views and feature point information. While binocular cameras only capture two parallel-view images, resulting in improved accuracy, the MISF method employs three cameras that can cross-verify each other, thereby increasing the reliability of measurement results.
(3) When compared to high equipment measurement methods, the MISF method is characterized by a lower cost. This approach eliminates the need for repeated measurements by humans in the field; instead, the camera position only needs to be fixed once to achieve repeatable measurements.

2.2. Multiple Vision Image Acquisition

2.2.1. Single Camera Calibration

The MISF method adopts Zhang’s [18,19] camera calibration approach using a checkerboard for calibration. This method accounts for both radial and tangential distortion and facilitates the capture of multiple sets of images from various angles and directions by moving the calibration board. This flexible, convenient, and accurate technique requires a checkerboard image containing 9 x 12 angle points, with each checkerboard measuring 15mm in physical length, as shown in Figure 2.
The calibration process for a single camera primarily comprises four steps. First, the physical coordinates of the feature points on the calibration board are optimized to solve the homologous matrix. Second, the stack matrix B is calculated during the calibration process, which allows for the acquisition of the camera’s intrinsic parameter matrix. Third, the extrinsic parameters of the camera are solved, followed by the derivation of the camera distortion coefficients.. Fourth, the maximum likelihood method is utilized to optimize the intrinsic parameters and improve the estimation accuracy. Each step is described in detail below.
(1) Solving homologous matrix
After the physical coordinates of the characteristic points on the calibration plate were transposed, the homologous matrix was solved. As for computer vision, a homologous matrix is defined as a two-dimensional coordinate point mapped onto a camera image. Assuming that the coordinate of the feature point is Q = X m Y m Z m T , the coordinates of the image on the camera are q = u v 1 T . The monograph relationship between the calibration plate plane and the camera image plane is shown as Equation (1):
s q = P R , T X
In Equation (1), s is the scale factor, P represents the parameter matrix of the camera, R represents the rotation matrix, and T denotes the translation vector. Suppose that R , T = r 1 r 2 r 3 t , the world coordinate system is established on the checkerboard in the calibration process, namely, Z = 0 in the checkerboard plane, and then the homologous matrix D can be derived by s u v 1 = D X Y 1 .
(2) Calculation of camera internal parameter matrix
Through the above, the homologous matrix D follows the expression of Equation (2):
D = h 1 h 2 h 3 = P r 1 r 2 t s
The camera intrinsic parameter matrix contains five unknowns, and the homologous matrix can generate two equations, so at least three calibration plates are required for calibration. For the convenience of later calculation, the stacking matrix B is assumed as Equation (3), where there are six unknowns. The vector b is assumed as Equation (4):
B = P T P 1 = B 11 B 12 B 13 B 21 B 22 B 23 B 31 B 32 B 33
b = B 11 B 12 B 22 B 13 B 23 B 33 T
Then, the following Equation (5) is reached:
h i T B h j = v i j T b
Because of the constraint of Equation (6) on the calibration process, it is necessary to obtain the matrix B through multiple checkerboard images, and finally to obtain the intrinsic parameter matrix P through Cholesky decomposition:
v 12 T v 11 v 22 T b = 0
(3) Solving the camera extrinsic parameters
According to Equation (2), let λ = 1 s , then, Equation (7) can be derived:
r 1 = λ P 1 h 1 r 2 = λ P 1 h 2 r 3 = r 1 × r 2 t = λ P 1 h 3
In Equation (7), λ = 1 P 1 h 1 = 1 P 1 h 2 . Then, the extrinsic parameter matrix [R, T] of the camera can be calculated.
(4) Results optimization
According to the derivation of above equations, the results can be obtained under ideal conditions. However, in reality, the results are often affected by Gaussian noise, which usually leads to suboptimal results. Therefore, it is necessary to optimize the results. In MISF, the maximum likelihood estimation method is adopted to optimize the results, mainly optimizing the intrinsic parameters of the camera for single-camera calibration results.
Assuming that n calibration board images were acquired, each of which contains m corner points, the j-th corner point in the i-th image can be calculated using the above projection equations and the following Equation (8) can be obtained:
m P , R i , t i , M i j = P R t M i j
In Equation (8), P denotes the camera internal parameter matrix, Ri denotes the rotation matrix of the i-th image, and ti denotes the translation vector of the i-th image. Thus, the corresponding likelihood function is established as Equation (9), and the Levenberg–Marquardt [20] algorithm is used to find the optimal solution.
L P , R i , t i , M i j = 1 2 π e i 1 n j 1 m m ^ P , R i , t i , M i j m i j 2 σ 2
The form of the distortion coefficient matrix is shown as Equation (10), where (u, v) is the ideal pixel coordinates, (u′, v′) represents the pixel coordinates that produce distortion in the real state, (u0, v0) represents the position of the main point on the image, (x, y) represents the continuous coordinates in the optimal state, and k1, k2 are the distortion coefficients of the first two terms.
u u 0 x 2 + y 2 u u 0 x 2 + y 2 2 v v 0 x 2 + y 2 v v 0 x 2 + y 2 2 k 1 k 2 = u u v v
Transform Equation (10) to the form Hk = d, which leads to Equation (11):
k = k 1 k 2 T = H T H 1 d
Finally, the distortion coefficient k is obtained, and the maximum likelihood estimation method is used to estimate the distortion coefficient accurately, as is shown in Equation (12).
i = 1 n j = 1 m m ^ P , k 1 , k 2 , R i , t i , M i j m i j 2

2.2.2. Trinocular Vision Calibration

In MISF, the calibration method by Zhang [18,19] is adopted, which is used in the previous monocular camera calibration to obtain the intrinsic and extrinsic parameters of three cameras. However, in order to calculate the three-dimensional coordinates of feature points, it is also necessary to obtain the inter-camera visual parameters. Based on the results of individual camera calibration, coordinate transformation is used to achieve mutual calibration among the three cameras, as shown in Figure 3.
In the trinocular vision measurement model, the external parameters of camera 1, camera 2, and camera 3 are Ri and Ti (i = 1, 2, 3), respectively, and represent the relative position relationship between their respective cameras and their corresponding world coordinate system. For a point E to be measured, En represents its world coordinate position, E1 represents the coordinates of camera 1 in the self-coordinate system, E2 represents the coordinates of camera 2 in the self-coordinate system, and E3 represents the coordinates of camera 3 in the self-coordinate system. They comply with Equation (13):
E 1 = R 1 E n + T 1 E 2 = R 2 E n + T 2 E 3 = R 3 E n + T 3
The rotation matrix and translation vector between camera 1 and camera 2 are R12 and T12, respectively, the rotation matrix and translation vector between camera 2 and camera 3 are R23 and T23, respectively, and the rotation matrix and translation vector between camera 3 and camera 1 are R13 and T13, respectively, so Equations (14)–(16) can be obtained:
R 12 = R 2 R 1 1 T 12 = T 2 R 2 R 1 1 T 1
R 23 = R 3 R 2 1 T 23 = T 3 R 3 R 2 1 T 2
R 13 = R 3 R 1 1 T 13 = T 3 R 3 R 1 1 T 1
After knowing the rotation matrix and translation vector of the three cameras, the position structure parameters of the trinocular vision measurement model can be calculated by Equations (14)–(16).
The calibration of the trinocular vision model is similar to the single camera calibration method. As the last step, the parameter values of three cameras (R12, T12), (R23, T23), and (R13, T13) are obtained using the Levenberg–Marquardt method for optimization. In the solution of the optimization, the internal parameters and distortion coefficients of the three cameras’ self-calibration will also change. That is to say, in the first measurement group, the original internal parameters of camera 1 and camera 2 are P1 and P2, and they are changed to P 12 1 and P 12 2 after optimization. Similarly, the internal parameters of the second and third measurement groups are optimized as P 23 2 , P 23 3 and P 13 1 , P 13 3 , respectively. After nonlinear optimization, all parameters required for the calibration of the trinocular vision measurement model are obtained.

2.2.3. Image Acquisition

To ensure the inclusion of a common target standing-tree subject in all three images, we acquired images of the standing tree using a camera at three different locations. In contrast to the traditional binocular parallel measurement approach that requires keeping the optical axes parallel to each other, the MISF method proposed in this paper requires orienting the camera optical axes towards the target tree. To obtain images of standing trees from multiple directions, we recommend positioning the three cameras in a triangular configuration with a certain horizontal angle (30°–45°) between the optical axes of two adjacent cameras. The shooting schema is depicted in Figure 4.

2.3. Multiple Vision Image Processing

2.3.1. Semantic Segmentation of Tree Images Based on Deep Learning

In the real-world environment, photos often exhibit complex backgrounds with significant amounts of noise. To effectively extract the required size information of standing trees from images, image segmentation is crucial. Within MISF, we leverage SEMD, a lightweight model for standing-tree image segmentation that we previously proposed in [21]
The SEMD segmentation model utilizes MobileNetV2 + SENet as the backbone network in the encoder section. This structure can calculate and extract features with any resolution, and the ratio of the input image spatial resolution to the output image spatial resolution is used as the output stride. To perform multi-scale fusion, different dilated convolutions are employed to process features calculated by MobileNetV2 + SENet, which effectively reduces the running time without compromising segmentation accuracy. In the decoder section, a similar convolutional network is utilized to reduce the number of information channels and address difficulties encountered during training due to an excess of low-feature channels.

2.3.2. Image Feature Point Extraction and Matching

Feature point extraction and matching between two images is the basis and prerequisite for measuring the size of standing trees. In MISF, the algorithm SURF [22] is adopted and improved to obtain feature point information.
In SURF, feature points are mainly achieved through three aspects: (1) setting a threshold: remove feature points smaller than the threshold; (2) suppressing non-maximum values by only selecting the total attack of the set scale and the upper and lower adjacent scales 26; (3) using the fitted three-dimensional quadratic function to obtain the sub-pixel position of the feature point after interpolation.
For the feature point extraction of the original SURF algorithm, the color image needs to be converted to grayscale [23], which may lead to a loss of the color information of the picture. In view of this, the MISF algorithm improves SURF by adding a 48-dimensional RGB color descriptor to preserve the color information and increase the robustness of the algorithm.
First, change the direction of the coordinate axis of the coordinate system so that the coordinate axis is consistent with the main direction of the feature point. Then, divide the area of 16s × 16s (s is the scale where the feature point is located) centered on the feature point into 4 × 4 = 16 sub-regions. Calculate the RGB color mean value of the image in these 16 sub-regions, so that each small region will generate a three-dimensional color description vector, as shown in Equation (17), where Ravg, Gavg, and Bavg represent the means of the RGB colors. In this way, 4 × 4 small regions become 3 × 4 × 4 = 48-dimensional color descriptors:
C = R a v g , G a v g , B a v g
Finally, the 112-dimensional descriptor composed of the 64-dimensional feature descriptor obtained by the SURF algorithm and the 48-dimensional color descriptor are used to represent the feature point information [24]. An example of the matching result is shown in Figure 5.

2.4. Measurement of Multi-Vision Standing Tree Size

2.4.1. Trinocular Convergence Vision Model

After the feature points in two images are obtained and matched using the improved SURF algorithm as described in Section 2.3 and the internal and external parameters of the orientation obtained as described in Section 2.2, the 3D information of the feature points of the standing trees can be obtained with following presentation calculations.
The photogrammetry technique is used to obtain the size attributes of the standing trees, which require the interconversion of the physical coordinate system, camera coordinate system, and world coordinate system, i.e., to complete the mapping between two-dimensional information and three-dimensional information.
In MISF, an additional camera is added according to the theory of the binocular vision ranging model, so that the respective optical axes of the three cameras form a certain angle, which can also be interpreted as three groups of binocular vision model. The schematic diagram is shown in Figure 6.
Suppose E is a point in space, and e1, e2, and e3 are its corresponding points on the imaging plane of cameras 1, 2, and 3, respectively. Without considering the existence of any error, the measurement value E1 measured by the binocular vision measurement composed of camera 1 and camera 2, the measurement value E2 measured by the binocular vision measurement composed of camera 2 and camera 3, and the measurement value E3 measured by the binocular vision measurement composed of camera 3 and camera 1 should coincide with the actual measurement value E at one point. That is, the three lines O1e1, O2e2, and O3e3 converge to the same point E. However, in the real environment, there may be an error in binocular vision ranging, which leads to the three points not coinciding with the actual measurement points, and then the generated points are three new points E1, E2, and E3 formed by the intersection of the three lines in pairs. Their 3D coordinates can be calculated by the binocular vision ranging algorithm.
Assuming that the coordinates of point E to be measured are E(Xn, Yn, Zn), the actual coordinates of the point E are optimally estimated by the objective function of Equation (18):
F = min E E 1 + E E 2 + E E 3
The coordinates of the three points e1, e2, and e3 projected by point E on the imaging plane of cameras 1, 2, and 3 are calculated. Then, since the cameras have been calibrated and M1, M2, and M3 are their projection matrices, respectively, the coordinates can be converted as Equations (19)–(21):
Z c 1 u 1 v 1 1 = M 1 X n Y n Z n 1 = m 11 1 m 12 1 m 13 1 m 14 1 m 21 1 m 22 1 m 23 1 m 24 1 m 31 1 m 32 1 m 33 1 m 34 1 X n Y n Z n 1
Z c 2 u 2 v 2 1 = M 2 X n Y n Z n 1 = m 11 2 m 12 2 m 13 2 m 14 2 m 21 2 m 22 2 m 23 2 m 24 2 m 31 2 m 32 2 m 33 2 m 34 2 X n Y n Z n 1
Z c 3 u 3 v 3 1 = M 3 X n Y n Z n 1 = m 11 3 m 12 3 m 13 3 m 14 3 m 21 3 m 22 3 m 23 3 m 24 3 m 31 3 m 32 3 m 33 3 m 34 3 X n Y n Z n 1
In Equations (19)–(21), Mi represents the respective projection matrices of the three cameras, where Mi = Ai [Ri Ti] (i = 1, 2, 3), (ui, vi, 1), (i = 1, 2, 3) are the corresponding homogeneous coordinates of the pixel coordinates on the imaging of the three projection points, respectively. (Xn, Yn, Zn, 1) represents the homogeneous coordinates of the point E to be measured. Suppose the coordinates of the three measured points are Ei(Xni, Yni, Zni, 1), i = 1, 2, 3, respectively. The least square result of this coordinate can be calculated using the binocular vision measurement algorithm as Equations (22)–(25):
E i : X n i Y n i Z n i = K i T K i 1 K i T J i , i = 1 , 2 , 3
K 1 = u 1 m 31 1 m 11 1 u 1 m 32 1 m 12 1 u 1 m 33 1 m 13 1 v 1 m 31 1 m 21 1 v 1 m 32 1 m 22 1 v 1 m 33 1 m 23 1 u 2 m 31 2 m 21 2 u 2 m 32 2 m 12 2 u 2 m 33 2 m 13 2 v 2 m 31 2 m 21 2 v 2 m 32 2 m 22 2 v 2 m 33 2 m 23 2 B 1 = m 14 1 u 1 m 34 1 m 24 1 v 1 m 34 1 m 14 2 u 2 m 34 2 m 24 2 v 2 m 34 2
K 2 = u 2 m 31 2 m 11 2 u 2 m 32 2 m 12 2 u 2 m 33 2 m 13 2 v 2 m 31 2 m 21 2 v 2 m 32 2 m 22 2 v 2 m 33 2 m 23 2 u 3 m 31 3 m 11 3 u 3 m 32 3 m 12 3 u 3 m 33 3 m 13 3 v 3 m 31 3 m 21 3 v 3 m 32 3 m 22 3 v 3 m 33 3 m 23 3 B 2 = m 14 2 u 2 m 34 2 m 24 2 v 2 m 34 2 m 14 3 u 3 m 34 3 m 24 3 v 3 m 34 3
K 3 = u 1 m 31 1 m 11 1 u 1 m 32 1 m 12 1 u 1 m 33 1 m 13 1 v 1 m 31 1 m 21 1 v 1 m 32 1 m 22 1 v 1 m 33 1 m 23 1 u 3 m 31 3 m 11 3 u 3 m 32 2 m 12 3 u 3 m 33 3 m 13 3 v 3 m 31 3 m 21 3 v 3 m 32 3 m 22 3 v 3 m 33 3 m 23 3 B 3 = m 14 1 u 1 m 34 1 m 24 1 v 1 m 34 1 m 14 3 u 3 m 34 3 m 24 3 v 3 m 34 3
Then, Equation (26) is reached, and the three conditions shown in Equation (27) should be satisfied for the validity of Equation (26):
F = min E E 1 + E E 2 + E E 3 = X n X n 1 2 + Y n Y n 1 2 + Z n Z n 1 2 + X n X n 2 2 + Y n Y n 2 2 + Z n Z n 2 2 + X n X n 3 2 + Y n Y n 3 2 + Z n Z n 3 2
f 1 = min X n X n 1 2 + X n X n 2 2 + X n X n 3 2 f 2 = min Y n Y n 1 2 + Y n Y n 2 2 + Y n Y n 3 2 f 3 = min Z n Z n 1 2 + Z n Z n 2 2 + Z n Z n 3 2
The sum of squared deviations between each variable and their arithmetic square number is the smallest, so the optimal value of the measured point E(Xn, Yn, Zn) can be obtained as follows:
X n = X n 1 + X n 2 + X n 3 3 Y n = Y n 1 + Y n 2 + Y n 3 3 Z n = Z n 1 + Z n 2 + Z n 3 3
Using this method, the coordinate values of the point to be measured can be optimized so as to improve the accuracy of measurement.

2.4.2. Solving the Initial Value in 3D Space

For the convenience of expression, the combination of camera 1 and camera 2 is called the first measurement group, the combination of camera 2 and camera 3 is called the second measurement group, and the combination of camera 3 and camera 1 is called the third measurement group. For the first and third measurement groups, the coordinate system of camera 1 is considered as the world coordinate system, and the coordinate system of camera 2 is considered as the world coordinate system for the second measurement group.
When calibrating the trinocular vision measurement system, the calibration results of the internal and external parameters of camera 1, 2, and 3 have been optimized to a certain degree. According to Mi = Ai [Ri Ti] (i =1, 2, 3) mentioned above, the projection matrices of camera 1, 2, and 3 are composed of the internal and external parameters of the cameras. So, M1, M2, and M3 will also change along with the optimization of the internal and external parameters of the camera. In other words, for each binocular measurement model composed of two cameras, their respective internal projection matrix results are optimized into A i j i , A i j j , and M ij i , M ij j , where both i and j represent the index of camera, 1, 2, or 3.
Let S be a matrix of third-order units and O = [0 0 0]T, then, the respective internal parameters and projection matrices of the two cameras in each measurement group should satisfy Equations (29) and (30), and they can be expressed as Equations (31) and (32) in the model, respectively:
M i j i = A i j i S   O T
M i j j = A i j j R ij   T ij
Z i j c i u i v i 1 = M i j i X n Y n Z n 1
Z i j c j u j v j 1 = M i j j X n Y n Z n 1
Here, Z ij ci and Z ij cj represent the depth direction values in the camera coordinate system of the binocular measurement model composed of pairwise cameras (i, j). Combined with Equations (19)–(25) and (29)–(32) mentioned above, the initial values of the feature points of the three groups of binocular vision models can be solved.

2.4.3. Coordinate Fusion

The camera 1 coordinate system is used for measurement group 1 and measurement group 3 and is considered as the world coordinate system, but the world coordinate system in measurement group 2 is based on camera 2. In MISF, camera 1 is selected as the reference camera of the whole system, and its coordinate system is regarded as the global coordinate system. In order to merge the three sets of initial value information into the same global coordinate system, the original measurement values of the second measurement group need to be converted into the coordinate system of camera 1.
In the second measurement group, the coordinate system of camera 2 is used as the world coordinate system, and the point to be measured satisfies E 2 = R 12 E 1 + T 12 in the coordinate systems of camera 1 and 3. Therefore, these three sets of initial values can be unified in the world coordinate system by converting the initial value information provided by the second measurement group into the camera 1 coordinate system. As a result, these three sets of initial values are all fused into the world coordinate system of camera 1.

2.4.4. Measurement of Standing Tree Size

(1) Height of tree
In traditional methods, altimeters are used for the measurement of the height of trees. This paper analyzes and measures the factors of standing trees by taking pictures of standing trees from different perspectives. Based on the above description, standing trees are segmented accurately. The three-dimensional homogeneous coordinates of the two points are calculated by searching for the highest feature point of the standing tree whose coordinates are X n 1 Y n 1 Z n 1 and the root feature point whose coordinates are X n 2 Y n 2 Z n 2 . As shown in Figure 7, the tree height is calculated by Equation (33):
H = X n 1 X n 2 2 + Y n 1 Y n 2 2 + Z n 1 Z n 2 2
(2) Diameter at breast height (DBH)
The DBH of standing trees is also an important attribute of standing trees in forestry. It refers to the cross-sectional diameter of the standing tree at a distance of 1.3 m from the ground. For those short standing trees with a trunk less than 1.3 m, the ground diameter at a distance of 0.2 m from the root on the ground is measured. In MISF, based on an image in which the trunk of the standing tree is clearly segmented, the two-dimensional ordinate corresponding to the breast height of the standing tree at a distance of 1.3 m from the ground and the coordinates of two intersection points of the trunk contour of the standing tree under the same axis are calculated first. Then, the three-dimensional coordinates corresponding to the two intersection coordinates are calculated, and finally the distance between the two three-dimensional coordinates are calculated, which is the DBH of the standing tree, as shown in Figure 8.
(3) Crown width
Using the three binocular measurement groups in the trinocular vision measurement, the two pixels points on the leftmost and rightmost sides of the standing tree crown are determined, and the pixel points are converted into points in the three-dimensional space using the camera imaging principle so as to obtain the distance, and the greatest distance among the three sub-models is considered as the final result of the crown width, as shown in Figure 9.

2.5. Experimental Design

2.5.1. Datasets

The data for the following experiments were collected from standing trees in the campus or the surrounding area of Zhejiang A&F University, within about five kilometers (Figure 10).
The dataset for the training of semantic segmentation consists of images of 778 standing trees captured by digital cameras under different backgrounds and different shooting angles. The dataset includes standing tree images in simple and complex backgrounds. Every image includes a single tree or multiple standing trees, so as to improve the accuracy of the semantic segmentation results of standing tree images.
The datasets for the experiments of measurement of three attributes of standing tree size include images of 25 standing trees for the measurement of height, DBH, and crown width, respectively.

2.5.2. Evaluation Indicators

The experiment evaluated the measurement accuracy of the three size attributes of standing tree, namely height, DBH, and crown width. The relative error was used as the evaluation metric, which was calculated using Equation (34).
In Equation (34), V′ means the value of measured, V means the reference value:
δ = V V V × 100 %

2.5.3. Comparison Methods

In the conducted experiments, both the method MISF proposed in this paper and the binocular measurement method [25] were utilized to measure and compare the three size attributes (tree height, DBH, and crown width) of each standing tree sample, with the aim of evaluating the performance of MISF. The binocular parallel measurement method, which utilizes two cameras with identical configurations arranged in a parallel orientation, was employed as the most commonly used structure of the vision model for comparison purposes. This approach is referred to as the parallel stereo model, also known as the standard vision model.
Furthermore, in order to evaluate the effectiveness of MISF in measuring tree height, a comparison was made with an alternative method proposed by Ganz et al. [26] that utilizes UAV images. For the measurement of DBH, the performance of MISF was compared against another method proposed by Putra et al. [27] that employs optical sensors. Lastly, for the measurement of crown width, the performance of MISF was compared with that of Wu et al. [28], who proposed a method using a single image obtained via mobile phone and triangulation calculations to derive crown width values.

2.5.4. Experimental Scheme

The multi-vision measuring model employed in this study utilizes three cameras positioned at different angles to extract the primary information related to standing trees. By integrating the multi-dimensional information obtained, accurate measurements of the height, crown width, and DBH can be achieved, with the final result being the maximum value determined across the three measurement models. The experiments conducted in this paper are primarily focused on the following three aspects.
(1) Standing tree segmentation
SEMD, a lightweight model proposed in our previous study [21] for standing tree image segmentation, is utilized in this work for the segmentation of standing tree from images. The model was trained using the dataset described in Section 2.5.1.
(2) Measurement of tree size
To measure the size attributes (e.g., height, DBH, and crown width) of standing trees, three datasets, each comprising 25 standing trees, were utilized as described in Section 2.5.1. Image processing was performed using the deep learning framework PyTorch on a computer running Windows 11 64-bit operating system, equipped with an AMD Ryzen 7 5800H CPU with Radeon Graphics @3.20 GHz, and 16 GB of memory. The development environment used was Python 3.6. The images were of the size 2736 × 2736 and were captured using cameras with a focal length of 27 mm (camera type: BMH-AN10), positioned at a height of 1.4 m.
(3) Accuracy comparison of tree size
The measurement accuracy of tree height, DBH, and crown width was compared against other methods described in Section 2.5.3. The performance of the three attributes was evaluated by combining the measurement results and regression analysis, with a focus on analyzing the underlying reasons for any observed differences.

3. Results

The principal objective of the MISF, as proposed in this paper, is to achieve automatic and precise measurement of the size of standing trees. Through experimental validation, MISF has demonstrated high accuracy and efficiency. The approach to measuring standing tree size according to MISF is comprised of two stages: image segmentation and calculation of tree size attributes.

3.1. Results of Standing Tree Segmentation

The results of segmentation of some standing tree image using SEMD are shown in Figure 11.
The SEMD exhibits noteworthy segmentation performance, enabling the precise detection of standing trees and facilitating successful segmentation within real-world scenarios, as depicted in Figure 11.

3.2. Measurement of Tree Height

In order to assess the efficacy of the proposed method, MISF, a comparative analysis was conducted against the binocular measuring method. Table 1 presents the results obtained from both methods with respect to tree height measurements.
Table 1 illustrates that the trees selected for measurement range in height from 3.15 m to 8.76 m. The results obtained using MISF exhibit measurement errors ranging from 0.76% to 2.91%, with an average error of 1.89%. Conversely, when employing the binocular method, the measurement errors range from 1.28% to 5.53%, with an average error of 3.72%. Therefore, it can be concluded that the proposed approach, MISF, offers an improvement in accuracy of 1.83% compared to the binocular measuring method.
Table 1 indicates that there is no significant correlation between tree height and measurement error. This can be attributed to the fact that the error of measurement primarily hinges on the complexity of the tree background when employing the MISF method. In this regard, simpler backgrounds correspond to higher segmentation accuracy and smaller errors. Conversely, the binocular measuring method relies on two-dimensional images taken under parallel vision, with measurement errors predominantly caused by occlusion of the highest point of the tree or the invisibility of pixels from other angles. By contrast, MISF employs a convergent imaging approach, capturing multiple images of standing trees from diverse angles. This expands the common field of view, providing greater amounts of relevant characteristic information pertaining to standing trees, ultimately leading to an improved measurement accuracy. The results of the linear regression of measured and actual tree height values are shown in Figure 12.
Figure 12a depicts the linear correlation between actual tree height values and those obtained through the binocular method. The correlation coefficient of tree height is 0.985, with an average square root error of 0.179 m. Conversely, Figure 12b demonstrates a significant linear correlation between reference values of tree height and those measured using MISF, with the variance analysis indicating no statistically significant differences between groups (p > 0.05). The correlation coefficient of tree height is 0.994, with an average square root error of 0.112 m. In comparison to the binocular measurement, this represents an improvement of 0.067 m.
In the context of tree trunk images captured by UAV in the literature [22], high levels of accuracy with respect to individual tree height estimation can be achieved by ensuring a low flight altitude, small camera lens angle, and accurate orientation. The resulting correlation coefficient R2 of tree height is 0.97. However, in utilizing MISF for this purpose, a tree height correlation coefficient of 0.994 is attained due to the convergent imaging approach employed, which captures images from multiple directions and yields a superior field of view, while also providing a simpler operation compared to the single-image approach used by UAV.

3.3. Measurement of DBH

The results of DBH measurements using MISF and the binocular measurement method are shown in Table 2.
Table 2 demonstrates that the DBH of the measured trees ranges from 10.71 cm to 30.63 cm. The resulting measurement errors using MISF range from 1.95% to 3.12%, with an average error of 2.42%. Conversely, when utilizing binocular measurement, errors range from 3.25% to 4.54%, with an average error of 4.07%. Thus, compared to binocular measurement, MISF offers an improvement in accuracy of 1.65%.
Table 2 indicates that the size of the DBH exerts a significant influence on measurement accuracy, with larger DBH values corresponding to higher levels of accuracy. This can be attributed to the fact that trunks possessing a large DBH are more prominent in images, resulting in improved segmentation accuracy. Meanwhile, trees with a smaller DBH are more susceptible to the impact of the background and other factors during segmenting, ultimately leading to reduced accuracy. The source of error may stem from the influence of tree background factors during image capture, which results in inaccurate segmentation, thereby causing feature points to become incorrect and producing errors in calculation.
Figure 13 depicts the results of linear regression between measured and actual DBH values.
Figure 13a illustrates the linear correlation between the actual DBH values and those obtained using binocular measurement. The correlation coefficient of DBH is 0.974, with a mean square root error of 0.817 cm. Conversely, Figure 13b demonstrates a significant linear correlation between the reference DBH values and those measured using MISF, with variance analysis indicating no statistically significant differences between groups (p > 0.05). The correlation coefficient of DBH is 0.991, with a mean square root error of 0.486 cm.
The correlation coefficient of DBH as measured by the proposed MISF method has achieved a score of 0.991. This accuracy surpasses that of the standing tree DBH measurement using a single image captured by a smartphone, as outlined in the literature [23]. Additionally, the R2 correlation coefficient for DBH reaches 0.95. This improved accuracy may be attributed to the expanded receptive field of the MISF method, which allows for more accurate feature point detection at a distance of 1.3 meters from the standing tree and subsequently more accurate distance calculations.

3.4. Measurement of Crown Width

The results of crown width measurements using MISF and binocular method are shown in Table 3.
Table 3 displays a range of tree crown widths between 2.03 m and 4.89 m, with measurement errors ranging from 1.31% to 5.23%, with an average error of 3.15%. The binocular measurements exhibit errors ranging from 3.42% to 6.86%, with an average error of 5.17%.
By comparison, the MISF method yields an accuracy improvement of 2.02%. Notably, Table 3 reveals that the size of crown width does not significantly impact measurement error. However, the crown width exhibits larger errors than the other two attributes, height and DBH, due to the relatively sparse and irregular leaves at the edge, which can affect segmentation results.
The results of linear regression between measured and actual crown width are shown in Figure 14.
Figure 14a illustrates the linear correlation between reference and binocular measurements of crown width. The crown width exhibits a correlation coefficient of 0.951, with an average square root error of 0.179 m. In contrast, Figure 14b shows a significant linear correlation between the reference and MISF measurements of crown width, with no statistically significant differences observed between groups (p > 0.05) upon variance analysis. The correlation coefficient of the crown width is 0.980, with an average square root error of 0.114 m.
Furthermore, the correlation coefficient of crown width measured by the proposed MISF method in this study is 0.980, surpassing the coefficient score of 0.961 obtained using a single standing tree image taken by a mobile phone, as reported in the literature [24]. This can be attributed to the limited field of view of the single image method which may result in relatively inaccurate feature point identification, leading to measurement errors. Conversely, MISF leverages multi-vision images to overcome this issue and deliver a superior performance.

4. Conclusions

This paper proposes a novel method for measuring the size of standing trees, called MISF. The method leverages multi-vision image segmentation and coordinate fusion techniques to achieve accurate recognition and segmentation of the main body of standing trees. Specifically, in the process of image segmentation and measurement, we utilize the SEMD segmentation model based on deep learning that we previously proposed to perform multi-scale fusion of the convolutional features of the standing tree image. This approach reduces the loss of the edge details of the standing tree and enhances the effective reading of feature point information. Finally, we use the measurement information from the three cameras to obtain the final measurements of the size attributes of the standing tree.
The experimental results demonstrate that the proposed method, MISF, achieves errors of 1.89%, 2.42%, and 3.15% for tree height, DBH, and crown width measured, respectively. These values represent significant improvements over traditional binocular measurement results, with no statistically significant differences between groups as determined by variance analysis (p > 0.05). The experiments verified the three contributions of this paper. Firstly, the results demonstrate that the proposed method, MISF, outperforms binocular measurement in terms of measurement accuracy. Secondly, the improved image matching algorithm, SURF, successfully handles multi-vision images. Thirdly, real environment experiments using MISF confirm the method’s ability to successfully obtain height, DBH, and crown width measurements.
Despite the promising results of MISF, it is worth noting that some errors may still exist in the measurement data. Possible reasons for these errors include: (1) the occlusion of feature information in standing trees, leading to the imprecise acquisition of feature points; (2) there may be some errors in the camera calibration values, resulting in deviations in final calculation results. To achieve better measurement accuracy, future work will explore improvements in camera calibration methods or more convenient measurement techniques.

Author Contributions

Conceptualization, L.S. and L.M.; methodology, L.S., G.W., X.Y., X.W. and P.W.; software, L.S.; validation, L.S.; formal analysis, L.S., X.Y., X.W. and P.W.; investigation, L.S., G.W., X.Y., X.W. and P.W.; data curation, L.S.; writing—original draft. L.S.; resources, G.W. and L.M.; writing—review and editing, G.W. and L.M.; visualization, G.W., X.Y. and X.W.; supervision, G.W. and L.M.; project administration. G.W.; funding acquisition, L.M. and P.W. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the Key Research and Development Program of Zhejiang Province (Grant number: 2021C02005) and the Natural Science Foundation of China (Grant number: U1809208).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data may be available on request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mokroš, M.; Liang, X.; Surový, P.; Valent, P.; Čerňava, J.; Chudý, F.; Tunák, D.; Saloň, Š.; Merganič, J. Evaluation of close-range photogrammetry image collection methods for estimating tree diameters. ISPRS Int. J. Geo-Inf. 2018, 7, 93. [Google Scholar] [CrossRef]
  2. Yang, Z.; Liu, Q.; Luo, P.; Ye, Q.; Duan, G.; Sharma, R.P.; Zhang, H.; Wang, G.; Fu, L. Prediction of individual tree diameter and height to crown base using nonlinear simultaneous regression and airborne LiDAR data. Remote Sens. 2020, 12, 2238. [Google Scholar] [CrossRef]
  3. Cabo, C.; Ordóñez, C.; López-Sánchez, C.A.; Armesto, J. Automatic dendrometry: Tree detection, tree height and diameter estimation using terrestrial laser scanning. Int. J. Appl. Earth Obs. Geoinf. 2018, 69, 164–174. [Google Scholar] [CrossRef]
  4. Chu, T.; Starek, M.J.; Brewer, M.J.; Murray, S.C. MULTI-platform uas imaging for crop height estimation: Performance analysis over an experimental maize field. In Proceedings of the 2017 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Fort Worth, TX, USA, 23–28 July 2017; pp. 4338–4341. [Google Scholar]
  5. Dyce, D.R.; Voogt, J.A. The influence of tree crowns on urban thermal effective anisotropy. Urban Clim. 2018, 23, 91–113. [Google Scholar] [CrossRef]
  6. Indirabai, I.; Nair, M.H.; Jaishanker, R.N.; Nidamanuri, R.R. Terrestrial laser scanner based 3D reconstruction of trees and retrieval of leaf area index in a forest environment. Ecol. Inform. 2019, 53, 100986. [Google Scholar] [CrossRef]
  7. Shi, Y.; Wang, S.; Zhou, S.; Kamruzzaman, M. Study on modeling method of forest tree image recognition based on CCD and theodolite. IEEE Access 2020, 8, 159067–159076. [Google Scholar] [CrossRef]
  8. Li, S.; Fang, L.; Sun, Y.; Xia, L.; Lou, X. Development of Measuring Device for Diameter at Breast Height of Trees. Forests 2023, 14, 192. [Google Scholar] [CrossRef]
  9. Olofsson, K.; Holmgren, J.; Olsson, H. Tree stem and height measurements using terrestrial laser scanning and the RANSAC algorithm. Remote Sens. 2014, 6, 4323–4344. [Google Scholar] [CrossRef]
  10. Yuan, F.; Fang, L.; Sun, L.; Zheng, S.; Zheng, X. Development of a portable measuring device for diameter at breast height and tree height Entwicklung eines tragbaren Messgerätes für Durchmesser in Brusthöhe und Baumhöhe. Aust. J. Forensic Sci. 2021, 138, 25–50. [Google Scholar]
  11. Wang, J.; Wang, X.; Liu, F.; Gong, Y.; Wang, H.; Qin, Z. Modeling of binocular stereo vision for remote coordinate measurement and fast calibration. Opt. Lasers Eng. 2014, 54, 269–274. [Google Scholar] [CrossRef]
  12. Yang, Y.; Tang, D.; Wang, D.; Song, W.; Wang, J.; Fu, M. Multi-camera visual SLAM for off-road navigation. Robot. Auton. Syst. 2020, 128, 103505. [Google Scholar] [CrossRef]
  13. Berveglieri, A.; Tommaselli, A.; Liang, X.; Honkavaara, E. Photogrammetric measurement of tree stems from vertical fisheye images. Scand. J. For. Res. 2016, 32, 737–747. [Google Scholar] [CrossRef]
  14. Krause, S.; Sanders, T.G.M.; Mund, J.-P.; Greve, K. UAV-Based Photogrammetric Tree Height Measurement for Intensive Forest Monitoring. Remote Sens. 2019, 11, 758. [Google Scholar] [CrossRef]
  15. Ramli, M.F.; Tahar, K.N. Homogeneous tree height derivation from tree crown delineation using Seeded Region Growing (SRG) segmentation. Geo-Spat. Inf. Sci. 2020, 23, 195–208. [Google Scholar] [CrossRef]
  16. Wu, X.; Zhou, S.; Xu, A.; Chen, B. Passive measurement method of tree diameter at breast height using a smartphone. Comput. Electron. Agric. 2019, 163, 104875. [Google Scholar] [CrossRef]
  17. Yang, L.; Wang, B.; Zhang, R.; Zhou, H.; Wang, R. Analysis on location accuracy for the binocular stereo vision system. IEEE Photonics J. 2017, 10, 1–16. [Google Scholar] [CrossRef]
  18. Zhang, Z. Flexible camera calibration by viewing a plane from unknown orientations. In Proceedings of the Seventh Ieee International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999; pp. 666–673. [Google Scholar]
  19. Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef]
  20. Wu, C.; Agarwal, S.; Curless, B.; Seitz, S.M. Multicore bundle adjustment. In Proceedings of the CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; pp. 3057–3064. [Google Scholar]
  21. Shi, L.; Wang, G.; Mo, L.; Yi, X.; Wu, X.; Wu, P. Automatic Segmentation of Standing Trees from Forest Images Based on Deep Learning. Sensors 2022, 22, 6663. [Google Scholar] [CrossRef]
  22. Huang, L.; Chen, C.; Shen, H.; He, B. Adaptive registration algorithm of color images based on SURF. Measurement 2015, 66, 118–124. [Google Scholar] [CrossRef]
  23. Fan, P.; Men, A.; Chen, M.; Yang, B. Color-SURF: A surf descriptor with local kernel color histograms. In Proceedings of the 2009 IEEE International Conference on Network Infrastructure and Digital Content, Beijing, China, 6–8 November 2009; pp. 726–730. [Google Scholar]
  24. Bay, H.; Ess, A.; Tuytelaars, T.; Van Gool, L. Speeded-up robust features (SURF). Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar] [CrossRef]
  25. Zhou, Y.; Li, Q.; Chu, L.; Ma, Y.; Zhang, J. A measurement system based on internal cooperation of cameras in binocular vision. Meas. Sci. Technol. 2020, 31, 065002. [Google Scholar] [CrossRef]
  26. Ganz, S.; Käber, Y.; Adler, P. Measuring tree height with remote sensing—A comparison of photogrammetric and LiDAR data with different field measurements. Forests 2019, 10, 694. [Google Scholar] [CrossRef]
  27. Putra, B.T.W.; Ramadhani, N.J.; Soedibyo, D.W.; Marhaenanto, B.; Indarto, I.; Yualianto, Y. The use of computer vision to estimate tree diameter and circumference in homogeneous and production forests using a non-contact method. For. Sci. Technol. 2021, 17, 32–38. [Google Scholar] [CrossRef]
  28. Xinmei, W.; Aijun, X.; Tingting, Y. Passive measurement method of tree height and crown diameter using a smartphone. IEEE Access 2020, 8, 11669–11678. [Google Scholar] [CrossRef]
Figure 1. Main framework of MISF.
Figure 1. Main framework of MISF.
Forests 14 01054 g001
Figure 2. Checkerboard calibration board.
Figure 2. Checkerboard calibration board.
Forests 14 01054 g002
Figure 3. Trinocular vision measurement calibration. (a) Camera self-calibration. (b) Position relationship of three cameras.
Figure 3. Trinocular vision measurement calibration. (a) Camera self-calibration. (b) Position relationship of three cameras.
Forests 14 01054 g003
Figure 4. Shooting schema.
Figure 4. Shooting schema.
Forests 14 01054 g004
Figure 5. Feature point matching of standing trees using the improved SURF.
Figure 5. Feature point matching of standing trees using the improved SURF.
Forests 14 01054 g005
Figure 6. Schematic representation of the measurement model.
Figure 6. Schematic representation of the measurement model.
Forests 14 01054 g006
Figure 7. Tree height measurement diagram.
Figure 7. Tree height measurement diagram.
Forests 14 01054 g007
Figure 8. DBH measurement diagram.
Figure 8. DBH measurement diagram.
Forests 14 01054 g008
Figure 9. CW measurement diagram.
Figure 9. CW measurement diagram.
Forests 14 01054 g009
Figure 10. Sampling area.
Figure 10. Sampling area.
Forests 14 01054 g010
Figure 11. Standing Tree segmentation results. (a) Original image. (b) Ground-truth. (c) SEMD segmentation results.
Figure 11. Standing Tree segmentation results. (a) Original image. (b) Ground-truth. (c) SEMD segmentation results.
Forests 14 01054 g011
Figure 12. Linear regression analysis plot of measured tree height and true tree height. (a) Binocular measurement. (b) MISF.
Figure 12. Linear regression analysis plot of measured tree height and true tree height. (a) Binocular measurement. (b) MISF.
Forests 14 01054 g012
Figure 13. Linear regression analysis of measured and true DBH values. (a) Binocular measurement. (b) MISF.
Figure 13. Linear regression analysis of measured and true DBH values. (a) Binocular measurement. (b) MISF.
Forests 14 01054 g013
Figure 14. Linear regression analysis of measured crown width versus true crown width. (a) Binocular measurement. (b) MISF.
Figure 14. Linear regression analysis of measured crown width versus true crown width. (a) Binocular measurement. (b) MISF.
Forests 14 01054 g014
Table 1. Comparison of tree height measurement results.
Table 1. Comparison of tree height measurement results.
SampleTrue Value (m)Binocular MeasurementMISF
Results (m)Relative Error (%)Results (m)Relative Error (%)
18.769.13 4.278.512.83
27.838.052.837.96 1.63
37.627.42 2.697.44 2.35
46.376.552.836.49 1.96
56.356.08 4.266.49 2.15
65.665.731.285.61 0.91
75.645.86 3.985.55 1.56
85.525.73 3.895.61 1.69
95.495.26 4.135.38 1.98
105.415.603.515.54 2.46
115.335.57 4.535.45 2.22
125.125.30 3.465.02 1.89
134.984.79 3.815.04 1.25
144.694.88 4.124.60 1.86
154.624.44 3.894.68 1.23
164.564.69 2.834.46 2.11
174.484.68 4.384.59 2.56
184.394.58 4.254.47 1.86
194.254.454.674.35 2.36
203.894.03 3.693.82 1.77
213.693.57 3.223.76 1.89
223.663.772.893.63 0.76
233.363.555.533.43 2.05
243.253.394.323.16 2.91
253.153.273.863.18 0.95
Average--3.72-1.89
Table 2. Comparison of DBH measurement results.
Table 2. Comparison of DBH measurement results.
SampleTrue Value (cm)Binocular MeasurementMISF
Results (cm)Relative Error (%)Results (cm)Relative Error (%)
130.6331.65 3.3231.23 1.95
227.5626.66 3.2526.96 2.19
326.9227.91 3.6927.58 2.47
426.3925.48 3.4625.78 2.31
525.6926.61 3.5926.28 2.28
625.4626.39 3.6525.97 1.99
723.5822.67 3.8823.05 2.23
822.6921.83 3.7923.14 1.98
921.5820.69 4.1221.11 2.18
1020.3119.58 3.5920.76 2.21
1120.1621.02 4.2520.65 2.42
1219.5518.73 4.1920.01 2.36
1319.2318.39 4.3618.77 2.41
1418.6517.86 4.2219.12 2.53
1518.5719.34 4.1718.11 2.49
1617.6916.89 4.5418.13 2.51
1717.6316.84 4.4617.19 2.48
1817.5616.79 4.3918.00 2.52
1916.7517.46 4.2516.35 2.39
2016.2416.95 4.3716.71 2.88
2115.3614.68 4.4515.76 2.59
2214.5215.17 4.5114.13 2.68
2313.6713.06 4.4814.01 2.46
2413.2713.87 4.5412.87 2.98
2510.7111.19 4.4611.04 3.12
Average--4.07-2.42
Table 3. Comparison of crown width measurement results.
Table 3. Comparison of crown width measurement results.
SampleTrue Value (m)Binocular MeasurementMISF
Results (m)Relative Error (%)Results (m)Relative Error (%)
14.895.145.115.06 3.56
24.694.435.624.84 3.12
34.64.423.814.44 3.46
44.564.805.234.69 2.86
54.434.254.054.28 3.32
64.284.494.794.40 2.69
74.184.405.224.33 3.58
83.733.574.283.63 2.76
93.693.874.813.83 3.69
103.633.513.423.70 2.06
113.623.825.493.79 4.64
123.593.414.993.47 3.46
133.513.725.953.58 2.13
143.473.295.073.57 2.98
153.463.686.363.28 5.23
163.423.196.673.52 2.83
173.363.545.363.25 3.19
183.143.336.123.25 3.36
193.122.974.773.04 2.69
202.933.084.963.03 3.46
212.682.545.062.75 2.79
222.612.755.432.52 3.35
232.392.255.692.45 2.49
242.141.996.862.17 1.31
252.031.944.261.95 3.82
Average--5.17-3.15
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mo, L.; Shi, L.; Wang, G.; Yi, X.; Wu, P.; Wu, X. MISF: A Method for Measurement of Standing Tree Size via Multi-Vision Image Segmentation and Coordinate Fusion. Forests 2023, 14, 1054. https://doi.org/10.3390/f14051054

AMA Style

Mo L, Shi L, Wang G, Yi X, Wu P, Wu X. MISF: A Method for Measurement of Standing Tree Size via Multi-Vision Image Segmentation and Coordinate Fusion. Forests. 2023; 14(5):1054. https://doi.org/10.3390/f14051054

Chicago/Turabian Style

Mo, Lufeng, Lijuan Shi, Guoying Wang, Xiaomei Yi, Peng Wu, and Xiaoping Wu. 2023. "MISF: A Method for Measurement of Standing Tree Size via Multi-Vision Image Segmentation and Coordinate Fusion" Forests 14, no. 5: 1054. https://doi.org/10.3390/f14051054

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop