Next Article in Journal
Multi-Step Attack Detection Based on Pre-Trained Hidden Markov Models
Previous Article in Journal
Reliability of Repeated Isometric Neck Strength in Rugby Union Players Using a Load Cell Device
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

3D Global Localization in the Underground Mine Environment Using Mobile LiDAR Mapping and Point Cloud Registration

Mineral Resources Division, Korea Institute of Geoscience and Mineral Resources, Daejeon 34132, Korea
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(8), 2873; https://doi.org/10.3390/s22082873
Submission received: 22 March 2022 / Revised: 6 April 2022 / Accepted: 7 April 2022 / Published: 8 April 2022
(This article belongs to the Section Remote Sensors)

Abstract

:
This study proposes a 3D global localization method that implements mobile LiDAR mapping and point cloud registration to recognize the locations of objects in an underground mine. An initial global point cloud map was built for an entire underground mine area using mobile LiDAR; a local LiDAR scan (local point cloud) was generated at the point where underground positioning was required. We calculated fast point feature histogram (FPFH) descriptors for the global and local point clouds to extract point features. The match areas between the global and the local point clouds were searched and aligned using random sample consensus (RANSAC) and iterative closest point (ICP) registration. The object’s location on the global coordinate system was measured using the LiDAR sensor trajectory. Field experiments were performed at the Gwan-in underground mine using three mobile LiDAR systems. The local point cloud dataset formed for the six areas of the underground mine precisely matched the global point cloud, with a low average error of approximately 0.13 m, regardless of the type of mobile LiDAR system used. In addition, the LiDAR senor trajectory was aligned on the global coordinate system to confirm the change in the dynamic object’s position over time.

1. Introduction

Recently, digital twin technology, which has emerged as a representative industry 4.0 technology, has been introduced in the mining industry to increase mining productivity, efficiency, and safety [1,2,3]. Digital twins are digital replications of physical objects, models, and systems into the virtual world [4]. In particular, digital twins simultaneously link physical and digital models; they enable real-time monitoring, analysis, prediction, optimization, and automation of the physical model using artificial intelligence, internet of things, cloud, big data, and mobile (i.e., AICBM) technology [5]. Digital twins enable decisions for optimal mine operation by monitoring, controlling, designing, and evaluating the static and dynamic assets (e.g., ores, equipment, workers, and facilities), mining processes (e.g., explosions, hauling, and processing), and environment (e.g., geology and exploration) [6].
To increase the replication precision and prediction accuracy of digital twins, it is important to recognize and measure the 3D spatial coordinates required for accurate digital transformation of static and dynamic assets, work processes, and environmental data. When errors exist in the spatial positioning of digital data, the prediction uncertainty of the digital twins may increase. For example, inaccurate spatial coordinates of 3D geological survey data (e.g., rock type, ore grade, and borehole data) can lead to erroneous ore reserve estimation and mine layout design. In addition, incorrect position coordinates of a dynamic object may degrade the prediction accuracy of mining operations. Furthermore, a workplace accident of the automated path guidance system may occur because of incorrect location recognition and destination setting.
However, unlike open-pit mines, underground mines are global positioning system (GPS)-denied areas; it is thus difficult to recognize, measure, or track accurate location coordinates within underground mines [7]. To overcome this problem, three approaches can be used to recognize locations in underground mines. The first approach involves the use of a total station instrument. The total station begins designating coordinates from the mine portal where GPS functions. It uses laser and artificial landmarks to designate coordinates sequentially from close areas (near the portal) to deep areas (distant from the portal). However, the use of the total station in underground mines has a few limitations. First, the total station calculates the new coordinates by only measuring the distance and vertical/horizontal angles from the known location. This increases the localization uncertainty because the cumulative error increases as the drift deepens. Second, installing a large number of artificial landmarks at dense intervals can be a good option to make autonomous driving equipment continuously recognize 3D coordinates. However, it is costly and time-consuming. Third, frequent drift blasting and excavation causes artificial landmarks to be lost, relocated, or reinstalled.
The second approach involves the use of wireless communication technologies. A wireless signal is transmitted by a sensor attached to an underground mine wall. To identify the location coordinates, the signal is recognized by a receiver mounted on the equipment. Until recently, localization methods used in underground mines were developed mainly by using Wi-Fi [8,9,10,11], ZigBee [12,13,14,15,16], radio frequency identification [17,18,19,20,21,22], and Bluetooth [23,24,25]. However, to enable continuous location recognition and improve recognition accuracy, sensors must be attached to the wall in a concentrated manner, which may lead to high installation costs. In addition, the excavations with crossings inclined at 90 degrees may generate diffraction, reflection, and refraction of wireless signals.
The third approach involves an indoor localization method that employs light detection and ranging (LiDAR)-based simultaneous localization and mapping (SLAM) technology. The LiDAR estimates the distance by converting the travel time required for the laser signal in the infrared frequency band to strike the object and return [26]. LiDAR SLAM technology recognizes the robot’s location while simultaneously generating a point cloud map [27]. Recently, a few studies have been conducted with the goal of recognizing location using LiDAR SLAM in an underground mine. Jacobson [28] developed a new multisensor-based 2D SLAM technology for automatic vehicle operation in an underground mine, while Losch et al. [29] developed a new 3D LiDAR SLAM platform for the underground localization of automatic driving robots.
To improve the accuracy of LiDAR mapping and pose estimation, it is preferable to use auxiliary sensors such as a vision camera and inertial measurement unit. However, underground mines have a high magnetic field intensity and dark working spaces, which prevent the use of auxiliary sensors. Thus, LiDAR odometry and mapping (LOAM) [30] was developed to recognize a robot’s location without auxiliary sensors through edge/plane feature extraction and consecutive 3D scan matching. In addition, lightweight and ground-optimized (LEGO) LOAM [31] has been suggested to complement the scan matching accuracy and cumulative drift error of LOAM via loop detection. A few researchers have proposed an improved LOAM or LEGO-LOAM for real-time localization in underground mines with high precision and low cost. Ren et al. [32] combined LOAM and a 3D generalized-ICP (GICP) algorithm for accurate loop detection and optimization with limited computing resources for an underground mine. Xue et al. [33] suggested an improved LEGO-LOAM fused with scan context for real-time localization in an underground coal mine. Ren and Wang [34] improved the 3D GICP-SLAM to implement real-time localization by introducing an inverse distance weighting (IDW) into the error correction.
There are some limitations to the application of existing LiDAR SLAMs for measurement of the 3D spatial coordinates of the assets, objects, and equipment required for data transformation to the digital twin. The SLAM simultaneously builds a point cloud map and estimates each object’s location in a local coordinate system. The LiDAR begins mapping at the origin of the local coordinate system. Meanwhile, the digital twin platform requires data on a global coordinate system; an object’s location coordinate must be transformed into the global coordinate system using a transformation matrix. However, the transformation matrix is variable because the origin of the local coordinate system is initialized when SLAM builds a new map. Hence, the local point cloud map should be compared with a pre-built global point cloud map of the entire mine area to identify a match area between maps by extracting geometric features of points. Then, the transformation matrix can be obtained by calculating the fine alignment between the local and global point cloud maps using the point cloud registration technique. This process is known as global localization. Global localization estimates the location in the global coordinate system by matching and aligning the pre-built point cloud map and the local LiDAR scan via point feature extraction and point cloud registration [35]. Thus far, studies have been conducted to estimate the location of a self-driving vehicle through 3D map matching between a pre-defined point cloud map and a local LiDAR scan in urban road [36] and outdoor environments [37]. However, no studies have applied LiDAR-based 3D global localization to recognize an object’s location in an underground mine.
The purpose of this study was to present a 3D global localization method that used mobile LiDAR mapping and point cloud registration for the recognition of an object’s location in an underground mine. This method searches the match area between the pre-built point cloud map of the underground mine and the local point cloud through random sample consensus (RANSAC) and iterative closest point (ICP) registration. In addition, it estimates the object’s location in the global coordinate system. We present the results of field experiments using three mobile LiDAR systems in the Gwan-in underground mine in South Korea.

2. Methods

The global localization method for an underground mine follows five steps. Figure 1 is a schematic diagram that describes the global localization procedure. First, mobile LiDAR is used to generate a global cloud map of the entire underground mine and transform the coordinates to the Universal Transverse Mercator (UTM) coordinate system in meters. A local point cloud with a range of ≥10 m is generated using the mobile LiDAR in an area that requires location recognition. Second, to increase the point cloud processing efficiency, the resolution of the global point cloud and local point cloud is reduced through octree-based subsampling. Third, through feature-based coarse registration, the matching area between the global and the local point clouds is searched and the transformation matrix is calculated. In this step, the correspondences between the global and local point clouds are iteratively searched using the RANSAC technique. Fourth, the alignment between the global and local point clouds is precisely matched using the ICP algorithm. Fifth, the position coordinates of the object are estimated from the LiDAR sensor trajectory.

2.1. Global and Local Point Clouds Generation

For global localization, the formation of a global point cloud for the entire area of an underground mine and the establishment of a global coordinate system must be performed. In this study, the global point cloud is assumed to be produced in advance using mobile LiDAR. In general, point cloud generation using LiDAR follows the scan matching method. Scan matching is a technique used to calculate the relative transformation between correspondences in which geometric features match each other in a pair of overlapping point clouds [38]. As the LiDAR sensor moves, it forms point cloud sets for continuous scenes. It explores the correspondence points in a pair of point clouds and matches their alignment.
The pose and location of the LiDAR sensor can be estimated by calculating the relative transformation between correspondence points. Equation (1) is the equation used for calculating the LiDAR sensor location [30]. Here, Lτ denotes the three-dimensional position coordinates of the LiDAR sensor at time τ; ΔTτ−1,τ denotes the relative transformation according to the time change (from τ −1 to τ). The relative transformation consists of a rotation matrix and a translation matrix. Figure 2 is a schematic that explains the principle of forming a point cloud map through continuous scan matching and estimation of the LiDAR sensor trajectory.
L τ = L τ 1 Δ T τ 1 , τ
The initial global point cloud is set as the local frame, in which the coordinates of the starting point of LiDAR scanning are assigned as the origin (0,0,0). Conversion of the initial global point cloud into the UTM coordinate system requires ground control points with known real-world coordinates. The initial global point cloud map is transformed into the UTM coordinate system by calculating the transformation matrix between ground control points and reference points.
The user operates the mobile LiDAR at a point where location coordinates in the mine are required. The local point cloud is generated to have a distance of ≥10 m around the object. Because there is no ground control point information in the local point cloud, automatic conversion to the global coordinate system is impossible. Therefore, global localization can be achieved through point feature extraction and point cloud registration, which are discussed in the next sections.

2.2. Point Cloud Down-Sampling

To increase the operating speed and efficiency of the global localization, the resolution of the global and local point clouds is lowered. For down-sampling, an octree structure-based subdivision method is applied [39]. The octree method is one of the methods used for indexing 3D point cloud data. The octree node exists in the form of a rectangular cuboid and is subdivided into eight children. No further subdivision is necessary when the octree node is not occupied by points. Figure 3 describes the subdivision method based on the octree structure. The octree level represents the number of subdivisions. Octree level 1 constitutes the division of the entire point cloud data area by eight octree nodes, while octree level 2 constitutes the division of the occupied octree nodes (gray volumetric cubic area in Figure 3) by the eight children. The octree structure is suitable for effective storage and retrieval of point cloud data that are not fully volumetric.
Global and local point clouds are subdivided by setting the appropriate octree level; only points closest to the center of the octree voxels are sampled. Increasing the octree level reduces the size of the octree cell, allowing more accurate maintenance of drift geometry characteristics. However, because of the increased number of points, considerable time is needed to calculate geometric features. Therefore, the appropriate octree level must be selected after sufficient adjustments.

2.3. Match Area Search between Global and Local Point Clouds

In this step, the match area between the global and local point clouds is searched, and they are matched with each other. The wall surface of the underground mine is irregular and inconsistent because of the blasted faces, rough shapes, joint structures, cavities, and ground supports generated during mine excavation. In addition, the drift geometry is not constant locally because of the orebody shape (e.g., scale, slope, and bedding direction) and geological engineering constraints (e.g., joints, faults, and safety ratio). If the irregular geometric and structural characteristics of the underground mine drift are considered, the matching efficiency, accuracy, and quality between the global and local point clouds can be improved. Therefore, in this study, the global and local point clouds were matched with each other by calculation and comparison of the surface normal features of the point clouds.

2.3.1. Surface Feature Extraction Using the FPFH Descriptor

The fast point feature histogram (FPFH) is calculated to visualize the geometrical properties of the global and local point clouds. FPFH is a local feature descriptor proposed by Rusu et al. [40] that expresses surface normal variations between k-nearest neighbors in histograms. The three-step procedure for FPFH calculation is as follows. The following processes apply equally to the global point cloud (Qinit).
  • Surface normal estimation: A down-sampled local point cloud ( P i n i t D ) is generated through subdivision of the local point cloud (Pinit) by voxel grids with a specific voxel size. The k-nearest neighbor points enclosed in a sphere with a specific radius (r) are extracted for a point (p) in the local point cloud ( P i n i t D ). Using principal component analysis, the best-fitting surface including the point p and the k-nearest neighbors is extracted. Then, the surface normal for point p is calculated by estimation of the smallest eigenvector through singular value decomposition (SVD). Equation (2) is used to calculate the surface normal (ni) of an ith point (pi), where V0 is the eigenvector that corresponds to the smallest eigenvector (λ0). The surface normal is iteratively computed for all N points in the down-sampled local point cloud (computational complexity: O(N)).
    n i = V 0 , λ 0 λ 1 λ 2
  • Computation of angular variations in the surface normal: Angular variations in the surface normal are calculated for the k-nearest neighbors of point p. The point pi and its nearest neighbor pj are defined in the Darboux uνw frame (u = ni, v = (pipj) × u, w = u × v ) . Figure 4a depicts the points pi and pj in the Darboux frame. The angular variation (α, ϕ, θ) for all nearest neighbors is calculated as shown in Equations (3)–(5). The computational complexity is O(k).
    α = v × n j
    ϕ = u × p j p i p j p i
    θ = arctan ( w × n j , u × n j )
  • Calculation of the FPFH: The angular variation (α, ϕ, θ) is binned into a multi-value histogram known as the simplified point feature histogram (SPFH) via the d3 binning scheme. Here, d is the number of subdivision intervals in the feature value range. FPFH is calculated as the sum of the SPFH of pi and the weighted SPFH of the k-nearest neighbor, as shown in Equation (6). Here, ωi is the Euclidean distance between pj and the jth k-nearest neighbor. Figure 4b shows the effect area for the FPFH calculation of pi. The computational complexity of FPFH is O(N × k).
    FPFH ( p i ) = SPFH ( p i ) + 1 k j = 1 k 1 ϖ j SPFH ( p j )
Figure 5 shows the FPFH descriptors with 33-dimensional bins calculated for any three points (point 1: mouth, point 2: body, point 3: foot) of the dragon point cloud dataset and the average FPFH descriptor calculated for all points. In Figure 5b, the x-axis shows that all SPFH values calculated for point 1 and its k-nearest neighbors are expressed as 33 bins. The y-axis represents the ratio of the number of points included in one bin. The FPFH descriptor is calculated according to the geometric characteristics of the point.

2.3.2. Match Area Search Using RANSAC

The match area between global and local point clouds is searched using the FPFH descriptor. Then, the transformation matrix is calculated. The RANSAC technique is used to match the two point clouds. The RANSAC is a model fitting algorithm that estimates the optimal solution while removing outliers via repeated model prediction and verification from experimental data [41]. The RANSAC is used effectively for 3D point cloud registration because it can gradually update the relative transformation that minimizes the distance between the source point cloud and the target point cloud while removing any outliers of the corresponding points [42]. In this study, the global and local point clouds were matched using the modified RANSAC [43] with pruning steps. The process was as follows.
  • Corresponding point set extraction: The down-sampled local point cloud ( P i n i t D ), down-sampled global point cloud ( Q i n i t D ), and their FPFH descriptors are input into the algorithm. Three or more sample points, pi {p1, p2, p3, …, pn}, are randomly extracted from the down-sampled local point cloud. The corresponding points, qi {q1, q2, q3, …, qn}, of pi are extracted by querying the nearest neighbors in the d-dimensional FPFH descriptor.
  • Pruning steps: In these steps, the calculation complexity of RANSAC is reduced by extracting only the inlier points via two steps of pruning. First, the relative dissimilarity vector between corresponding points (pi, qi) is calculated. The relative dissimilarity vector constitutes the ratio of the difference between the two corresponding edge lengths to the larger edge length calculated from the n polygon of corresponding points (pi, qi). Figure 6 is a schematic diagram that explains the relative dissimilarity vector computation. Equation (7) is the calculation of the relative dissimilarity vector ( η ). Here, d 1 , n p denotes the edge length between p1 and pn, while d 1 , n q denotes the edge length between q1 and qn. If the relative dissimilarity vector is greater than the vector threshold (εvector), the process returns to step 1.
    η = d 12 p d 12 q max ( d 12 p , d 12 q ) , d 23 p d 23 q max ( d 23 p , d 23 q ) , d n 1 , n p d n 1 , n q max ( d n 1 , n p , d n 1 , n q ) , d 1 , n p d 1 , n q max ( d 1 , n p , d 1 , n q )
  • Transformation matrix update using RANSAC: A temporary transformation matrix between the correspondence points (pi, qi) is calculated, and the down-sampled local point cloud ( P i n i t D ) is transformed to the global point cloud ( Q i n i t D ). Euclidean distances of the nearest neighbors between P i n i t D and Q i n i t D are calculated; only k point pairs are extracted as inlier points (pk, qk) whose Euclidean distance is lower than the Euclidean distance threshold (εdistance). Equation (8) is the calculation of a required iteration number N of RANSAC; p denotes the desired success probability, w denotes an expected inlier fraction, and n constitutes the number of sample points. Model fitting and inlier point update continue until an optimal transformation matrix that minimizes the sum of the squared distances between the inlier points is obtained or the maximum iteration number Nmax is reached. The initial local point cloud moves to the match area (PinitPRANSAC) by transforming the coordinates using the rotation matrix (RRANSAC) and the movement matrix (tRANSAC) (Equation (9)).
    N = log 1 p log 1 w n
    P R A N S A C = R R A N S A C P i n i t + t R A N S A C

2.4. Alignment Matching between Global and Local Point Clouds Using ICP

In this step, the local point cloud is precisely aligned on the global point cloud through the ICP algorithm. The ICP algorithm is used to find the optimal rotation matrix and translation matrix that match the alignment of the two point cloud data.
  • Corresponding point set extraction: A random sample point set, pi {p1, p2, p3, …, pn}, is extracted in the local point cloud (PRANSAC), and the correspondence point set, qi {q1, q2, q3, …, qn}, is extracted from the global point cloud (Qinit).
  • Rotation and translation matrix estimation using SVD: Through SVD, a rotation matrix and a translation matrix are calculated for moving the sample point set p to q. First, the cross-covariance matrix (W) for n correspondence point pairs (pi, qi) is calculated by Equation (10). Here, qiT constitutes the transpose matrix of qi. Next, an m × n cross-covariance matrix W is decomposed into an m × m orthogonal matrix U, an m × n diagonal matrix Σ, and an m × n orthogonal matrix V through SVD. Equation (11) is an expression for decomposing the cross-covariance matrix W through SVD, where VT is the transpose matrix of the orthogonal matrix V. Through Equations (12) and (13), the rotation matrix RICP and the translation matrix tICP are calculated.
    W = C o v ( p , q ) = 1 n i = 1 n p i q i T
    W = U Σ V T
    R I C P = V U T
    t I C P = q R I C P p
  • Threshold judging: As shown in Equation (14), the random sample point set p is transformed to the correspondence point set q. Here, p′ constitutes a transformed random sample point set. Next, the error function (E) is calculated. The error function constitutes the sum of the Euclidean distances between the transformed point cloud data set p′ and the correspondence point set q; it is calculated using Equation (15). Here, pi′ constitutes the ith transformed random sample point.
p = R I C P p + t I C P
E = i = 1 n q i p i
Steps 1–3 are repeatedly executed until the error function is minimized. When the error function becomes less than the error threshold (εICP), the algorithm is terminated. When the algorithm is finished, the entire local point cloud (PRANSAC) is transformed to the global coordinate system (PGlobal), as shown in Equation (16).
P G l o b a l = R I C P P R A N S A C + t I C P

2.5. Estimation of the Object’s Precise Location in the Global Coordinate System

The last process involves measurement of the object’s 3D position coordinates. The LiDAR sensor trajectory obtained during generation of the local point cloud in Section 2.1 is used in this process. The LiDAR sensor trajectory (L) can be expressed in three-dimensional X, Y, Z coordinates for time τ, as shown in Equation (17). The interval and range of time τ depend on the scanning frequency and LiDAR operating time. The LiDAR sensor trajectory is transformed into the global coordinate system using the rotation matrix and translation matrix calculated in Section 2.3 and Section 2.4. Equation (18) presents the formula that converts the LiDAR sensor trajectory into the global coordinate system. Here, LGlobal denotes the final LiDAR sensor trajectory of the global coordinate system.
L τ = X τ ,   Y τ ,   Z τ  
L G l o b a l = R I C P ( R R A N S A C L + t R A N S A C ) + t I C P

3. Field Experiment

3.1. Study Area

The study area comprised the Gwan-in underground mine (38°07′00″ N, 127°13′24″ E), which is located in Gwanin-myeon, Pocheon-si, Gyeonggi-do, Korea. Figure 7 shows an aerial photograph of the study area. The underground mine produces 280,000 tons of iron ore annually via sublevel stoping. The slope of the rampway is approximately 6° to 8°; the width and height of the tunnel are approximately 4.5 m. Iron ore is produced in the 3rd, 5th, 6th, and 7th level drifts.

3.2. Global and Local Point Cloud Datasets

Global and local point clouds of the study area were generated using mobile LiDAR. The mobile LiDAR used in the experiment was GeoSLAM’s ZEB-REVO, A.M. Autonomy’s MAPTORCH, and Apple’s iPad Pro. Three LiDAR systems were used in the field experiment for evaluation of the registration accuracy between point clouds formed from different LiDAR products. The iPad Pro LiDAR was used for the experiment because of its suitability for use in underground mining sites (low cost, light weight, and ease of carrying) in comparison to the other commercialized LiDAR systems. Furthermore, the iPad Pro can be used for data collection, data transmission and reception, and communication between workers in underground mines. Figure 8 shows the appearance of each device and the scene of LiDAR scanning in the study area.
Table 1 presents the specifications of each system. The ZEB-REVO is equipped with Hokuyo’s URM-30LX-F product, which transmits laser signals up to 30 m through a horizontal range of 270° and rotates 360° about the horizontal direction [44]. The MAPTORCH is equipped with Velodyne’s PUCK-LITE 3D LiDAR; it can transmit 16 channels of laser signals through a horizontal 360° area up to a range of 100 m [45]. Both systems form continuous 3D point cloud data using the scan matching algorithm developed by the manufacturer. The iPad Pro uses the flash LiDAR scanning method, in which the photodetector measures the time-of-flight for each pixel in a 2D array field of view [46]. A vertical cavity surface emitting laser sensor with a 10-µm pixel size and 30-K resolution emits the laser; single photon avalanche diodes measure the corresponding distance [47].
A global point cloud was developed for the 6th level drift using the ZEB-REVO. Figure 9 shows the global point cloud of this drift. The 6th level drift is divided into zones A and B according to the ore body distribution; the average elevation of the bottom surface is 65 m above mean sea level. The total number of points in the global point cloud was 9.2 × 107, and the average surface density was 3516 pts/m2. The global point cloud was converted into the Korea 2000 Korea Central Belt 2010 TM coordinate system using 10 ground control points of the 6th level drift.
Three point cloud maps were formed for zone B of the 6th level drift using the ZEB-REVO, MAPTORCH, and iPad Pro. The total numbers of points were 1.9 × 107, 9.7 × 107, and 4.5 × 107, respectively; the mean surface density values were 2073, 6412, and 4662 pts/m2, respectively. MAPTORCH has a longer scanning range and higher data acquisition rate compared to the other systems; thus, it had the highest number of points in the point cloud map and the highest mean surface density.
Six local point cloud datasets with a length of 20 m were extracted from the point cloud maps in consideration of the drift shape, geometric features, and presence or absence of structures. Figure 10, Figure 11 and Figure 12 illustrate the point cloud maps formed using ZEB-REVO, MAPTORCH, and iPad Pro, along with the local point cloud datasets extracted from each map. The local point cloud datasets include sections with irregular ceiling shapes (section a), curved tunnels (sections b and c), intersections (section d), regular shapes (section e), and railing structures near the catchment area (section f). In Figure 11, there are noise data in the local point cloud datasets at the bottom because of an error in the mapping algorithm. Table 2 presents the amount of point cloud data for each section.

3.3. Results of Field Experiments

By setting the octree level to 10, the resolution of the global point cloud and local point cloud datasets was reduced. The number of the down-sampled global point cloud was approximately 7.0 × 105, and the mean surface density was 15.37 pts/m2. Figure 13a shows the amount of data of the down-sampled local point cloud datasets. Table 3 presents parameters utilized for FPFH descriptor calculation and RANSAC and ICP registration. A desktop PC with the Intel(R) Core(TM) i9-9900KF CPU (3.60 GHz) and 16 GB RAM running Windows 10 was used to run the algorithm.
Figure 13 illustrates the amount of the local point cloud datasets and the computation time required to calculate the FPFH descriptor for each section. The computation time of the FPFH descriptor increased as the amount of the point cloud increased. Figure 14a shows the average FPFH descriptor calculated for the global point cloud down-sampled to a voxel size of 0.5 m. Figure 14b–d show the average FPFH descriptor of the local point cloud datasets (sections a–f) obtained from three mobile LiDAR systems. The average FPFH descriptors for each section showed an almost identical histogram form, irrespective of the type of LiDAR. Moreover, they showed a form similar to the form of the average FPFH descriptor of the global point cloud.
Figure 15 is the result of matching the global point cloud and local point cloud datasets through the RANSAC and ICP registration algorithms. All local point cloud datasets are exactly aligned to sections a–f in zone B of the 6th level drift. Irrespective of the type of mobile LiDAR system, it was possible to accurately identify the match area between the global point cloud and the local point cloud datasets.
Figure 16 illustrates the computation time and accuracy of RANSAC and ICP registration. An average interval of approximately 3.71 s was needed for RANSAC registration in the local point cloud datasets, except for section a generated by MAPTORCH. The high computation time of approximately 10 s in section a was caused by the low precision of the local point cloud, as shown in Figure 11b; many iterations were required for algorithm convergence.
To evaluate the accuracy of RANSAC and ICP registration, the inlier root mean square error (inlier RMSE, IRMSE) was calculated, as in Equation (19). Here, C is the set of all inlier correspondences {p, q}, NC is the number of all inlier correspondences, and dp,q is the Euclidean distance between inlier correspondences.
I R M S E = 1 N C ( p , q ) C N C d p , q 2
In the local point cloud datasets of ZEB-REVO and iPad Pro, the average IRMSE values of RANSAC registration were calculated as approximately 0.19 m and 0.2 m, respectively (Figure 16b). In contrast, the average IRMSE of the local point cloud datasets of MAPTORCH was high (approximately 0.36 m). The noise points present in sections b, c, d, and f deteriorated the RANSAC matching accuracy.
ICP registration required an average computation times of approximately 0.5 s (ZEB-REVO), 1.17 s (MAPTORCH), and 0.73 s (iPad Pro), respectively (Figure 16c). Compared to RANSAC registration, ICP registration required a short computation time because the transformation matrix is calculated by consideration of only the distance error between random correspondences. In addition, the computation time of ICP registration was proportional to the amount of the local point cloud dataset. In the local point cloud datasets of ZEB-REVO, MAPTORCH, and iPad Pro, the average IRMSE values of ICP registration were calculated as approximately 0.12 m, 0.12 m, and 0.14 m, respectively (Figure 16d). The low IRMSE values confirmed that precise matching between global point cloud and local point cloud datasets was possible through ICP registration. In addition, MAPTORCH’s local point cloud datasets including noise points could be matched with the global point cloud with high accuracy without additional noise filtering.
Using the coordinate transformation matrix obtained via RANSAC and ICP registration, LiDAR sensor trajectory data for each section were transformed into the global coordinate system. Figure 17 shows the LiDAR sensor trajectories of sections c and e among the local point cloud datasets of ZEB-REVO. The global coordinate changes could be confirmed according to the scanning time of the LiDAR. Through RANSAC and ICP registration, global localization could be achieved by recognition of the user’s location coordinates within the entire underground mining area. In addition, the position change could be observed according to travel time.

4. Discussion

4.1. Global Localization Considering the Variation in Local Point Cloud Resolution

The sensitivity of RANSAC and ICP registration to the variation in local point cloud resolution was analyzed. The local point cloud datasets (sections a–f) of ZEB-REVO were down-sampled by setting the octree level to values of 6, 8, or 12. The global point cloud was down-sampled by setting the octree level to 10. The drift length of the local point cloud datasets was set to 20 m. The parameters used for the algorithms were identical to the parameters in Table 3.
Figure 18 shows the result of matching the down-sampled local point cloud datasets to the global point cloud via RANSAC and ICP registration by setting the octree level to 6, 8, and 12, respectively. Global localization was possible, regardless of the local point cloud resolution. Figure 19 expresses the computation time and IRMSE values of RANSAC and ICP registration. As the octree level increased, the point resolution of the local point cloud datasets increased; the computation time of the FPFH descriptor calculation also proportionally increased (Figure 19a,b). The computation time and IRMSE of RANSAC registration were almost constant for each section, regardless of point resolution (Figure 19c). While the computation time of ICP registration increased as the point resolution increased, the average IRMSE was constant (approximately 0.12 m) (Figure 19d). Therefore, global localization in the study area was possible with the lowest computation time and highest matching accuracy when down-sampling the local point cloud datasets of ZEB-REVO by setting the octree level to 6.

4.2. Global Localization Considering Variation in Local Point Cloud Length

The sensitivities of RANSAC and ICP registration, according to the variation in local point cloud length, were analyzed. New local point cloud datasets were formed from the point cloud obtained with ZEB-REVO by setting the drift length to 10 m and 30 m, respectively. The octree-level to sample the local point cloud datasets was set to 10. Algorithm parameters were set as shown in Table 3.
Figure 20 illustrates the results of matching the local point cloud datasets with drift lengths of 10 m and 30 m to the global point cloud. In particular, the global point cloud and the local point cloud were accurately matched even under the condition of the 10 m drift length, which generally lacked characteristics of drift shape. The average IRMSE values were approximately 0.19 m (RANSAC) and 0.12 m (ICP), regardless of drift length variation. The average total computation times required for FPFH descriptor calculation and RANSAC and ICP registration were approximately 8.5 s, 10.3 s, and 10.4 s when the drift length condition was 10 m, 20 m, and 30 m, respectively. Therefore, in the study area, global localization could be achieved with the least time and high accuracy when forming a local point cloud with a length of ≥10 m.

4.3. Comparative Analysis with Global Localization Methods

We analyzed and compared the proposed global localization method with an existing method developed by Ren and Wang [34]. Table 4 represents the differences in global localization techniques between the existing method and our method. The existing method generates an offline point cloud map, referred to as the global point cloud in our method, using the GICP-SLAM [32]. On the other hand, we used the graph SLAM [48] technique. Then, the offline map is converted into a distance-weight map (DWM) to correct the offline map error by calculating the coordinates of the point according to the distance between the point and the current LiDAR center. It finds a match area between the DWM and the current local point cloud to estimate an initial LiDAR pose using spatial descriptors including the scan context [49] features of the current local point cloud, key frame, and pose information. Finally, it predicts and corrects the current LiDAR pose on the DWM in real-time using an Unscented Kalman Filter (UKF) by matching the edge and plane features of the underground roadway.
Comparative analysis revealed two major differences in the global localization techniques of the two methods. First, the existing method has focused on real-time implementation of global localization and pose estimation while continuously matching the DWM with the current local point cloud using the UKF. To reduce the computation complexity, it considered the point cloud roadway only. In contrast, our method has only focused on the recognition and measurement of exact 3D spatial coordinates of the current location in the underground mine using point cloud registrations without further and gradual pose estimations.
Next, it has been mentioned that the initial position has a great influence on the real-time positioning [34]. This is because it follows the UKF of predicting the next pose based on the initial position. For this reason, to estimate the exact initial pose, it was preferred that the scan context, key frame, and pose information were used rather than considering the characteristics of drift shape. In contrast, our method used whole point cloud features including ceilings, roadways, and sidewalls to find the exact 3D coordinates of the current location. It would be advantageous for global localization at the underground mine with a wide area on a single level and multiple levels.
For a more quantitative comparison of the two methods, further field experiments can be performed to compare the performances, processing time, relative position error, etc. Experiments may include various technique comparisons between (1) FPFH + RANSAC and scan context, (2) FPFH and other point cloud descriptors, and (3) the use of only roadways and whole underground spaces. In addition, we are planning to advance our method by reflecting (1) further and continuous pose estimation after initial global localization, (2) real-time global localization in the underground mine sites, and (3) cumulative map error compensation using the DWM.

5. Conclusions

In this study, we proposed a 3D global localization method that could be implemented using mobile LiDAR mapping and point cloud registration to recognize the location of objects in an underground mine. A prior global point cloud map was generated for the entire underground mine, and the map frame was converted into a global UTM coordinate system. A local point cloud with a minimum length of 10 m was developed using mobile LiDAR. The global point cloud and local point cloud were down-sampled through an octree-based sub-sampling technique; FPFH descriptors were calculated for point feature extraction. The global point cloud and local point cloud were precisely matched via RANSAC and ICP registration; the transformation matrix was calculated to transform the local point cloud to the global UTM coordinate system. Finally, the LiDAR sensor trajectory was converted into the global UTM coordinate system and the object locations were measured. This application of the proposed method to the Gwan-in underground mine confirmed that global matching of local point cloud datasets is possible, regardless of the type of mobile LiDAR system and geometry of the underground drift. Through the RANSAC and ICP registration, the result was precisely aligned with an average inlier RMSE error of approximately 0.13 m between global point cloud and local point cloud datasets. In addition, by converting the LiDAR sensor trajectory data into a global UTM frame, the spatial location coordinates of an object could be recognized according to travel time.
The proposed method can provide innovative solutions for 3D position recognition in the Gwan-in underground mine. In the study area, managers have adopted only traditional localization methods by using the total station instrument and installing artificial landmarks. Only about ten or more artificial landmarks are installed on the ceiling on a single level. Because they are memorizing all locations of the mine, they are not obligated to adopt advanced localization methods such as installing wireless signal sensors and implementing the LiDAR SLAM. However, they are increasingly seeing the need to configure digital twins and install mine automation systems to improve mine productivity and sustainability. Using cost-effective and novel tools including 3D mobile LiDAR mapping systems and point cloud registration methods, they can immediately and accurately track the 3D spatial coordinates of assets, processes, and environments in the study area.
An analysis of the accuracy and computation time of the 3D global localization method through field experiments confirmed that this method can be utilized effectively in an underground mine site. The proposed 3D global localization method has the following contributions for the mining industry to become automated, intelligent, and digitalized.
  • It enables improvement of replication precision and analysis/prediction/automation accuracy of the digital twin platform. In particular, accurate analysis and prediction for overall mining operations may be possible through accurate spatial conversion of geological survey (e.g., ore mapping and borehole), production (e.g., blasting patterns and ore volume), and geotechnical (e.g., rock fall and road conditions) data into the digital twin platform.
  • In addition, the operating accuracy of autonomous driving equipment can be increased via sophisticated 3D position recognition in underground mines composed of multiple drift levels.
  • To generate the local point cloud map, low-cost, open-source 3D LiDAR can be used. Recently, various low-cost 3D LiDAR products have been commercialized for self-driving robots. They can build the local point cloud with a minimum length of 10 m and appropriate point resolution. They are also sufficiently small and rigid to attach to a worker’s helmet, equipment bodies, or facilities.
  • Furthermore, there is no requirement for auxiliary sensors (i.e., inertial measurement units) for continuous mapping and pose estimation in an underground mine with high magnetic field intensity.
In the future, we plan to advance the proposed system so that it can be operated in the underground mine in real-time. At the current level, it is insufficient to implement the system in real-time due to the computation complexity of the algorithms, the vastness of point cloud data, and the use of individual 3D mobile LiDAR and processors. The hardware and software would be configured to simultaneously enable LiDAR mapping, point cloud registration, and visualization on one processor. In addition, it would be possible to estimate continuous pose estimation through continuous scan-to-global point cloud map matching after initial global localization. To evaluate the applicability of the 3D global localization method, additional field experiments will be performed on underground mines that have various drift geometries and are of different types (e.g., coal mines, limestone mines, etc.).

Author Contributions

Conceptualization, S.C.; methodology, J.B. and C.L.; software, C.L.; formal analysis, J.B. and J.P.; investigation, J.B. and J.P.; resources, J.B., J.P. and C.L.; data curation, J.B. and C.L.; writing—original draft preparation, J.B.; writing—review and editing, J.B., J.P. and C.L.; supervision, S.C.; project administration, S.C.; funding acquisition, S.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Energy Efficiency & Resources of the Korea Institute of Energy Technology Evaluation and Planning (KETEP) grant funded by the Korea government Ministry of Trade, Industry, and Energy (No. 20216110100040). This research was also supported by the basic research project of the Korea Institute of Geoscience and Mineral Resources, Daejeon, South Korea, funded by the Ministry of Science and ICT of Korea (GP2020-007).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

In this paper, publicly archived datasets were used to show how FPFH descriptor works for point feature extraction and analysis. The dataset is available at the following link: http://graphics.stanford.edu/data/3Dscanrep/ (accessed on 6 April 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mining Magazine, Technology of Mining in the Future. Available online: https://www.miningmagazine.com/innovation/news/1393966/technology-of-mining-in-the-future (accessed on 21 March 2022).
  2. Mining Magazine, Digital Twins to Emerge as ESG Tool. Available online: https://www.miningmagazine.com/plant/news/1421789/digital-twins-to-emerge-as-esg-tool (accessed on 21 March 2022).
  3. Barnewold, L.; Lottermoser, B.G. Identification of digital technologies and digitalisation trends in the mining industry. Int. J. Min. Sci. Technol. 2020, 30, 747–757. [Google Scholar] [CrossRef]
  4. Forbes, What the Rise of Digital Twin Technology Can Mean for Your Business. Available online: https://www.forbes.com/sites/forbestechcouncil/2021/07/15/what-the-rise-of-digital-twin-technology-can-mean-for-your-business/?sh=61a809df251e (accessed on 21 March 2022).
  5. World Economic Forum, Future Technologies That Will Drive Industry 4.0. Available online: https://www.weforum.org/agenda/2019/01/future-technologies-will-drive-industry-4-0/ (accessed on 21 March 2022).
  6. Mining Magazine, Slow Build Can Enhance Mining’s Digital Future. Available online: https://www.miningmagazine.com/innovation/opinion/1396228/slow-build-can-enhance-minings-digital-future (accessed on 21 March 2022).
  7. Kim, H.; Choi, Y. Location estimation of autonomous driving robot and 3D tunnel mapping in underground mines using pattern matched LiDAR sequential images. Int. J. Min. Sci. Technol. 2021, 31, 779–788. [Google Scholar] [CrossRef]
  8. Zhang, Y.; Li, L.; Zhang, Y. Research and Design of Location Tracking System Used in Underground Mine Based on WiFi Technology. In Proceedings of the 2009 International Forum on Computer Science-Technology and Applications, Chongqing, China, 25–27 December 2009. [Google Scholar]
  9. Tao, P.; Xiaoyang, L. Hybrid Wireless Communication System Using ZigBee and WiFi Technology in the Coalmine Tunnels. In Proceedings of the 2011 Third International Conference on Measuring Technology and Mechatronics Automation, Shanghai, China, 6–7 January 2011. [Google Scholar]
  10. Sun, J.; Li, C. Tunnel personnel positioning method based on TOA and modified location-fingerprint positioning. Int. J. Min. Sci. Technol. 2016, 26, 429–436. [Google Scholar] [CrossRef] [Green Version]
  11. Mohapatra, A.G.; Keswani, B.; Nanda, S.; Ray, A.; Khanna, A.; Gupta, D.; Keswani, P. Precision local positioning mechanism in underground mining using IoT-enabled WiFi platform. Int. J. Comput. Appl. 2020, 42, 266–277. [Google Scholar] [CrossRef]
  12. Huang, X.; Zhu, W.; Lu, D. Underground Miners Localization System Based on ZigBee and WebGIS. In Proceedings of the 2010 18th International Conference on Geoinformatics, Beijing, China, 18–20 June 2010. [Google Scholar]
  13. Bedford, M.D.; Kennedy, G.A. Evaluation of ZigBee (IEEE 802.15.4) Time-of-Flight-Based Distance Measurement for Application in Emergency Underground Navigation. IEEE Trans. Antennas Propag. 2012, 60, 2502–2510. [Google Scholar] [CrossRef]
  14. Maity, T.; Das, P.S.; Mukherjee, M. A wireless surveillance and safety system for mine workers based on Zigbee. In Proceedings of the 2012 1st International Conference on Recent Advances in Information Technology (RAIT), Dhanbad, India, 15–17 March 2012. [Google Scholar]
  15. Xu, H.; Li, F.; Ma, Y. A ZigBee-based miner localization system. In Proceedings of the 2012 IEEE 16th International Conference on Computer Supported Cooperative Work in Design (CSCWD), Wuhan, China, 23–25 May 2012. [Google Scholar]
  16. Moridi, M.A.; Kawamura, Y.; Sharifzadeh, M.; Chanda, E.K.; Wagner, M.; Jang, H.; Okawa, H. Development of underground mine monitoring and communication system integrated ZigBee and GIS. Int. J. Min. Sci. Technol. 2015, 25, 811–818. [Google Scholar] [CrossRef] [Green Version]
  17. Zhang, Q.; Wang, B.; Cheng, G.; Wang, Z.; Wei, J.; Yan, D. Object position tracking based on e-map and RFID in coal mine. In Proceedings of the 2009 4th IEEE Conference on Industrial Electronics and Applications, Xi’an, China, 25–27 May 2009. [Google Scholar]
  18. Ji, Y.; Xu, Z.; Feng, Q.; Sang, Y. Concurrent collision probability of RFID tags in underground mine personnel position systems. Min. Sci. Technol. 2010, 20, 734–737. [Google Scholar] [CrossRef]
  19. Rusu, S.R.; Hayes, M.J.D.; Marshall, J.A. Localization in large-scale underground environments with RFID. In Proceedings of the 2011 24th Canadian Conference on Electrical and Computer Engineering (CCECE), Niagara Falls, ON, Canada, 8–11 May 2011. [Google Scholar]
  20. Mishra, P.K.; Stewart, R.F.; Bolic, M.; Yagoub, M.C.E. RFID in Underground-Mining Service Applications. IEEE Pervasive Comput. 2014, 13, 72–79. [Google Scholar] [CrossRef]
  21. Jung, J.; Choi, Y. Collecting Travel Time Data of Mine Equipments in an Underground Mine using Reverse RFID Systems. Tunn. Undergr. Space 2016, 26, 253–265. [Google Scholar] [CrossRef] [Green Version]
  22. Wang, J.; Guo, Y.; Guo, L.; Zhang, B.; Wu, B. Performance Test of MPMD Matching Algorithm for Geomagnetic and RFID Combined Underground Positioning. IEEE Access 2019, 7, 129789–129801. [Google Scholar] [CrossRef]
  23. Jung, J.; Choi, Y. Measuring Transport Time of Mine Equipment in and Underground Mine Using a Bluetooth Beacon System. Minerals 2017, 7, 1. [Google Scholar] [CrossRef] [Green Version]
  24. Baek, J.; Choi, Y.; Lee, C.; Suh, J.; Lee, S. BBUNS: Bluetooth Beacon-Based Underground Navigation System to Support Mine Haulage Operations. Minerals 2017, 7, 228. [Google Scholar] [CrossRef] [Green Version]
  25. Park, S.; Choi, Y. Bluetooth beacon-based mine production management application to support ore haulage operations in underground mines. Sustainability 2021, 13, 2281. [Google Scholar] [CrossRef]
  26. Donges, A.; Noll, R. Laser Measurement Technology; Springer: Berlin/Heidelberg, Germany, 2015; pp. 1–684. [Google Scholar]
  27. Thrun, S.; Burgard, W.; Fox, D. Probabilistic Robotics; The MIT Press: Cambridge, MA, USA, 2005; pp. 1–672. [Google Scholar]
  28. Jacobson, A.; Zeng, F.; Smith, D.; Boswell, N.; Peynot, T.; Milford, M. What localizes beneath: A metric multisensor localization and mapping system for autonomous underground mining vehicles. J. Field Robot. 2021, 38, 5–27. [Google Scholar] [CrossRef]
  29. Losch, R.; Grehl, S.; Donner, M.; Buhl, C.; Jung, B. Design of an Autonomous Robot for Mapping, Navigation, and Manipulation in Underground Mines. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018. [Google Scholar]
  30. Zhang, J.; Singh, S. LOAM: Lidar Odometry and Mapping in Real-time. In Proceedings of the Robotics: Science and Systems Conference (RSS), Berkeley, CA, USA, 12–16 July 2014. [Google Scholar]
  31. Shan, T.; Englot, B. LeGO-LOAM: Lightweight and Ground-Optimized Lidar Odometry and Mapping on Variable Terrain. In Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Madrid, Spain, 1–5 October 2018. [Google Scholar]
  32. Ren, Z.; Wang, L.; Bi, L. Robust GICP-based 3D LiDAR SLAM for underground mining environment. Sensors 2019, 19, 2915. [Google Scholar] [CrossRef] [Green Version]
  33. Xue, G.; Wei, J.; Li, R.; Cheng, J. LeGO-LOAM-SC: An Improved Simultaneous Localization and Mapping Method Fusing LeGO-LOAM and Scan Context for Underground Coalmine. Sensors 2022, 22, 520. [Google Scholar] [CrossRef]
  34. Ren, Z.; Wang, L. Accurate Real-Time Localization Estimation in Underground Mine Environments Based on a Distance-Weight Map (DWM). Sensors 2022, 22, 1463. [Google Scholar] [CrossRef]
  35. Ghallabi, F. Precise Self-Localization of Autonomous Vehicles Using Lidar Sensors and Highly Accurate Digital Maps on Highway Roads. Ph.D. Thesis, Université Paris Sciences et Lettres, Paris, France, 10 December 2020. [Google Scholar]
  36. Nagy, B.; Benedek, C. Real-Time Point Cloud Alignment for Vehicle Localization in a High Resolution 3D Map. In Proceedings of the European Conference on Computer Vision—ECCV 2018 Workshops, Munich, Germany, 8–14 September 2018. [Google Scholar]
  37. Koide, K.; Miura, J.; Menegatti, E. A portable three-dimensional LIDAR-based system for long-term and wide-area people behavior measurement. Int. J. Adv. Robot. Syst. 2019, 16, 1729881419841532. [Google Scholar] [CrossRef]
  38. Houshiar, H. Documentation and Mapping with 3D Point Cloud Processing. Ph.D. Thesis, Universität Würzburg, Würzburg, Germany, 20 September 2017. [Google Scholar]
  39. Elseberg, J.; Borrmann, D.; Nüchter, A. One billion points in the cloud—An octree for efficient processing of 3D laser scans. ISPRS J. Photogramm. Remote Sens. 2013, 76, 76–88. [Google Scholar] [CrossRef]
  40. Rusu, R.B.; Blodow, N.; Beetz, M. Fast Point Feature Histograms (FPFH) for 3D Registration. In Proceedings of the 2009 IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009. [Google Scholar]
  41. Fischler, M.A.; Bolles, R.C. Random sample consensus: A Paradigm for Model Fitting with Applications to Image Analysis and Automated Cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  42. Xu, G.; Pang, Y.; Bai, Z.; Wang, Y.; Lu, Z. A fast point clouds registration algorithm for laser scanners. Appl. Sci. 2021, 11, 3426. [Google Scholar] [CrossRef]
  43. Buch, A.G.; Kraft, D.; Kamarainen, J.K.; Petersen, H.G.; Kruger, N. Pose estimation using local structure-specific shape and appearance context. In Proceedings of the 2013 IEEE International Conference on Robotics and Automation (ICRA), Karlsruhe, Germany, 6–10 May 2013. [Google Scholar]
  44. GeoSLAM. Specifications. Available online: https://geoslam.com/solutions/zeb-revo-rt/ (accessed on 21 March 2022).
  45. Autonomy, A.M. Specifications. Available online: https://www.amatonomy.com/map-torch/ (accessed on 21 March 2022).
  46. Wang, D.; Watkins, C.; Xie, H. MEMS mirrors for LiDAR: A review. Micromachines 2020, 11, 456. [Google Scholar] [CrossRef] [PubMed]
  47. Forbes. VCSEL Technology for Next Generation 3D Sensing Applications—Part 1. Available online: https://www.forbes.com/sites/sabbirrangwala/2021/10/28/vcsel-technology-for-next-generation-3d-sensing-applicationspart-1/?sh=10c999e118e5 (accessed on 21 March 2022).
  48. Zlot, R.; Bosse, M. Efficient large-scale three-dimensional mobile mapping for underground mines. J. Field Robot. 2014, 31, 758–779. [Google Scholar] [CrossRef]
  49. Kim, G.; Kim, A. Scan context: Egocentric spatial descriptor for place recognition within 3d point cloud map. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 4802–4809. [Google Scholar]
Figure 1. Schematic diagram describing the procedure of 3D global localization.
Figure 1. Schematic diagram describing the procedure of 3D global localization.
Sensors 22 02873 g001
Figure 2. Description of point cloud generation and LiDAR sensor trajectory via scan matching.
Figure 2. Description of point cloud generation and LiDAR sensor trajectory via scan matching.
Sensors 22 02873 g002
Figure 3. Description of octree-based subdivision.
Figure 3. Description of octree-based subdivision.
Sensors 22 02873 g003
Figure 4. Schematic diagram for FPFH descriptor computation and RANSAC registration: (a) a graphical formulation of the Darboux uvw frame and (α, ϕ, θ) angular variation; (b) the effect area (red solid line) for computation of the FPFH descriptor of pi with five nearest neighbors. Colored dotted lines present the effect area of each nearest neighbors.
Figure 4. Schematic diagram for FPFH descriptor computation and RANSAC registration: (a) a graphical formulation of the Darboux uvw frame and (α, ϕ, θ) angular variation; (b) the effect area (red solid line) for computation of the FPFH descriptor of pi with five nearest neighbors. Colored dotted lines present the effect area of each nearest neighbors.
Sensors 22 02873 g004
Figure 5. Example of FPFH descriptor: (a) dragon point cloud dataset (provided by Stanford University Computer Graphics Laboratory); FPFH descriptor of (b) point 1 (mouth), (c) point 2 (body), and (d) point 3 (foot); and (e) average FPFH descriptor of whole dragon point cloud data.
Figure 5. Example of FPFH descriptor: (a) dragon point cloud dataset (provided by Stanford University Computer Graphics Laboratory); FPFH descriptor of (b) point 1 (mouth), (c) point 2 (body), and (d) point 3 (foot); and (e) average FPFH descriptor of whole dragon point cloud data.
Sensors 22 02873 g005
Figure 6. Graphical diagram of relative dissimilarity vector computation when including three corresponding points {(p1, q1), (p2, q2), (p3, q3)}.
Figure 6. Graphical diagram of relative dissimilarity vector computation when including three corresponding points {(p1, q1), (p2, q2), (p3, q3)}.
Sensors 22 02873 g006
Figure 7. Graphical diagram Aerial view of the study area (Image source: National Geographic Information Institute).
Figure 7. Graphical diagram Aerial view of the study area (Image source: National Geographic Information Institute).
Sensors 22 02873 g007
Figure 8. The mobile LiDAR systems used for field experiments in the study area: (a) ZEB-REVO (GeoSLAM; London, UK) (b) MAPTORCH (A.M. Autonomy; Seoul, Korea); (c) iPad Pro (Apple; Santa Clara, CA, USA; image source: https://www.apple.com (accessed on 6 April 2022); (d) scene of the mobile LiDAR scanning in the study area.
Figure 8. The mobile LiDAR systems used for field experiments in the study area: (a) ZEB-REVO (GeoSLAM; London, UK) (b) MAPTORCH (A.M. Autonomy; Seoul, Korea); (c) iPad Pro (Apple; Santa Clara, CA, USA; image source: https://www.apple.com (accessed on 6 April 2022); (d) scene of the mobile LiDAR scanning in the study area.
Sensors 22 02873 g008
Figure 9. Global point cloud map of 6th level drift generated using ZEB-REVO: (a) Plan view of 6th level drift. Reference grid is in meters. The 6th level drift is divided into two zones (A and B) due to the ore body distribution. Local point cloud datasets were generated for the zone B. The origin of the local transverse Mercator coordinate system is 38°00′00″ N, 127°00′00″ E (map datum: GRS 80); (b) 3D top view of 6th level drift.
Figure 9. Global point cloud map of 6th level drift generated using ZEB-REVO: (a) Plan view of 6th level drift. Reference grid is in meters. The 6th level drift is divided into two zones (A and B) due to the ore body distribution. Local point cloud datasets were generated for the zone B. The origin of the local transverse Mercator coordinate system is 38°00′00″ N, 127°00′00″ E (map datum: GRS 80); (b) 3D top view of 6th level drift.
Sensors 22 02873 g009
Figure 10. Local point cloud datasets generated using ZEB-REVO: (a) point cloud map of zone B of 6th level drift; (b) section a; (c) section b; (d) section c; (e) section d; (f) section e; (g) section f.
Figure 10. Local point cloud datasets generated using ZEB-REVO: (a) point cloud map of zone B of 6th level drift; (b) section a; (c) section b; (d) section c; (e) section d; (f) section e; (g) section f.
Sensors 22 02873 g010
Figure 11. Local point cloud datasets generated using MAPTORCH: (a) point cloud map of zone B of 6th level drift; (b) section a; (c) section b; (d) section c; (e) section d; (f) section e; (g) section f.
Figure 11. Local point cloud datasets generated using MAPTORCH: (a) point cloud map of zone B of 6th level drift; (b) section a; (c) section b; (d) section c; (e) section d; (f) section e; (g) section f.
Sensors 22 02873 g011
Figure 12. Local point cloud datasets generated using iPad Pro: (a) point cloud map of zone B of 6th level drift; (b) section a; (c) section b; (d) section c; (e) section d; (f) section e; (g) section f.
Figure 12. Local point cloud datasets generated using iPad Pro: (a) point cloud map of zone B of 6th level drift; (b) section a; (c) section b; (d) section c; (e) section d; (f) section e; (g) section f.
Sensors 22 02873 g012
Figure 13. Results of octree-based down-sampling and FPFH descriptor calculation: (a) amount of local point cloud after down-sampling; (b) computation time of FPFH descriptor calculation (s).
Figure 13. Results of octree-based down-sampling and FPFH descriptor calculation: (a) amount of local point cloud after down-sampling; (b) computation time of FPFH descriptor calculation (s).
Sensors 22 02873 g013
Figure 14. FPFH descriptor of the global point cloud and local point cloud datasets: (a) global point cloud; (b) local point cloud dataset generated by ZEB-REVO; (c) local dataset generated by MAPTORCH; and (d) local dataset generated by iPad Pro.
Figure 14. FPFH descriptor of the global point cloud and local point cloud datasets: (a) global point cloud; (b) local point cloud dataset generated by ZEB-REVO; (c) local dataset generated by MAPTORCH; and (d) local dataset generated by iPad Pro.
Sensors 22 02873 g014
Figure 15. Experimental results of RANSAC and ICP registration between the global point cloud and local point cloud datasets of section a–f: (a) ZEB-REVO; (b) MAPTORCH; (c) iPad Pro.
Figure 15. Experimental results of RANSAC and ICP registration between the global point cloud and local point cloud datasets of section a–f: (a) ZEB-REVO; (b) MAPTORCH; (c) iPad Pro.
Sensors 22 02873 g015
Figure 16. Experimental results of RANSAC and ICP registration: (a) computation time of RANSAC registration; (b) inlier RMSE of RANSAC registration; (c) computation time of ICP registration; (d) inlier RMSE of ICP registration.
Figure 16. Experimental results of RANSAC and ICP registration: (a) computation time of RANSAC registration; (b) inlier RMSE of RANSAC registration; (c) computation time of ICP registration; (d) inlier RMSE of ICP registration.
Sensors 22 02873 g016
Figure 17. Experimental results of estimating user’s location coordinates on the global coordinate system according to the travel time: (a) top view of section c; (b) 3D view of section e.
Figure 17. Experimental results of estimating user’s location coordinates on the global coordinate system according to the travel time: (a) top view of section c; (b) 3D view of section e.
Sensors 22 02873 g017
Figure 18. Experimental results of RANSAC and ICP registration considering octree level variation of the local point cloud datasets of section a–f: (a) octree level 6; (b) octree level 8; (c) octree level 12.
Figure 18. Experimental results of RANSAC and ICP registration considering octree level variation of the local point cloud datasets of section a–f: (a) octree level 6; (b) octree level 8; (c) octree level 12.
Sensors 22 02873 g018
Figure 19. Experimental results of FPFH descriptor calculation, RANSAC registration, and ICP registration considering variation in the local point cloud resolution: (a) amount of the local point cloud; (b) computation time of FPFH descriptor calculation; (c) computation time and inlier RMSE of RANSAC registration; (d) computation time and inlier RMSE of ICP registration.
Figure 19. Experimental results of FPFH descriptor calculation, RANSAC registration, and ICP registration considering variation in the local point cloud resolution: (a) amount of the local point cloud; (b) computation time of FPFH descriptor calculation; (c) computation time and inlier RMSE of RANSAC registration; (d) computation time and inlier RMSE of ICP registration.
Sensors 22 02873 g019
Figure 20. Experimental results of RANSAC and ICP registration considering variation in drift length of the local point cloud datasets: (a) 10 m and (b) 30 m lengths of local point cloud datasets.
Figure 20. Experimental results of RANSAC and ICP registration considering variation in drift length of the local point cloud datasets: (a) 10 m and (b) 30 m lengths of local point cloud datasets.
Sensors 22 02873 g020
Table 1. Specifications of mobile LiDAR systems used for field experiments.
Table 1. Specifications of mobile LiDAR systems used for field experiments.
SpecificationsZEB-REVOMAPTORCHiPad Pro 11
ManufacturerGeoSLAMA.M. AutonomyApple
Country of manufacturerLondon, UKSeoul, KoreaSanta Clara, CA, USA
LiDAR productHokuyo UTM-30LX-FVelodyne PUCK-LITEdToF sensors
Scanning methodMotorized optomechanical Motorized optomechanicalFlash LiDAR
Maximum scanning range30 m100 m5 m
Data acquisition rate43,200 points/s0.3 million points/s10 µm pitch pixel
Resolution0.625° horizon, 1.8° vertical0.1° horizon, 2° vertical0.03 MP resolution
Field of view (FOV)270° horizon × 360° vertical360° horizon × 30° vertical122° horizontal
Rotation rate0.5 Hz5–20 Hz-
Laser wavelength905 nm905 nm-
Additional sensorsMicrostrain 3DM-GX2Microstrain 3DM-GX5accelerometer, gyro sensor
Accuracy±3 cm±3 cm-
Table 2. Amount of point cloud data according to the local point cloud sections generated by ZEB-REVO, MAPTORCH, and iPad Pro.
Table 2. Amount of point cloud data according to the local point cloud sections generated by ZEB-REVO, MAPTORCH, and iPad Pro.
Type of SystemThe Amount of Point Cloud Data (unit: 1.0 × 106)
Section aSection bSection cSection dSection eSection f
ZEB-REVO1.051.050.540.890.630.40
MAPTORCH0.382.813.952.833.632.21
iPad pro6.021.060.920.911.091.16
Table 3. Parameters set for implementing FPFH descriptor calculation and RANSAC and ICP registration.
Table 3. Parameters set for implementing FPFH descriptor calculation and RANSAC and ICP registration.
ProcessParametersValue
FPFH descriptorDown-samplingVoxel size0.5 m
Surface normal estimationMaximum number of k-nearest neighbors1000
Sphere radius (r)1 m
FPFH feature calculationMaximum number of k-nearest neighbors1000
Sphere radius (r)2.5 m
RANSACPruningRelative dissimilarity vector threshold (εvector)0.9
Distance threshold (εdistance)2.5 m
The number of corresponding point sets to fit RANSAC4
Convergence criteriaMaximum number of iterations4.0 × 108
Maximum number of validations500
ICPPruningMaximum correspondence points-pair distance0.5 m
Convergence criteriaMaximum number of iterations30
Relative RMSE1.0 × 10−6
Table 4. Comparative analysis of global localization techniques used in our method and the existing method proposed by Ren and Wang [34].
Table 4. Comparative analysis of global localization techniques used in our method and the existing method proposed by Ren and Wang [34].
TechniquesRen and Wang [34]Ours
MappingGlobal point cloud generationGICP-SLAM [32], DWMGraph SLAM [48]
Initial pose estimationDescriptorScan Context [49],
Key frame and pose information
FPFH
EstimatorPoint cloud matching methodRANSAC
LocalizationDescriptorEdge and plane features of roadway-
EstimatorUnscented Kalman FilterICP
Real-time or notOX
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Baek, J.; Park, J.; Cho, S.; Lee, C. 3D Global Localization in the Underground Mine Environment Using Mobile LiDAR Mapping and Point Cloud Registration. Sensors 2022, 22, 2873. https://doi.org/10.3390/s22082873

AMA Style

Baek J, Park J, Cho S, Lee C. 3D Global Localization in the Underground Mine Environment Using Mobile LiDAR Mapping and Point Cloud Registration. Sensors. 2022; 22(8):2873. https://doi.org/10.3390/s22082873

Chicago/Turabian Style

Baek, Jieun, Junhyeok Park, Seongjun Cho, and Changwon Lee. 2022. "3D Global Localization in the Underground Mine Environment Using Mobile LiDAR Mapping and Point Cloud Registration" Sensors 22, no. 8: 2873. https://doi.org/10.3390/s22082873

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop