Next Article in Journal
Deep Learning Framework for Liver Segmentation from T1-Weighted MRI Images
Next Article in Special Issue
Exploring Adversarial Robustness of LiDAR Semantic Segmentation in Autonomous Driving
Previous Article in Journal
A Novel Densely Packed 4 × 4 MIMO Antenna Design for UWB Wireless Applications
Previous Article in Special Issue
AFTR: A Robustness Multi-Sensor Fusion Model for 3D Object Detection Based on Adaptive Fusion Transformer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Level Optimization for Data-Driven Camera–LiDAR Calibration in Data Collection Vehicles

School of Resource and Environmental Sciences, Wuhan University, Wuhan 430079, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(21), 8889; https://doi.org/10.3390/s23218889
Submission received: 13 September 2023 / Revised: 20 October 2023 / Accepted: 27 October 2023 / Published: 1 November 2023
(This article belongs to the Special Issue Multi-modal Sensor Fusion and 3D LiDARs for Vehicle Applications)

Abstract

:
Accurately calibrating camera–LiDAR systems is crucial for achieving effective data fusion, particularly in data collection vehicles. Data-driven calibration methods have gained prominence over target-based methods due to their superior adaptability to diverse environments. However, current data-driven calibration methods are susceptible to suboptimal initialization parameters, which can significantly impact the accuracy and efficiency of the calibration process. In response to these challenges, this paper proposes a novel general model for the camera–LiDAR calibration that abstracts away the technical details in existing methods, introduces an improved objective function that effectively mitigates the issue of suboptimal parameter initialization, and develops a multi-level parameter optimization algorithm that strikes a balance between accuracy and efficiency during iterative optimization. The experimental results demonstrate that the proposed method effectively mitigates the effects of suboptimal initial calibration parameters, achieving highly accurate and efficient calibration results. The suggested technique exhibits versatility and adaptability to accommodate various sensor configurations, making it a notable advancement in the field of camera–LiDAR calibration, with potential applications in diverse fields including autonomous driving, robotics, and computer vision.

1. Introduction

The fusion of LiDAR (Light Detection and Ranging) data and camera image data is an essential step in many fields such as autonomous driving, 3D reconstruction, urban planning, and environmental monitoring [1,2,3]. The significance of data fusion arises from the dissimilarities between LiDAR and camera image data, which vary in terms of spatial resolution, velocity and distance estimation capacities, resistance to adverse weather conditions, and sensor sizes, among other factors [4]. LiDAR sensors are capable of accurately capturing 3D spatial information while lacking color and texture data that images can provide [5,6]. By fusing these two types of data, it is possible to create more accurate and detailed maps of the environment, which is crucial for applications like autonomous driving [7]. What is more, combining LiDAR and image data can improve the detection and recognition of objects in a scene. LiDAR can detect the presence and position of objects while images can provide details about their appearance and texture. The fusion of these data sources can lead to better object detection and tracking, which is vital for applications such as robotics and autonomous vehicles. Lastly, LiDAR is capable of measuring distances with high accuracy but it cannot provide depth perception of objects that are hidden from view. Combining LiDAR and image data can help overcome this limitation by providing a more comprehensive understanding of the scene’s depth and structure. Overall, the fusion of LiDAR and image data provides a more comprehensive understanding of the environment, which is essential for a wide range of applications. It enables improved accuracy [8], object detection [9,10], depth perception [11,12,13], and 3D modeling [14,15], making it an essential technique for many fields.
When it comes to the fusion between LiDAR and image data, the precise calibration between the two sensors plays a key role [16,17,18]. LiDAR sensors capture 3D point clouds of the environment, and camera sensors capture 2D images. To create a comprehensive 3D model, it is necessary to transform the camera image data into 3D coordinates that match the LiDAR point cloud data. Calibration ensures that the transformation is accurate, which is essential for generating accurate point clouds. Accurate calibration of LiDAR and camera sensors also enables better object detection and tracking [9,10]. By precisely aligning the two sensor data streams, it is possible to accurately locate objects in 3D space. What is more, calibration reduces measurement errors and noise, which can improve the overall accuracy of the data fusion process. This is particularly important for LiDAR data, which can be affected by noise caused by reflections and other factors.
Precise calibration is essential for achieving accurate data fusion between LiDAR and camera sensors. However, calibration is not a one-time process as the sensors may shift and change over time. This is particularly prevalent in data collection vehicles where the sensors are affixed to a mobile vehicle, resulting in a greater likelihood of modification to their relative pose. It will be much more time-consuming and resource-intensive if the data collection vehicle has to be sent back to the calibration field to do a thorough calibration. Accordingly, the implementation of data-driven precise calibration serves as a valuable means to account for such fluctuations [19], where the calibration parameters are updated using the data acquired by LiDAR and camera sensors. By updating the calibration parameters as necessary, data-driven calibration can adjust for changes in the environment and improve the accuracy of the data fusion process without the need for extensive calibration in a dedicated calibration field. This agile calibration approach saves time and resources while maintaining the integrity of the calibration process.
Currently, a considerable body of research is dedicated to the calibration of LiDAR and camera systems. Some approaches, as described in [20,21,22], rely on predefined targets that are visible in both the camera and LiDAR data to estimate calibration parameters. However, to eliminate the need for pre-deployed targets, several calibration methods leverage feature extraction and matching techniques. These methods utilize various types of features, including point features [23,24,25], line features [26,27,28], surface features [29], semantic features [30], and 3D structure features [31]. Instead of establishing explicit feature correspondence between the camera image and the point cloud, certain methods [19,32,33,34,35,36,37,38] employ general appearance similarity as a metric to evaluate calibration quality, formulating the calibration problem as a nonlinear optimization task. Additionally, a few alternative approaches [39,40,41,42,43,44,45,46] based on deep learning have emerged as promising paradigms for addressing this calibration challenge.
Despite the variety of approaches employed, a general framework for gaining insight into the calibration process is still needed. Additionally, current methods tend to be highly sensitive to the initial calibration parameters. Thus, poor initial calibration parameters can lead to suboptimal calibration results using these approaches. Therefore, the primary objective of this research paper is to propose a novel and practical optimization method for the data-driven calibration of camera–LiDAR systems in data collection vehicles. Specifically, this approach aims to address the challenges associated with poor initial calibration parameters.
The paper presents several contributions:
(1)
A comprehensive and general model is proposed for the camera–LiDAR calibration;
(2)
An objective function that takes poor calibration parameters into account is introduced;
(3)
A multi-level optimization technique that achieves both calibration precision and computational efficiency is developed.
These contributions enhance the state-of-the-art in camera–LiDAR calibration and provide valuable insights for future research in this field.
In the subsequent sections of this paper, we present a comprehensive discussion of the related work, proposed methodology, experimental results, and conclusion. Specifically, Section 2 delves into the relevant literature and prior research in the domain of camera–LiDAR calibration. Section 3 delineates the proposed method, which is divided into several subsections for clarity. Section 3.1 offers a high-level overview of the devised model, which aims to achieve data-driven calibration between a camera and a LiDAR system. In Section 3.2, we elucidate the data encoding technique employed to extract edge features from dense point clouds and camera images, with an emphasis on its specific characteristics. Subsequently, Section 3.3 presents the proposed objective function, which serves as a metric for evaluating the alignment between the camera and LiDAR data. In Section 3.4, we expound on the multi-level optimization process, which refines the calibration parameters in a systematic manner. Following the detailed explanation of the proposed method, Section 4 showcases the experimental results and engages in pertinent discussions to highlight the efficacy of the approach. Finally, Section 5 provides a conclusive summary of the paper and outlines potential avenues for future research.

2. Related Work

Research on the calibration problem between LiDAR and cameras has been ongoing since LiDAR began to be used in vehicles. Calibration techniques can be broadly classified as either offline or online. Offline methods require a predefined target and are typically carried out in an offline setting. Online methods, on the other hand, rely on LiDAR and camera data and are more suitable for on-road applications. Online methods can be further classified into feature matching-based and statistical-based approaches. Statistical-based methods, also known as direct methods, use all the available information without finding corresponding points. Feature-based methods, also known as indirect methods, involve finding the corresponding points and utilizing that information. Features may include points, lines, or surfaces. In recent years, several calibration approaches utilizing deep learning techniques have also emerged. In general, calibration approaches can be classified into four categories: target-based approaches, feature matching-based approaches, statistics-based approaches, and deep learning-based approaches, as shown in Table 1.

2.1. Target-Based Approaches

Target-based approaches for camera–LiDAR calibration rely on a predefined target that is visible by both sensors. Typically, the target is designed to have a known geometric structure and can be represented in both sensor data, albeit in different forms.
The offline calibration method, using a calibration board as described in [20], can accurately calculate the relative pose between a laser rangefinder and a camera by placing the calibration board indoors. However, this method cannot be performed in real-time as the relative pose between the laser rangefinder and camera is constantly changing during vehicle operation, rendering this method ineffective. Similarly, the method of using pre-positioned ground control points for the registration of unmanned aerial vehicle images and onboard laser point clouds, as described in [21], also faces this problem.

2.2. Feature Matching-Based Approaches

Feature matching-based approaches typically involve first converting the LiDAR and camera data into a common coordinate system using the initial calibration parameters. Next, salient features are extracted from the LiDAR and camera data, such as corners or edges, using feature detection algorithms. These features are then matched between the LiDAR and camera data based on their descriptors, which are high-dimensional representations of the features. Once the matching features are identified, the calibration parameters can be estimated using optimization methods, such as the PnP algorithm or bundle adjustment. These methods compute the transformation between the LiDAR and camera coordinate systems that minimize the reprojection error between the matched features.
A feature-based method that uses Harris corner points of road markings for matching is described in [23]. This method projects the point cloud onto a plane to form an intensity image, which is then matched with the image data. Similarly, [24] extracts Harris corner points from images and performs an exhaustive search for corresponding points in the LiDAR data, with the use of the Fourier transform for computational acceleration. In [25], the authors utilize SIFT [49] to extract intensity features from point cloud images for point cloud registration.
In [26], skyline features are extracted from both the point cloud projection and the image, and an ICP [50] algorithm considering the point normal vectors is used to find the corresponding points on the skyline. Finally, the camera pose is calculated based on the corresponding points. Similarly, [27] uses a brute-force search to iteratively solve the registration parameters, and the search range is reduced by half after each iteration. Furthermore, [28] is also based on line matching. Canny edge lines are extracted from both the image and the point cloud projection, and the camera pose is calculated based on the correspondence relationship between the lines using the generalized collinearity equation.
The method proposed by [29] is based on surface matching. The method involves extracting features from both the point cloud and the digital image, then using a feature descriptor to match corresponding features. The matching is performed on planar surfaces, and the camera pose is estimated using an iterative closest point algorithm. This method can achieve high accuracy but it relies on the availability of planar surfaces in the scene.
Ref. [30] proposed a camera–LiDAR calibration method based on semantic segmentation of images. Specifically, they extracted feature objects through the semantic segmentation of images and constructed a cost function based on the matching degree of the LiDAR points projected into the feature object region. The proposed method utilized semantic information, which is a higher-level representation, and thus demonstrated robustness to scene noise compared to edge-based methods. However, this method requires specific scene requirements, such as recognizable objects with certain shapes like cars, which limits its applicability in mapping applications.
In [31], sparse point clouds are constructed through structure from motion (SFM) [51] from images, and rigid ICP is used to align the sparse point clouds with the LiDAR point clouds. However, this method is essentially an offline method since it mainly uses continuous image frames to construct sparse point clouds, and then performs ICP alignment and the joint BA solution with the LiDAR point clouds.

2.3. Statistics-Based Approaches

Statistics-based approaches typically involve projecting LiDAR point clouds onto the camera image plane using the initial calibration parameters. This creates a 2D projection image that can be compared to the actual camera image. To compare the projection image and the camera image, filtering methods are employed to process the two images separately. These methods may include edge detection, noise reduction, or other image processing techniques. After filtering, the two images are overlapped, and specific statistical measures, such as correlation coefficients or mutual information, can be calculated to measure the similarity between the two images. Once the similarity measures are computed, non-linear optimization techniques can be used to refine the calibration parameters. These techniques aim to minimize the difference between the projection image and the camera image by adjusting the calibration parameters. This optimization process can be iterative, with the calibration parameters updated after each iteration until convergence is reached.
Ref. [32] utilizes the mutual information between image pixel values and laser reflectance intensity, and ref. [33] computes the mutual information between the image pixel values and both the reflectance and the depth maps from the LiDAR data. One drawback of mutual information methods is that they heavily rely on local features, which results in a significant dependence on the initial registration parameters. Moreover, using reflectance values for the mutual information method has a drawback in that it requires calibration of the laser reflectance values, as uncalibrated reflectance values are considered invalid, which can lead to inaccurate similarity measurements. In their work, ref. [19] proposed an approach that extracts edge points from both the camera images and the LiDAR data. The method utilizes an objective function that integrates the information of camera intensity and depth discontinuity in a product sum fashion. It can detect and correct miscalibration between the two data sources through a grid search optimization.

2.4. Deep Learning-Based Approaches

Deep learning-based approaches have emerged as a promising method for LiDAR–camera calibration. These approaches aim to replace the manual feature extraction step with neural networks to better handle the complex data involved in LiDAR–camera calibration. By leveraging the powerful representation learning capabilities of neural networks, these approaches can automatically extract features that are more relevant to the calibration task. Moreover, the subsequent feature matching and parameter calculation process can also be implemented using neural networks. This enables the entire calibration process to be performed in an end-to-end fashion, with the neural network taking raw data as input and directly outputting the calibrated parameters.
In a recent study by [39], an end-to-end approach is proposed to tackle the calibration problem. This approach employs convolutional neural networks (CNNs) to extract feature information from both camera and LiDAR-projected images, and subsequently, another CNN block is utilized to establish correspondence between the features. Finally, a fully connected network is employed to output the calibration parameters. Subsequent research [40,41,42,43] has also employed neural networks as a tool to tackle the problem of calibration. Despite their impressive performance in various applications, deep learning models are known to suffer from limitations when it comes to applying them to arbitrary configurations. In such cases, conventional calibration techniques may be more practical and efficient than re-training the models. Moreover, the lack of interpretability of deep learning models makes it difficult to perform failure case analysis and estimate the operational limits analytically, which poses a significant challenge for these black box approaches.

3. Materials and Methods

This paper proposes a novel and comprehensive general framework for the online calibration of LiDAR and camera data, which abstracts out the technical details of different methods currently adopted. It also contains specific implementations used for data-driven calibration of dense LiDAR data and camera images in data collection vehicles, which includes a data encoding process that can effectively extract informative and discriminative features from dense LiDAR and camera data, an objective function that takes very poor calibration parameters into account, and a multi-level optimization process that can perform well with poor initial calibration parameters and provide rapid optimization speed. The proposed method addresses several challenges associated with data-driven calibration, especially the requirement for robustness against initial calibration parameter inaccuracies.
The overview of the proposed method is shown in Figure 1.

3.1. The Camera–LiDAR Calibration General Model

As discussed above, numerous approaches were developed for online calibration of LiDAR and camera data. However, these methods lack a unified model that can effectively summarize the common techniques utilized. Therefore, there is a need to develop a general model that can integrate the various approaches and provide a comprehensive overview of the online calibration process. Such a model can enable researchers and practitioners to understand the underlying principles and mechanisms of online calibration and provide them with a systematic and efficient approach to achieve optimal results. This model can also facilitate the development of new techniques and methods that can further improve the accuracy and reliability of LiDAR–camera calibration.
The proposed model for LiDAR–camera calibration is composed of three essential stages: data encoding, objective definition, and parameter optimization.
The data encoding stage is responsible for extracting informative and discriminative features from both LiDAR and camera data, which are then utilized for the subsequent calibration process. In feature-based methods, the data encoding pertains to the explicit feature extractor, which may include corner point extractor, skyline extractor, and other similar techniques. In statistical-based methods, the data encoding pertains to the specific statistics utilized in the encoding process. In Equation (1), F stands for the encoding function, its input includes camera image data C , LiDAR data L , and calibration parameters P . It outputs the encoded representation of camera data E n c C and LiDAR data E n c L :
E n c C , E n c L = F ( C , L , P )
The objective definition stage defines the objective function to optimize during the calibration process. This objective function seeks to minimize the difference between the measurements obtained from the two sensors, with the ultimate goal of achieving precise calibration between them. The objective function can be designed based on various metrics, such as the mean squared error or the mutual information between the two sensor measurements. In Equation (2), G stands for the objective function and O b j stands for the resulting objective score:
O b j = G ( E n c C , E n c L )
Finally, the parameter optimization stage performs the actual optimization process, whereby the parameters that maximize the objective function are iteratively adjusted. This process can be carried out using various optimization algorithms, such as gradient descent [53], the Levenberg–Marquardt algorithm [54], etc. The choice of optimization algorithm can depend on the nature of the criterion and the specific requirements of the calibration task. In Equation (3), P * stands for the resulting optimal calibration parameters:
P * = a r g m a x ( O b j )
Putting them together, the camera–LiDAR calibration general model can be expressed as Equation (4):
P * = a r g m a x ( G ( F ( C ,   L ,   P ) ) )
A synopsis of the representative prevailing literature utilizing the proposed general model is presented in Table 2 to facilitate a better understanding of the overall model.
Overall, the proposed model for LiDAR–camera calibration provides a systematic approach for achieving precise calibration between these two sensors. This model not only facilitates comprehension of existing methods but also establishes a paradigm for the development of future approaches.

3.2. Data Encoding in Road Scenes

The purpose of data encoding is to extract informative and discriminative features from the data collected by LiDAR and camera sensors.
Within road scenes, notable visual features, including lane markings and traffic signs, are captured by cameras and also have corresponding representations in LiDAR point clouds. The edges identified from camera images and point clouds demonstrate a significant level of correlation. Consequently, for effective information encoding, it is recommended to extract edges from both modalities. Specifically, in the case of camera images, edge detection should be based on pixel values, while in the case of point clouds, edge extraction should rely on point intensities.
The following details outline the specific implementation approach for the data encoding, designed for both camera image and point cloud modalities.
Image data encoding is divided into two steps:
(1)
Edge filtering
To align the image with the point cloud, edge information is used as a reference, therefore here we perform edge filtering on the image to extract the edge information in the image. In simple terms, after converting the original image to grayscale, each pixel value is set to the maximum difference between the original grayscale value of that point and the grayscale values of its eight neighboring points. The image after edge filtering is referred to as E (it should be noted that other filtering methods can also be used for edge filtering here).
(2)
Inverse distance transform
After the edge filtering process, an inverse distance transform is applied to smooth the edge image. This helps to make the final objective function smoother and rewards transformations that are closer to the true value, while avoiding getting stuck in local optima. The purpose of this step is to spatially smooth out the distribution of edge information, which in turn helps to make the objective function smoother and ensures a more stable optimization process. The specific approach is described by Equation (5).
D i , j = α · E i , j + ( 1 α ) · max x , y E x , y · γ m a x ( x i , y j )
Equation (5) employs L1 distance as a computational expedient, allowing for a linear-time application of the transformation with respect to the number of pixels. Consequently, the neighboring pixels are subject to an exponential spillover effect from each edge. The weighing factor, denoted by the symbol α , determines the relative contribution of the central pixel value and its surrounding pixels to the calculation of the resultant smoothed pixel value. Specifically, we set α = 1 3 and γ = 0.98 , consistent with the values used in [19].
Figure 2 shows the results of the camera image encoding process, encompassing the transformation from the original camera image to the resulting smoothed edge image.
In addition to camera image data encoding, the encoding of point clouds also holds significant importance. A thorough analysis of point clouds in road scenes reveals numerous discernible features, including lane markings, guardrails, poles, traffic signs, etc. Besides that, it should also be noted that the camera–LiDAR calibration requires precise geometric accuracy in both the longitudinal and the latitudinal directions. Longitudinal direction refers to the direction along the road, and latitudinal direction refers to the direction across the road. When considering the latitudinal geometric accuracy, it is noteworthy that line features on the road, such as lane markings, guardrails, and curbs, are well distributed along the latitudinal direction, rendering them effective for controlling latitudinal accuracy. As for the longitudinal direction, vertical features such as poles and traffic signs have a significant impact on controlling longitudinal accuracy due to their clear geometric contours. However, such features are relatively sparse and susceptible to occlusion and incomplete scanning in real-world scenarios, making them suitable only as supplementary features to play an auxiliary role. What is more, the line features on the road and the vertical features also exhibit notable edges in the camera image. Given these considerations, the utilization of both line features on the road and vertical features appears to be a judicious option. The overall process is illustrated in Figure 3.
(1)
Trajectory Filtering
To ensure the accuracy and reliability of calibration and eliminate potential noise in point cloud data, it is often necessary to filter out points that lie outside of the road range. One effective method involves shifting each trajectory point to the left by approximately half of the road’s width, thus generating a series of left boundary points which, when connected, define the left road boundary. Similarly, shifting each point to the right by half of the road’s width produces a set of right boundary points that define the right road boundary. Connecting these left and right boundaries generates a road surface, which can be used to filter the point cloud data and remove any points that lie outside of the defined road range.
(2)
Stratification based on Trajectory Elevation
To facilitate further analyses, the collected data points are segregated into distinct upper and lower layers based on the elevation of the corresponding trajectory points.
(3)
Ground Filtering
The process of extracting the road surface from the lower layer data involves the application of several techniques. First, the CSF filter is employed to identify and isolate the relevant data points on the road. Next, the Otsu threshold method is utilized to perform background removal, which separates the points of interest from background points on the road surface. The efficacy of this approach can be attributed to the material differences between the points of interest and background points, which result in a large discrepancy in reflection intensity. By effectively utilizing thresholding techniques, the relevant data points can be accurately extracted while minimizing the impact of background noise. Once the data are appropriately thresholded, linear feature edges such as lane lines, guardrails, and curbstones can be extracted easily.
The CSF filter [56] and Otsu [57] thresholding techniques are combined to effectively isolate the relevant data points and remove extraneous background noise. Following this, the normal vectors of the point cloud are extracted and carefully analyzed to determine whether each point falls on an edge, based on the angle between the normal vectors.
(4)
Poles Extraction
To analyze the upper layer data, the principal component analysis (PCA) [58] technique is utilized to extract point cloud clusters along the z-axis. This process is effective in identifying and isolating vertical elements such as poles and traffic signs. PCA is a statistical approach that identifies the major axis of variance in a given dataset, and the associated eigenvectors can be utilized to determine the orientation of the object under consideration. By analyzing the orientation of the eigenvectors associated with a point cloud, vertical objects can be distinguished from the surrounding environment. The edge of the vertical objects can then be extracted based on the normal vectors calculated from the point cloud.
The point cloud under different stages is demonstrated in Figure 4.

3.3. Enhanced Objective Definition via Suppression of Identical Pixel Contributions

The edges extracted from the camera image and the dense point cloud are two different representations of the same edge in the real world. If the calibration parameters are accurate enough, it is reasonable to assume that the two edges will perfectly overlap with each other. Based on this assumption, the basic idea of our camera–LiDAR calibration algorithm is to define an objective function to describe the degree of overlap between the edges in the image and the edges in the point cloud, and then search for the optimal calibration parameters that maximize the value of the objective function.
The calibration parameters are expressed as a rigid body transformation consisting of six degrees of freedom. The rotational component is represented by three Euler angles, while the translational component is represented by three lengths. This spatial transformation enables the precise alignment and positioning of objects in three-dimensional space. To obtain the corresponding objective function value O C for a given set of calibration parameters C , we can project the point cloud R i onto the image plane D i using the parameters C ; then, we can calculate the objective function value by summing the pixel values of the corresponding image edges overlapped with the projected point cloud edges. Here, we do not sum up the product of the camera pixel value and the corresponding pixel value from point cloud edges projection. That is because pixel value from point cloud edge projection does not represent the real edginess of the point cloud. Therefore, simply summing up all the camera image pixel values overlapped with the projected point cloud edges is enough and reasonable. For multiple frames, we simply sum up the results obtained for each frame. This process is illustrated in Figure 5.
The objective function can be expressed mathematically using Equation (6).
O C = f = n w n p = 1 R f D i , j f
In Equation (6), f denotes the frame number and w represents the data window size used for calculating the objective function value. The outermost sum operator means the sum of multiple frames. For example, w = 1 means only considering the most recent frame. p denotes the index of a point in the point cloud data, while i and j represent the image coordinates corresponding to the point after projection onto the image plane from the point cloud.
Despite the plausibility and potential effectiveness of the objective function, an unusual phenomenon was observed whereby, under highly suboptimal calibration parameters, the objective function attains an unusually large value. This has the potential to cause the subsequent optimization process to converge towards the suboptimal calibration parameter. Upon investigation, it was discovered that this occurrence was due to a substantial proportion of points within the point cloud being projected onto identical pixels on the image plane, specifically the edges of the camera image. This, in turn, resulted in the objective function acquiring an exceedingly undesired high value.
To avoid situations where multiple points in the point cloud are projected onto a single pixel with a high pixel value, a mechanism was introduced where a pixel only contributes to the objective function once, which will be referred to as the suppression of identical pixel contributions. In other words, if a pixel already has a point cloud projection, subsequent projections of the point cloud onto that pixel will not contribute to the objective function anymore. During the optimization process, there may be cases where incorrect calibration parameters cause the point cloud to project onto the image in a cluster of points, resulting in a significantly large objective function value. However, this calibration is obviously incorrect. Therefore, to avoid such abnormally large objective function values, we suppress the contributions of duplicate projections onto the same pixel when calculating the objective function. The suppression of identical pixel contributions can be seen in the pseudocode of Algorithm 1.
Algorithm 1: Pseudocode of objective function definition
Input: Point cloud edges data, camera image edges data, calibration parameters.
Output: Objective function value.
1. Initialize a pixel coordinates pool P o o l to store the pixel coordinates that have already contributed in the objective function calculation, initialize the objective function value as O b j .
2. Use the calibration parameters to project the point cloud edges data onto the camera image plane, which results in a list containing the corresponding pixel coordinates L i s t P .
for pixel ( x , y ) in L i s t P do
3. if ( x , y ) not in P o o l then
Add the pixel value with coordinate ( x , y ) in the camera image edges data I c x , y to the objective function value O b j .
Add the coordinate ( x , y ) into the pool P o o l .
4. Return the objective function value O b j .

3.4. Multi-Level Parameter Optimization

Prior research has demonstrated that the objective function resulting from these detailed textures exhibits prominent non-convex characteristics. Consequently, the impractical approach would entail conducting a comprehensive exploration of the calibration parameters to identify the parameter combination that maximizes the objective function. However, a global search of the parameters is time-consuming and unrealistic. A feasible method is to utilize the property of the objective function having a locally convex nature and perform an iterative local grid search optimization of the parameters. The search process for optimizing calibration parameters typically involves iteratively perturbing each parameter by a small increment, referred to as a step size, in order to explore the parameter space and identify the optimal parameters. If a search process takes r steps in one search round, the search space for one parameter will encompass ( 2 r + 1 ) candidate values, where r is referred to as the search radius henceforth. Considering we have 6 parameters in total to perturb, there will be ( 2 r + 1 ) 6 parameter combinations in the whole parameter search space. During one round of search, the objective value is calculated for each parameter candidate in the search space. After finding out the maximum objective value, the corresponding parameter will be used as the initial parameter for the next search round. The iterative process will continue until the parameter value reaches a point of convergence, defined as the maximum objective value no longer exhibiting significant changes. The computational burden of a single round of iterative grid search can be quantified by Equation (7).
L = C ( 2 r + 1 ) 6
The equation incorporates a factor denoted as C , which accounts for the iterative nature of the process. This factor may be considered constant, as it is not dependent upon the method proposed in this paper.
In fact, the parameter search space is a continuous space. And in practice, the continuous search space is discretized to make the search process feasible. Two important concepts in this discretization process are the search step size and the search radius. The search step size is the discretization unit and determines the discretization precision and hence the optimization granularity. If the search step size is set relatively large, the final optimization result may not be precise enough and deviate more or less from the theoretical optimal value. On the other hand, the search radius is the steps taken in each search round. The product of the search step size and the search radius is called the parameter search range. The parameter search range determines the “field of view” for each iteration during the optimization process, which determines the sensitivity of the optimization process to local non-convexity and hence the robustness of the optimization result.
The parameter search space is typically continuous, but in practice, it is discretized to enable feasible search processes. Two important concepts in this discretization process are the search step size and the search radius. The search step size determines the discretization precision and optimization granularity, and if set too large, may result in imprecise optimization results. In contrast, the search radius is the steps taken in each search round. The product of the search step size and the search radius is called the parameter search range, which determines the “field of view” for each iteration during the optimization process. This field of view affects the sensitivity of the optimization process to local non-convexity and hence the robustness of the optimization result.
Both the search range and search step size are important for the exploration capabilities of a search process. A larger search range can help prevent the search process from getting stuck in a local maximum, while a smaller search step size ensures that the search space is not discretized too coarsely, thus avoiding the oversight of many optimal results. However, if the search range is fixed, a smaller search step size will result in a larger search radius, which can exponentially increase the computational burden of the search process, as discussed previously.
To address this quandary, a multi-level optimization approach is proposed, inspired by the image pyramid model. This approach employs a coarse-to-fine strategy, beginning with a large search step size and a small search radius for the initial iterative grid search. This combination provides a large search range but with a relatively small search radius. The resultant parameters are then used as initial parameters for the next level of iterative grid search, where the search step size is decreased to a smaller value while the search radius remains unchanged. This updated configuration is used to conduct the new level of iterative grid search. The search process continues in this manner until the search step size is small enough to indicate that the result is precise enough. Figure 6 provides a visual illustration of the entire process.
Assuming a desired search range of R and a desired search step size of s ~ , in the conventional single-level case, the search radius r s should satisfy condition (8):
R r s < s ~
Using condition (8), we can derive the following Equation (9), given that the search radius r s should be an integer:
r s = R s ~
By applying Equation (7), we can derive the following Equation (10) for the computational burden of a single-level search:
L s = C ( 2 R s ~ + 1 ) 6
For the multi-level search method, assuming a search radius of r 0 , the initial search step size should be set as R r 0 . After each level, the search step size is divided by a constant K , resulting in a search step size of R r 0 K l after l levels. The search process terminates when the search step size is smaller than the desired search step size of s ~ . Therefore, we can derive condition (11):
R r 0 K l < s ~
Using condition (11), we can derive the following Equation (12), given that the level number l should be an integer:
l = l o g K R r 0 s ~
By applying Equation (7), we can derive the following Equation (13) for the computational burden of a multi-level search:
L m = C ( 2 r 0 + 1 ) 6 l o g K R r 0 s ~
By comparing Equations (10) and (13), we can observe that the computational burden of the single-level search method L s increases exponentially with the desired search range R , while the computation burden of the multi-level search method L m increases at a much slower rate due to the logarithmic effect. In scenarios where the initial calibration parameters exhibit significant deviations from the optimal values, a broader search range may be necessary, thereby favoring the adoption of a multi-level search method over a single-level search method.

4. Results and Discussion

4.1. Experimental Platform

To validate the efficacy of the proposed approach in this study, a data collection vehicle was utilized, which was equipped with a Z+F PROFILER 9012 2D LiDAR (Zoller & Fröhlich GmbH, Wangen im Allgäu, Germany, 360° field-of-view, scan rate of more than 1 million points per second, maximum scanning speed of 200 profiles/s, maximum range of 119 m) for the acquisition of high-density point cloud data, GPS/IMU devices (Northrop Grumman Corporation, West Falls Church, VA, USA) for the fusion and rectification of the data obtained from the single-beam LiDAR, as well as a Point Gray industrial camera (Teledyne FLIR LLC, Wilsonville, OR, USA) for image data acquisition. The sensors were mounted on a Nissan Qashqai vehicle (Dongfeng Motor Company Limited, Wuhan, Hubei, China) for data collection purposes. In particular, a metal frame was securely mounted on top of the vehicle to hold the 2D LiDAR and camera. The camera was positioned in a forward-facing direction, while the LiDAR was tilted at a vertical angle of 45° for enhanced environmental scanning capabilities.
The acquired dataset comprises 100 Individual frames of data, wherein each frame comprises a camera image and its corresponding LiDAR point cloud. The data were captured along a 5 km stretch of a highway road located in Zhejiang, China. Under sunny weather conditions with a temperature of approximately 15 °C and no significant wind, the data collection vehicle maintained a consistent cruising speed of approximately 20 km/h throughout the data collection process.
The calibration parameters for the alignment of the image and point cloud are initially determined by utilizing the calibration parameters provided by the vehicle manufacturer. To obtain the ground truth calibration parameters, a manual point selection process is employed, wherein ten corresponding point pairs are identified in each frame. Subsequently, the PnP method from the OpenCV library is applied using these point pairs to determine the ground truth calibration parameters with greater precision and accuracy.
Our algorithm is implemented in C++, and the experimental results described below were obtained on a laptop computer with an Intel i7-10875H CPU (Intel Corporation, Santa Clara, CA, USA), 16 GB RAM, and a Nvidia GeForce RTX 2080 Super with Max-Q Design GPU (Nvidia Corporation, Santa Clara, CA, USA). The objective function calculation was accelerated using the CUDA framework for GPU computation.

4.2. Evaluation and Discussion

A crucial factor affecting the optimization process is the properties of the objective function. Therefore, it is imperative to undertake a thorough analysis of the proposed objective function’s properties.
Figure 7 shows the distributions of the objective function values obtained by perturbing the six parameters on the basis of a pre-registered parameter. The distributions correspond to the perturbations of yaw, pitch, roll, and x, y, z, respectively. The perturbation unit is (0.125°, 5 cm).
The depicted figure displays that the optimal value of the objective function is attained at the actual value (located at 0 on the abscissa), indicating that maximizing the objective function would lead to obtaining the correct calibration parameters. Furthermore, the function displays prominent local convex properties in the vicinity of the actual value. However, the objective function does not exhibit global convex properties, which implies that it possesses multiple local maximums and the conventional local grid search approach might be insufficient and infeasible. Specifically, each parameter shows varying degrees of non-convexity at a global level. The rotation-related parameters, such as yaw, pitch, and roll angles, tend to exhibit a greater degree of convexity than that of translation-related parameters, including x, y, and z translations. This can be attributed, in part, to the fact that rotation can result in a more pronounced alteration of the visual features within the field of view compared to translation alone. For instance, a camera that is translated along a roadway for a limited distance may capture nearly identical visual features within its field of view, whereas a rotational adjustment can induce a more noticeable transformation of the scene.
To investigate the effectiveness of the proposed multi-level optimization approach, we conducted a quantitative analysis of the proposed approach by comparing it with the method proposed in [19]. It should be noted that the latter method employs a distinct data processing technique and objective function as compared to the approach proposed in this study. For the sake of comparison, we employed the same data processing technique and objective function as the one utilized in this paper. We perturbed the ground truth parameters and utilized them as initial values, after which we applied different optimization methods to obtain the calibration parameters. Subsequently, we evaluated the disparities between the optimized parameters and the ground truth parameters.
In this study, the primary assessment metric employed is the Mean Absolute Error (MAE), which provides a clear indication of the disparities between the calibration outcomes and the corresponding ground truth values concerning rotation angles, encompassing yaw, pitch, and roll, as well as translation distances, encompassing x, y, and z translations. Moreover, the standard deviation values of the aforementioned parameters are also calculated to effectively demonstrate the precision of the methods used in the calibration process.
To validate the effectiveness of the proposed approach when faced with poor initial calibration parameters, we utilized uniform random sampling to generate perturbations within a range of [−10°, 10°] and [−1.0 m, 1.0 m], similar to the range in [39]. This range represents a state of poor calibration parameters, which we deemed crucial to the comprehensive assessment of the approach. In each frame, 10 samples were taken. The effectiveness of the approach was assessed by calculating the mean value and standard deviation of the absolute error for yaw, pitch, roll, x, y, and z over all samples in the entire sequence of frames. These metrics were indicative of both the accuracy and precision of the approach. In addition, the average runtime was measured as a means to compute the calibration cost, thereby enabling the assessment of the approaches’ efficiency.
The Levinson method [19] was employed with two different parameter configurations: one with a search radius of 1 (denoted as Lev-1) and the other with a search radius of 2 (denoted as Lev-2). The rotation and translation step sizes for both configurations were set at (0.125° and 5 cm), respectively. These configurations were set up to investigate the impact of different search ranges on calibration accuracy and runtime, and to serve as control groups for the proposed approaches. The proposed approach utilized a desired search range R of (1°, 40 cm), a desired search step size s ~ of (0.125°, 5 cm), a search radius r 0 of 1, and a step size decrease factor K of 2, where the search step size is reduced to half of the previous search step size at each level. These numerical values were chosen based on current methods [19,39,40,42], and the proposed method configuration is denoted as Pro-1. To evaluate the impact of the enhanced objective definition via suppression of identical pixel contributions, the identical pixel suppression was turned off for a second configuration, denoted as Pro-2. The results of these configurations are presented in detail in Table 3.
Comparing the first two columns of the results, we observe that increasing the search radius can significantly enhance the calibration performance, especially when the initial calibration parameters are poor. However, this comes at the cost of significantly longer average runtimes, making larger search radii impractical due to the computational overhead. The third column demonstrates that the proposed approach in this study can greatly improve optimization performance while maintaining a reasonable computational cost. Comparing the last two columns shows that the identical pixel suppression mechanism helps to achieve a more accurate and precise calibration result.
To provide a visual representation of the proposed approach, a set of sample frames were selected and analyzed using both the Levinson method [19] and the proposed method. The calibration results obtained from each method are presented in Figure 8 for comparison. It is important to note that the Levinson method utilized a search radius of 1, ensuring that the runtime of both methods was comparable and that the calibration performance could be evaluated fairly.
The first column in the figure displays the initial overlap of the point cloud on the camera image, which indicates a significant misalignment between these two data sources. The second column shows the overlap achieved by utilizing ground truth calibration parameters, in which a perfect alignment between the two data sources is achieved. The last two columns demonstrate the calibration results obtained by applying the Lev-1 configuration and the proposed method presented in this paper, respectively. The outcomes clearly indicate that the proposed method resulted in a more accurate calibration result.

4.3. Ablation Study

The proposed calibration approach incorporates several hyperparameters that require ad hoc determination. In order to assess the significance of various components within the calibration approach, we systematically varied these components and measured the resulting impact on performance. The findings of these experiments are presented in Table 4.
In Table 4 row A, the increase in the search radius r 0 positively influences both the accuracy and precision of the calibration; however, this improvement is accompanied by a substantial increase in calibration cost, rendering the search radius r 0   the parameter with the most substantial influence on the calibration cost. Although a larger search radius expands the field of view during the search process, ensuring a more robust exploration, it is generally recommended to utilize a relatively smaller value to mitigate the associated calibration cost burden. In row B, we investigate the effect of adjusting the desired search step size s ~ for rotation and translation. Generally, a smaller desired search step size leads to more accurate and precise calibration results, albeit at the expense of longer execution time and higher calibration costs. We posit that this is due to the finer granularity of the search process, which enhances the quality of the solution. Nevertheless, the increase in search levels within the multi-level search process prolongs the overall search time. In practical scenarios, we recommend adapting the desired search step size based on the specific calibration requirements. This approach ensures that the calibration process is tailored to meet the desired level of accuracy and precision, thereby optimizing the calibration effectiveness in real-world applications.
In Table 4 row C, we further investigated the impact of the step size decrease factor K on the calibration process. Our observations indicate that a larger value of K generally compromises the quality of calibration while offering the advantage of reduced calibration cost. This phenomenon can be attributed to the fact that a larger K value accelerates the focus on a smaller search area, potentially overlooking certain regions that may contain superior solutions. Simultaneously, the calibration cost decreases due to the expedited convergence of the calibration process. In row D, we conducted an experiment by disabling the suppression of identical pixel contributions, resulting in the improved effectiveness of the calibration approach, as anticipated.

5. Limitations and Future Works

The LiDAR utilized in this study pertains to a 2D LiDAR commonly employed in data collection vehicles. However, it is important to note that in the context of autonomous driving vehicles, 3D LiDAR sensors are typically utilized [59]. Present-day 3D LiDAR devices are characterized by sparse point distributions, resulting in relatively low data density compared to camera images. This discrepancy presents challenges when comparing and associating the two data modalities. Moreover, the edge extraction technique presented in this paper may not be directly applicable when the point density is insufficient. Nevertheless, alternative feature extraction techniques [23,26,29,30] that account for the characteristics of low-density point clouds can be employed to address this limitation. Furthermore, as LiDAR sensors continue to advance in terms of capabilities [60], with the emergence of FLASH LiDAR capable of capturing point densities comparable to cameras, the proposed method holds promise for its effective application in the future.
The camera model employed in this study is based on the conventional perspective camera model. However, it is important to acknowledge that wide-angle/fisheye lenses are extensively utilized in real-world applications, particularly in autonomous vehicles, due to their ability to capture wide fields of view, including blind spots [61]. These lenses introduce various forms of visual distortion, with radial distortion being the most prominent geometric effect. Other distortions, such as inaccurate center of distortion estimation and uneven illumination, should also be taken into consideration. These distinctive characteristics of wide-angle/fisheye lenses differentiate them from conventional perspective lenses. Nevertheless, various distortion correction techniques, such as those for radial distortion correction [62,63], center of distortion [64], tangential distortion [65], and uneven illumination [66], are available to mitigate these distortions. Furthermore, several general camera models and calibration methods were proposed to account for the differences between wide-angle/fisheye lenses and perspective lenses [67,68,69]. By employing appropriate correction methods, a fisheye image can be transformed into an approximation of the desired rectilinear model, which is suitable for numerous applications. Similarly, in [27], an equidistant projection for lens projection and a spherical model for panoramic stitching were utilized, resulting in the successful registration of panoramic/fisheye image sequences and LiDAR points using skyline features. The edge-matching technique presented in this paper primarily relies on the pixel differences between adjacent pixels to extract edges from camera images. This approach is less susceptible to the challenges associated with wide-angle/fisheye lenses. Therefore, we argue that by thoroughly considering the characteristics of wide-angle/fisheye lenses, the edge-matching technique described in this paper can be applied without significant obstacles.
The proposed method in this paper demonstrates the potential for further improvement. The current methodology relies on limited data, specifically dense point clouds, and necessitates accurate extraction of road marking point cloud data. A more universal indicator derived from point cloud data could be explored to correspond with image edges. Furthermore, the multi-level optimization algorithm involves numerous hyperparameters, as discussed in Section 4.3, which are currently set on an ad hoc basis. Future research should focus on developing an automatic approach for determining these hyperparameters.

6. Conclusions

Achieving accurate calibration between camera and LiDAR sensors is critical for effective information fusion. However, current data-driven calibration methods are susceptible to poor initial calibration parameters, which can lead to suboptimal performance.
This paper proposes a novel camera–LiDAR calibration general model that offers a comprehensive and systematic framework to achieve optimal calibration results, while abstracting away technical details associated with existing methods, enabling researchers and practitioners to comprehend the underlying principles and mechanisms of data-driven calibration. It also provides an objective function that exploits the unique properties of camera image data and LiDAR point clouds. Notably, the proposed objective function takes poor calibration parameters into account, and although it exhibits global non-convexity, it possesses local convexity properties that enable iterative searches for calibration parameters and the optimization of calibration parameters. Balancing precision, robustness, and efficiency during iterative optimization within a local range is a challenging task, but the proposed multi-level optimization method successfully addresses this issue.
In the case of suboptimal initial calibration settings characterized by rotation angle deviations of up to 10° and translation deviations of up to 1.0m from the ground truth, our proposed method achieves a notable enhancement in calibration accuracy. Specifically, the mean calibration error achieved by our method is 0.3° for rotation and 5 cm for translation. This represents a substantial improvement compared to the performance metrics of existing methods, where rotation error reached 1.3° and translation error amounted to 28 cm. These results demonstrate the efficacy and effectiveness of our method in mitigating the adverse effects of poor initial calibration settings and achieving more accurate calibration estimates. In real-world applications, it should be noted that some hyperparameters can be modified to ensure a better balance between better calibration quality and high efficiency of the process. In particular, the search radius should not be increased unless the desired search step size should be determined according to the practical calibration requirements.

Author Contributions

Conceptualization, Z.J. and Z.C.; methodology, Z.J., N.H. and B.L.; software, Z.J.; validation, N.H. and B.L.; formal analysis, Z.J. and Z.C.; investigation, N.H.; resources, Z.C.; data curation, B.L.; writing—original draft preparation, Z.J.; writing—review and editing, Z.J., Z.C., N.H. and B.L.; visualization, B.L.; supervision, Z.C.; project administration, Z.C.; funding acquisition, Z.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Program of China, grant number 2021YFB2501101.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yeong, D.J.; Velasco-Hernandez, G.; Barry, J.; Walsh, J. Sensor and Sensor Fusion Technology in Autonomous Vehicles: A Review. Sensors 2021, 21, 2140. [Google Scholar] [CrossRef] [PubMed]
  2. Silva, C.S.; Wimalaratne, P. Sensor Fusion for Visually Impaired Navigation in Constrained Spaces. In Proceedings of the 2016 IEEE International Conference on Information and Automation for Sustainability (ICIAfS), Galle, Sri Lanka, 16–19 December 2016; pp. 1–6. [Google Scholar]
  3. Elmenreich, W.; Leidenfrost, R. Fusion of Heterogeneous Sensors Data. In Proceedings of the 2008 International Workshop on Intelligent Solutions in Embedded Systems, Regensburg, Germany, 10–11 July 2008; pp. 1–10. [Google Scholar]
  4. Khurana, A.; Nagla, K.S. Extrinsic Calibration Methods for Laser Range Finder and Camera: A Systematic Review. MAPAN 2021, 36, 669–690. [Google Scholar] [CrossRef]
  5. Remondino, F.; Barazzetti, L.; Nex, F.; Scaioni, M.; Sarazzi, D. UAV Photogrammetry for Mapping and 3d Modeling–Current Status and Future Perspectives. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2012, 38, 25–31. [Google Scholar] [CrossRef]
  6. Kadobayashi, R.; Kochi, N.; Otani, H.; Furukawa, R. Comparison and Evaluation of Laser Scanning and Photogrammetry and Their Combined Use for Digital Recording of Cultural Heritage. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2004, 35, 401–406. [Google Scholar]
  7. Wei, L.; Cappelle, C.; Ruichek, Y. Camera/Laser/GPS Fusion Method for Vehicle Positioning under Extended NIS-Based Sensor Validation. IEEE Trans. Instrum. Meas. 2013, 62, 3110–3122. [Google Scholar] [CrossRef]
  8. Wu, T.-E.; Tsai, C.-C.; Guo, J.-I. LiDAR/Camera Sensor Fusion Technology for Pedestrian Detection. In Proceedings of the 2017 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Kuala Lumpur, Malaysia, 12–15 December 2017; pp. 1675–1678. [Google Scholar]
  9. Budzan, S.; Kasprzyk, J. Fusion of 3D Laser Scanner and Depth Images for Obstacle Recognition in Mobile Applications. Opt. Lasers Eng. 2016, 77, 230–240. [Google Scholar] [CrossRef]
  10. Klimentjew, D.; Hendrich, N.; Zhang, J. Multi Sensor Fusion of Camera and 3D Laser Range Finder for Object Recognition. In Proceedings of the 2010 IEEE Conference on Multisensor Fusion and Integration, Salt Lake City, UT, USA, 5–7 September 2010; pp. 236–241. [Google Scholar]
  11. Khan, F.; Salahuddin, S.; Javidnia, H. Deep Learning-Based Monocular Depth Estimation Methods—A State-of-the-Art Review. Sensors 2020, 20, 2272. [Google Scholar] [CrossRef]
  12. Hossain, S.; Lin, X. Efficient Stereo Depth Estimation for Pseudo-LiDAR: A Self-Supervised Approach Based on Multi-Input ResNet Encoder. Sensors 2023, 23, 1650. [Google Scholar] [CrossRef]
  13. Park, K.; Kim, S.; Sohn, K. High-Precision Depth Estimation Using Uncalibrated LiDAR and Stereo Fusion. IEEE Trans. Intell. Transp. Syst. 2019, 21, 321–335. [Google Scholar] [CrossRef]
  14. Wang, R. 3D Building Modeling Using Images and LiDAR: A Review. Int. J. Image Data Fusion 2013, 4, 273–292. [Google Scholar] [CrossRef]
  15. Remondino, F. Heritage Recording and 3D Modeling with Photogrammetry and 3D Scanning. Remote Sens. 2011, 3, 1104–1138. [Google Scholar] [CrossRef]
  16. Kim, E.; Park, S.-Y. Extrinsic Calibration between Camera and LiDAR Sensors by Matching Multiple 3D Planes. Sensors 2019, 20, 52. [Google Scholar] [CrossRef] [PubMed]
  17. Pusztai, Z.; Eichhardt, I.; Hajder, L. Accurate Calibration of Multi-Lidar-Multi-Camera Systems. Sensors 2018, 18, 2139. [Google Scholar] [CrossRef] [PubMed]
  18. Gong, X.; Lin, Y.; Liu, J. 3D LIDAR-Camera Extrinsic Calibration Using an Arbitrary Trihedron. Sensors 2013, 13, 1902–1918. [Google Scholar] [CrossRef] [PubMed]
  19. Levinson, J.; Thrun, S. Automatic Online Calibration of Cameras and Lasers. In Proceedings of the Robotics: Science and Systems IX.; Robotics: Science and Systems Foundation, Berlin, Germany, 23 June 2013. [Google Scholar]
  20. Zhang, Q.; Pless, R. Extrinsic Calibration of a Camera and Laser Range Finder (Improves Camera Calibration). In Proceedings of the 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE Cat. No.04CH37566), Sendai, Japan, 28 September–2 October 2004; Volume 3, pp. 2301–2306. [Google Scholar]
  21. Szabó, S.; Enyedi, P.; Horváth, M.; Kovács, Z.; Burai, P.; Csoknyai, T.; Szabó, G. Automated Registration of Potential Locations for Solar Energy Production with Light Detection And Ranging (LiDAR) and Small Format Photogrammetry. J. Clean. Prod. 2016, 112, 3820–3829. [Google Scholar] [CrossRef]
  22. Geiger, A.; Moosmann, F.; Car, O.; Schuster, B. Automatic Camera and Range Sensor Calibration Using a Single Shot. In Proceedings of the 2012 IEEE International Conference on Robotics and Automation, St. Paul, MN, USA, 14–18 May 2012; pp. 3936–3943. [Google Scholar]
  23. Gao, Y.; Huang, X.; Zhang, F.; Fu, Z.; Yang, C. Automatic Geo-Referencing Mobile Laser Scanning Data to Uav Images. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2015, XL-1/W4, 41–46. [Google Scholar] [CrossRef]
  24. Wong, A.; Orchard, J. Efficient FFT-Accelerated Approach to Invariant Optical–LIDAR Registration. IEEE Trans. Geosci. Remote Sens. 2008, 46, 3917–3925. [Google Scholar] [CrossRef]
  25. Böhm, J.; Becker, S. Automatic Marker-Free Registration of Terrestrial Laser Scans Using Reflectance. In Proceedings of the 8th Conference on Optical 3D Measurement Techniques, Zurich, Switzerland, 9–13 July 2007; pp. 9–12. [Google Scholar]
  26. Hofmann, S.; Eggert, D.; Brenner, C. SKYline Matching Based Camera Orientation from Images and Mobile Mapping Point Clouds. ISPRS Ann. Photogramm. Remote Sens. Spatial Inf. Sci. 2014, II-5, 181–188. [Google Scholar] [CrossRef]
  27. Zhu, N.; Jia, Y.; Ji, S. Registration of Panoramic/Fish-Eye Image Sequence and LiDAR Points Using Skyline Features. Sensors 2018, 18, 1651. [Google Scholar] [CrossRef]
  28. Fei, D.; HaiYan, G. Automatic Registration Between Lidar and Digital Images. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2008, 37, 487–490. [Google Scholar]
  29. González-Aguilera, D.; Rodríguez-Gonzálvez, P.; Gómez-Lahoz, J. An Automatic Procedure for Co-Registration of Terrestrial Laser Scanners and Digital Cameras. ISPRS J. Photogramm. Remote Sens. 2009, 64, 308–316. [Google Scholar] [CrossRef]
  30. Zhu, Y.; Li, C.; Zhang, Y. Online Camera-LiDAR Calibration with Sensor Semantic Information. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–15 June 2020; pp. 4970–4976. [Google Scholar]
  31. Swart, A.; Broere, J.; Veltkamp, R.; Tan, R. Refined Non-Rigid Registration of a Panoramic Image Sequence to a LiDAR Point Cloud. In Photogrammetric Image Analysis; Lecture Notes in Computer Science; Stilla, U., Rottensteiner, F., Mayer, H., Jutzi, B., Butenuth, M., Eds.; Springer: Berlin/Heidelberg, Germany, 2011; Volume 6952, pp. 73–84. ISBN 978-3-642-24392-9. [Google Scholar]
  32. Pandey, G.; McBride, J.R.; Savarese, S.; Eustice, R.M. Automatic Targetless Extrinsic Calibration of a 3d Lidar and Camera by Maximizing Mutual Information. In Proceedings of the Twenty-Sixth AAAI Conference on Artificial Intelligence, Toronto, ON, Canada, 22–26 July 2012. [Google Scholar]
  33. Mastin, A.; Kepner, J.; Fisher, J. Automatic Registration of LIDAR and Optical Images of Urban Scenes. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 2639–2646. [Google Scholar]
  34. Taylor, Z.; Nieto, J.; Johnson, D. Multi-Modal Sensor Calibration Using a Gradient Orientation Measure: Multi-Modal Sensor Calibration Using a Gradient Orientation Measure. J. Field Robot. 2015, 32, 675–695. [Google Scholar] [CrossRef]
  35. Scott, T.; Morye, A.A.; Pinies, P.; Paz, L.M.; Posner, I.; Newman, P. Exploiting Known Unknowns: Scene Induced Cross-Calibration of Lidar-Stereo Systems. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September—3 October 2015; pp. 3647–3653. [Google Scholar]
  36. Napier, A.; Corke, P.; Newman, P. Cross-Calibration of Push-Broom 2D LIDARs and Cameras in Natural Scenes. In Proceedings of the 2013 IEEE International Conference on Robotics and Automation, Karlsruhe, Germany, 6–10 May 2013; pp. 3679–3684. [Google Scholar]
  37. Taylor, Z.; Nieto, J. Automatic Calibration of Lidar and Camera Images Using Normalized Mutual Information. In Proceedings of the 2013 IEEE International Conference on Robotics and Automation (ICRA), Karlsruhe, Germany, 6–10 May 2013. [Google Scholar]
  38. Taylor, Z.; Nieto, J. A Mutual Information Approach to Automatic Calibration of Camera and Lidar in Natural Environments. In Proceedings of the Australian Conference on Robotics and Automation, Wellington, New Zealand, 3–5 December 2012; pp. 3–5. [Google Scholar]
  39. Schneider, N.; Piewak, F.; Stiller, C.; Franke, U. RegNet: Multimodal Sensor Registration Using Deep Neural Networks. In Proceedings of the 2017 IEEE Intelligent Vehicles Symposium (IV), Redondo Beach, CA, USA, 11–14 June 2017; pp. 1803–1810. [Google Scholar]
  40. Wu, S.; Hadachi, A.; Vivet, D.; Prabhakar, Y. NetCalib: A Novel Approach for LiDAR-Camera Auto-Calibration Based on Deep Learning. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021. [Google Scholar] [CrossRef]
  41. Iyer, G.; Ram, R.K.; Murthy, J.K.; Krishna, K.M. CalibNet: Geometrically Supervised Extrinsic Calibration Using 3D Spatial Transformer Networks. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 1110–1117. [Google Scholar]
  42. Li, J.; Lee, G.H. DeepI2P: Image-to-Point Cloud Registration via Deep Classification. arXiv 2021, arXiv:2104.03501. [Google Scholar]
  43. Lv, X.; Wang, B.; Ye, D.; Wang, S. LCCNet: LiDAR and Camera Self-Calibration Using Cost Volume Network. arXiv 2021, arXiv:2012.13901. [Google Scholar]
  44. Lv, X.; Wang, S.; Ye, D. CFNet: LiDAR-Camera Registration Using Calibration Flow Network. Sensors 2021, 21, 8112. [Google Scholar] [CrossRef]
  45. Jing, X.; Ding, X.; Xiong, R.; Deng, H.; Wang, Y. DXQ-Net: Differentiable Lidar-Camera Extrinsic Calibration Using Quality-Aware Flow. In Proceedings of the 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Kyoto, Japan, 23–27 October 2022; pp. 6235–6241. [Google Scholar]
  46. Ye, C.; Pan, H.; Gao, H. Keypoint-Based LiDAR-Camera Online Calibration with Robust Geometric Network. IEEE Trans. Instrum. Meas. 2021, 71, 1–11. [Google Scholar] [CrossRef]
  47. Zhao, Y.; Wang, Y.; Tsai, Y. 2D-Image to 3D-Range Registration in Urban Environments via Scene Categorization and Combination of Similarity Measurements. In Proceedings of the 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 16–20 May 2016; pp. 1866–1872. [Google Scholar]
  48. Jiang, P.; Osteen, P.; Saripalli, S. Semcal: Semantic Lidar-Camera Calibration Using Neural Mutual Information Estimator. In Proceedings of the 2021 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI), Karlsruhe, Germany, 23–25 September 2021; pp. 1–7. [Google Scholar]
  49. Lowe, D.G. Object Recognition from Local Scale-Invariant Features. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999; Volume 2, pp. 1150–1157. [Google Scholar]
  50. Chen, Y.; Medioni, G. Object Modelling by Registration of Multiple Range Images. Image Vis. Comput. 1992, 10, 145–155. [Google Scholar] [CrossRef]
  51. Ullman, S. The Interpretation of Structure from Motion. Proc. R. Soc. Lond. B Biol. Sci. 1979, 203, 405–426. [Google Scholar] [CrossRef]
  52. Girardeau-Montaut, D. CloudCompare; EDF R&D Telecom ParisTech: Paris, France, 2016; Volume 11. [Google Scholar]
  53. Lemaréchal, C. Cauchy and the Gradient Method. Doc Math Extra 2012, 251, 10. [Google Scholar]
  54. Levenberg, K. A Method for the Solution of Certain Non-Linear Problems in Least Squares. Q. Appl. Math. 1944, 2, 164–168. [Google Scholar] [CrossRef]
  55. Barzilai, J.; Borwein, J.M. Two-Point Step Size Gradient Methods. IMA J. Numer. Anal. 1988, 8, 141–148. [Google Scholar] [CrossRef]
  56. Zhang, W.; Qi, J.; Wan, P.; Wang, H.; Xie, D.; Wang, X.; Yan, G. An Easy-to-Use Airborne LiDAR Data Filtering Method Based on Cloth Simulation. Remote Sens. 2016, 8, 501. [Google Scholar] [CrossRef]
  57. Otsu, N. A Threshold Selection Method from Gray-Level Histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef]
  58. Pearson, K. LIII. On Lines and Planes of Closest Fit to Systems of Points in Space. Lond. Edinb. Dublin Philos. Mag. J. Sci. 1901, 2, 559–572. [Google Scholar] [CrossRef]
  59. Royo, S.; Ballesta-Garcia, M. An Overview of Lidar Imaging Systems for Autonomous Vehicles. Appl. Sci. 2019, 9, 4093. [Google Scholar] [CrossRef]
  60. McManamon, P.F.; Banks, P.; Beck, J.; Fried, D.G.; Huntington, A.S.; Watson, E.A. Comparison of Flash Lidar Detector Options. Opt. Eng. 2017, 56, 031223. [Google Scholar] [CrossRef]
  61. Hughes, C.; Glavin, M.; Jones, E.; Denny, P. Wide-Angle Camera Technology for Automotive Applications: A Review. IET Intell. Transp. Syst. 2009, 3, 19. [Google Scholar] [CrossRef]
  62. Shah, S.; Aggarwal, J.K. Aggarwal, J.K. A Simple Calibration Procedure for Fish-Eye (High Distortion) Lens Camera. In Proceedings of the 1994 IEEE international Conference on Robotics and Automation, San Diego, CA, USA, 8–13 May 1994; pp. 3422–3427. [Google Scholar]
  63. Brauer-Burchardt, C.; Voss, K. A New Algorithm to Correct Fish-Eye-and Strong Wide-Angle-Lens-Distortion from Single Images. In Proceedings of the 2001 International Conference on Image Processing (Cat. No. 01CH37205), Thessaloniki, Greece, 7–10 October 2001; Volume 1, pp. 225–228. [Google Scholar]
  64. Asari, K.V. Design of an Efficient VLSI Architecture for Non-Linear Spatial Warping of Wide-Angle Camera Images. J. Syst. Archit. 2004, 50, 743–755. [Google Scholar] [CrossRef]
  65. Shah, S.; Aggarwal, J.K. Intrinsic Parameter Calibration Procedure for a (High-Distortion) Fish-Eye Lens Camera with Distortion Model and Accuracy Estimation. Pattern Recognit. 1996, 29, 1775–1788. [Google Scholar] [CrossRef]
  66. Yu, W.; Chung, Y.; Soh, J. Vignetting Distortion Correction Method for High Quality Digital Imaging. In Proceedings of the Proceedings of the 17th International Conference on Pattern Recognition, 2004. ICPR 2004, Cambridge, UK, 26 August 2004; Volume 3, pp. 666–669. [Google Scholar]
  67. Urban, S.; Leitloff, J.; Hinz, S. Improved Wide-Angle, Fisheye and Omnidirectional Camera Calibration. ISPRS J. Photogramm. Remote Sens. 2015, 108, 72–79. [Google Scholar] [CrossRef]
  68. Scaramuzza, D.; Martinelli, A.; Siegwart, R. A Flexible Technique for Accurate Omnidirectional Camera Calibration and Structure from Motion. In Proceedings of the Fourth IEEE International Conference on Computer Vision Systems (ICVS’06), New York, NY, USA, 4–7 January 2006; p. 45. [Google Scholar]
  69. Kannala, J.; Brandt, S.S. A Generic Camera Model and Calibration Method for Conventional, Wide-Angle, and Fish-Eye Lenses. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 1335–1340. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The overview of the proposed data-driven camera–LiDAR calibration method. The camera image captured originally is in size 1920 × 1200, and is cropped to 730 × 442 for better display. The point cloud is viewed in CloudCompare v2.11.3 [52] software and zoomed to the corresponding area for easier comparison. The image titled “Projected Edges” is obtained by projecting the point cloud in “Point Cloud Edges” onto the camera image coordinate system.
Figure 1. The overview of the proposed data-driven camera–LiDAR calibration method. The camera image captured originally is in size 1920 × 1200, and is cropped to 730 × 442 for better display. The point cloud is viewed in CloudCompare v2.11.3 [52] software and zoomed to the corresponding area for easier comparison. The image titled “Projected Edges” is obtained by projecting the point cloud in “Point Cloud Edges” onto the camera image coordinate system.
Sensors 23 08889 g001
Figure 2. Camera image encoding process. The camera image captured originally is in size 1920 × 1200, and is cropped to 730 × 442 for better display. (a) Original camera image; (b) image after edge filtering; (c) image after inverse distance transform.
Figure 2. Camera image encoding process. The camera image captured originally is in size 1920 × 1200, and is cropped to 730 × 442 for better display. (a) Original camera image; (b) image after edge filtering; (c) image after inverse distance transform.
Sensors 23 08889 g002
Figure 3. Point cloud encoding diagram.
Figure 3. Point cloud encoding diagram.
Sensors 23 08889 g003
Figure 4. Point cloud encoding process. The point cloud is viewed in CloudCompare [52] software and is zoomed in to achieve an almost bird’s-eye viewpoint, providing a clearer and more detailed representation. (a) Raw point cloud; (b) point cloud after filtering; (c) point cloud after edge extraction.
Figure 4. Point cloud encoding process. The point cloud is viewed in CloudCompare [52] software and is zoomed in to achieve an almost bird’s-eye viewpoint, providing a clearer and more detailed representation. (a) Raw point cloud; (b) point cloud after filtering; (c) point cloud after edge extraction.
Sensors 23 08889 g004
Figure 5. Objective function definition. The LiDAR edge image was generated by projecting the edges of the point cloud onto the coordinate system of the camera image. This projection can be understood as visualizing the point cloud from the perspective of the camera’s optical center.
Figure 5. Objective function definition. The LiDAR edge image was generated by projecting the edges of the point cloud onto the coordinate system of the camera image. This projection can be understood as visualizing the point cloud from the perspective of the camera’s optical center.
Sensors 23 08889 g005
Figure 6. Multi-level optimization process using search radius 1.
Figure 6. Multi-level optimization process using search radius 1.
Sensors 23 08889 g006
Figure 7. Distribution of the proposed objective function.
Figure 7. Distribution of the proposed objective function.
Sensors 23 08889 g007
Figure 8. Results of calibration on different frames. The results presented in the latter three columns correspond to the specific region delineated by the yellow boxes in the initial image, in order to enhance visual presentation.
Figure 8. Results of calibration on different frames. The results presented in the latter three columns correspond to the specific region delineated by the yellow boxes in the initial image, in order to enhance visual presentation.
Sensors 23 08889 g008
Table 1. Classification of camera–LiDAR calibration approaches presented in the literature, dissected into four main categories that are target-based approaches, feature matching based approaches, statistics-based approaches, and deep learning-based approaches.
Table 1. Classification of camera–LiDAR calibration approaches presented in the literature, dissected into four main categories that are target-based approaches, feature matching based approaches, statistics-based approaches, and deep learning-based approaches.
CategorySubcategoryReferences
Target-Based-[20,21,22]
Feature Matching-BasedPoint Features[23,24,25]
Line Features[26,27,28]
Surface Features[29]
Semantic Features[30]
3D Structure Features[31]
Statistics-BasedReflectivity—Grayscale intensity[32,47]
Surface normal—Grayscale intensity[37,38]
Gradient magnitude and orientation—Gradient magnitude and orientation[34]
3D semantic label—2D semantic label[48]
Deep Learning-BasedRegression[39,40,41,42,43]
Calibration Flow[44,45]
Keypoints[46]
Table 2. Summary of related work using the proposed general model.
Table 2. Summary of related work using the proposed general model.
MethodData EncodingObjective DefinitionParameter Optimization
[20]Predefined calibration board featuresCombination of reprojection error and laser to calibration plane errorsLevenberg–Marquardt method
[23]Harris corner extractor and canny edge extractorSum of squared error between points from template matchingLeast mean squares optimization with RANSAC
[26]Skyline pixels extractorSum of squared error between points from ICP considering the point normal vectorsLeast mean squares optimization with RANSAC
[27]Skyline pixels extractorNumber of matching skyline pixelsBrute force optimization
[31]Sparse point clouds using SFMReprojection error, point-to-plane error, and the difference between camera pose and the pose given by IMUBundle adjustment
[32]Reflectance of LiDAR data and intensity of camera imageMutual information between image pixel values and laser reflectance intensityBarzilai–Borwein steepest gradient ascent algorithm [55]
[30]Semantic segmentation from camera imageMatching degree of the LiDAR points projected into the feature object regionNon-monotonic subgradient ascent algorithm
[39]Convolutional neural networksThe difference between the real perturbation and network output perturbationStochastic gradient descent
Table 3. Quantitative result comparison. The assessment of the calibration approach’s effectiveness involves analyzing the mean value and standard deviation of the absolute error for parameters such as yaw, pitch, roll, x, y, and z. The evaluation of efficiency entails quantifying the calibration cost, which is determined by measuring the average runtime of the calibration process.
Table 3. Quantitative result comparison. The assessment of the calibration approach’s effectiveness involves analyzing the mean value and standard deviation of the absolute error for parameters such as yaw, pitch, roll, x, y, and z. The evaluation of efficiency entails quantifying the calibration cost, which is determined by measuring the average runtime of the calibration process.
ConfigurationsLev-1Lev-2Pro-1Pro-2
Mean (°)Yaw1.21280.17320.21140.3432
Pitch1.44320.29730.32540.3965
Roll1.12830.33430.38630.4230
X0.34530.03280.04370.1032
Y0.24760.06120.06730.1187
Z0.23090.04190.04410.0872
Standard
Deviation (m)
Yaw0.86350.16820.27280.2983
Pitch0.91630.21370.31460.3879
Roll1.27340.25140.48290.5654
X0.29030.06810.09000.1019
Y0.28290.06110.07350.0716
Z0.30380.03580.08380.1012
Calibration Cost (TFLOPs) 17.93161.31410.68010.932
1 We used values of 5.99 TFLOPS for RTX 2080 Super with Max-Q Design GPU.
Table 4. Variations of several hyper-parameters of the proposed approach. Unlisted values are identical to those of the base settings. MR represents the mean value of the absolute error pertaining to rotation angles, namely yaw, pitch, and roll. Similarly, MT denotes the mean value of the absolute error associated with positional coordinates, specifically x, y, and z. DR and DT stand for the corresponding standard deviations of the aforementioned errors. Furthermore, CC represents the calibration cost, which is quantified by measuring the average duration of the calibration process.
Table 4. Variations of several hyper-parameters of the proposed approach. Unlisted values are identical to those of the base settings. MR represents the mean value of the absolute error pertaining to rotation angles, namely yaw, pitch, and roll. Similarly, MT denotes the mean value of the absolute error associated with positional coordinates, specifically x, y, and z. DR and DT stand for the corresponding standard deviations of the aforementioned errors. Furthermore, CC represents the calibration cost, which is quantified by measuring the average duration of the calibration process.
r 0 s ~ K MR (°)MT (m)DR (°)DT (m)CC (TFLOPs) 1
base1(0.125°, 5 cm)20.3077 20.05170.35680.082410.680
A2 0.29730.05080.30130.073397.382
3 0.29520.04910.25500.04985523.103
B (0.0625°, 2.5 cm) 0.28460.04770.30320.060513.436
(0.25°, 10 cm) 0.32120.06590.37100.08319.452
(0.5°, 20 cm) 0.41340.09150.43340.10168.170
C 1.50.30240.04670.30300.053613.939
1.750.30210.04890.33780.051211.884
2.50.33240.06190.39840.09639.213
30.37120.09870.47700.12897.194
Dwithout suppression of identical pixel contributions0.38760.10300.41720.091610.932
1 We used values of 5.99 TFLOPS for RTX 2080 Super with Max-Q Design GPU. 2 The metrics associated with the base settings are highlighted in bold for more convenient comparisons with other variations.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jiang, Z.; Cai, Z.; Hui, N.; Li, B. Multi-Level Optimization for Data-Driven Camera–LiDAR Calibration in Data Collection Vehicles. Sensors 2023, 23, 8889. https://doi.org/10.3390/s23218889

AMA Style

Jiang Z, Cai Z, Hui N, Li B. Multi-Level Optimization for Data-Driven Camera–LiDAR Calibration in Data Collection Vehicles. Sensors. 2023; 23(21):8889. https://doi.org/10.3390/s23218889

Chicago/Turabian Style

Jiang, Zijie, Zhongliang Cai, Nian Hui, and Bozhao Li. 2023. "Multi-Level Optimization for Data-Driven Camera–LiDAR Calibration in Data Collection Vehicles" Sensors 23, no. 21: 8889. https://doi.org/10.3390/s23218889

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop