Next Article in Journal
Multi-Mode Data Generation and Fault Diagnosis of Bearings Based on STFT-SACGAN
Next Article in Special Issue
An Improved Method for Photovoltaic Forecasting Model Training Based on Similarity
Previous Article in Journal
A Non-Intrusive Load Monitoring Method Based on Feature Fusion and SE-ResNet
Previous Article in Special Issue
An Efficient Framework for Accurate Liver Segmentation in Abdominal CT Images with Low Knowledge Requirement
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dynamic Vehicle Pose Estimation with Heuristic L-Shape Fitting and Grid-Based Particle Filter

1
School of Internet of Things, Nanjing University of Posts and Telecommunication, Nanjing 210023, China
2
School of Computer, Software and Cyberspace Security, Nanjing University of Posts and Telecommunication, Nanjing 210023, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(8), 1903; https://doi.org/10.3390/electronics12081903
Submission received: 28 February 2023 / Revised: 6 April 2023 / Accepted: 15 April 2023 / Published: 18 April 2023
(This article belongs to the Special Issue Advances in Image Processing and Detection)

Abstract

:
Vehicle pose estimation with LIDAR plays a crucial role in autonomous driving systems. It serves as the fundamental basis for functions such as tracking, path planning, and decision-making. However, the majority of current vehicle pose estimation techniques struggle to produce satisfactory results when faced with incomplete observation measurements, such as L-shaped point cloud clusters without side contours or those including side-view mirrors. In addition, the requirement for real-time results further increases the difficulty of the pose estimation task. In this paper, we present a vehicle Pose Estimation method with Heuristic L-shape fitting and grid-based Particle Filter (PE-HL-PF). We design a geometric shape classifier module to divide clusters into symmetrical and asymmetrical ones according to their shape features. Furthermore, a contour-based heuristic L-shape fitting module is introduced for asymmetrical clusters, and a structure-aware grid-based particle filter is used to estimate the pose of symmetrical clusters. PE-HL-PF first utilizes a heuristic asymmetrical module that selects dominant contours fitting orientation in a heuristic manner, thereby avoiding the need for a complex traversal search. Additionally, a symmetrical module based on particle filtering is incorporated to enhance the stability of orientation estimation. This method achieves significant improvements in both the runtime efficiency and pose estimation accuracy of incomplete point clouds. Compared with state-of-the-art pose estimation methods, our PE-HL-PF demonstrates a notable performance improvement. Our method can estimate the pose of thousands of objects in less than 1 millisecond, a significant improvement over previous methods. The results of experiments performed on the KITTI dataset validate the effectiveness of our approach.

1. Introduction

The field of autonomous driving has gained significant attention worldwide, leading to extensive research efforts in recent years [1,2]. The autonomous driving system consists of several components, including environmental perception, decision-making and planning, and control execution [3]. A critical aspect of a safe autonomous driving system is accurate vehicle tracking, which enables the ego-car system to plan an appropriate path and avoid collisions with other vehicles. As a critical pre-processing step for object tracking, vehicle pose estimation has recently become a hot area of research.
Light Detection and Ranging (LIDAR) scanners have been spotlighted as one of the most important sensors in autonomous driving systems. Over the past few years, numerous studies have been conducted on vehicle pose estimation using LIDAR data due to its unique advantages, e.g., range field of view, lighting invariance, and high data accuracy. In general, the methods used for vehicle pose estimation can be classified into three categories: deep-learning-based methods [4,5], model-based methods [6,7], and feature-based methods [8,9].
Deep-learning-based methods for vehicle pose estimation are similar to those widely applied in computer vision [10]. These methods extract features from a vast amount of labeled data to train the model, which can then output the estimated pose in an end-to-end manner. Although the effectiveness of these methods has been verified in certain situations, achieving satisfactory performance requires substantial computational resources and large-scale labeled datasets. These limitations limit the scalability of this kind of method for large-scale commercial applications.
In contrast to deep-learning-based methods that use an end-to-end approach, several studies have proposed step-by-step methods to process LIDAR data [7,11,12]. These methods first remove the ground points from the point cloud using the ground segmentation algorithm [11], then cluster the remaining points that belong to the same targets, and finally employ either model-based or feature-based pose estimation methods to fit the pose of targets. Unlike deep-learning-based methods, these methods do not require handcrafted labeled data, and the computational loads are usually low, making them widely used by industry.
Model-based vehicle pose estimation methods [6,7,12,13] typically use rectangles or cuboids to represent vehicles, and then employ optimization or sampling techniques to generate the most probable pose of the vehicle. As shown in Figure 1, the side-view mirror of (a) causes a skew when estimating the orientation of the vehicle, where the orange rectangle is the predicted bounding box with model-based methods and the red rectangle denotes the ground truth.
Feature-based vehicle pose estimation methods typically rely on edge features and line fitting algorithms, such as L-shape fitting or Principal Component Analysis (PCA).
However, these methods rely heavily on the features of the side edge. When the side-edge is displayed incompletely, they struggle to accurately regress the orientation. As shown in Figure 1, the boundaries of the side-edge in target (b) are unclear, and measurements are approximately symmetrical about a certain ray of LIDAR. Hence, a single treatment for all measurements could encounter ambiguities if the estimated targets are of various shapes. In addition, it is crucial to ensure the limits of real-time performance for vehicle pose estimation, which can help save time for highly computational tasks such as tracking and path planning.
We present a novel solution, named the Pose Estimation Algorithm with Heuristic L-Shape Fitting and grid-based Particle Filter (PE-HL-PF), to address the challenges encountered in vehicle pose estimation in real-world traffic scenarios. PE-HL-PF leverages the divide-and-conquer approach to classify the shape of targets into symmetrical-like and asymmetrical-like types. For targets that are symmetrical to a LIDAR ray from the ego-car, we regress the orientation using a structure-aware weighted sum of the grid-wise velocity, which is achieved using the verified particle filter algorithm. Otherwise, asymmetrical-like ones are processed with the contour-based heuristic L-shape fitting method (HL). It extracts length-aware contours with visible constraints. Furthermore, it then selects the dominant contours and fit orientation with RANSAC. The proposed approach tackles the challenge of incomplete edge features that can lead to ambiguity in existing methods while also ensuring real-time performance, which is essential for computational tasks such as tracking and path planning. Our source code can be found at https://github.com/sunjingnjupt/PE-HL-PF (accessed on 27 February 2023).
We summarize our contributions as follows.
(1)
We propose a novel pose estimation method in this paper that combines heuristic L-shape fitting and grid-based particle filter techniques for the first time.
(2)
For symmetrical clusters, the target’ heading is derived by applying a structure-aware weighted average of grid-wise velocities. This manner improves the stability of pose estimation for sparse measurements.
(3)
For asymmetrical targets, we propose a contour-based heuristic L-Shape fitting module, which heuristically learns the length-aware contours and heading of the target. It enhances the model’s robustness to outliers while reducing its running time.
(4)
We conduct extensive experiments on the KITTI dataset to assess the performance of our PE-HL-PF model. Our PE-HL-PF model outperforms existing methods in terms of both runtime speed (10-fold increase) and accuracy (51% reduction in orientation estimation error and 27% decrease in position estimation error).

2. Related Works

In autonomous driving systems, both the ego vehicle pose estimation and the pose estimation of surrounding dynamic vehicles are important components of the perception system, and they interact with each other. In this section, we will first introduce the relevant work on ego vehicle pose estimation. Furthermore, we then review the dynamic vehicle pose estimation methods in termsof deep-learning-based vehicle pose estimation methods and traditional machine-learning-based vehicle pose estimation methods.

2.1. Ego Vehicle Pose Estimation Methods

In the field of vehicle pose estimation, determining the ego vehicle pose is of utmost importance. To achieve this, numerous scholars [14,15,16,17,18] have resorted to integrating various systems, such as the Global Navigation Satellite System (GNSS), cameras, and Inertial Measurement Unit (IMU) systems.
In [14], the authors proposed an integrated localization system that combines the GNSS, IMU, and wheel speed sensor (WSS). By using the Kalman filter to integrate various information, the localization accuracy is greatly improved. Ref. [15] utilizes the IMU and automotive onboard sensors to estimate the yaw misalignment. Thus, the measurement deviation of the IMU can be reduced, and the accuracy of state estimation can be improved. Ref. [16] fuses the information from the GNSS and an IMU and proposes a new vehicle slip angle (VSA) estimation method. Ref. [17] proposes a sideslip angle estimation that incorporates a dynamic model and visual information while considering acceleration error and camera measurement error. In ref. [18], authors proposed an algorithm for estimating the sideslip angle of an autonomous vehicle, which uses consensus and vehicle kinematics and dynamics synthesis. All the above methods have improved the estimated accuracy of ego vehicle pose estimation.

2.2. Deep-Learning-Based Vehicle Pose Estimation Methods

Recently, many famous deep-learning-based vehicle pose estimation methods have been proposed, which can be assigned to LIDAR-based methods or fusion-based methods.
Lots of LIDAR-based vehicle pose estimation methods have been proposed. PointPillars [19] converts the raw point data into a stacked pillar tensor and a pillar index tensor, then the features that can be learned from the pillars are put into a convolutional neural network (CNN), and the features from the last step are used by the detection head to predict 3D bounding boxes for objects. In order to reduce the computational cost, ref. [4] proposes a task-oriented and instance-aware down-sampling strategy to sample the interesting foreground points and perform the task of object detection. Similarly, Chen et al. [20] also designed a semantic-guided point sampling algorithm to help retain more useful foreground points. PillarNet [19] can reduce noise in depth estimation tasks by discarding redundant voxels and mapping sparse voxels back to the image space. It can effectively suppress noise and improve accuracy. Furthermore, VoxelNeXt [21] is a 3D object detection and tracking method that directly predicts objects based on sparse voxel features without relying on hand-crafted proxies. BTCDet [22] predicts the probability of occupancy in these regions. Therefore, with the learned occlusion information, the detection model can generate high-quality 3D proposals and refine the final bounding box.
Besides the LIDAR-based methods, other researchers aim to fuse the RGB data and the 3D point cloud data to improve the detector’s performance. Wang et al. [23] proposed sliding Frustums to aggregate local point-wise features for the task of 3D object detection. Liang et al. [24] designed a learnable architecture that mainly includes two streams: one for extracting image features, and another for extracting features from LIDAR Bird’s Eye View (BEV). They then designed the continuous fusion layers to bridge multiple intermediate layers on both streams. The detection header was used to generate the final detection results in BEV space. Yang et al. [25] proposed a map-aware detector to detect 3D point cloud objects. They exploited information on the road layout from the High-Definition (HD) maps and rasterized it onto the BEV, and then they connected the road mask representation with the LIDAR representation to produce the input representation. After that, they adopted a Fully Convolutional Network (FCN) to detect the point cloud objects. In order to handle the inferior image conditions, Bai et al. [26] proposed a soft-association mechanism with the famous transformer technology to fuse the LIDAR point and image pixel, which can adaptively fuse the object queries with useful image features. DeepFusion [27] designs the inverseAug and learnableAlign to align the transformed features from the LIDAR and camera data. InverseAug projects the key points obtained after the data augmentation via inversing geometric-related augmentations and fuses the LIDAR and camera data. Meanwhile, learnableAlign leverages cross-attention to dynamically capture the correlations between the image and LIDAR feature during fusion.
In general, deep-learning-based methods do well in vehicle detection with sufficient labeled point cloud data. As we all know, the precision of the deep-learning-based methods depends heavily on the training dataset [28]. However, it is hard to provide a wealth of training data sets for all kinds of scenarios with various complex obstacles in the actual scene, which is a main limitation for the wide applications of deep-learning-based methods. Another critical issue that exists in deep-learning-based methods is the high computational complexity. Highly computational resources are needed to meet the real-time constraint, limiting the broader application in the real world.

2.3. Traditional Machine-Learning-Based Vehicle Pose Estimation Methods

Traditional machine-learning-based methods can extract usable information about the object, such as trajectory and orientation, without considerable labeled training data. Due to their ease of deployment on the vehicle side, they have received increasing attention from industry and academia. These methods can be grouped into two separate directions: the first one is the model-based method, and the other one is the feature-based method.

2.3.1. Model-Based Methods

Model-based methods typically parameterize the rectangle of point cloud objects, and the most likely vehicle pose is determined through model iteration. Ref. [6] employs a likelihood field model to construct rectangles representing the vehicles, followed by using the Scaling Series algorithm to sample the estimated pose. On the other hand, Ding et al. [29] proposed an efficient convex hull-based pose estimation model that uses the minimum occlusion area as the criterion for deciding the best orientation.
The convex-hull-based methods usually calculate the 2D convex hull of the top-view projection of the objects. They then derive the 2D bounding boxes directly from the convex hull with various strategies, such as the minimum distance-based strategy [30] and the minimum area-based strategy [29]. The prior strategy chooses the optimal bounding box by minimizing the average distance between the points of the convex hull and the fitted rectangle with the convex hull edges. The other strategy selects the optimal bounding box by minimizing the area of the rectangle fitted with the convex hull edges. He et al. [31] presented a 3D pose estimation model based on the heuristic rules. It considers both the edge length and edge visibility to filter the convex hull-based orientations, and the tracking results are also used to smooth the orientation estimates. Benjiamin Naujoks et al. [13] proposed an orientation-corrected bounding box fitting method that uses rotating calipers to construct the minimum area rectangle of a convex hull. Then it fits the orientation of the rectangle with a heuristic search approach, and the computational time is significantly reduced. Liu et al. [32] used the convex-hull model to acquire a rectangle of the vehicular point clouds and then inferred the center position and the visible sides of the object rectangle, and they estimated the pose with the bounding box and position inference. In addition, ref. [33] proposes a novel orientation normalization vehicle modeling method containing an orientation normalization bounding box fitting model and a view-dependent orientation normalization model.
While model-based methods have achieved encouraging results in vehicle pose estimation, they are prone to being highly sensitive to noise and outliers in the data. Additionally, overfitting of the convex hull can lead to inaccurate or unstable results. For instance, side view mirrors can significantly influence the convex hull of a vehicle, especially when they extend beyond the main body. Therefore, there is a need for more accurate pose estimation models that can account for these issues and provide reliable and robust results.

2.3.2. Feature-Based Methods

The other common type of pose estimation approach is the feature-based method [34,35,36], they rely on the edge features to deduce the pose of the objects. The most famous type among the feature-based methods is the method based on the L-shape fitting. [34] partitions a cluster into two disjoint sets optimally using the least square error and then fits them with two perpendicular lines. Zhang et al. [35] proposed an efficient search-based L-shape fitting method to produce the bounding box of the vehicle detection. Firstly, they iterated through all the possible directions of the rectangle; for each rectangle oriented in that direction, they calculated the distance of all points to the rectangle’s edges as the corresponding square errors. They can find the optimal direction with the least squared error and fit the rectangle based on this direction. This method reaches a state-of-the-art performance in pose estimation. The method in [9] innovatively decomposes the L-shape fitting problems into two steps: L-shape vertex searching and L-shape corner point locating with 2D laser data, which can significantly reduce the computational complexity. Darms et al. [36] extracted the edge feature of the target from the raw data and estimated the target as a box, which is used in the tracking process. The aforementioned research has shown that these methods perform effectively when using laser data. With the development of the LIDAR, some researchers applied the feature-based method to fit the 3D point cloud objects.
Himmelsbach et al. [37] proposed a method that fits the dominant line of the point cloud segments in the x y -plane with random sampling consistency detection (RANSAC) to determine the orientation of the vehicle. In [38], a corner-edge-based algorithm is designed to fit the pose of the vehicle. It first finds the nearest corner of the target and uses the RANSAC to fit the L-shape-based orientations. The LIDAR-based 3D object perception method proposed in [39] compresses the 3D LIDAR data into a 2.5D occupancy grid and fits the bounding box of the point cloud objects by finding connected components of grid cells. Lalonde et al. [40] used covariance analysis to capture the spatial distribution of points and derived the orientation of the point cloud object with principal components analysis.
In order to improve the efficiency of the model, Ye et al. [41] extracted the points on the border and used an L-shape algorithm to fit these points into a polyline. Finally, it selects the longest line of polylines as the orientation of the cluster with the multi-layer laser. A fast RANSAC-based approach [8] maps the 3D points into the 2D grids and selects the boundary cells of the targets, then selects the visible cells and fits every possible orientation line with the selected cells by the RANSAC. Finally, it validates the best orientation from the all-fitted oriented lines.
When most feature-based methods deal with targets with specific geometric shapes, such as an L-shape, they have certain effectiveness. However, the performances become unsatisfactory when dealing with complex real-world traffic scenes that involve various shapes of point cloud targets. For instance, the L-shape-based methods [38,41] are limited to L-shaped targets and may not perform well with targets of other shapes, such as symmetrical ones. Besides, the L-shape-based methods [38,41] are also sensitive to noise and outliers when fitting the orientation.
Both model-based and feature-based methods for vehicle pose estimation have demonstrated promising results in certain scenarios. However, they may not be able to handle point cloud targets of various shapes simultaneously, which can lead to overfitting or underfitting of the models and result in inaccurate or unstable results. Moreover, most methods are susceptible to the presence of noise and outliers in the data, which can affect the accuracy and consistency of the results. Meanwhile, they are computationally expensive, especially for large datasets, and may require tuning of various parameters to obtain optimal results. To address these limitations, there is an urgent need for a robust and efficient pose estimation method that can handle point cloud targets of various shapes simultaneously, is resistant to noise and outliers, and has a reduced computational burden.

3. A Pose Estimation Algorithm with Heuristic L-Shape Fitting and Grid-Based Particle Filter (PE-HL-PF)

Given a point cloud cluster P = p 1 , p 2 , , p i , , p N where N denotes the number of the cluster, and the p i = x i , y i , z i T denotes the Euclidean coordinate with regard to the ego-coordinate system, we will firstly fit a bounding box to estimate the vehicle pose in x y -plane. Afterward, we use the z value of the points to fit a 3D bounding box for the object. The overall diagram of the proposed Pose Estimation algorithm with Heuristic L-shape fitting and a grid-based particle filter (PE-HL-PF) are shown in Figure 2. Adopting the thought of divide and conquer, PE-HL-PF introduces a geometric shape classifier to divide clusters into symmetrical-like or asymmetrical-like ones for the first time. Furthermore, for asymmetrical-like ones, a contour-based heuristic L-shape fitting method (HL) is introduced. It extracts the length-aware contours and regresses orientation with dominant contours. Furthermore, the symmetrical clusters regressed the orientations with a verified particle filter algorithm (PF). The details about our PE-HL-PF are given in the following subsections.

3.1. Geometric Shape Classifier & Symmetrical Estimation Model for Vehicles

In real-world autonomous driving scenes, there exist a number of point cloud clusters with various shapes. Such as clusters in Figure 2, the symmetrical point cloud clusters that are labeled with orange ellipses, are rightly symmetrical about a specific LIDAR ray from ego-car. Additionally, the L-shape clusters, which are marked with rectangles in Figure 2, are asymmetrical relative to the ego-car. Many of the currently available methods treat all measurements with a uniform approach, such as convex-hull-based or L-shape-based pose estimation methods, which may result in inaccuracies when attempting to estimate the pose of clusters with varying shapes. To alleviate this, we introduce the Geometric Shape Classifier (GSC) to categorize clusters as either symmetrical or asymmetrical. Subsequently, we develop separate modules for estimating symmetrical and asymmetrical poses to process these clusters.

3.1.1. Geometric Shape Classifier

Clusters are categorized as symmetrical if they meet two criteria: the calculated width falls within a predefined threshold and the number of symmetrical contours exceeds a specified threshold. It should be noted that the symmetrical designation refers to approximate symmetry rather than strict symmetry and is employed for the purpose of point cloud processing by type. Figure 3 provides an illustration of the Geometric Shape Classifier. Further details regarding the Geometric Shape Classifier and symmetrical pose estimation model are provided in this subsection.
To calculate the width of the cluster in the first criteria, we define two endpoints p m a x and p m i n in a cluster according to the observation azimuths described with Equation (1).
θ o b z p i = a t a n 2 y i , x i
We define p m a x and p m i n as the points that produce the maximum observation azimuth θ o b z p m a x and the minimum observation azimuth θ o b z p m i n on the target, respectively. Besides, the angle covered by the point cloud cluster is defined as Δ θ v e h i c l e = θ o b z p m i n θ o b z p m a x . As shown in Figure 3a, the 1 2 width of the cluster is obtained by selecting the longer distance between the endpoints and the angle bisector l b of Δ θ v e h i c l e . If the width of the cluster falls within the predefined range w i d t h s , w i d t h l , we proceed to the second step of selecting symmetrical clusters by calculating the number of symmetrical contours, as demonstrated in Figure 3b.
To extract the contours of cluster P, the GSC algorithm divides P into a set of segments S S y m m e t r i c = s 1 , s 2 , s 3 , , s j , , s M evenly, where M represents the number of segments. The contour is then obtained by selecting the point in each segment that has the smallest distance to the origin point. Once the contours have been extracted, their distances are compared with determine if the cluster is symmetrical about the l b . For instance, we compare the distance values of the contour points in s j i = 1 , 2 , , M and s M j + 1 j = 1 , 2 , , M separately. If the distance difference between the corresponding contour points is less than the predefined threshold D s y m we will define the two contours as symmetrical. If the number of symmetrical contours on the cluster exceeds a specified threshold N s y m , we will define the cluster as symmetrical.

3.1.2. Symmetrical Estimation Model

For the symmetrical clusters, the LIDAR measurements are modeled as a dynamic occupancy grid with particles, as proposed in [42]. Specifically, the point cloud scene is represented by 2-D discrete grids, and the vehicles are represented by a set of particles T = t i t i = x i , y i , v x i , v y i , a i , i = 1 , , N T , where x i and y i denote the position of particle t i along x -axis and y -axis, v x i , v y i denotes the velocity along x -axis and y -axis, a i denotes the age of the particle existence, and the N T denotes the number of particles. By utilizing the coordinate mapping function f · , the particle’s position and velocity on the occupancy grid can be denoted as T = t i t i = c i , r i , v r i , v c i , a i , i = 1 , , N T . The dynamic occupancy grid map can be represented as a binary grid, where grids occupied by particles are assigned a value of 1, while unoccupied grids are set to 0.
Subsequently, the particles are transferred from the previous frame t 1 to the current frame t, utilizing the estimated velocity and motion parameters of the ego vehicle. As a result, the set O ^ G can be used to represent the index of predicted occupied grids in the current frame. In addition, the index of measured occupied cells of the vehicles on the current frame can be represented as O ¯ G . The particles within overlapping grids in O G are resampled, while those in non-overlapping cells are destroyed. The overlapping map can be represented with Equation (2):
O G = O ^ G O ¯ G
Upon completion of a cycle of estimation for each particle, the velocity of each grid can be estimated as the average velocity of the particles whose position coincides with the position of the grid. The grid-wise velocity can be computed with Equation (3). Where v c g a n d v r g denote the velocity of the grid along column and row, and ( r g , c g ) represents the index for the grid along row and column.
v c g , v r g = t i T , f x i = r g , f y i = c g v c i , v r i t i T r i = r g , c i = c g
Next, the velocity of the vehicles is computed using a structure-aware weighted average of velocities for the corresponding grid. Specifically, greater weight is assigned to the velocity of contour grids relative to that of inner grids, based on statistical findings indicating that the speed predicted by the edge grid more closely approximates the true value. Furthermore, we apply the direction of the velocity as the estimated heading of symmetrical clusters. Lastly, the oriented bounding box of symmetrical clusters is fitted, referring to the work in [13].

3.2. Asymmetrical Pose Estimation Model for Vehicle

Asymmetrical clusters are typically characterized by the following geometric features in contrast to symmetrical clusters. (a) Asymmetrical clusters usually have two visible sides with different lengths, whereas symmetrical clusters have one or two sides with symmetrical lengths. (b) On the side view of asymmetrical clusters, there are many irregularly raised points, such as a side view mirror on the side surface, making it difficult to use all contours fitting orientation with traditional L-shaped-based methods directly.
Regarding characteristic (a), the asymmetrical model extracts contours based on the visible sides’ length. Specifically, it divides the clusters into multiple segments with different coverage angles, where the coverage angles correspond to the length of the located sides. This approach allows for learning more geometric features around the longer sides.
For the characteristic of (b), the asymmetrical model has conducted itself in the following two manners. Firstly, it directly selected the optimal contours around the rotated longer side instead of dealing with the L-shape fitting problem as a complex optimization problem. Secondly, the asymmetrical model fits the orientation line with RANdom SAmple Consensus (RANSAC) algorithm, which is robust when coping with irregular raised contours on the side-view. The illustration of the process for the asymmetrical model is shown in Figure 4. The details of the asymmetrical model are discussed in the following subsections, which contain two parts: length-aware contour extraction and orientation fitting with optimal contours.

3.2.1. Length-Aware Contours Extraction

As mentioned above, the asymmetrical clusters contain two visible sides with different lengths, and we extract contours according to the length of the visible sides. The first step is to find the corner point of the asymmetrical cluster and obtain two visible sides by connecting the line between the endpoints and the corner point. Specifically, the process of acquiring the corner point is shown as follows.
Firstly, as shown in Figure 4, we utilize l to indicate the vector between endpoints p m a x and p m i n , then, the foot point f o o t l p i of p i corresponding to l can be calculated through Equation (4)
f o o t l p i = p max p max p i , l l , l · l
where · denotes the inner product. Consequently, the relative distance d p i , l between point p i and vector l is obtained using Equation (5)
d p i , l = n ˜ , p i p max
where n ˜ is the normal vector of l . We design two constraints for corner point p c extraction: The distance d p c , l should be the maximum among d p i , l , i = 1 , 2 , , N , where N is the number of points in the cluster.
To cope with the outliers and ensure the validity of the p c , p c and the origin O e of ego-car coordinate system should be on the same side of l . The process of finding the corner point can be formulated using the following Equation (6).
max p i P d p i , l s . t . d p i , l · d O e , l > 0
Having found the corner point p c , we connect the corner point p c and p max with the l max , and connect the p min and p c with l min .
Contours are obtained in different sector segments with different angular resolutions around the longer side and shorter side, based on the lengths of l m a x and l m i n . Specifically, the contour point is the point nearest to the ego-car among all points within a segment.

3.2.2. Orientation Fitting with the Optimal Contours

In general, the orientation line of a vehicle is often parallel to its longest side. According to this common sense, we heuristically choose the contours around the longer side of the edges to fit the orientation of the target. Nevertheless, apart from that, there are lots of contours that stay away from the outline, such as the contour point in segment s 3 in Figure 5, which may result in skewing of predicted orientation lines. We overcome this issue by selecting optimal contours in a range that excludes unqualified contours.
Specifically speaking, we use the longer edge l m a x as the dominant line l d o m i n a n t . Furthermore, the rotated dominant l d o m i n a n t can be achieved by rotating l d o m i n a n t around p c from the direction of l d o m i n a n t to the direction away from the ego-car with the angle δ , as shown in Figure 5b. The slope value k d o m i n n a t of l d o m i n a n t can be calculated with Equation (7)
k d o m i n n a t = tan θ l + δ = tan θ l + tan δ 1 tan θ l tan δ
where θ l represents the slope observation angle of l d o m i n a n t . The optimal contours c j * can be obtained with the following constraint:
d c j * , l d o m i n a n t · d p dominant , l d o m i n a n t > 0
where p dominant denotes the other endpoint of the l d o m i n a n t except p c .
Figure 5 shows the schematic illustration for optimal contours extraction. As we can see, the points around the shorter edge l m i n are divided into two segments s 1 , s 2 , and the points around the longer edge l m a x are divided into nine segments s 3 , s 4 , s 5 , , s 11 . We select the contour points in each segment, which are marked with blue.
The RANSAC [8] algorithm is used to fit the orientation line l v of the vehicle because it is robust to outliers and has a lower computational load. After obtaining the yaw angle of orientation, the position of the target is inferred according to the method in [13]. The obtained tight bounding box is shown as the black rectangles in Figure 4. To handle occlusions, such as objects in front of or behind the ego-car, the width and length of the estimated bounding box are inferred by setting a predefined length.
Since the above processes are based on the X Y plane, In order to acquire the height of the estimated bounding box, we use the following Equation (9)
h = z max z min
where z max denotes the maximum value in the z dimension of the original X Y Z coordinate system. Similarly, z min indicates the minimum value of the z dimension of the original X Y Z coordinate system.

4. Experimental Results and Analysis

The details of the dataset and evaluation metrics are given in Section 4.1, we also give the implementation details and compare methods in Section 4.2. Then we report and analyze the experimental results with aspects of estimated results (angle error and position error) and running time. The example video can be found online (https://youtu.be/FMX1Pya6qwg (accessed on 27 February 2023)).

4.1. Dataset and Evaluate Metrics

To evaluate the performance of our proposed method, PE-HL-PF, we conducted experiments using the KITTI object tracking evaluation dataset [43], which consists of 21 training sequences and 29 test sequences. Similar to [12], we selected the most common category, “car”, to evaluate our performance.
As the ground truth of the test sequences was unavailable, we utilized the training sequences for training and evaluating our methods. Specifically, following the standard protocol as outlined in [23], we partitioned the training data into two subsets in a 1:1.1 ratio. Specifically, one subset was utilized for adjusting the model’s parameters, while the other was used for evaluating its performance. The evaluation subset consisted of sub-dataset IDs 0000, 0002, 0006, 0007, 0008, 0009, 0010, 0013, 0019, and 0020. The selected sub-datasets comprise a wide range of complex scenarios, including highways, cities, and suburban areas. And the measurements exhibit various shapes due to different measurement views and mutual occlusion between objects in these datasets. Further details about the selected sub-datasets are provided in Table 1.
We employed widely used evaluation metrics, including position error and angle error, to evaluate the accuracy of PE-HL-PF. Furthermore, the running time was also measured to verify that our PE-HL-PF is computationally efficient.
Position error is defined as the Euclidean distance between the center of the ground truth bounding box and the center of the estimated bounding box. And the ground truth is provided by the KITTI website. In light of the fact that the estimation model aims to accurately estimate the location of the center of the predicted bounding box, it is clear that a good estimate is reflected by a small position error.
Angle error refers to the deviation between the ground truth and the estimated orientation. PE-HL-PF pursues the precisely oriented bounding boxes for the surrounding clusters since a precise predicted orientation is crucial to downstream tasks such as tacking. Thus, a good estimated model has a small angle error.

4.2. Implementation Details and Compared Methods

We use a standard office computer with a 2.60 GHz Intel Core i5 CPU and 7.5 Gb of RAM. Our approach and comparison methods are implemented in the C++ framework on the Ubuntu system. In order to facilitate the visualization of the algorithm’s performance, we adopted the GUI library Qt5, open-source OpenGL library, and the QGLViewer library. Furthermore, we also added numerous interactive buttons to facilitate the parameter adjustment. Furthermore, there is no parallelization enabled.
In this paper, the segmentation and ground plane line extraction algorithm proposed in [11] is applied to remove the ground points, and residual obstacle points are clustered into many clusters with a seed-filling connected components labeling algorithm [44]. Moreover, we search a wide range of the following values and select those that yield the best estimation.
During learning the optimal parameters, we considered various optimization algorithms, such as the genetic algorithm [3], grid search method, etc. As in ref. [45], we determined the predefined values of each parameter based on the statistics of the training dataset. Subsequently, we performed an iterative optimization of individual parameters to achieve the best results. Specifically, the wide range for width threshold is w i d t h s , w i d t h l , which is determined as [1–2, 2–3] m, and we find the best value of the range to be [1.3, 2.3] meters. The value of N s y m is in the range of 10, 20, 30, and 40, and we set it at 20. For compared segment groups, the threshold values D s y m are set in the range of [0, 0.5]. Besides, the threshold value of the symmetrical segment number N s y m is in the range of [5, 20]. The value of δ is in the range of [1, 60] degrees.
To verify the effectiveness of the PE-HL-PF algorithm, we compare its estimation performance with state-of-the-art vehicle pose estimation methods such as PE-CPD [12], PE-CHM [32]. PE-CPD [12] proposes a vehicle detection model that uses coherent point drift. It consists of two parts: a dynamic vehicle detection component with an adaptive threshold, and a dynamic vehicle confirmation component that employs the scaling series algorithm. PE-CHM [32] fits the point cloud object into a polygon with the convex-hull algorithm based on the minimum distance and finds the pose of the vehicle with a fixed side model. We compare our estimated angle error and location error with the above methods. Furthermore, we also evaluate the components of our PE-HL-PF with different settings.

4.3. Experimental Results

In this subsection, we evaluate the performance of the proposed PE-HL-PF with the compared vehicle pose estimation methods in terms of estimated orientation error, position error, and running time. Furthermore, we design ablation studies to verify the effectiveness of different components.

4.3.1. Estimated Performance of PE-HL-PF

As for the evaluation of pose estimation, we first show the estimated results of orientation and position errors. We show the angle errors (degree) for ten selected datasets in Table 2. In addition to angle errors, we also compare the estimated position error (meter) of our PE-HL-PF with that of the state-of-the-art methods in Table 3.
As shown in Table 2, the performance of orientation estimation for PE-CHM [32] is not as good as that of our PE-HL-PF method. The results indicate that the orientation error estimated by PE-HL-PF is 1.41 degrees lower than that of PE-CHM. PE-CHM requires polygon fitting before orientation fitting and iterates over all polygon edges to fit the orientation. However, this approach can lead to the over-fitting of convex hulls due to irregular objects on the vehicle surface, such as mirrors, resulting in inaccurate orientation estimation. In contrast, our PE-HL-PF method uses RANSAC to fit the orientation using optimal length-aware contours, ensuring that the fitting of the orientation is not affected by the irregular object. In this manner, our method reaches a robust performance in the orientation fitting when dealing with the asymmetrical measurements.
Additionally, we also compare the PE-HL-PF with the PE-CPD [12], which involves knowledge of motion coherence. By comparing the mean of the orientation error in 10 selected sequences, we find that our method improves by 1.75 degrees. Figure 6 illustrates the distribution of orientation error among the state-of-the-art pose estimation methods in dataset 19. As shown in Figure 6a, most results for objects are distributed from 0 to 3 degrees with PE-HL-PF. Conversely, there is a higher proportion of results between 3 and 20 degrees for PE-CPD and PE-CHM. These findings suggest that the performance of PE-HL-PF is more consistent and stable when compared with the other methods.
Except for angle errors, we also evaluate our methods with a center location error, which evaluates the accuracy of the estimated bounding box. The performance of our proposed PE-HL-PF method is compared with PE-CHM and PE-CPD in Table 2. The results indicate an improvement of at least 0.14 m (0.52-0.38) compared with the other methods. That is due to the fact that our PE-HL-PF derives position from the estimated orientation, and the combination of the L-shape fitting module and grid-based particle filter fitting module significantly improves the orientation estimation performance on point clouds with diverse shapes.
Besides, Figure 7 shows the position error distribution of the PE-HL-PF, PE-CPD, and PE-CHM. It can be observed that the position errors of most objects are less than 0.15 m. Based on the aforementioned observations, it is evident that our proposed PE-HL-PF method is more precise and stable in vehicle pose estimation when compared with PE-CPD and PE-CHM.
Qualitative results of PE-HL-PF are shown in Figure 8. It contains six roadway scenarios in the KITTI Vision Benchmark Suite. As for Figure 8, the upper three scenarios are on the highways, and the lower parts are the scenarios of the city and suburb. In addition, each scenario consists of an RGB image and a LIDAR point cloud image, where the black rectangles are the fitted bounding boxes of the objects. As shown in Figure 8, even for some difficult shapes of clusters, such as clusters with side-view mirrors or symmetrical-like clusters, an accurate estimation result can be achieved using our method. We can conclude that our method achieves an encouraging estimated performance in different scenarios.

4.3.2. Estimated Result of Different Components of PE-HL-PF

To analyze individual components of our proposed method, such as the geometric shape classifier and the asymmetrical and symmetrical pose estimation module, we also conduct ablation experiments to estimate the pose of vehicles without the three components. As the orientation error is directly influenced by the above components and the estimated orientation line determines the oriented bounding box, we apply angle error to estimate the effectiveness of the various components. All models are evaluated on the selected KITTI tracking datasets.
The results of mean angle errors with different individual components are shown in Table 4. Setting one verifies PE-HL-PF with all components, and setting two estimates the objects’ pose only with our asymmetrical estimation module. Upon comparing the results of setting one and setting two, it is evident that setting one yields a 0.51-degree improvement (3.28–2.77). This outcome could be attributed to the utilization of the particle filter algorithm in setting one, which employs the low-level velocity of grids to fit the orientation of the sparse symmetrical clusters. This manner is more robust than the shape of point cloud clusters.
Setting three extracts the visible length-aware contours and fits all possible perpendicular lines with RANSAC for the asymmetrical-like clusters. The processing manner for the symmetrical-like cluster is the same as setting one. By comparing setting one and setting three, we can find that the strategy of dominant contour extraction boosts a 0.14-degree improvement. That may be because the dominant contour extraction can cope with the outliers, such as points on the side-view mirror, thus improving the accuracy of the estimated orientation. In setting four, we extract contours with a regular angular resolution for asymmetrical-like clusters. The result of setting four illustrates that more detailed geometric feature extraction can assist the orientation fitting.

4.3.3. Running Time Comparison

Table 5 presents a comparison of the mean running time (in milliseconds) for 100 point cloud objects in various scenarios. PE-HL-PF has improved the running efficiency by dozens of times compared with existing models, such as PE-CHM and PE-CPD. This improvement is primarily due to the fact that our algorithm directly fits the orientation line using optimal dominant length-aware contours, as opposed to fitting all possible perpendicular lines to the object boundaries. This leads to a significant reduction in computational load.
Based on the data presented in Table 5, we can conclude that our PE-HL-PF algorithm is capable of estimating 100 objects in approximately 1 millisecond. In conclusion, the PE-HL-PF algorithm has demonstrated real-time performance in the estimation of the pose of point cloud objects in real-world scenarios.

5. Conclusions

This paper introduces a novel pose estimation algorithm, the PE-HL-PF, which employs heuristic L-shape fitting and a particle filter algorithm for bounding box fitting. To improve accuracy, a classification module is employed to distinguish between symmetrical and asymmetrical point cloud clusters. The algorithm includes separate modules for estimating the pose of symmetrical and asymmetrical clusters, taking advantage of the differential treatment of the two types. Additionally, the proposed method selects optimal contours around the longer edge to fit the vehicle’s orientation, thereby reducing computational complexity. The results show that the proposed method achieves high estimation performance compared with existing approaches.
The results of extensive experiments on the KITTI dataset have demonstrated the effectiveness of our proposed PE-HL-PF algorithm, enabling us to accurately generate an oriented bounding box for vehicle pose estimation while meeting real-time constraints. In our future work, we aim to further validate the effectiveness of our algorithm on additional datasets and explore its applicability to a broader range of scenarios.

Author Contributions

Methodology, writing—review and editing, J.S.; supervision and writing—review, Y.-M.J. and S.-D.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No. 62176264), Natural Science Foundation of Jiangsu Province (20KJA520001), Open Research Project of Zhejiang Lab (No. 2021KF0AB05).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, Z.; Wang, L.; Xiao, L.; Dai, B. Unsupervised Subcategory Domain Adaptive Network for 3D Object Detection in LiDAR. Electronics 2021, 10, 927. [Google Scholar] [CrossRef]
  2. Li, C.; Gao, F.; Han, X.; Zhang, B. A New Density-Based Clustering Method Considering Spatial Distribution of Lidar Point Cloud for Object Detection of Autonomous Driving. Electronics 2021, 10, 2005. [Google Scholar] [CrossRef]
  3. Liu, W.; Hua, M.; Deng, Z.; Huang, Y.; Hu, C.; Song, S.; Gao, L.; Liu, C.; Xiong, L.; Xia, X. A Systematic Survey of Control Techniques and Applications: From Autonomous Vehicles to Connected and Automated Vehicles. arXiv 2023, arXiv:2303.05665. [Google Scholar]
  4. Zhang, Y.; Hu, Q.; Xu, G.; Ma, Y.; Wan, J.; Guo, Y. Not all points are equal: Learning highly efficient point-based detectors for 3d lidar point clouds. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 18953–18962. [Google Scholar]
  5. Shi, S.; Guo, C.; Jiang, L.; Wang, Z.; Shi, J.; Wang, X.; Li, H. Pv-rcnn: Point-voxel feature set abstraction for 3d object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10529–10538. [Google Scholar]
  6. Chen, T.; Dai, B.; Liu, D.; Fu, H.; Song, J. Likelihood-field-model-based dynamic vehicle detection with velodyne. In Proceedings of the International Conference on Intelligent Human-Machine Systems and Cybernetics, Hangzhou, China, 26–27 August 2015; Volume 2, pp. 497–502. [Google Scholar]
  7. Chen, T.; Wang, R.; Dai, B.; Liu, D.; Song, J. Likelihood-field-model-based dynamic vehicle detection and tracking for self-driving. IEEE Trans. Intell. Transp. Syst. 2016, 17, 3142–3158. [Google Scholar] [CrossRef]
  8. Oniga, F.; Nedevschi, S. A Fast Ransac Based Approach for Computing the Orientation of Obstacles in Traffic Scenes. In Proceedings of the International Conference on Intelligent Computer Communication and Processing, Cluj-Napoca, Romania, 6–8 September 2018; pp. 209–214. [Google Scholar]
  9. Qu, S.; Chen, G.; Ye, C.; Lu, F.; Wang, F.; Xu, Z.; Gel, Y. An efficient L-shape fitting method for vehicle pose detection with 2D lidar. In Proceedings of the IEEE International Conference on Robotics and Biomimetics, Kuala Lumpur, Malaysia, 12–15 December 2018; pp. 1159–1164. [Google Scholar]
  10. Deng, S.; Liang, Z.; Sun, L.; Jia, K. VISTA: Boosting 3D Object Detection via Dual Cross-VIew SpaTial Attention. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 8448–8457. [Google Scholar]
  11. Himmelsbach, M.; Hundelshausen, F.V.; Wuensche, H.J. Fast segmentation of 3D point clouds for ground vehicles. In Proceedings of the IEEE Intelligent Vehicles Symposium, La Jolla, CA, USA, 21–24 June 2010; pp. 560–565. [Google Scholar]
  12. Liu, K.; Wang, W.; Tharmarasa, R.; Wang, J. Dynamic vehicle detection with sparse point clouds based on PE-CPD. IEEE Trans. Intell. Transp. Syst. 2018, 20, 1964–1977. [Google Scholar] [CrossRef]
  13. Naujoks, B.; Wuensche, H.J. An Orientation Corrected Bounding Box Fit Based on the Convex Hull under Real Time Constraints. In Proceedings of the IEEE Intelligent Vehicles Symposium, Suzhou, China, 26–30 June 2018; pp. 1–6. [Google Scholar]
  14. Gao, L.; Xiong, L.; Xia, X.; Lu, Y.; Yu, Z. Vehicle localization with vehicle dynamics during gnss outages. In Advances in Dynamics of Vehicles on Roads and Tracks: Proceedings of the 26th Symposium of the International Association of Vehicle System Dynamics, IAVSD 2019, August 12-16, 2019, Gothenburg, Sweden; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 1071–1079. [Google Scholar]
  15. Xia, X.; Xiong, L.; Huang, Y.; Lu, Y.; Gao, L.; Xu, N.; Yu, Z. Estimation on IMU yaw misalignment by fusing information of automotive onboard sensors. Mech. Syst. Signal Process. 2022, 162, 107993. [Google Scholar] [CrossRef]
  16. Liu, W.; Xia, X.; Xiong, L.; Lu, Y.; Gao, L.; Yu, Z. Automated vehicle sideslip angle estimation considering signal measurement characteristic. IEEE Sens. J. 2021, 21, 21675–21687. [Google Scholar] [CrossRef]
  17. Liu, W.; Xiong, L.; Xia, X.; Lu, Y.; Gao, L.; Song, S. Vision-aided intelligent vehicle sideslip angle estimation based on a dynamic model. IET Intell. Transp. Syst. 2020, 14, 1183–1189. [Google Scholar] [CrossRef]
  18. Xia, X.; Hashemi, E.; Xiong, L.; Khajepour, A. Autonomous Vehicle Kinematics and Dynamics Synthesis for Sideslip Angle Estimation Based on Consensus Kalman Filter. IEEE Trans. Control Syst. Technol. 2022, 31, 179–192. [Google Scholar] [CrossRef]
  19. Shi, G.; Li, R.; Ma, C. Pillarnet: Real-time and high-performance pillar-based 3d object detection. In Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022, Proceedings, Part X; Springer Nature: Cham, Switzerland, 2022; pp. 35–52. [Google Scholar]
  20. Chen, C.; Chen, Z.; Zhang, J.; Tao, D. Sasa: Semantics-augmented set abstraction for point-based 3d object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 22 February–1 March 2022; Volume 1. [Google Scholar]
  21. Chen, Y.; Liu, J.; Zhang, X.; Qi, X.; Jia, J. VoxelNeXt: Fully Sparse VoxelNet for 3D Object Detection and Tracking. arXiv 2023, arXiv:2303.11301. [Google Scholar]
  22. Xu, Q.; Zhong, Y.; Neumann, U. Behind the curtain: Learning occluded shapes for 3D object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 22 February–1 March 2022; Volume 36, pp. 2893–2901. [Google Scholar]
  23. Wang, Z.; Jia, K. Frustum convnet: Sliding frustums to aggregate local point-wise features for amodal 3d object detection. arXiv 2019, arXiv:1903.01864. [Google Scholar]
  24. Liang, M.; Yang, B.; Wang, S.; Urtasun, R. Deep continuous fusion for multi-sensor 3d object detection. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 641–656. [Google Scholar]
  25. Yang, B.; Liang, M.; Urtasun, R. Hdnet: Exploiting hd maps for 3d object detection. In Proceedings of the Conference on Robot Learning, Zurich, Switzerland, 29–31 October 2018; pp. 146–155. [Google Scholar]
  26. Bai, X.; Hu, Z.; Zhu, X.; Huang, Q.; Chen, Y.; Fu, H.; Tai, C.L. Transfusion: Robust lidar-camera fusion for 3d object detection with transformers. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1090–1099. [Google Scholar]
  27. Li, Y.; Yu, A.W.; Meng, T.; Caine, B.; Ngiam, J.; Peng, D.; Shen, J.; Lu, Y.; Zhou, D.; Le, Q.V.; et al. Deepfusion: Lidar-camera deep fusion for multi-modal 3d object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17182–17191. [Google Scholar]
  28. Zhu, X.; Vondrick, C.; Ramanan, D.; Fowlkes, C.C. Do We Need More Training Data or Better Models for Object Detection? In Proceedings of the British Machine Vision Conference, Surrey, UK, 3–7 September 2012; Volume 3. [Google Scholar]
  29. Ding, N. An Efficient Convex Hull-Based Vehicle Pose Estimation Method for 3D LiDAR. arXiv 2023, arXiv:2302.01034. [Google Scholar]
  30. Börcs, A.; Nagy, B.; Benedek, C. Dynamic 3D environment perception and reconstruction using a mobile rotating multi-beam Lidar scanner. In Handling Uncertainty and Networked Structure in Robot Control; Springer: Berlin/Heidelberg, Germany, 2015; pp. 153–180. [Google Scholar]
  31. He, Y.; Zhang, W.; Yang, M. Pose Estimation of Moving Vehicles Based on Heuristic Rules for Autonomous Driving. In Proceedings of the International Conference on Robotics and Biomimetics, Xishuangbanna, China, 5–9 December 2022; pp. 729–734. [Google Scholar]
  32. Liu, K.; Wang, J. Fast dynamic vehicle detection in road scenarios based on pose estimation with convex-hull model. Sensors 2019, 19, 3136. [Google Scholar] [CrossRef] [PubMed]
  33. Xu, F.; Wang, Z.; Wang, H.; Lin, L.; Liang, H. Dynamic vehicle pose estimation and tracking based on motion feedback for LiDARs. Appl. Intell. 2023, 53, 2362–2390. [Google Scholar] [CrossRef]
  34. Shen, X.; Pendleton, S.; Ang, M.H. Efficient L-shape fitting of laser scanner data for vehicle pose estimation. In Proceedings of the International Conference on Cybernetics and Intelligent Systems and IEEE Conference on Robotics, Automation and Mechatronics, Siem Reap, Cambodia, 15–17 July 2015; pp. 173–178. [Google Scholar]
  35. Zhang, X.; Xu, W.; Dong, C.; Dolan, J.M. Efficient L-shape fitting for vehicle detection using laser scanners. In Proceedings of the IEEE Intelligent Vehicles Symposium, Los Angeles, CA, USA, 11–14 June 2017; pp. 54–59. [Google Scholar]
  36. Darms, M.; Rybski, P.; Urmson, C. Classification and tracking of dynamic objects with multiple sensors for autonomous driving in urban environments. In Proceedings of the IEEE Intelligent Vehicles Symposium, Suzhou, China, 26–30 June 2018; pp. 1197–1202. [Google Scholar]
  37. Himmelsbach, M.; Wuensche, H.J. Tracking and classification of arbitrary objects with bottom-up/top-down detection. In Proceedings of the IEEE Intelligent Vehicles Symposium, Madrid, Spain, 3–7 June 2012; pp. 577–582. [Google Scholar]
  38. Zhao, C.; Fu, C.; Dolan, J.M.; Wang, J. L-shape fitting-based vehicle pose estimation and tracking using 3d-lidar. IEEE Trans. Intell. Veh. 2021, 6, 787–798. [Google Scholar] [CrossRef]
  39. Yuan, C.; Yu, X.; Luo, Z. 3D point cloud matching based on principal component analysis and iterative closest point algorithm. In Proceedings of the International Conference on Audio, Language and Image Processing, Shanghai, China, 11–12 July 2016; pp. 404–408. [Google Scholar]
  40. Lalonde, J.F.; Vandapel, N.; Huber, D.F.; Hebert, M. Natural terrain classification using three-dimensional ladar data for ground robot mobility. J. Field Robot. 2006, 23, 839–861. [Google Scholar] [CrossRef]
  41. Ye, Y.; Fu, L.; Li, B. Object detection and tracking using multi-layer laser for autonomous urban driving. In Proceedings of the International Conference on Intelligent Transportation Systems, Rio de Janeiro, Brazil, 1–4 November 2016; pp. 259–264. [Google Scholar]
  42. Steyer, S.; Tanzmeister, G.; Wollherr, D. Grid-based environment estimation using evidential mapping and particle tracking. IEEE Trans. Intell. Veh. 2018, 3, 384–396. [Google Scholar] [CrossRef]
  43. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The kitti dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef]
  44. Hennequin, A.; Couturier, B.; Gligorov, V.V.; Lacassagne, L. SparseCCL: Connected components labeling and analysis for sparse images. In Proceedings of the 2019 Conference on Design and Architectures for Signal and Image Processing (DASIP), Montreal, QC, Canada, 16–18 October 2019; pp. 65–70. [Google Scholar]
  45. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
Figure 1. Illustration of point cloud map with different shapes. (a) displays the asymmetrical clusters, where the ground truth bounding box is denoted by the green rectangle, and the estimated bounding box is indicated by the red rectangle. (b) depicts the symmetrical clusters.
Figure 1. Illustration of point cloud map with different shapes. (a) displays the asymmetrical clusters, where the ground truth bounding box is denoted by the green rectangle, and the estimated bounding box is indicated by the red rectangle. (b) depicts the symmetrical clusters.
Electronics 12 01903 g001
Figure 2. Schematic illustration of PE-HL-PF. It mainly has three components: contour-based geometric shape classifier, Pose estimation model for Asymmetrical-like targets, and Pose estimation model for Symmetrical-like targets.
Figure 2. Schematic illustration of PE-HL-PF. It mainly has three components: contour-based geometric shape classifier, Pose estimation model for Asymmetrical-like targets, and Pose estimation model for Symmetrical-like targets.
Electronics 12 01903 g002
Figure 3. Illustration of the process for the geometric shape classifier. (a) shows the processing of endpoints extraction and width calculation, (b) shows the step of cluster classifying according to extracted contours, the solid blue points in each segment represent contours. (c) shows a real-world traffic scene on the road, there are various clusters on it, the cluster in the black rectangle represents a symmetrical-like cluster.
Figure 3. Illustration of the process for the geometric shape classifier. (a) shows the processing of endpoints extraction and width calculation, (b) shows the step of cluster classifying according to extracted contours, the solid blue points in each segment represent contours. (c) shows a real-world traffic scene on the road, there are various clusters on it, the cluster in the black rectangle represents a symmetrical-like cluster.
Electronics 12 01903 g003
Figure 4. Illustration of the process for HL. (a) denotes a raw point cloud cluster, (b) shows the points segmentation, (c) displays the length-aware contours extraction in each segment, (d) exploits optimal contours extraction, (e) shows orientation fitting and position inference.
Figure 4. Illustration of the process for HL. (a) denotes a raw point cloud cluster, (b) shows the points segmentation, (c) displays the length-aware contours extraction in each segment, (d) exploits optimal contours extraction, (e) shows orientation fitting and position inference.
Electronics 12 01903 g004
Figure 5. Schematic illustration for optimal contour extraction. The red dot represents the corner point, and the blue dots represent the contours in each segment, and the dashed lines divide the points on the target into different segments.
Figure 5. Schematic illustration for optimal contour extraction. The red dot represents the corner point, and the blue dots represent the contours in each segment, and the dashed lines divide the points on the target into different segments.
Electronics 12 01903 g005
Figure 6. Error distribution of three methods in the aspect of orientation at dataset 19. (a) shows the estimation result of our PE-HL-PF, (b) denotes the result of PE-CPD, and (c) shows the result of PE-CHM.
Figure 6. Error distribution of three methods in the aspect of orientation at dataset 19. (a) shows the estimation result of our PE-HL-PF, (b) denotes the result of PE-CPD, and (c) shows the result of PE-CHM.
Electronics 12 01903 g006
Figure 7. Error distribution of three methods in the aspect of position at dataset 19. (a) shows the estimation result of our PE-HL-PF, (b) denotes the result of PE-CPD, and (c) shows the result of PE-CHM.
Figure 7. Error distribution of three methods in the aspect of position at dataset 19. (a) shows the estimation result of our PE-HL-PF, (b) denotes the result of PE-CPD, and (c) shows the result of PE-CHM.
Electronics 12 01903 g007
Figure 8. Visualizations of PE-HL-PF results on KITTI set. (best viewed in color with zoom-in) (a) represents the result in the highway scenario, (b) denotes the result in the city and suburb scenario.
Figure 8. Visualizations of PE-HL-PF results on KITTI set. (best viewed in color with zoom-in) (a) represents the result in the highway scenario, (b) denotes the result in the city and suburb scenario.
Electronics 12 01903 g008
Table 1. Detail of the selected sub-datasets.
Table 1. Detail of the selected sub-datasets.
ID of Sub-DatasetNumber of FramesNumber of Dynamic Objects
0000153535
00022321142
0006269661
00077992488
00083891339
00098033135
0010293673
0013339124
001910591413
00208376259
Total517317,769
Table 2. Comparison of the mean orientation errors with ground truth object (deg).
Table 2. Comparison of the mean orientation errors with ground truth object (deg).
ID of Sub-DatasetPE-CPDPE-CHMPE-HL-PF
00003.874.751.83
00024.314.042.87
00064.013.652.49
00075.324.972.85
00085.214.033.42
00094.593.693.59
00104.833.872.88
00134.624.471.82
00193.974.023.10
00204.514.312.89
Average4.524.182.77
Table 3. Comparison of the mean position errors with ground truth object (meter).
Table 3. Comparison of the mean position errors with ground truth object (meter).
ID of Sub-DatasetPE-CHMPE-CPDPE-HL-PF
00000.630.550.25
00020.700.610.42
00060.430.520.37
00070.540.630.36
00080.560.470.51
00090.590.550.52
00100.550.510.40
00130.510.420.30
00190.430.440.32
00200.400.500.34
Average0.530.520.38
Table 4. Mean angle error with different individual components of PE-HL-PF (deg).
Table 4. Mean angle error with different individual components of PE-HL-PF (deg).
Classifier &
Symmetrical Estimation Module
Asymmetrical Estimation ModuleAngle Error
Length-Aware ContoursDominant Contours Extraction
12.77
2 3.28
3 2.91
4 3.01
Table 5. Comparison of the mean runtime for 100 objects (ms).
Table 5. Comparison of the mean runtime for 100 objects (ms).
ID of Sub-DatasetPE-CHMPE-CPDPE-HL-PF
000015.3166.011.12
000215.264.981.11
000613.767.011.01
00079.8565.320.97
00089.5664.320.66
000913.5663.800.93
001011.0566.340.78
001314.2166.211.31
001919.4567.011.39
002016.0365.311.22
Average13.7965.631.05
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sun, J.; Ji, Y.-M.; Liu, S.-D. Dynamic Vehicle Pose Estimation with Heuristic L-Shape Fitting and Grid-Based Particle Filter. Electronics 2023, 12, 1903. https://doi.org/10.3390/electronics12081903

AMA Style

Sun J, Ji Y-M, Liu S-D. Dynamic Vehicle Pose Estimation with Heuristic L-Shape Fitting and Grid-Based Particle Filter. Electronics. 2023; 12(8):1903. https://doi.org/10.3390/electronics12081903

Chicago/Turabian Style

Sun, Jing, Yi-Mu Ji, and Shang-Dong Liu. 2023. "Dynamic Vehicle Pose Estimation with Heuristic L-Shape Fitting and Grid-Based Particle Filter" Electronics 12, no. 8: 1903. https://doi.org/10.3390/electronics12081903

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop