Next Article in Journal
Modern Electrochemical Biosensing Based on Nucleic Acids and Carbon Nanomaterials
Previous Article in Journal
A Novel Approach for Simulation of Automotive Radar Sensors Designed for Systematic Support of Vehicle Development
Previous Article in Special Issue
Guided Depth Completion with Instance Segmentation Fusion in Autonomous Driving Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PTA-Det: Point Transformer Associating Point Cloud and Image for 3D Object Detection

1
School of Automation, Northwestern Polytechnical University, Xi’an 710129, China
2
Institute of Photonics & Photon Technology, Northwest University, Xi’an 710127, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(6), 3229; https://doi.org/10.3390/s23063229
Submission received: 7 February 2023 / Revised: 10 March 2023 / Accepted: 16 March 2023 / Published: 17 March 2023
(This article belongs to the Special Issue Robust Multimodal Sensing for Automated Driving Systems)

Abstract

:
In autonomous driving, 3D object detection based on multi-modal data has become an indispensable perceptual approach when facing complex environments around the vehicle. During multi-modal detection, LiDAR and a camera are simultaneously applied for capturing and modeling. However, due to the intrinsic discrepancies between the LiDAR point and camera image, the fusion of the data for object detection encounters a series of problems, which results in most multi-modal detection methods performing worse than LiDAR-only methods. In this investigation, we propose a method named PTA-Det to improve the performance of multi-modal detection. Accompanied by PTA-Det, a Pseudo Point Cloud Generation Network is proposed, which can represent the textural and semantic features of keypoints in the image by pseudo points. Thereafter, through a transformer-based Point Fusion Transition (PFT) module, the features of LiDAR points and pseudo points from an image can be deeply fused under a unified point-based form. The combination of these modules can overcome the main obstacle of cross-modal feature fusion and achieves a complementary and discriminative representation for proposal generation. Extensive experiments on KITTI dataset support the effectiveness of PTA-Det, achieving a mAP (mean average precision) of 77.88% on the car category with relatively few LiDAR input points.

1. Introduction

3D object detection is a fundamental perception task for autonomous driving, which usually takes point clouds or images as input. It aims to estimate the 3D bounding boxes of objects and recognize their categories. Due to the success of convolution neural network (CNN) in 2D object detection, image-based 3D object detection has emerged to obtain spatial clues about the objects. Although the image elaborates the front-view projection of objects in a 3D scene, accurate depth measurement is still required in order to localize 3D objects. In the last decade, 3D sensors such as LiDAR have developed rapidly from the survey [1]. With these devices, researchers can obtain point clouds that reflect the relative positions of the sensor and the obstacles for 3D object detection. Early works relying on LiDAR points (e.g., PointRCNN [2] and VoxelNet [3]) have achieved superior results over image-based detection methods. However, they suffer from the poor semantic information of point clouds, as shown in Figure 1a. In addition, the objects in Figure 1b are difficult to detect using LiDAR-only methods because the distant point cloud is extremely sparse. In contrast, an image is an array of pixels and can provide continuous textural structure, which is in favor of distinguishing false-positive instances. Intuitively, it is essential to design a multi-modal 3D object detection method to exploit both the geometry clues in the point cloud and the textural clues in the image.
As can be seen from the KITTI [4] leaderboard, there is still a gap between the mean Average Precision (mAP) of multi-modal methods and LiDAR-only methods. The first performance bottleneck encountered by multi-modal methods is the capability to extract intra-modal features. According to the related works [5,6,7,8,9,10,11,12,13,14,15,16,17,18], normally point cloud features are extracted via PointNet++ [19]/3D sparse convolution [20], while image features are extracted through 2D convolution [21]. However, the useful long-distance dependencies in each modality are difficult to capture due to the local receptive fields of these building blocks. To balance speed and accuracy, multi-modal methods usually reduce the number of point clouds and image size of the input. The approach will cause serious long-distance information loss and reduce the detection accuracy.
The second performance bottleneck of multi-modal methods is restricted by the fusion mode of inter-modal features. In order to utilize mature 2D object detectors and deep learning methods for point cloud, Frustum-PointNet [7] generated 2D proposals for images and lifted them into frustums. Then it predicted the bounding box from points in the extruded frustum. As proposed in the investigation [22], using multiple data separately does not take their complementarity into account. Later, researchers attempted to fuse point cloud and image features at the high-resolution feature level. AVOD [6] as a pioneering work was proposed to project point clouds into Bird’s eye view (BEV) and aggregate BEV features and image features in anchors to generate proposals. Owing to the quantized errors during projection, the result is affected by an inaccurate alignment between the two features. Some works turned to more fine-grained multi-modal feature fusion, namely feature fusion at the point level. PointPainting [10] performed image semantic segmentation and appends the semantic score of the projected position on the image to the LiDAR points. The enhanced features are sent to a point-based method for proposal generation. Although the mAP is relatively improved, the simple feature concatenation is not enough to effectively fuse image and point cloud features. In summary, the main reason limiting their accuracy is the problematic fusion of multi-modal data.
To conquer the first performance bottleneck, we strive to learn more expressive point cloud features and image features. To extract point cloud features, the transformer [23] proposed in machine translation is used to construct the feature extraction module. Recent studies [24,25] have shown that transformer is capable of point cloud classification and segmentation. Compared with CNNs, transformer is developed based on a self-attention mechanism and can capture more distant dependencies. Therefore, relying on self-attention, Point Transition Down (PTD) and Point Transition Up (PTU) modules are designed to extract point cloud features. In contrast, the extraction of image features in multi-modal detection is still an open problem. Inspired by the application of pseudo point cloud (PPC) in 3D object detection by MDPPC [26], a Pseudo Point Cloud Generation network that converts image keypoints into PPCs is developed, and then the PPCs are used to acquire the features of image keypoints from a novel feature frustum. Attributed to the unified form of PPC and LiDAR points, both image features and point cloud features can be learned in the form of points.
To address the second performance bottleneck, a two-stream feature extraction network, based solely on transformer, is developed to solve the fusion of inter-modal features. Specifically, the two-stream structure consists of a point cloud branch and a PPC branch. The two branches independently learn high-level point cloud features and PPC features. In particular, the image features of the object keypoints are used as the initial PPC features, and these features are further analyzed and encoded in the PPC branch. Benefiting from the unified feature representation, a Point Fusion Transition (PFT) module is developed, which accurately fuses the two features at the point level to highlight the key cues across the two modalities.
In general, we present a multi-modal 3D object detection method PTA-Det, constructed on the basis of point-level feature learning and fusion. Accompanied by a series of modules, the mAP of multi-modal 3D object detection can be improved with better robustness and accuracy. Our main contributions are summarized as follows:
  • The PPC generated by a Pseudo Point Cloud Generation network, a point-based representation of image feature, has been leveraged for multi-modal 3D object detection.
  • A two-stream feature extraction network entirely relying on transformer has been developed, to learn intra-modal features and inter-modal features at the point level.
  • Competitive results on KITTI dataset have been achieved. Results demonstrate that our model is compatible with most LiDAR-only detectors and can be easily upgraded to a multi-modal detector.

2. Related Work

PPC-based 3D Object Detection. From the investigation by Reading et al. [27], image-based detection methods show unsatisfactory results owing to lacking direct depth measurement. However, Wang et al. [28] argued that image-based methods are mainly affected by the representation of the data rather than its quality. They converted the depth image to the PPCs and applied a point-based method to detect objects. Pseudo-LiDAR [26] performed depth estimation and proposal prediction on image. For each proposal, a point cloud frustum is extracted from the PPCs obtained from the depth image transformation. Nevertheless, both of these methods ignore the depth error introduced into the PPC, which further affects their detection accuracy. To solve the problem, Pseudo-LiDAR++ [29] utilized extremely sparse LiDAR points to correct their nearby PPCs to achieve accurate depth prediction. Wang et al. [30] proposed a PnP module that integrated sparse depth values into an intermediate feature layer to correct depth prediction. In addition to depth correction, End-to-End Pseudo-LiDAR [31] jointly trained depth prediction and object detection for accurate proposals. Hence, in our proposed Pseudo Point Cloud Generation network, we not only dynamically generate PPCs using depth prediction, but also apply corresponding image features for the subsequent detection pipeline.
Multi-modal based 3D Object Detection. According to the different fusion strategies, the existing multi-modal detection methods can be divided into three categories: result-level, proposal-level and point-level methods. In the result-level methods [5,7,11,12,17], it is common to utilize the feature of one modality to generate the proposal, and utilize the feature of the other modality in the proposal to generate the bounding box. These methods have high recall even when the object is far or occluded, but their accuracy is limited by ignoring the complementarity between different data. The proposal-level methods [6,9,16,32,33] take the encoded features of image and point cloud as inputs, and fuse the two features in anchors to generate proposals. These methods benefit from multi-modal data and can generate high-quality proposals. However, their performance is affected by irrelevant information mixed in the anchors and inaccurate feature alignment. The point-level methods [10,13,14,15,34] have shown promising results. ImVoteNet [14] fused 2D votes of image and point cloud features in a point-wise concatenation manner, but the approach is insufficient to fuse the two features. To address the drawback, EPNet [13] proposed the LI-Fusion module that adaptively fuses point cloud features with image features according to the importance of the image feature channel. EPNet++ [15] proposed the CB-Fusion module that added the fusion direction from the point domain to the image domain. It showed that bi-direction interaction approach leads to a more comprehensive and discriminative feature representation. Recently, several transformer-based multi-modal detection methods [34,35] have been proposed. Dosovitskiy et al. [36] demonstrated that transformers have comparable expressive power in CNN. Due to the powerful feature aggregation ability of transformer, we aim to design a two-stream feature extraction backbone based solely on its attention mechanism. Unlike existing multi-modal backbones, the proposed backbone can handle both point clouds and pseudo points and capture the useful patterns about objects in each modality. A PFT module, as a submodule of the backbone, is also proposed to perform feature interactions at multiple scales to supplement the missing object information in each modality.

3. Method

In this research, we present a multi-modal 3D object detection method named PTA-Det. As shown in Figure 2, the PTA-Det mainly consisted of a Pseudo Point Cloud Generation Network, a Two-stream Feature Extraction Network, a 3D Region Proposal Network (RPN), and a 3D Box Refinement Network. The Pseudo Point Cloud Generation Network comprised a Frustum Feature module and a Frustum-to-point Transition module. The Two-stream Feature Extraction Network included a point cloud branch and a PPC branch. The former contained a stacked PTD encoder and a stacked PTU decoder, and the latter contained a stacked pseudo PTD (PPTD) encoder and a stacked FP decoder. Meanwhile, the stacked PFT module was used to connect the two branches at multiple levels. In the following, the four subnetworks utilized in the investigation were elucidated in sequence.

3.1. Pseudo Point Cloud Generation Network

In this network, image was transformed into PPCs which were further utilized to represent image features. During processing, the image depth was predicted in a semi-supervision manner, and LiDAR points were projected onto the image to obtain sparse depth labels, which were used to supervise depth prediction. With the help of the foreground mask from Mask-RCNN [37] and the predicted depth image, the foreground pixels can be converted into pseudo points. At the same time, adhering to CaDDN [27], the Frustum Feature Module was used to construct the frustum feature. Then, the PPC features were obtained by interpolating the frustum feature in the Frustum-to-point Transition module.
Frustum Feature Module. In order to make full use of the image information, a Frustum Feature module was constructed to generate frustum feature. In Figure 3, extracting image features and predicting image depth were two fundamental steps. Similar to CaDDN [27], ResNet-101 [38] was utilized as the backbone to process images and the output of its Block1 was used to collect image features F I R H F × W F × C , where H F , W F were the height and width of the image feature, and C was the number of feature channels.
On the other hand, a depth prediction head was applied to the output of the image backbone to predict image depth. The depth prediction was viewed as a bin-based classification problem and the depth range was discretized into D bins by the discretization strategy LID [39]. Then the depth distribution D b i n R H F × W F × D and depth residual D r e s R H F × W F × 1 can be obtained.
Early depth estimators [27,39,40] computed the loss over the entire image including a large number of background pixels. These methods placed over-emphasis on background regions in depth prediction. According to Qian et al. [31], background pixels can occupy about 90% of all pixels in the KITTI dataset. Therefore, instead of calculating the loss of all image pixels, the off-the-shelf image segmentation network Mask-RCNN [37] was employed to select N foreground points from LiDAR points by distinguishing their 2D projection positions. The N points were re-projected onto the image to acquire sparse depth label for calculating the depth loss of foreground pixels. In addition, the foreground loss will be given more weight to balance the contributions of foreground and background pixels.
With the image feature and image depth, the frustum feature F T R H F × W F × D × C can be constructed as follows
F T = S M ( D b i n ) F I
where ⊗ was the outer product and S M represented the S o f t M a x function. Equation (1) stated that at each image pixel, the image features were weighted by the depth distribution values along the depth axis. CNN was known to extract image features in convolutional kernels, where object pixels may be surrounded by the pixels of the background or other objects. In contrast, the frustum feature network lifted image features onto depth bins along the depth axis, which enabled the model to discriminate misaligned features in 3D space.
Frustum-to-point Transition Module. The submodule aims to extract the PPC features from frustum feature. There are two issues to be addressed regarding the choice of PPC. First, due to the presence of depth errors, the PPCs converted from image may not be consistent with the distribution of the object in space. Second, the number of PPC is proportional to the image resolution, and the number is generally large. Nevertheless, only in the area where the point cloud is relatively sparse can PPC play an important role by compensating for the missing object information.
For the first issue, we applied the farthest point sampling (FPS) algorithm to select M of the previous N foreground points as the initial PPCs in Figure 4. Foreground points are used because they have more accurate depth values near the projected positions in the image. Accordingly, the projected coordinates { c i = ( u i , v i ) | i = 1 , , M } of M foreground points can be obtained via calibration matrix.
As for the second issue, the object keypoints that focus on more representative object parts are introduced as the final PPCs. Keypoints are defined as locations that reflect the local geometry of an object, such as points on mirrors and wheels. To determine the locations of keypoints in 3D space, inspired by Deformable Convolutional Networks [41], a 2D keypoint offset was predicted which represented the offset of each pixel on the image to its nearest keypoint. For the M projected coordinates, M keypoint offsets were acquired as follows
O I ( c i ) = ( Δ u i , Δ v i ) = q G ( c i , q ) O I ( q )
where q enumerated the nearby integral locations of c i on the image and G ( · , · ) was the bilinear interpolation kernel. Keypoint offset O I R H F × W F × 2 was predicted when generating image features illustrated in Figure 3.
Then, the locations of the 2D keypoints can be obtained as { c i = ( u i + Δ u i , v i + Δ v i ) | i = 1 , , M } by moving the M pixels according to the corresponding keypoint offsets. With the depth value d e p t h ( c i ) of the updated positions, the final PPCs can be determined in camera space. As shown in Figure 4, the features of the PPCs F p s e u R M × C can be extracted from the frustum feature F T using the trilinear interpolation. Subsequently, in order to process the PPCs features and LiDAR points features simultaneously, the PPC p i = ( u i + Δ u i , v i + Δ v i , d e p t h ( c i ) ) was re-projected to LiDAR space from the camera space by the transformation function f r e p r o j in KITTI
p i p s e u = ( x , y , z ) = f r e p r o j ( p i ) = T L i D A R r e f e r · T r e f e r c a m e r a · p i
where p i p s e u was the final coordinate of the ith PPC, T r e f e r c a m e r a was the transformation matrix from the coordinate of color camera to the reference camera, and T L i D A R r e f e r was the transformation matrix from the reference camera to LiDAR. To verify the effectiveness of Frustum-to-point Transition module, an alternative directly using M initial foreground points as PPCs and extracting their features in the same way was provided. In Section 4, the comparison between two strategies on the KITTI dataset will be presented.
Overall, the multi-modal detection task is transformed into single-modal detection by using PPC instead of image to convey object information. The unified point-based representation helps to make subsequent interactions across multi-modal features easier.

3.2. Two-Stream Feature Extraction Network

Multiple multi-modal methods [10,13,15] used a two-stream structure to process image and point cloud separately. Limited by the local receptive field of traditional building blocks, e.g., CNN and sparse convolution, these methods struggled to capture all useful feature relationships. In addition, feature alignment and fusion between image and point cloud were still tricky problems.
Based on the unified point-based representation described above, a two-stream feature extraction network was developed to learn the features of point cloud and image at the point level. The two-stream network was mainly built on a transformer for better feature learning and fusion. It had the inputs of the coordinates of point clouds P r a w R N × 3 and the coordinates of PPCs P p s e u R M × 3 , and the corresponding features were F r a w R N × C 1 and F p s e u R M × C 2 . Here, the feature of raw point p j was represented as f j r a w = v j , r j , g j , b j , r j R C 1 , where v j was a one-hot class vector indicating the confidence score of specific class, r j , g j , b j was the normalized RGB pixel-values of projected location of p j , and r j was the reflectance. The feature channel C 2 of pseudo point p i was the same as the image feature channel C.
Point Transition Down. In the two-stream network, a stacked PTD encoder was responsible for iteratively extracting multilevel point-based representations. Based on recent attempts [24,25] at object classification, PTD integrated the feature sampling and grouping, self-attention feature extraction and forward-feedback network into a independent module. In Figure 5, PTD first subsampled M points from the input point P I (here P r a w or P p s e u can act as P I ) and use k-NN algorithm to construct a neighbor embedding for each point. Then, an LBR (Linear layer, BatchNorm layer and ReLU function) operator and a max-pooling operator (MP) were used to encode local features as follows
f l o c a l ( p ) = MP ( LBR ( c o n c a t q k n n ( p , P I ) · f I ( q ) ) )
where f I ( q ) was the feature of point q which belonged to the neighbor of point p, k n n ( p , p I ) was k-nearest neighbors of point p in P I .
Next, we sent the local feature F l o c a l R M × C O into self-attention feature extraction network to learn long-range dependencies of the features. The relationship between the query (Q), key (K), value (V) matrices and self-attention was as follows
( Q , K , V ) = L B R ( F l o c a l ) · W e Q , K , V R M × C O , W e R C O × 3 C O Q = r g ( Q ) , K = r g ( K ) , V = r g ( V ) A = S M ( α ( Q K + δ ) )
where W e was the learnable weights of the linear layer and r g ( · ) represented repeat and grouping operation. Q , K , V R M × L × C O were the outputs after repeat and grouping operation related to the input Q, K, and V. Furthermore, a position encoding defined as δ = θ ( p i p j ) was added to the attention, where p i , p j were the coordinates of points i and j. θ and α both consisted of two linear layers and a ReLU function. Thereafter, the output of PTD could be derived as
F O = β ( s u m ( A · ( V + δ ) ) ) + F l o c a l
where s u m ( · ) represented the element-wise product, + denoted channel-wise summation along the neighborhood axis, and β was an LBR operator.
In the point cloud branch, the stacked PTD encoder (including four PTD modules) was used to learn point cloud features. In the PPC branch, the PPTD encoder adopted the same structure to extract image features.
Point Transition Up. In the point cloud branch, the stacked PTU decoder aimed to restore the point cloud to its initial number and obtained the multi-scale features for proposal generation. PTU can be easily constructed by replacing the feature sampling and grouping in PTD with the inverse distance-weighted average operation while keeping the other structures intact. The inverse distance-weighted average operation was proposed as the skip connection in PointNet++ [19]
f i n t ( p i ) = j = 1 k w j · f e ( p j ) j = 1 k w j
where w j = 1 d ( p i , p j ) p , j = 1 , , k , p i was the coordinate of the interpolated point, p j was the coordinate of the neighboring point of p i , d ( · , · ) denoted the Euclidean distance between two points, and f i n t ( p i ) denoted the interpolated features of p i . Let p = 2 , k = 3 be the same settings in PointNet++ [19], then, the interpolated features were added by the skip connection features as
F I n = F i n t n + LBR ( F s k i p n ) , n = 1 , 2 , 3 , 4
where F s k i p n was the n-th output of the PTD and F i n t n was the interpolated feature of the n-th PTU. F I n was used as the input of the remaining structure of the n-th PTU. On the contrary, in the PPC branch, a stacked FP decoder with four FP layers was used to recover the initial PPCs. Since the position of the PPC was defined on the object keypoint, the distribution of the PPC was more focused on the object surface than the point cloud directly sampled from the LiDAR. Meanwhile, considering the large memory and time overhead of PTU itself, the FP layer was selected to handle the PPC that did not require a large receptive field.
Point Fusion Transition. According to the above introduction, the stacked PTD encoders of the two branches simultaneously extracted point-based features layer-by-layer. However, the features from the point cloud branch lacked the semantic and textural information about the object, and the features from the PPC branch lacked the geometry information for locating the object. Moreover, both the point cloud provided by LiDAR and the image-generated PPC were inevitably contaminated by noise. To address these problems, a dual input and dual output PFT module was designed for feature fusion in Figure 6. PFT fused two input features based on cross-attention and produced two enhanced features as the inputs to the next level. Finally, an additional PFT was used to fuse the outputs of the two branches (see Figure 2) to obtain the final point representations.
PFT module was also based on transformer and the Q, K, and V matrices were generated separately for the two inputs
( Q r a w , K r a w , V r a w ) = F r a w · W r a w ( Q p s e u , K p s e u , V p s e u ) = F p s e u · W p s e u Q r a w , K r a w , V r a w R N × C e Q p s e u , K p s e u , V p s e u R M × C e W r a w R C 1 × 3 C e , W p s e u R C 2 × 3 C e
where W r a w and W p s e u were both learnable weights. Then, the cross-attention for each data is defined as
A r a w = S M ( σ ( K r a w · Q p s e u T ) ) A p s e u = S M ( ε ( K p s e u · Q r a w T ) ) A r a w R N × M , A p s e u R M × N
where σ and ε both comprised two linear layers and a ReLU function. Here, we multiplied the K matrix of one modality by the Q matrix of the other modality to generate cross-attention. It differed from the way computed in PTD. This practice was inspired by HVPR [42] which took voxel-based features as queries and computed matching probabilities between the voxel-based features and the memory items through dot product. In Section 4, we conducted ablation experiments to compare the effects of different attention calculation ways. Finally, the enhanced features as the outputs of PFT can be expressed as
F r a w e n h = LBR ( F r a w ( A r a w · V p s e u ) ) F p s e u e n h = LBR ( F p s e u ( A p s e u · V r a w ) )
It was worth mentioning that Zhang et al. [34] proposed a similar structure to PFT. However, they had the limitations that the information can only flow from the image domain to the point domain. In contrast, PFT conducted bidirectional information exchange which provided semantic information for point cloud and geometry information for PPC.

3.3. RPN and Refinement Network

The two-stream feature extraction network described in Section 3.2 aimed to learn expressive features for every LiDAR point. After that, the features will be sent to the RPN to generate proposals. To obtain high-quality proposals, 3D votes were computed as suggested by ImVoteNet [14], since 3D votes can help narrow down the search space from point to proposal center. The votes were then concatenated with the output of the two-stream network. Finally, the enhanced features were fed into the RPN that included a classification head and a regression head. After acquiring the proposals, non-maximum suppression (NMS) was applied to eliminate redundant proposals. The remaining proposals were sent to the refinement network for generating bounding boxes. In the experiments, two refinement strategies, including point cloud region pooling and RoI-aware point cloud feature pooling were adopted, as proposed by PointRCNN [2] and Part-A2 [43], respectively. Actually, our PTA-Det can be plugged into most point-based detectors as multi-modal detectors.

3.4. Overall Loss Function

The model is optimized by a multi-task loss which can be formulated as
L t o t a l = λ d e p t h L d e p t h + λ r p n L r p n + λ r c n n L r c n n
where the L d e p t h denotes the loss of depth prediction for generating the PPCs in Pseudo Point Cloud Generation network. L r p n is the loss of the two-stream feature extraction network to generate the proposal. L r c n n is the loss of the refinement network. λ d e p t h , λ r p n , and λ r c n n are fixed loss weighting factors. L d e p t h can be computed as
L d e p t h = L b i n + λ 1 L r e s
where λ 1 is the balance weight for depth residual with the setting of λ 1 = 10 . L b i n and L r e s are defined as
L b i n = 1 N i = 1 N FL ( D b i n ( u i , v i ) , D g t _ b i n i )
L r e s = 1 N i = 1 N SML ( D r e s ( u i , v i ) , D g t _ r e s i )
where FL denotes focal loss [44] and SML is S m o o t h L 1 loss. D g t _ b i n i and D g t _ r e s i denote bin-based index and normalized residual value of the ith foreground point’s depth. D b i n ( u i , v i ) and D r e s ( u i , v i ) have been introduced in Section 3.1, and the focal loss is adopted in L b i n with the setting of α = 0.25 and λ = 2.0 . L r p n consists of a classification loss and a regression loss as
L r p n = L c l s + λ 2 L r e g
with
L c l s = α t ( 1 c t ) γ l o g ( c t )
and
c t = c i f p i s t h e f o r e g r o u n d p o i n t 1 c o t h e r w i s e
L r e g = u ( x , y , z , l , w , h , s i n θ , c o s θ ) SML ( r e s ^ u , r e s u )
where λ 2 is the balance weight, c denotes the classification confidence for the point p, and L c l s is supervised by focal loss. r e s ^ u and r e s u are the predicted residuals and residual labels of the foreground point. S m o o t h - L 1 loss is used to regress the offsets of the location, size, and direction. The loss of the refinement network is the same as that of PointRCNN [2] or Part- A 2 [43].

4. Experiment

The model was evaluated on KITTI, a commonly used benchmark dataset for 3D object detection. PTA-Det was built on the basis of the OpenPCDet [45] which was an open-source project for 3D object detection.

4.1. Dataset and Evaluation Metric

KITTI Dataset. The KITTI dataset consists of 7481 training samples and 7518 test samples, focusing on the categories of car, pedestrian and cyclist. Following the investigations [5,7], the original training samples are further separated into a training set (3712 frames) and a validation set (3769 frames). The Average Precision (AP) is calculated using 40 recall positions as the validation metric according to Geiger et al. [4]. All the objects are classified into easy, moderate, and hard levels based on their sizes, occlusion, and truncation. In the experiments, the results on the validation set are reported for all difficulty levels.
nuScenes Dataset. The nuScenes dataset is a multi-modal dataset for 3D object detection. It includes 1000 scenes that is composed of 700 scenes for training and 150 scenes for validation. The remaining scenes are used for testing. For each scene, it captures about 20 s of video material with 6 cameras, 1 LiDAR and 5 radars, and is annotated with 3D bounding boxes every 10 frames. According to official evaluation metrics for 3D detection, mean Average Precision (mAP) and nuScenes detection score (NDS) are reported by our model.

4.2. Implementation Details

Network settings. As a multi-modal 3D object detection method, LiDAR points, RGB image, and camera calibration matrices were taken as inputs. We assumed that the 3D scene was constrained to [(0,70.4), (−40,40), (−3,1)] meters along the X (forward), Y (left), and Z (up) axes in the LiDAR coordinate, respectively. During depth prediction, the depth range was discretized into 80 bins. Unlike LiDAR coordinate, the camera coordinate was set along the X (left), Y (down), and Z (forward) axes. The transformation between two coordinates can be achieved by a calibration matrix. For experiments on nuScenes, we set the 3D scene to [(−51.2, 51.2), (−51.2, 51.2), (−5,3)] meters along the X, Y and Z axes.
For each 3D scene, 16,000 LiDAR points and image with a resolution of 1280 × 384 were used as the initial inputs to the model. In the Pseudo Point Cloud Generation Network, Mask-RCNN implemented by detectron2 [46] was used to generate the foreground mask, and 4096 foreground points were selected through the mask to guide the depth prediction. In a scene where the number of foreground points was less than 4096, the remaining points were randomly selected from the background points. Then, 1600 points were further sampled from the foreground points as the input to the point cloud branch, where the stacked PTD encoder had the point numbers set to 800, 400, 200, 100, respectively. By contrast, the Pseudo Point Cloud Generation Network produced 480 PPCs for the PPC branch, and the PPTD encoder iteratively extracted the features of PPCs, whose numbers were 240, 120, 60, and 30, respectively.
Training scheme. After generating proposals in RPN, redundant proposals were eliminated using NMS. The thresholds were set to 0.8 and 0.85 in the training and testing stages. In the refinement network, we utilized the IoU between the proposal and the ground truth to distinguish between positive and negative proposals. Following PointRCNN [2], different thresholds were selected for classification and regression. Specifically, the proposals with IoU scores higher than 0.6 were considered positive samples for classification. In contrast, the proposals with IoU scores lower than 0.45 were considered negative samples. The proposals with IoU scores higher than 0.55 were used to calculate the regression loss.
We trained the model with a batch size of 2 for 80 epochs and adopted the Adaptive Moment Estimator (Adam) optimizer with an initial learning rate, weight decay, and momentum at 0.01, 0.01, and 0.9, respectively. All experiments were conducted on two RTX 3090 GPUs using the deep learning framework PyTorch [47]. It is worth noting we did not use any data augmentation techniques during training.

4.3. Main Results

Results on KITTI dataset. PTA-Det was compared with several LiDAR-only and multi-modal 3D object detection methods, and the results were summarized in Table 1. Two versions of PTA-Det were given, one of which adopted the point cloud region pooling strategy and the other adopted the RoI-aware point cloud pooling strategy. The mAP of the former was 1.54% and 0.72% higher than that of the latter for the car and pedestrian categories, respectively. However, in the cyclist category, the latter outperformed the former by 1.06%. The results showed that the refinement strategy employed by the former was sufficient to accurately localize the object when image features were used as a complement. Although the latter can better capture the point cloud distribution in the proposals, its advantage was not obvious in our model.
Table 1 showed that PTA-Det exhibited better performance than a variety of previous multimodal methods by about 2% to 14% mAP in the car category. However, the current PTA-Det showed less mAP in the pedestrian and cyclist categories. The reason was that, considering the large memory overhead in the transformer, the number of point clouds input to the model was reduced. This made the points on the surface of small objects more sparse, which in turn led to poor detection performance of the model on small objects.
To reveal the reason for the degradation, we studied the performance of PointRCNN [2] under different sampling strategies, as summarized in Table 2. PointRCNN [2] abstracted a set of point representations using iterative set-abstraction (SA) blocks. In the default case, it sampled 16,384 points from the scene as input, and used four SA layers in sequence to sample the points with group sizes of 4096, 1024, 256, and 64, respectively. Three different sampling strategies were also presented, the last of which used the same number of points as PTA-Det.
As shown in Table 2, the mAP of PointRCNN declined significantly in all categories as the number of input points decreased. Comparing the fourth strategy with PTA-Det-1, it was proved that PTA-Det can reach a higher accuracy than PointRCNN with the same number of points. These investigations supported our conjecture above on the reason for the poor performance of PTA-Det when detecting small objects. Therefore, although Table 1 showed that PTA-Det performed worse than the LiDAR-only method, we believe that our model will achieve more competitive performance as long as a reasonable memory reduction strategy is developed to reduce the memory overhead while increasing the number of point clouds.
In order to further verify the effectiveness of PTA-Det, we provided PointRCNN and Part- A 2 with the same foreground point input as PTA-Det through the Pseudo Point Cloud Generation Network introduced in Section 3.1, and then compared their detection performances again. The results in Table 3 showed that PTA-Det outperformed the two methods in BEV detection and 3D detection of the car category. In addition, to illustrate the superiority of using pseudo point clouds to represent image features, we have added multiple multimodal detection methods for further comparison, which used the same data processing approach as the other methods. The results in Table 3 showed that PTA-Det still showed competitive results in 3D Detection of car category under the easy and moderate level. At the same time, we also noticed that the performance of PTA-Det was slightly worse than that of the latest multimodal detectors under the BEV Detection and the difficulty level of 3D Detection. The possible reason for this phenomenon was that we only used the feature information near the keypoints of the image to interact with the point cloud features. Compared with other methods extracting the image features in the whole image or 2D ROIs, we may lose some key object information. Then, we showed the comparisons of the detection performance in multiple scene instances in Figure 7 to further prove the advantage of our model in car category. Finally, we made several qualitative investigations to illustrate the effectiveness of PTA-Det on KITTI test set, as visualized in Figure 8.
Results on nuScenes dataset. We conducted experiments on the popular nuScenes dataset for 3D object detection to further validate the effectiveness of PTA-Det. We presented detection results of two versions of PTA-Det on the nuScenes validation set. From Table 4, PTA-Det-1 obtained 59.45 mAP and 65.23 NDS, and PTA-Det-2 obtained 57.32 mAP and 63.57 NDS. We also reported detailed detection results for all car-related categories for both methods. Except for the construction vehicle category, PTA-Det-1 surpassed the performance of PTA-Det-2 in all other metrics. Meanwhile, PTA-Det-1 outperformed the previous best baseline CenterPoint [52] by +3.84 mAP and +0.51 NDS, and our model also outperformed many other single- or multi-modal detection methods with fewer input point clouds.

4.4. Ablation Studies

We conducted the ablation studies on KITTI dataset to evaluate the influence of each module or strategy on the final results, including PTD, PTU, PFT, two selection strategies for PPC, and the calculation way of attention in PFT. For comparison, a baseline two-stream structure was designed to replace the two-stream network of PTA-Det. It also contained a point cloud branch and a PPC branch, both of which were built with PointNet++ block and FP layer. In addition, the interaction between the two branches was achieved through simple feature concatenation at multiple levels. Other network structures and parameter settings remained unchanged in the baseline model. Both PTA-Det and the baseline selected the RoI-aware point cloud feature pooling strategy in the refinement network. To trade off the speed and accuracy of PTA-Det, we chose ResNet-50 as the image backbone and followed the research of Pan et al. [54] to use the computational cost reduction strategy in PFT.
Effects of PTD and PTU modules. From the first three rows in Table 5, the baseline obtained a 43.25% mAP. After replacing all PointNet++ blocks with PTD, the mAP was improved by 5.59%. If all FP layers in the point cloud branch were replaced with PTU, the mAP was improved by 11.13%. When we used the two strategies together in the 5th row, the mAP was improved by 15.52% relative to that of the baseline. The improvements were attributed to the self-attention mechanism, which can aggregate long-distance dependencies better than Pointnet++ [19] blocks and FP layers [19].
Effects of PFT and Fusion Operation in PFT. Table 5 also showed that after introducing PFT into the baseline network, although the mAP of the pedestrian category was improved by 0.24%, the performance of the other categories became worse. This was because it was difficult for PFT to directly compute the attention of multi-scale features from SA blocks in Pointnet++. However, as can be found in the 5th and 6th rows, PFT can clearly improve the mAP if accompanied by PTD and PTU. A total improvement up to 16.88% had been realized when PTD, PTU, and PFT were used simultaneously.
To analyze the impact of the structure of the PFT on the results, three schemes had been studied in Table 6. In PFT, by default, the cross-modal features of each modality were subtracted from its input features, and attached it with an LBR as the output. This module was denoted as PFT . We then replaced subtraction with summation and concatenation, denoted as PFT + and PFT c , respectively. As in Table 6, PFT exhibited better performance than the two schemes. The improvements were +3.13% and +1.05% on mAP, showing the advantage of subtraction in PFT.
Influence of the Calculation Way of Attention in PTD, PTU and PFT. In PTD and PTU, the subtraction between the query matrix and the key matrix was used to compute the self-attention of the point features, while the multiplication between the two matrices was used in PFT to compute the cross-modal attention. In order to investigate the impact of the two attention computation ways on mAP, we combined the PTD, PTU, and PFT modules using subtraction or multiplication operation to form four different schemes for comparison. Table 7 showed that the fourth way had 4.69% to 7.68% higher mAP than the other three ways. The subtraction between point-based features helped capture the relationship between different features in the same dimension, since it provided the attention computation for each channel. This was crucial for PTD and PTU to obtain intra-modal features. In contrast, inter-modal features varied greatly and a larger perspective was required to capture their relationships. Multiplication can produce a scalar to measure the distance between features across channels. Thus, multiplication was more suitable for computing cross-attention than subtraction for PFT.
Effect of Sampling Strategy in Pseudo Point Cloud Generation Network. In Section 3.1, we mentioned two strategies to obtain PPCs: (1) apply FPS algorithm to sample the foreground points and the sampled points were used as the final PPCs, which was denoted as FPS; (2) apply the keypoint sampling strategy based on 2D keypoint offset, which was denoted as KPS. Table 8 showed that KPS performed better. The results showed that using the object keypoints as PPCs can provide more information about the object than PPCs directly sampled from the foreground points.

5. Conclusions

In this paper, a method named PTA-Det is proposed, which uses pseudo points as an intermediate modality between the image and the point cloud to solve the multi-modal 3D object detection problem. The pseudo points generated by a Pseudo Point Cloud Generation network not only contain representative semantic and textural information, but also compensate for the missing information of the object. The generated PPC and point cloud are then fed into a two-stream attention-based feature extraction network to learn intra-modal features. Simultaneously, multiple PFT modules in the backbone fuse the two features layer by layer using cross-attention. PTA-Det aims to explore a more reasonable fusion method for camera image and LiDAR points and form a plug-and-play module that can be combined with LiDAR-only methods. Extensive experiments are conducted on the KITTI and nuScenes datasets and competitive results are given. PTA-Det shows better performance on car category than most existing multimodal detection methods on multiple datasets. It is worth mentioning that our method can achieve better accuracy than the LiDAR-only methods under the same number of input points. In short, the experimental results indicate that PTA-Det could be a robust approach for 3D object detection in autonomous driving and many other applications.

Author Contributions

Conceptualization, T.Z. and R.W.; Methodology, R.W.; Supervision, T.Z.; Writing—original draft, R.W.; Writing—review & editing, W.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. These data can be found here: https://www.cvlibs.net/datasets/kitti (accessed on 7 December 2022) and https://www.nuscenes.org/nuscenes (accessed on 7 December 2022).

Acknowledgments

We thank Qi Xiao for her help in the experiment.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Alaba, S.Y.; Ball, J.E. A Survey on Deep-Learning-Based LiDAR 3D Object Detection for Autonomous Driving. Sensors 2022, 22, 9577. [Google Scholar] [CrossRef] [PubMed]
  2. Shi, S.; Wang, X.; Li, H. PointRCNN: 3D Object Proposal Generation and Detection From Point Cloud. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), CVPR 2019, Long Beach, CA, USA, 16–20 June 2019; pp. 770–779. [Google Scholar] [CrossRef] [Green Version]
  3. Zhou, Y.; Tuzel, O. VoxelNet: End-to-End Learning for Point Cloud Based 3D Object Detection. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  4. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? The KITTI vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar] [CrossRef]
  5. Chen, X.; Ma, H.; Wan, J.; Li, B.; Xia, T. Multi-view 3D Object Detection Network for Autonomous Driving. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 6526–6534. [Google Scholar] [CrossRef] [Green Version]
  6. Ku, J.; Mozifian, M.; Lee, J.; Harakeh, A.; Waslander, S.L. Joint 3D Proposal Generation and Object Detection from View Aggregation. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2018, Madrid, Spain, 1–5 October 2018; pp. 1–8. [Google Scholar] [CrossRef] [Green Version]
  7. Qi, C.R.; Liu, W.; Wu, C.; Su, H.; Guibas, L.J. Frustum PointNets for 3D Object Detection From RGB-D Data. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 918–927. [Google Scholar] [CrossRef] [Green Version]
  8. Xu, D.; Anguelov, D.; Jain, A. PointFusion: Deep Sensor Fusion for 3D Bounding Box Estimation. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 244–253. [Google Scholar] [CrossRef] [Green Version]
  9. Sindagi, V.A.; Zhou, Y.; Tuzel, O. MVX-Net: Multimodal VoxelNet for 3D Object Detection. In Proceedings of the International Conference on Robotics and Automation, ICRA 2019, Montreal, QC, Canada, 20–24 May 2019; pp. 7276–7282. [Google Scholar] [CrossRef] [Green Version]
  10. Vora, S.; Lang, A.H.; Helou, B.; Beijbom, O. PointPainting: Sequential Fusion for 3D Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, 13–19 June 2020; pp. 4603–4611. [Google Scholar] [CrossRef]
  11. Wang, Z.; Jia, K. Frustum ConvNet: Sliding Frustums to Aggregate Local Point-Wise Features for Amodal 3D Object Detection. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019. [Google Scholar]
  12. Xie, L.; Xiang, C.; Yu, Z.; Xu, G.; Yang, Z.; Cai, D.; He, X. PI-RCNN: An Efficient Multi-sensor 3D Object Detector with Point-based Attentive Cont-conv Fusion Module. In Proceedings of the 2020 AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020. [Google Scholar]
  13. Huang, T.; Liu, Z.; Chen, X.; Bai, X. EPNet: Enhancing Point Features with Image Semantics for 3D Object Detection. In Proceedings of the ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020. [Google Scholar]
  14. Qi, C.R.; Chen, X.; Litany, O.; Guibas, L.J. ImVoteNet: Boosting 3D Object Detection in Point Clouds With Image Votes. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, 13–19 June 2020; pp. 4403–4412. [Google Scholar] [CrossRef]
  15. Liu, Z.; Huang, T.; Li, B.; Chen, X.; Wang, X.; Bai, X. EPNet++: Cascade Bi-directional Fusion for Multi-Modal 3D Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2022, Early Access. [Google Scholar] [CrossRef]
  16. Wang, C.; Ma, C.; Zhu, M.; Yang, X. PointAugmenting: Cross-Modal Augmentation for 3D Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, Virtual, 19–25 June 2021; pp. 11794–11803. [Google Scholar] [CrossRef]
  17. Chen, W.; Li, P.; Zhao, H. MSL3D: 3D object detection from monocular, stereo and point cloud for autonomous driving. Neurocomputing 2022, 494, 23–32. [Google Scholar] [CrossRef]
  18. Liu, M.; Ma, J.; Zheng, Q.; Liu, Y.; Shi, G. 3D Object Detection Based on Attention and Multi-Scale Feature Fusion. Sensors 2022, 22, 3935. [Google Scholar] [CrossRef] [PubMed]
  19. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. In Proceedings of the Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017; pp. 5099–5108. [Google Scholar]
  20. Yan, Y.; Mao, Y.; Li, B. SECOND: Sparsely Embedded Convolutional Detection. Sensors 2018, 18, 3337. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the 26th Annual Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1106–1114. [Google Scholar]
  22. Wang, W.; Tran, D.; Feiszli, M. What Makes Training Multi-Modal Classification Networks Hard? In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, 13–19 June 2020; pp. 12692–12702. [Google Scholar] [CrossRef]
  23. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. In Proceedings of the NIPS 2017: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  24. Zhao, H.; Jiang, L.; Jia, J.; Torr, P.H.S.; Koltun, V. Point Transformer. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Virtual, 11–17 October 2021. [Google Scholar]
  25. Guo, M.; Cai, J.; Liu, Z.; Mu, T.; Martin, R.R.; Hu, S. PCT: Point cloud transformer. Comput. Vis. Media 2021, 7, 187–199. [Google Scholar] [CrossRef]
  26. Weng, X.; Kitani, K. Monocular 3D Object Detection with Pseudo-LiDAR Point Cloud. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshops, ICCV Workshops 2019, Seoul, Republic of Korea, 27–28 October 2019; pp. 857–866. [Google Scholar] [CrossRef] [Green Version]
  27. Reading, C.; Harakeh, A.; Chae, J.; Waslander, S.L. Categorical Depth Distribution Network for Monocular 3D Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, Virtual, 19–25 June 2021; pp. 8555–8564. [Google Scholar] [CrossRef]
  28. Wang, Y.; Chao, W.; Garg, D.; Hariharan, B.; Campbell, M.E.; Weinberger, K.Q. Pseudo-LiDAR From Visual Depth Estimation: Bridging the Gap in 3D Object Detection for Autonomous Driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, 16–20 June 2019; pp. 8445–8453. [Google Scholar] [CrossRef] [Green Version]
  29. You, Y.; Wang, Y.; Chao, W.; Garg, D.; Pleiss, G.; Hariharan, B.; Campbell, M.; Weinberger, K.Q. Pseudo-LiDAR++: Accurate Depth for 3D Object Detection in Autonomous Driving. arXiv 2019, arXiv:1906.06310. [Google Scholar]
  30. Wang, T.; Wang, F.; Lin, J.; Tsai, Y.; Chiu, W.; Sun, M. Plug-and-Play: Improve Depth Prediction via Sparse Data Propagation. In Proceedings of the International Conference on Robotics and Automation, ICRA 2019, Montreal, QC, Canada, 20–24 May 2019; pp. 5880–5886. [Google Scholar] [CrossRef]
  31. Qian, R.; Garg, D.; Wang, Y.; You, Y.; Belongie, S.J.; Hariharan, B.; Campbell, M.; Weinberger, K.Q.; Chao, W. End-to-End Pseudo-LiDAR for Image-Based 3D Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 14–19 June 2020. [Google Scholar]
  32. Liang, M.; Yang, B.; Chen, Y.; Hu, R.; Urtasun, R. Multi-Task Multi-Sensor Fusion for 3D Object Detection. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  33. Zhai, Z.; Wang, Q.; Pan, Z.; Gao, Z.; Hu, W. Muti-Frame Point Cloud Feature Fusion Based on Attention Mechanisms for 3D Object Detection. Sensors 2022, 22, 7473. [Google Scholar] [CrossRef] [PubMed]
  34. Zhang, Y.; Chen, J.; Huang, D. CAT-Det: Contrastively Augmented Transformer for Multi-modal 3D Object Detection. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022. [Google Scholar]
  35. Bai, X.; Hu, Z.; Zhu, X.; Huang, Q.; Chen, Y.; Fu, H.; Tai, C. TransFusion: Robust LiDAR-Camera Fusion for 3D Object Detection with Transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, 19–24 June 2022; pp. 1080–1089. [Google Scholar] [CrossRef]
  36. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  37. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R.B. Mask R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, ICCV 2017, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar] [CrossRef]
  38. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  39. Tang, Y.; Dorn, S.; Savani, C. Center3D: Center-based Monocular 3D Object Detection with Joint Depth Understanding. In Proceedings of the 42nd DAGM German Conference, DAGM GCPR 2020, Tubingen, Germany, 28 September–1 October 2020. [Google Scholar]
  40. Lee, J.H.; Han, M.; Ko, D.W.; Suh, I.H. From Big to Small: Multi-Scale Local Planar Guidance for Monocular Depth Estimation. arXiv 2019, arXiv:1907.10326. [Google Scholar]
  41. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable Convolutional Networks. In Proceedings of the IEEE International Conference on Computer Vision, ICCV 2017, Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar] [CrossRef] [Green Version]
  42. Noh, J.; Lee, S.; Ham, B. HVPR: Hybrid Voxel-Point Representation for Single-Stage 3D Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, Virtual, 19–25 June 2021; pp. 14605–14614. [Google Scholar] [CrossRef]
  43. Shi, S.; Wang, Z.; Shi, J.; Wang, X.; Li, H. From Points to Parts: 3D Object Detection From Point Cloud with Part-Aware and Part-Aggregation Network. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 2647–2664. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Lin, T.; Goyal, P.; Girshick, R.B.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 318–327. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Team, O.D. OpenPCDet: An Open-source Toolbox for 3D Object Detection from Point Clouds. 2020. Available online: https://github.com/open-mmlab/OpenPCDet (accessed on 1 August 2022).
  46. Wu, Y.; Kirillov, A.; Massa, F.; Lo, W.Y.; Girshick, R. Detectron2. 2019. Available online: https://github.com/facebookresearch/detectron2 (accessed on 1 August 2022).
  47. Paszke, A.; Gross, S.; Chintala, S.; Chanan, G.; Yang, E.; Devito, Z.; Lin, Z.; Desmaison, A.; Antiga, L.; Lerer, A. Automatic dif ferentiation in PyTorch. In Proceedings of the NIPS 2017: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  48. Zheng, W.; Tang, W.; Jiang, L.; Fu, C. SE-SSD: Self-Ensembling Single-Stage Object Detector From Point Cloud. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, Virtual, 19–25 June 2021; pp. 14494–14503. [Google Scholar] [CrossRef]
  49. Zhao, X.; Liu, Z.; Hu, R.; Huang, K. 3D Object Detection Using Scale Invariant and Feature Reweighting Networks. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence, AAAI 2019, The Thirty-First Innovative Applications of Artificial Intelligence Conference, IAAI 2019, The Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019, Honolulu, HI, USA, 2019, 27 January–1 February; pp. 9267–9274. [CrossRef] [Green Version]
  50. Chen, Z.; Li, Z.; Zhang, S.; Fang, L.; Jiang, Q.; Zhao, F. AutoAlignV2: Deformable Feature Aggregation for Dynamic Multi-Modal 3D Object Detection. arXiv 2022, arXiv:2207.10316. [Google Scholar]
  51. Wu, X.; Peng, L.; Yang, H.; Xie, L.; Huang, C.; Deng, C.; Liu, H.; Cai, D. Sparse Fuse Dense: Towards High Quality 3D Detection with Depth Completion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, 18–24 June 2022; pp. 5408–5417. [Google Scholar] [CrossRef]
  52. Yin, T.; Zhou, X.; Krähenbühl, P. Center-Based 3D Object Detection and Tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, Virtual, 19–25 June 2021; pp. 11784–11793. [Google Scholar] [CrossRef]
  53. Lang, A.H.; Vora, S.; Caesar, H.; Zhou, L.; Yang, J.; Beijbom, O. PointPillars: Fast Encoders for Object Detection from Point Clouds. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, 16–20 June 2019; pp. 12697–12705. [Google Scholar] [CrossRef] [Green Version]
  54. Pan, X.; Xia, Z.; Song, S.; Li, L.E.; Huang, G. 3D Object Detection with Pointformer. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, Virtual, 19–25 June 2021; pp. 7463–7472. [Google Scholar] [CrossRef]
Figure 1. (a) The background obstacle has a similar 3D structure to a car. It is difficult to distinguish them in the point cloud. (b) The points of the distant car are too sparse to determine the bounding box.
Figure 1. (a) The background obstacle has a similar 3D structure to a car. It is difficult to distinguish them in the point cloud. (b) The points of the distant car are too sparse to determine the bounding box.
Sensors 23 03229 g001
Figure 2. Illustration of the architecture of PTA-Det. It is composed of four parts: (1) Pseudo Point Cloud Generation Network for generating the coordinates and features of PPCs, (2) Two-stream Feature Extraction Network for learning point-based intra-modal and inter-modal features, (3) 3D RPN for proposal generation, (4) 3D Box Refinement for rectifying proposals.
Figure 2. Illustration of the architecture of PTA-Det. It is composed of four parts: (1) Pseudo Point Cloud Generation Network for generating the coordinates and features of PPCs, (2) Two-stream Feature Extraction Network for learning point-based intra-modal and inter-modal features, (3) 3D RPN for proposal generation, (4) 3D Box Refinement for rectifying proposals.
Sensors 23 03229 g002
Figure 3. Illustration of Frustum Feature Module.
Figure 3. Illustration of Frustum Feature Module.
Sensors 23 03229 g003
Figure 4. Illustration of Frustum-to-point Transition module. The white point is the foreground point, yellow is sampled foreground point, and red is the keypoint of the object.
Figure 4. Illustration of Frustum-to-point Transition module. The white point is the foreground point, yellow is sampled foreground point, and red is the keypoint of the object.
Sensors 23 03229 g004
Figure 5. Illustration of PTD. The letters near the box indicate its shape, e.g., N × C, N is the number of points and C is the dimension numbers of features. LBR combines Linear and BatchNorm layers and a ReLU function. KNN-group indicates the grouping operation with k-Nearest Neighbors algorithm.
Figure 5. Illustration of PTD. The letters near the box indicate its shape, e.g., N × C, N is the number of points and C is the dimension numbers of features. LBR combines Linear and BatchNorm layers and a ReLU function. KNN-group indicates the grouping operation with k-Nearest Neighbors algorithm.
Sensors 23 03229 g005
Figure 6. Illustration of PFT. It is employed to perform bi-directional information exchange between two point-based features from different modalities.
Figure 6. Illustration of PFT. It is employed to perform bi-directional information exchange between two point-based features from different modalities.
Sensors 23 03229 g006
Figure 7. Visual comparisons with two baseline methods. The second row shows the detection results of the modified PointRCNN, the third row shows Part- A 2 and the fourth row shows ours. The detection results of the first, second, and four scenes show our method can detect distant objects than these LiDAR-only methods under the same point cloud input. From the detection results of the third scene, for the van which is very close to the car structure, PTA-Det can effectively reduce false positive instances due to the addition of image semantic information.
Figure 7. Visual comparisons with two baseline methods. The second row shows the detection results of the modified PointRCNN, the third row shows Part- A 2 and the fourth row shows ours. The detection results of the first, second, and four scenes show our method can detect distant objects than these LiDAR-only methods under the same point cloud input. From the detection results of the third scene, for the van which is very close to the car structure, PTA-Det can effectively reduce false positive instances due to the addition of image semantic information.
Sensors 23 03229 g007
Figure 8. Visualized results by PTA-Det on KITTI test set. Green/turquoise/yellow bounding box indicates car/pedestrian/cyclist category, respectively. The scenes are arranged from left to right. The first and fourth show that PTA-Det performs well against distant objects. The second successfully detects all cars and excludes the object of van category. From the third scene, we observe that PTA-Det performs a little worse with multiple pedestrians and overlooks a car due to occlusion.
Figure 8. Visualized results by PTA-Det on KITTI test set. Green/turquoise/yellow bounding box indicates car/pedestrian/cyclist category, respectively. The scenes are arranged from left to right. The first and fourth show that PTA-Det performs well against distant objects. The second successfully detects all cars and excludes the object of van category. From the third scene, we observe that PTA-Det performs a little worse with multiple pedestrians and overlooks a car due to occlusion.
Sensors 23 03229 g008
Table 1. Comparison with popular 3D object detection methods on the KITTI validation split. The available results are evaluated by mAP for each category. ‘L’ and ‘R’ stand for LiDAR and image, respectively. indicates no data augmentation technique is applied. PTA-Det-1 and PTA-Det-2 represent our method adopts the refinement network proposed in PointRCNN and Part- A 2 , respectively. The optimal results are marked in bold font.
Table 1. Comparison with popular 3D object detection methods on the KITTI validation split. The available results are evaluated by mAP for each category. ‘L’ and ‘R’ stand for LiDAR and image, respectively. indicates no data augmentation technique is applied. PTA-Det-1 and PTA-Det-2 represent our method adopts the refinement network proposed in PointRCNN and Part- A 2 , respectively. The optimal results are marked in bold font.
MethodModalityCarPedestrianCyclist
EasyMod.HardmAPEasyMod.HardmAPEasyMod.HardmAP
SECOND [20]L88.6178.6277.2281.4856.5552.9847.7352.4280.5867.1563.1070.28
PointRCNN [2]L89.4178.1075.5181.0170.3960.4151.4860.7981.8158.1053.8664.59
Part- A 2 [43]L85.2874.2269.8576.4553.2446.8041.0645.5066.5141.7839.3066.54
SE-SSD [48]L90.2186.2579.2285.23--------
MV3D [5]L+R71.2962.6856.5663.51--------
AVOD-FPN [6]L+R84.4174.4468.6575.83-58.80---49.70--
F-PointNet [7]L+R83.7670.9263.6572.7870.0061.3253.5961.6477.1556.4953.3762.34
SIFRNet [49]L+R85.6272.0564.1973.9569.3560.8552.9561.0580.9760.3456.6965.97
EPNET [13]L+R88.7678.6578.3281.9166.7459.2954.8260.2883.8865.5062.7070.69
PTA-Det-1L+R86.3177.0670.2877.8861.7751.8446.9853.5370.6149.0245.5455.06
PTA-Det-2L+R84.7274.4569.8676.3460.8452.4845.1152.8172.4349.1746.7556.12
Table 2. Detection performance of PointRCNN on the validation set under different sampling strategies.
Table 2. Detection performance of PointRCNN on the validation set under different sampling strategies.
Sampling Strategy3D Object Detection (%)
Input PointGroup SizeCarCyc.Ped.mAP
16,3844096–1024–256–6481.0164.5960.7968.80
80002000–1000–500–25076.5548.2852.8659.23
40002000–1000–500–25068.6537.2745.0950.34
1600800–400–200–10038.2831.052.3623.90
Table 3. Performance comparisons on the KITTI validation set. All methods adopt the same foreground point sampling strategy proposed in Pseudo Point Generation Network, and they do not use any data augmentation technique for a fair comparison. All results are reported by the average precision with 0.7 IoU threshold and 40 recall positions for car category. * indicates our re-implementation. The optimal results are highlighted in bold front.
Table 3. Performance comparisons on the KITTI validation set. All methods adopt the same foreground point sampling strategy proposed in Pseudo Point Generation Network, and they do not use any data augmentation technique for a fair comparison. All results are reported by the average precision with 0.7 IoU threshold and 40 recall positions for car category. * indicates our re-implementation. The optimal results are highlighted in bold front.
MethodModalityBEV Detection (%)3D Detection (%)
EasyMod.HardEasyMod.Hard
PointRCNN [2]L89.8381.7576.8983.4172.6467.87
Part- A 2  [43]L89.2282.6278.3184.6271.7468.85
PointPainting [10]L+R90.3282.6476.9184.6774.6768.87
AutoAlignV2 * [50]L+R90.9586.0383.3186.2776.7270.81
EPNet++ * [15]L+R89.4685.2380.3283.2169.5868.51
SFD * [51]L+R92.7286.2881.7386.1774.9771.98
PTA-Det-1L+R90.5284.4776.3986.3177.0670.28
PTA-Det-2L+R91.3383.2578.6784.7274.4569.86
Table 4. Comparison of previous baseline works on the nuScenes validation set. None of the methods in the table use data augmentation techniques and all take the same point cloud input as our model. C.V. is the abbreviation of construction vehicle. mAP, NDS scores, and APs are reported for vehicle-related categories. The optimal results are highlighted in bold front.
Table 4. Comparison of previous baseline works on the nuScenes validation set. None of the methods in the table use data augmentation techniques and all take the same point cloud input as our model. C.V. is the abbreviation of construction vehicle. mAP, NDS scores, and APs are reported for vehicle-related categories. The optimal results are highlighted in bold front.
MethodmAPNDSCarTruckBusTrailerC.V.
PointPillar [53]43.5356.8377.4344.6553.7847.6115.76
PointPainting [10]45.6154.5278.6242.6452.4248.6416.12
SECOND [20]50.2261.9284.2951.5453.1350.3221.82
CenterPoint [52]55.6164.3284.3253.7462.4254.2423.62
PTA-Det-159.4565.2386.1255.9366.4357.3925.92
PTA-Det-257.3263.3885.4353.6362.3856.3226.43
Table 5. Contributions of PTD/PTU/PFT modules to PTA-Det.
Table 5. Contributions of PTD/PTU/PFT modules to PTA-Det.
Component3D Object Detection (%)
BaselinePTDPTUPFTCarCyc.Ped.mAP
68.9432.8427.9643.25
65.7742.6238.1248.84
72.6643.3847.1154.38
66.7432.4428.2042.46
75.0351.7349.5658.77
75.9852.0052.4160.13
Table 6. Comparison of fusion structure in PFT to PTA-Det.
Table 6. Comparison of fusion structure in PFT to PTA-Det.
Integrating Scheme3D Object Detection (%)
CarCyc.Ped.mAP
PFT c 75.2248.2647.5257.00
PFT + 75.4151.6950.1359.08
PFT 75.9852.0052.4160.13
Table 7. Comparison of calculation ways of attention. The three consecutive symbols represent the way of attention computation used in PTD, PTU, and PFT, where – stands for subtraction, × for multiplication.
Table 7. Comparison of calculation ways of attention. The three consecutive symbols represent the way of attention computation used in PTD, PTU, and PFT, where – stands for subtraction, × for multiplication.
Calculation of Attention3D Object Detection (%)
CarCyc.Ped.mAP
–  –  –72.4750.2643.5855.44
×××70.9845.2743.0353.09
×  ×  –69.3448.2239.7952.45
–  –  ×75.9852.0052.4160.13
Table 8. Comparison of sampling strategies of Pseudo Point Cloud Generation network to PTA-Det.
Table 8. Comparison of sampling strategies of Pseudo Point Cloud Generation network to PTA-Det.
Sampling Strategy3D Object Detection (%)
FPSKPSCarCyc.Ped.mAP
73.8353.7648.7958.79
75.9852.0052.4160.13
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wan, R.; Zhao, T.; Zhao, W. PTA-Det: Point Transformer Associating Point Cloud and Image for 3D Object Detection. Sensors 2023, 23, 3229. https://doi.org/10.3390/s23063229

AMA Style

Wan R, Zhao T, Zhao W. PTA-Det: Point Transformer Associating Point Cloud and Image for 3D Object Detection. Sensors. 2023; 23(6):3229. https://doi.org/10.3390/s23063229

Chicago/Turabian Style

Wan, Rui, Tianyun Zhao, and Wei Zhao. 2023. "PTA-Det: Point Transformer Associating Point Cloud and Image for 3D Object Detection" Sensors 23, no. 6: 3229. https://doi.org/10.3390/s23063229

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop