Next Article in Journal
A Class of Enhanced Nonparametric Control Schemes Based on Order Statistics and Runs
Previous Article in Journal
Farlie–Gumbel–Morgenstern Bivariate Moment Exponential Distribution and Its Inferences Based on Concomitants of Order Statistics
Previous Article in Special Issue
An ϵ-Greedy Multiarmed Bandit Approach to Markov Decision Processes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Point Cloud Registration via Heuristic Reward Reinforcement Learning

Data Mining Laboratory, Dalian University of Technology, Dalian 116000, China
Stats 2023, 6(1), 268-278; https://doi.org/10.3390/stats6010016
Submission received: 15 December 2022 / Revised: 30 January 2023 / Accepted: 31 January 2023 / Published: 6 February 2023
(This article belongs to the Special Issue Feature Paper Special Issue: Reinforcement Learning)

Abstract

:
This paper proposes a heuristic reward reinforcement learning framework for point cloud registration. As an essential step of many 3D computer vision tasks such as object recognition and 3D reconstruction, point cloud registration has been well studied in the existing literature. This paper contributes to the literature by addressing the limitations of embedding and reward functions in existing methods. An improved state-embedding module and a stochastic reward function are proposed. While the embedding module enriches the captured characteristics of states, the newly designed reward function follows a time-dependent searching strategy, which allows aggressive attempts at the beginning and tends to be conservative in the end. We assess our method based on two public datasets (ModelNet40 and ScanObjectNN) and real-world data. The results confirm the strength of the new method in reducing errors in object rotation and translation, leading to more precise point cloud registration.

1. Introduction

Point cloud registration is a primary task of high-quality 3D model reconstruction. A complete point cloud effectively captures the surface details of a detected object. However, due to the limitation of scanning equipment and environment, single-view point clouds with noise can inevitably be obtained. To get a complete point cloud of the object, multi-viewpoints need to be transformed into the same coordinate system, referring to the point cloud registration [1,2,3]. The iterative methods are applied in traditional algorithms, such as the Iterative Closest Point (ICP) algorithm. Despite being widely used, ICP is computationally expensive and demands the initial positions of two point clouds [4,5], which can sometimes lead to the local optimum.
In recent studies, some learning-based methods have been proposed to directly predict a transformation matrix for the source point cloud, e.g., ReAgent [6]. It handles point cloud registration using Imitation Learning (IL) and Reinforcement Learning (RL). An accurate initial policy can be obtained by imitating an expert, then fine-tuning the policy with a symmetry-invariant reward. ReAgent realized the registration step by step. This paper improves the model by addressing two limitations in ReAgent, including (1) the lack of the extraction of local features to point cloud in the state embedding stage, and (2) the fixed penalties for different states in the reward.
Specifically, with the purpose of getting the positional relation between the source and target point cloud accurately, and generating a more effective state representation, a feature extraction layer combined with EdgeConv is proposed, which enhances feature description in the state embedding. Furthermore, this paper defines a new reward function with time-varying penalties related to the current step. The new reward function allows more aggressive attempts at the early search stage while tending to be conservative over time. Finally, extensive experiments are conducted to assess the new method based on different public datasets and real-world data.
To sum up, the main contributions of this paper are threefold.
  • First, to enrich the encoding process, an improved state-embedding module is proposed. It combines EdgeConv to capture the local features of relative coordinates, reflecting the key information of two point cloud positions among the current state.
  • Second, a heuristic reward function is proposed. Unlike the invariable penalty in each step, the newly designed reward function allows aggressive attempts at the beginning when the environment is still unclear.
  • Finally, the new method with an improved state-embedding module and the heuristic reward function is evaluated on two public datasets as well as real-world data of train components. The experimental results show that the new method effectively reduces the errors in rotation and translation, and can lead to more precise point cloud registration.
The rest of this paper is organized as follows. Related work is discussed in Section 2, and Section 3 introduces the three-dimensional point cloud registration methods in detail, including some principles of point cloud registration, an improved state-embedding module, and the stochastic reward function. Section 4 shows the experimental results to prove the effectiveness and feasibility of our methods. Finally, Section 5 concludes the paper.

2. Related Work

Point cloud registration methods can be generally divided into traditional algorithms and learning-based methods.
In traditional algorithms, coarse registration is usually the first step to making two point clouds closer. Some local feature descriptors need to be generated in point clouds [2,7,8,9], then similar features between the two point clouds should be identified. The key point matching algorithms are used to find the corresponding key point pairs, so that the correspondence between two point clouds is built up. The transformation matrix can be generated by the Singular Value Decomposition (SVD) method [10]. Even if some algorithms eliminate the mismatched corresponding key point pairs [11,12], there are still errors in the transformation, especially when two point clouds partially overlap. Therefore, the Iterative Closest Point algorithm (ICP) has been widely used as the fine calibration, as it minimizes the Euclidean distance between the point pairs, making the registration more precise. Although the requirements of ICP about the initial position and overlap rate of two point clouds are strict, the result of ICP still easily converges to the local optimum. Thus, improved work based on ICP was developed, aiming at solving problems such as the object in movement and the slow convergence rate [13,14,15]. The 3D-NDT method uses probability density to replace feature extraction and key point matching [3]. The 4-point congruent sets (4PCS) algorithm selects four points in the same base from the source point cloud and detects other four points in the same base from the target point cloud in the range of errors. Along this line, the correspondence is generated [16]. The traditional algorithms usually need to build up the correspondence or iterative calculation between the source and target point cloud.
On the other hand, learning-based registration methods make registration based on neural networks. As the pointwise network PointNet is proposed [17,18], other deep learning models are put forward gradually [19,20,21]. To extract the local features from point to point, the EdgeConv layer in DGCNN was proposed [22]. PointNetLK uses PointNet to extract features, then, the high-dimensional features are considered as the image to make the image registration so that the point cloud can be aligned [23]. Deep Closest Point (DCP) combines the feature-embedded network and an attention module to get a rigid transformation matrix [24]. Moreover, given the broad implementations of reinforcement learning (RL), e.g., classic games [25], quantitative trading [26], and image registration [27,28], RL has also been found useful in handling point cloud registration, and the ReAgent [6] that combined imitation learning (IL) and RL is a typical example. This paper focuses on improving the state embedding and reward function.

3. Methodology

This section introduces the heuristic reward reinforcement learning (RL) framework for point cloud registration. In general, an improved state-embedding module and the heuristic reward function are proposed to reduce transformation errors, improving registration more precisely.

3.1. Point Cloud Registration Network and ReAgent

Suppose there are two point clouds, the source point cloud X and the target point cloud Y . The point set of X may be the same as Y , indicating they are the same object. If X is just a part of Y , then X partially overlaps Y . There is a point cloud X in observation which need to be transformed to X , so the current observation can be defined as O ( X , Y ) , and the registration between X and X can be written as follows:
X T = X ,
where T is the transformation matrix. In traditional algorithms, based on the correspondence of two point clouds, the transformation matrix is usually obtained by SVD. However, the transformation matrix would be the prediction as the output by the learning-based methods.
The iterative registration is a kind of fine-tuning method with higher accuracy and inefficiency; it would transform the point cloud step by step instead of directly transforming in one shot. Suppose that the X requires n steps to be aligned with X . That is:
X T 1 T 2 T n = X n = X
at step i , the point cloud X i can be represented as:
X i = X i 1 T i
Thus, the transformation of the point cloud in each step can be regarded as the discrete actions in rotation and translation, and this process is similar to the learning process in RL. If the representation of state and reward function can be defined appropriately, RL techniques should be able to implemented.
Figure 1 shows the architecture of ReAgent for one iteration in step i . First, the features of X i and Y are embedded and concatenated to the state vector S i as the representation of the state, then it updates steps by using discrete, limited step sizes in each iteration. The policy π ( S ) gives the probability of the actions that can be selected in the rotation and translation axis; it is computed by the action head of the agent. Additionally, it predicts the step sizes for this iteration. The disentanglement of rotation and translation is used to avoid errors in transformation.
Since point cloud registration is a complicated task, using RL to train the agent at the beginning may fall into the suboptimal policy, so the IL is used to train for the state embedding and the policy initialization.

3.2. Feature Extraction and an Improved State-Embedding Module

The representation of state is achieved by a PointNet-like architecture in ReAgent, and features of two point clouds are extracted separately with shared MLP layers.
In the process of state embedding, as shown in Figure 2, the target point cloud Y is the extracted feature at the first step, then generated as ϕ ( Y ) . The observed source point cloud X i is the embedded feature at the beginning of each step. For example, ϕ ( X i ) represents the feature vector of the source point cloud in step i. Finally, ϕ ( X i ) is concatenated with ϕ ( Y ) , and the state vector in step i can be generated as follows:
S i = c o n c a t e [ ϕ ( X i ) , ϕ ( Y ) ]
Specifically, in Figure 2b, the PointNet-like architecture is served as the feature embedding layer. The input is a point cloud with N points and three-dimension information x y z . It increases the dimension of features through the one-dimensional convolution layer {64, 128, 1024}, then a 1 × 1024 global feature can be generated after max pooling, and two global features from X i and Y are concatenated to a 1 × 2048 state vector.
In ReAgent, fewer embedding layers are considered to learn the expressive feature vector sufficiently. However, some works like PointNet++ and DGCNN have proved that the local features are important to improve the accuracy of point cloud recognition and segmentation. DCP also uses DGCNN in feature extraction for better registration results. Although these local feature extraction methods require additional computing costs, local features about the positional relations between points and neighboring points are necessary for feature embedding. Based on extensive experiments, the new framework in this paper replaces the PointNet-like architecture by feature extraction layers combined with EdgeConv. The improved state-embedding module is shown in Figure 3a.
Furthermore, the input point cloud will go through an EdgeConv layer. Figure 3b shows that k neighboring points can be detected by k-nn algorithm for each point in N, computing the difference of three-dimension coordinate values between the k neighboring points and their center point; k vector point to the center point can be obtained, and a local neighborhood graph can be generated. After that, the local features between the center point and k neighboring points can be extracted by two-dimension convolutional layers. After pooling in dimension k ,   N × 64   feature vector is obtained. The subsequent feature extraction architecture is similar to the ReAgent. Two 1 × 1024 feature vectors concatenate to a 1 × 2048 and serve as the state vector.

3.3. Heuristic Reward Function in RL

Only training the agent to imitate the expert policy does not guarantee consistently good performance in different datasets. Thus, ReAgent used RL to fine-tune and improve the model’s generality. First, as an important evaluation measure, the Chamfer distance (CD) can be represented by:
C D ( X , Y ) = 1 | X | x X m i n y Y x y 2 2 ,
where x X and y Y . Note that CD reflects the similarity of two point clouds in terms of the coordinate differences. In ReAgent, the reward function, denoted by r , is defined based on CD:
r = { ε , C D ( X i , X ) > C D ( X i 1 , X ) ε 0 , C D ( X i , X ) = C D ( X i 1 , X ) ε + ,   C D ( X i , X ) < C D ( X i 1 , X )
where X is the true source point cloud. It is identical to the target point cloud Y , or only partially overlaps Y . So, X is used to represent the point cloud that needs registration based on the observed point cloud X i . Note that, X i is a point cloud that is observed in step i , and X i 1 is the observed point cloud in step i 1 . The penalties ( ε + , ε 0 , ε ) would be given depending on the CD at step i compared with the CD at step i 1 . The three penalties correspond to three reward states: “better”, “same”, and “worse”, respectively. The values of ( ε + , ε 0 , ε ) are set to (0.5, 0.1, 0.6) in [6].
If the CD between the current observed point cloud X i and point cloud X is smaller than the last step, the transformation of point cloud X i in current step is considered as a “better” state, so the positive penalty ε + would be given. If the CD is larger or the same as last step, the transformation is considered as “worse” or “same”. In these cases, the negative penalties ε and ε 0 would be given respectively.
At the beginning of registration, there may be a large difference in rotation and distance between the point cloud X i and X at the initial position. In the process of iterative registration by RL, the actions selected by the agent’s policy π ( S ) in the first few steps may not reduce the values of CD, even making the CD increase.
Heuristic algorithms, such as simulated annealing that follow time-varying acceptance rates for new attempts, have been proven to be efficient in achieving global optimum. Examples of the implementations of heuristic methods in machine learning can be found in simulated annealing-based mobile sequential recommendation [29,30,31], stochastic deep learning [32], and stochastic subsampling RL [33,34].
Inspired by the simulated annealing algorithm, two parameters related to the current step are introduced to optimize the reward function:
θ m = t m α i
θ n = t n β i
where i is the current step number, and t m , α , t n , β are set according to the experimental results. Then, the following heuristic reward function is proposed:
r = { ε θ m , C D ( X i , X ) > C D ( X i 1 , X ) ε 0 θ n , C D ( X i , X ) = C D ( X i 1 , X ) ε + θ m ,   C D ( X i , X ) < C D ( X i 1 X )
where θ m is a growing exponential function, and θ n is a decreasing exponential function.
Therefore, in the first few steps, the actions selected by policy π ( S ) may cause the values of CD to increase. Given the fact that the penalties of “worse” and “better” are small, while the penalty of “same” is relatively large, this reward function encourages the agent to take aggressive movements and avoid staying in the “same” state. In the last few steps, the penalties of “worse” and “better” states increase, leading to a more careful and accurate transformation by policy π ( S ) .

4. Experimental Results

This section discusses the results from experiments based on different datasets and robustness checks.

4.1. Registration on ModelNet40

First, I demonstrate the results from the experiments evaluating the new method, based on the ModelNet40 dataset. Following the same setting in [6], ModelNet40 has been split into two parts, the 1–20 categories models and the 21–40 categories. In the experiments, all models have taken resample, rigid rotation, and translation, so that the source and target point cloud can be obtained.
Based on imitation learning, the agent would be pre-trained for 50 epochs on the first 20 categories without any noise. Then, based on RL, another 50 epochs are supplied for fine-tuning the policy on the first 20 categories with some Gaussian noise added.
All experiments are performed under Windows11 operating system, Intel i9-12900k and 32 GB RAM, and RTX3090ti with the simulation software. We followed the parameters in [6]. The Proximal Policy Optimization (PPO) is used to update the policy, and the formulation in [35] can be implemented as used in actor-critic architecture. The PPO loss and advantage A ^ are the same as the ReAgent. In the rotation and translation axis, there are 11 step sizes in each axis, [−0.27, −0.09, −0.03, −0.01, −0.0033, 0, 0.0033, 0.01, 0.03, 0.09, 0.27]. Note that the negative values indicate that the agent would take a transformation in negative directions of coordinate axes. The learning rate in pretraining by IL is set to 0.001 with halving it in each of the 10 epochs. The learning rate of RL is set to 0.0001. All the point cloud data would be pretreated according to the ReAgent.
We used several metrics that are commonly used in related work to evaluate performance.
Mean Absolute Error (MAE) is the error between the predicted vector v p and ground truth vector v g t , and it can be calculated as following:
M A E v = 1 3 | v p v g t |
where the vector can be a rotation or translation vector to calculate the errors.
Isotropic Error (ISO) only considers the values of rotation and translation matrix to calculate errors, so the ISO can be obtained as follows:
I S O r = arccos t r a c e ( R d R g t 1 1 ) 2
I S O t = T d T g t 2
where t r a c e is the sum of the diagonal elements of the matrix; R d and T d are the rotation and translation matrix in the end; R g t and T g t are ground truths for the point cloud to transformation.
The Chamfer Distance has been mentioned in the definition of the reward function. A Modified Chamfer Distance ( C D ˜ ) is proposed by Lee and Yew [36]. It is defined as follows:
C D ˜ = ( P s , P t ) = C D ( P s , P t , c l e a n ) + C D ( P t , P s , c l e a n )
where c l e a n means the point cloud with no noise.
Table 1 shows the experimental results on ModelNet40. Due to a different testing environment, the results of ReAgent were slightly different from the original paper, while the main patterns were found consistent. It can be seen that the new method and ReAgent obtained smaller errors of rotation and translation than the DCP-v2 [24] and PointNetLK [23]. Additionally, the errors of the new method are smaller than ReAgent in all 40 categories, while the running speed of ReAgent is faster. Although the running time of the new method is slower than ReAgent and DCP-v2, the accuracy of the new method in registration is better.

4.2. Robustness Test

To check the robustness of the model when noises exist, the different variance of Gaussian noise is respectively added to the point cloud, and the noise clipped to 0.05. Figure 4 shows the models with noise in different σ . The Chamfer Distances ( C D ˜ s) are calculated based on different results of registration.
The C D ˜ -noise curve in Figure 5 shows the C D ˜ with different noise magnitudes. It can be seen that the values of C D ˜ obtained by the new method are consistently smaller than ReAgent, indicating that the values of the state embedding used some local features to represent the point cloud, and the heuristic reward function proposed in this paper. Overall, the results confirm the robustness of the new method under different noise levels.

4.3. Experiment on ScanobjectNN and Other Real-World Data

Experiments have also been conducted based on the ScanObjectNN dataset [37], which is collected from the depth sensor as the real data. The point clouds are segmented objects in ScanObjectNN, including 15 categories and 581 models in total, and 2048 points for each point cloud.
Furthermore, an additional category containing train components was manually collected (see [38]). Figure 6 shows 8 component point clouds such as traction rods, bolts, and wheelsets. They also have 2048 points with some noise after resampling. So, there are 16 categories and 589 point cloud models in the evaluated dataset.
Figure 7 shows the process of registration in train component wheelsets; the source point cloud (red) is transformed into the target point cloud (blue) step by step. The green point cloud represents the initial position of the source point cloud. Figure 7a demonstrates the results based on the ReAgent registration, and Figure 7b shows the results based on the newly proposed method. As can be seen, ReAgent results are unstable. It transforms after two point clouds overlapped at step 7 and the final step. Compared with ReAgent, the new method was designed to transform more and more conservatively over time, hence led to a stabilized overlapping result.
As reported in Table 2, the errors of the new method are consistently smaller than DCP-v2, PointNetLK, and ReAgent. Importantly, the running time of the new method did not have significant changes compared with the running time in the ModelNet40-related experiments. It shows that the methods can be applied to practical applications on real-world data.

5. Discussion and Conclusions

Despite the overall outperformance of the new method which has been confirmed, there are two limitations that may lead to additional improvements in future work. First, since the EdgeConv has been used in embedding layers, the extraction of local features required an increasing computational complexity. Second, θ m and θ n related to the current step in the reward optimization were determined based on a series of experimental results. To address these two limitations, simplified but efficient embedding layers may be investigated so that the computing cost and the embedding effectiveness can be better balanced. Additionally, implementing the optimization process with parallel computing and high-performance computing techniques is also a possible research direction to enhance the computing efficiency while remaining the embedding quality. Furthermore, regarding the parameters, an adaptive time-dependent searching strategy may be developed for a more powerful optimization of the registration.
In conclusion, this paper introduces a point cloud registration method via heuristic reward reinforcement learning. An improved state-embedding module is also proposed to extract more local features about related positions from point to point. The heuristic reward function follows a time-dependent searching strategy, which allows aggressive attempts at the beginning and tends to be conservative in the end. The new method is evaluated on ModelNet40, ScanObjectNN, and additional real-world data, and the results confirm the improvements in terms of multiple evaluation metrics.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The ModelNet40 and ScanObjectNN are publicly available. ModelNet40 can be downloaded here: https://modelnet.cs.princeton.edu/, (accessed on 11 January 2023); The ScanObjectNN can be downloaded here: https://hkust-vgd.github.io/scanobjectnn/, (accessed on 11 January 2023).

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Li, H.; Hartley, R. The 3D-3D registration problem revisited. In Proceedings of the 2007 IEEE 11th International Conference on Computer Vision, Rio De Janeiro, Brazil, 14–21 October 2007; pp. 1–8. [Google Scholar]
  2. Rusu, R.B.; Blodow, N.; Beetz, M. Fast point feature histograms (FPFH) for 3D registration. In Proceedings of the 2009 IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; pp. 3212–3217. [Google Scholar]
  3. Magnusson, M.; Lilienthal, A.; Duckett, T. Scan registration for autonomous mining vehicles using 3D-NDT. J. Field Robot. 2007, 24, 803–827. [Google Scholar] [CrossRef]
  4. Yang, B.; Zang, Y. Automated registration of dense terrestrial laser-scanning point clouds using curves. ISPRS J. Photogramm. Remote Sens. 2014, 95, 109–121. [Google Scholar] [CrossRef]
  5. He, B.; Lin, Z.; Li, Y.F. An automatic registration algorithm for the scattered point clouds based on the curvature feature. Opt. Laser Technol. 2013, 46, 53–60. [Google Scholar] [CrossRef]
  6. Bauer, D.; Patten, T.; Vincze, M. Reagent: Point cloud registration using imitation and reinforcement learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 14586–14594. [Google Scholar]
  7. Rusu, R.B.; Blodow, N.; Marton, Z.C.; Beetz, M. Aligning point cloud views using persistent feature histograms. In Proceedings of the 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems, Nice, France, 22–26 September 2008; pp. 3384–3391. [Google Scholar]
  8. Johnson, A.E. Spin-Images: A Representation for 3-D Surface Matching. Ph.D. Thesis, Robotics Institute, Carnegie Mellon University, Pittsburgh, PA, USA, 1997. [Google Scholar]
  9. Tombari, F.; Salti, S.; Stefano, L.D. Unique signatures of histograms for local surface description. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2010; pp. 356–369. [Google Scholar]
  10. Oomori, S.; Nishida, T.; Kurogi, S. Point cloud matching using singular value decomposition. Artif. Life Robot. 2016, 21, 149–154. [Google Scholar] [CrossRef]
  11. Taati, B.; Greenspan, M. Local shape descriptor selection for object recognition in range data. Comput. Vis. Image Underst. 2011, 115, 681–694. [Google Scholar] [CrossRef]
  12. Papazov, C.; Haddadin, S.; Parusel, S.; Krieger, K.; Burschka, D. Rigid 3D geometry matching for grasping of known objects in cluttered scenes. Int. J. Robot. Res. 2012, 31, 538–553. [Google Scholar] [CrossRef]
  13. Hong, S.; Ko, H.; Kim, J. VICP: Velocity updating iterative closest point algorithm. In Proceedings of the 2010 IEEE International Conference on Robotics and Automation, Anchorage, AK, USA, 3–7 May 2010; pp. 1893–1898. [Google Scholar]
  14. Yang, J.; Li, H.; Jia, Y. Go-icp: Solving 3d registration efficiently and globally optimally. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 1457–1464. [Google Scholar]
  15. Censi, A. An ICP variant using a point-to-line metric. In Proceedings of the 2008 IEEE International Conference on Robotics and Automation, Pasadena, CA, USA, 19–23 May 2008; pp. 19–25. [Google Scholar]
  16. Aiger, D.; Mitra, N.J.; Cohen-Or, D. 4-points congruent sets for robust pairwise surface registration. In ACM SIGGRAPH 2008 Papers; ACM: New York, NY, USA, 2008; pp. 1–10. [Google Scholar]
  17. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  18. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. In Proceedings of the 31st Annual Conference on Neural Information Processing Systems (NIPS), Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  19. Li, Y.; Bu, R.; Sun, M.; Wu, W.; Di, X.; Chen, B. Pointcnn: Convolution on x-transformed points. In Proceedings of the 32nd Conference on Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 2–8 December 2018; Volume 31. [Google Scholar]
  20. Wu, W.; Qi, Z.; Fuxin, L. Pointconv: Deep convolutional networks on 3d point clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 9621–9630. [Google Scholar]
  21. Liu, Y.; Fan, B.; Xiang, S.; Pan, C. Relation-shape convolutional neural network for point cloud analysis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8895–8904. [Google Scholar]
  22. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic graph cnn for learning on point clouds. ACM Trans. Graph. 2019, 38, 1–12. [Google Scholar] [CrossRef]
  23. Aoki, Y.; Goforth, H.; Srivatsan, R.A.; Lucey, S. Pointnetlk: Robust & efficient point cloud registration using pointnet. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7163–7172. [Google Scholar]
  24. Wang, Y.; Solomon, J.M. Deep closest point: Learning representations for point cloud registration. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3523–3532. [Google Scholar]
  25. Zhu, T.; Ma, M.H. Deriving the Optimal Strategy for the Two Dice Pig Game via Reinforcement Learning. Stats 2022, 5, 805–818. [Google Scholar] [CrossRef]
  26. Zhu, T.; Zhu, W. Quantitative trading through random perturbation Q-network with nonlinear transaction costs. Stats 2022, 5, 546–560. [Google Scholar] [CrossRef]
  27. Liao, R.; Miao, S.; de Tournemire, P.; Grbic, S.; Kamen, A.; Mansi, T.; Comaniciu, D. An artificial agent for robust image registration. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  28. Ma, K.; Wang, J.; Singh, V.; Tamersoy, B.; Chang, Y.J.; Wimmer, A.; Chen, T. Multimodal image registration with deep context reinforcement learning. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Cham, Switzerland, 2017; pp. 240–248. [Google Scholar]
  29. Ye, Z.; Xiao, K.; Ge, Y.; Deng, Y. Applying simulated annealing and parallel computing to the mobile sequential recommendation. IEEE Trans. Knowl. Data Eng. 2018, 31, 243–256. [Google Scholar] [CrossRef]
  30. Ye, Z.; Xiao, K.; Deng, Y. A unified theory of the mobile sequential recommendation problem. In Proceedings of the 2018 IEEE International Conference on Data Mining (ICDM), Singapore, 17–20 November 2018; IEEE: New York, NY, USA, 2018; pp. 1380–1385. [Google Scholar]
  31. Xiao, K.; Ye, Z.; Zhang, L.; Zhou, W.; Ge, Y.; Deng, Y. Multi-user mobile sequential recommendation for route optimization. ACM Trans. Knowl. Discov. Data 2020, 14, 1–28. [Google Scholar] [CrossRef]
  32. Guo, P.; Ye, Z.; Xiao, K.; Zhu, W. Weighted aggregating stochastic gradient descent for parallel deep learning. IEEE Trans. Knowl. Data Eng. 2022, 34, 5037–5050. [Google Scholar] [CrossRef]
  33. Guo, P.; Xiao, K.; Ye, Z.; Zhu, H.; Zhu, W. Intelligent career planning via stochastic subsampling reinforcement learning. Sci. Rep. 2022, 12, 1–16. [Google Scholar] [CrossRef] [PubMed]
  34. Guo, P.; Xiao, K.; Ye, Z.; Zhu, W. Route optimization via environment-aware deep network and reinforcement learning. ACM Trans. Intell. Syst. Technol. 2021, 12, 1–21. [Google Scholar] [CrossRef]
  35. Schulman, J.; Wolski, F.; Dhariwal, P.; Radford, A.; Klimov, O. Proximal policy optimization algorithms. arXiv 2017, arXiv:1707.06347. [Google Scholar]
  36. Yew, Z.J.; Lee, G.H. Rpm-net: Robust point matching using learned features. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11824–11833. [Google Scholar]
  37. Uy, M.A.; Pham, Q.H.; Hua, B.S.; Nguyen, T.; Yeung, S.K. Revisiting point cloud classification: A new benchmark dataset and classification model on real-world data. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1588–1597. [Google Scholar]
  38. Li, J.; Chen, B.; Yuan, M.; Zhao, Q.; Luo, L.; Gao, X. Matching Algorithm for 3D Point Cloud Recognition and Registration Based on Multi-Statistics Histogram Descriptors. Sensors 2022, 22, 417. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Architecture of ReAgent for one step.
Figure 1. Architecture of ReAgent for one step.
Stats 06 00016 g001
Figure 2. The architecture of feature embedding in ReAgent.
Figure 2. The architecture of feature embedding in ReAgent.
Stats 06 00016 g002
Figure 3. Improved state-embedding module.
Figure 3. Improved state-embedding module.
Stats 06 00016 g003
Figure 4. Examples of point cloud with different variance of Gaussian noise. (a) Model without noise; (b) σ = 0.01 ; (c) σ = 0.03 ; (d) σ = 0.05 .
Figure 4. Examples of point cloud with different variance of Gaussian noise. (a) Model without noise; (b) σ = 0.01 ; (c) σ = 0.03 ; (d) σ = 0.05 .
Stats 06 00016 g004
Figure 5. Performance Comparisons under Different Noise Levels.
Figure 5. Performance Comparisons under Different Noise Levels.
Stats 06 00016 g005
Figure 6. The train component point cloud.
Figure 6. The train component point cloud.
Stats 06 00016 g006
Figure 7. Example of component registration (initial position, step 1, step 7, and the final step).
Figure 7. Example of component registration (initial position, step 1, step 7, and the final step).
Stats 06 00016 g007aStats 06 00016 g007b
Table 1. Registration Results on ModelNet40.
Table 1. Registration Results on ModelNet40.
The First 20 CategoriesThe Second 20 Categories
MAEISO C D ˜ MAEISO C D ˜ T
RTRT×0.001RTRT×0.001(ms)
DCP-v23.8760.0327.8260.0712.814.9120.0389.1380.0793.9521
PointNetLK1.9120.0133.8260.0281.121.8530.0173.8120.0321.6242
ReAgent IL + RL1.7830.0113.1890.0240.761.7600.0112.9960.0230.9919
Our method IL + RL1.5880.0113.1340.0240.781.5570.0102.8970.0221.0026
Table 2. Results on ScanObjectNN and train component.
Table 2. Results on ScanObjectNN and train component.
MAEISO C D ˜ T
RTRT×0.001ms
DCP-v28.7600.08117.3200.1635.0853
PointNetLK1.3210.0152.3140.0301.6246
ReAgent IL + RL1.4490.0122.7890.0250.7522
Our method IL + RL1.1530.0122.2760.0220.6827
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, B. Point Cloud Registration via Heuristic Reward Reinforcement Learning. Stats 2023, 6, 268-278. https://doi.org/10.3390/stats6010016

AMA Style

Chen B. Point Cloud Registration via Heuristic Reward Reinforcement Learning. Stats. 2023; 6(1):268-278. https://doi.org/10.3390/stats6010016

Chicago/Turabian Style

Chen, Bingren. 2023. "Point Cloud Registration via Heuristic Reward Reinforcement Learning" Stats 6, no. 1: 268-278. https://doi.org/10.3390/stats6010016

Article Metrics

Back to TopTop