Next Article in Journal
Early Fire Detection Using Long Short-Term Memory-Based Instance Segmentation and Internet of Things for Disaster Management
Previous Article in Journal
Practical Considerations for Laser-Induced Graphene Pressure Sensors Used in Marine Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

RST: Rough Set Transformer for Point Cloud Learning

Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650500, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(22), 9042; https://doi.org/10.3390/s23229042
Submission received: 26 September 2023 / Revised: 25 October 2023 / Accepted: 6 November 2023 / Published: 8 November 2023
(This article belongs to the Topic Artificial Intelligence in Sensors, 2nd Volume)

Abstract

:
Point cloud data generated by LiDAR sensors play a critical role in 3D sensing systems, with applications encompassing object classification, part segmentation, and point cloud recognition. Leveraging the global learning capacity of dot product attention, transformers have recently exhibited outstanding performance in point cloud learning tasks. Nevertheless, existing transformer models inadequately address the challenges posed by uncertainty features in point clouds, which can introduce errors in the dot product attention mechanism. In response to this, our study introduces a novel global guidance approach to tolerate uncertainty and provide a more reliable guidance. We redefine the granulation and lower-approximation operators based on neighborhood rough set theory. Furthermore, we introduce a rough set-based attention mechanism tailored for point cloud data and present the rough set transformer (RST) network. Our approach utilizes granulation concepts derived from token clusters, enabling us to explore relationships between concepts from an approximation perspective, rather than relying on specific dot product functions. Empirically, our work represents the pioneering fusion of rough set theory and transformer networks for point cloud learning. Our experimental results, including point cloud classification and segmentation tasks, demonstrate the superior performance of our method. Our method establishes concepts based on granulation generated from clusters of tokens. Subsequently, relationships between concepts can be explored from an approximation perspective, instead of relying on specific dot product or addition functions. Empirically, our work represents the pioneering fusion of rough set theory and transformer networks for point cloud learning. Our experimental results, including point cloud classification and segmentation tasks, demonstrate the superior performance of our method.

1. Introduction

Due to the significant advancements in autonomous driving and unmanned systems, LiDAR has emerged as a crucial component of three-dimensional perception systems. A point cloud comprises data points generated by 3D sensors, providing richer perceptual and interactive information compared to 2D scenes. Geometric information within point clouds forms the basis of numerous applications [1]. Three-dimensional sensors, such as LiDAR, generate irregular and unstructured reflective points that correspond to object surfaces [2]. Consequently, unlike 2D images, the disorderly and unstructured nature of 3D point clouds presents challenges in swiftly capturing semantic information directly from this data type [3].
Conventional point cloud learning methods [4,5,6] typically involve reordering the input point sequence or voxelating the point cloud to achieve canonical domain convolution. Nevertheless, as convolution-based models primarily focus on local structures, the disorder and irregularity of point cloud data significantly impact fine-grained localization accuracy. Recently, transformer-based models have garnered significant attention due to their superior ability to capture long-range dependencies. The core component for generating refined attention features based on global context is the self-attention module. Transformer-based models have addressed the limitations of convolution-based models and demonstrated significant advantages and potential in point cloud learning.
Owing to the disorder and lack of structure in point cloud data, there is a significant amount of uncertainty in the features, particularly evident at boundary points. Nevertheless, existing transformer models have not taken into account the potential misdirection and confusion caused by characterizing uncertainty in the attention mechanism. The self-attention mechanism relies on token-to-token relationships to guide the features directly through dot products or summation. This deterministic guided representation assumes that the features are correct and valid. Once there is a sudden occurrence of uncertainty token, the negative effect of this incorrect relationship is amplified in the global guidance. The current point cloud transformer clearly uses these deterministic guided representation methods to handle uncertain point cloud features, leading to a contradictory situation that is inappropriate. In fact, point clouds are not isolated; they share common features. It is feasible to create concepts by grouping similar point clouds and allowing these concepts to approximate each other to guide the representation, as depicted in Figure 1. However, there is insufficient research on more objective guidance mechanisms to accommodate the disorder and lack of structure in point cloud data. Therefore, it is essential to introduce a novel feature representation method to handle uncertain tokens.
In recent years, rough set theory has been considered as a powerful mathematical tool for the uncertainty issue [7]. Rough set theory forms concepts by the granulation of a cluster of tokens, then expresses the similarities and differences between concepts via an approximation operator. It can extract the most important features to achieve its concept induction [8]. Obviously, rough set is an effective method for studying the characterization of uncertain conceptual relationships. It is more capable of guiding the relationship between global information. Nevertheless, we empirically found that none of the current transformers constructed based on rough set methods improve the feature extraction capability of the model. The key reason is that the most popular existing rough set models are used for the consistency measurement of the issue of feature selection. In these research works, the samples are described by both conditional and decisional attributes. For our study, point cloud data are in the same feature space. It necessitates redesigning the process of approximation computation.
To tackle these problems, an attention mechanism grounded in rough set theory is proposed to accommodate the presence of uncertain tokens during the guidance process. The granulation and lower-approximation expressions are redefined for neighborhood rough sets to align with the fundamental rough set definition, thus enhancing their suitability for deep learning applications. The technique leverages neighborhood granulation to naturalize point cloud features, forming information granules. Feature guidance is subsequently achieved by quantifying the similarity between information granules through lower approximation. Unlike the traditional dot product attention in transformers, this approach demonstrates an improved resilience to the inherent uncertainty within point cloud data, resulting in a more objective representation.
As a result, it effectively addresses the limitations of the current point cloud transformer.
The major contributions are as follows:
1
We redefine the granulation and lower-approximation expressions for neighborhood rough set to conform to the fundamental definition of rough sets and enhance their applicability in deep learning. Through empirical investigation, we have determined that this marks the initial fusion of rough set theory and deep learning network models in the context of point cloud learning.
2
We propose a novel rough set-based attention mechanism to replace the dot product attention, thereby constructing a transformer network structure (RST) tailored for point cloud learning. This network directly takes point cloud data as inputs and extracts features using multi-head rough set attention. In comparison to the traditional transformer model, the RST network exhibits a stronger ability to provide an objective relationship guidance for uncertain point cloud data.
3
The model is evaluated through point cloud classification and segmentation experiments using the ModelNet40 [9] and ShapeNet [10] datasets. All the results demonstrate that our method outperforms the most advanced networks. Additionally, we conduct a visual analysis to elucidate the improvements over traditional attention mechanisms. The resource codes are validated at https://github.com/WinnieSunning/RST, (accessed on 7 November 2023).

2. Related Work

In this section, we will conduct an in-depth literature review of the state-of-the-art deep learning methods for 3D LiDAR point clouds, including traditional point cloud learning methods, transformer-based point cloud learning methods, and other advanced point cloud learning methods.

2.1. Traditional Point Cloud Learning Methods

Before the transformer method became popular, the deep learning methods on point clouds can be roughly categorized into two classes: PointNet-based [11,12,13] and convolution-based [4,5,14]. PointNet [11] uses the multi-layer perceptron (MLP) and max pooling operation to extract features of point clouds. PointNet can perform the classification task well, but the local features’ extraction ability is poor, which makes it difficult to analyze complex scenes. Inspired by CNN, PointNet++ [12] is proposed, which is able to extract local features at different scales and obtain deep features via a multi-layer network structure. In a recent study, Wijaya et al. proposed a new PointNet-based point cloud feature learning network: PointStack [15]. It features multi-resolution feature learning and learnable pooling, which further reduces the information loss during sampling. PointCNN [5] proposes a new convolution operator that converts input points and features to data on a regular grid and then applies convolution. Similarly, PointConv [6] treats convolution kernels as nonlinear functions of the local coordinates of points composed of weight and density functions.
The traditional approach has two main shortcomings. One is that its global or local information is easily lost, leading to poor generality and difficulty in adapting to complex tasks. Second, due to the disorderly nature of point clouds, the use of convolution methods must consider transforming their inputs [16], which leads to further information loss as well as computational cost.

2.2. Transformer-Based Point Cloud Learning Methods

The transformer-based methods are mainly divided into: local transformer and global transformer. Local transformers are designed to implement feature aggregation in local patches rather than the whole point cloud. PT [17] adopted the PointNet++ hierarchical architecture for point cloud classification and segmentation. It focused on local patch processing, and replaced the shared MLP modules in PointNet++ with local transformer blocks. Compared to the local transformer, the global transformer can connect each output feature to all input features and can learn global contextual features. PCT [18], as a pure global transformer network, was proposed in. PCT takes 3D coordinates as input, and employs a neighborhood-embedding structure to map the point cloud to a high-dimensional feature space. 3CROSSNet [19] is the first transformer network that can extract features across scales in point cloud learning. It consists of three components, which can perform self-attention operations on point clouds at different levels and scales, respectively, to extract rich features.
The existing transformer models above consider the point as an isolated element, which can be described as a feature vector. Then, the dot product or addition relations between two points are employed for guiding the generation of feature maps. As a matter of fact, these unordered points are not standalone, and they form a meaningful subset, which can be regarded as certain concepts. We should seek relationships between concepts in their commonalities to characterize them via approximate methods rather than specifying the dot product or addition function.

2.3. Other Advanced Point Cloud Learning Methods

In addition to research focusing on modifications to the overall feature extraction network, several state-of-the-art deep learning methods have been applied in practical 3D LiDAR point cloud applications. Li [20] examined a rapid deterministic point set registration method that utilizes gravity prior information, improving the speed and accuracy of point cloud registration. A self-attention and orientation coding network was constructed by Yan et al. [21], which integrates long-distance context into point local descriptors, fully explores the relationship between points, and solves the problem of position recognition of point cloud data. Wu and Tian et al. [22] proposed the Point Prompt Training framework to support 3D representation learning for multi-dataset collaborative learning and overcome the negative transfer related to collaborative learning. Zhu and Yang [23] introduced a comprehensive 3D pre-training framework aimed at enabling the acquisition of efficient 3D representations through the learning of point cloud representations via differentiable neural rendering.
Aligning with recent research trends in point cloud registration, scene recognition, large-scale 3D representation learning, and general pretraining paradigms. It is evident that fundamental theoretical disciplines like physics and mathematics offer substantial theoretical support and technical inspiration for deep learning methods. Therefore, addressing the challenges in existing point cloud learning through mathematical approaches holds significant research value and significance.

3. Method

In this section, we will begin by introducing the redefinition of granulation and lower-approximation operators to make them applicable to deep learning networks. Building upon this, a rough set-based attention mechanism is developed. Subsequently, based on this foundation, we construct a transformer network for point cloud learning.

3.1. The Rough Set-Based Attention Mechanism

For the neighborhood rough set that we employ, the mutually exclusive information granules generated by neighborhood relations serve as the fundamental units for approximation. Below, we first introduce the fundamental properties of the neighborhood relations adopted.
Definition 1 ([24]). 
Let U be a non-empty set, if, for any element x i , x j , x k in U, there exists a uniquely determined real function Δ corresponding to it, and Δ satisfies:
(1) Δ ( x i , x j ) 0. When and only when x i = x j , Δ ( x i , x j ) = 0.
(2) Δ ( x i , x j ) = Δ ( x j , x i ) .
(3) Δ ( x i , x k ) Δ ( x i , x j ) + Δ ( x j , x k ) .
where Δ is a distance function on U and < U , Δ > is a distance space, also called a metric space.
In the N-dimensional Euclidean space, given any two points x i = ( x 1 i , x 2 i , , x N i ) and x j = ( x 1 j , x 2 j , , x N j ) , the distances are:
Δ x i , x k = l = 1 N x l i x l j 2 1 2
The specific definition of the distance function can have various expressions due to the existence of multiple variables [25], but they all play the role of measuring the distance relationship between any two points in the Euclidean space.
Information granules generated through neighborhood relations can be further guided through upper and lower approximations. Below, we provide the definition of upper and lower approximations employed in our context.
Definition 2 ([24]). 
Given a nonempty finite set U = x 1 , x 2 , , x n on a real space and a neighborhood relation on U domain relation N, we call the NAS = < U , N > as a neighborhood approximation space.
Definition 3 ([24]). 
Given N A S = < U , N > and X ⊆ U, X is in the neighborhood approximation space, the lower and upper approximations of N A S = < U , N > are defined as:
N ̲ X = x i δ x i X , x i U
N ¯ X = x i δ x i X , x i U
It has the following properties:
X U , N ¯ X X N X
In the neighborhood rough set, mutually exclusive information particles generated by neighborhood relations are the basic units used for approximation. It can be said that the two modules of granulation and approximation form the rough set methodology’s cornerstone. Therefore, we will design the self-attention mechanism based on the two modules of granulation and approximation.
The following describes the granulation module. The fuzzy equivalence relation functions R used to measure between samples are diverse. To better extract the commonality among the high-dimensional features of the point cloud data, we generated the relational functions for the granulation operation with Gaussian functions.
R ( x , y ) = e x p ( | | x y | | 2 2 σ 2 )
The Gaussian function satisfies all the properties needed to be defined as a rough set relation function for n feature vectors X of length m.
The granulation matrix has the following properties:
(1) ∀x,yU, R ( x , x ) = 1;
(2) R ( x , y ) = R ( y , x ) ;
(3) R ( x , y ) ∈ [0,1].
The granulation matrix serves as an information bottleneck in this process. It reflects the relationships between objects, expresses the granular structure of the argument domain, and carries all the sample information available to the approximation operation. Essentially, once the relationships between samples are extracted, subsequent rough calculations are performed on the fuzzy information grains constituted between samples rather than on individual samples.
The following describes the approximation module. In the fuzzy calculation of rough sets, the classification of samples is no longer either 0 or 1 in the deterministic sense, but is evaluated in terms of fuzzy affiliation between 0 and 1. For all information grains generated by the relational function, the mutual approximation between different fuzzy information grains can guide the importance of each other, and this property makes the approximation matrix a natural global feature guidance matrix, while the affiliation of the approximation is reflected by the upper and lower approximations of the rough set.
For the relational function k ( x , y ) , the lower and upper approximation affiliations to the approximated information grain d i are:
k S ̲ d i ( x ) = inf y U S N ( k ( x , y ) ) , d i ( y )
k T ¯ d i ( x ) = sup y U T k ( x , y ) , d i ( y )
Among them, where N is the complementary operator, T is the trigonometric norm and S is the trigonometric remainder.
By approximating the different concepts ( d i ) formed through granulation to each other, the fuzzy lower-approximation affiliations of ∀xU affiliation to d through the relational function R, which we construct, can be expressed as:
k S ̲ d i ( x ) = inf y d i min 1 exp x y 2 2 σ 2 , R ( x , y )
The value of the lower approximation characterizes the necessary correlation affiliation between the information grains composed of two features with each other. The granulation operation and approximation operation are calculated as shown in Figure 2.
The approximation operation more objectively and accurately measures the degree of correlation between two information grains, which leads to the correlation of global features.
The rough set-based attention mechanism constructed based on granulation and approximation operations, where ( Q , K , V ) R N × D is generated by shared linear transformations and the input features F. The specific calculation of the rough set-based attention mechanism is as follows.
First, we can use the query and key matrices to calculate the granulation matrix (G) by R:
G = R ( Q , K T )
The granularity matrix is naturally normalized due to the design of the relationship function.
Then, G regenerates a weight matrix A of the same size as the granulation matrix via the approximation operation, whose values characterize the degree of necessary correlation between information grains and are used as the weights of attention. The rough set-based attention mechanism output features F s a are the weighted sums of the value vector using the corresponding attention weights:
F s a = A · V
The whole rough set self-attention computation process is carried out in the form of information grains, and both its granulation and weighted sum are permutation-independent operators. Therefore, the rough set-based attention mechanism is better adapted to the irregularity and disorder of the point cloud, while the correlation between the features is better measured. The overall approximate guided representation of rough set attention is shown in Algorithm 1.
Algorithm 1 Approximate guided representation methods based on rough set
1:
A feature matrix Q = X 1 , , X d (d tokens of length n)
2:
Granulation operation:
3:
for each token X do
4:
    R = X i , X j Calculate each token with the others by Formula (6);
5:
end for
6:
Obtain the granulation matrix G = R 11 , R 12 , , R n n (n information granule g of length n)
7:
Approximation operation:
8:
for each information granule g do
9:
    k S d i j ̲ = g i g j = inf 1 R i j R i j j j The lower approximation is calculated by Formula (9);
10:
end for
11:
Obtain the weight matrix A = k S d 11 ̲ , k S d 12 ̲ , , k S d m n ̲
12:
return A.

3.2. Transformer Network for Point Cloud Learning

The overall network structure is shown in Figure 3. The transformer network based on rough set attention consists of three components. In the Input Embedding module, the coordinate information and features of the point cloud are directly input to the high dimension, which is used to encode the coordinates and features of each point. We mainly refer to PCT for the design here, given an input point cloud ρ R N × d where each N point has a d-dimensional feature description, a d e - dimensional embedded feature F e R N × d e is first learned via the Input Embedding module. The embedded features are then entered into the encode module. And in encoder, multiple rough set-based multi-headed attentions constitute the main forward network for feature extraction. The output of each attention layer is aligned with the dimensionality of the input. Finally, the output of the attention layer is fed through multiple linear layers to the subsequent classification decoder. The architecture of the segmentation network decoder is almost the same as that for the classification network, following most other point cloud segmentation networks. The loss function uses the softmax cross entropy loss.
The rough set attention completes the feature bootstrapping of the global information and extracts the main features. Finally, classification and segmentation are performed by the decoder. Our transformer is framework-based, and it can implement a wide variety of point cloud learning tasks, with a modular self-attention to allow for the direct replacement of traditional self-attention.

4. Experiments

We discuss the performance of the rough set transformer net for two classical point cloud learning tasks: shape classification and part segmentation. The experiments were run on a server equipped with a Tesla V100 GPUs, and the operating system was Ubuntu 20.04.

4.1. Classification on ModelNet40

In this section, we have completed cloud classification experiments on the ModelNet40 dataset, which comprises 40 categories of 3D models. The ModelNet40 dataset includes 9843 training samples and 2468 testing samples. The same sampling strategy as used in PCT was adopted to uniformly sample each object to 1024 points. For the classification task, we employed two evaluation metrics: overall accuracy (OA) and mean accuracy (mA). We also compared the most representative models based on non-transformer and transformer approaches. During training, we utilized random translation, random anisotropic scaling, and random input dropout strategies to augment the input points’ data during training. During testing, no data augmentation or voting methods were used. The SGD optimizer was used for 400 epochs with the batch size 32. We set the initial learning rate to 0.0001 and adopted a cosine annealing schedule to adjust the learning rate at every epoch.
The experimental results are shown in Table 1. It is apparent that our method outperforms most previous models, especially the transformer-based models. Firstly, compared to the original transformer, we conducted two sets of experiments based on addition and dot product relationships. It can be observed from the experimental results that our method achieved a 1.3% and 0.9% improvement in mA, as well as a 2.8% and 1.8% improvement in OA. Compared to the baseline model PCT, our model makes a 0.5% improvement, respectively, on overall ACC. Additionally, it achieves the result of 90.8% mean accuracy, which is 0.8% higher than the PCT. The experimental results visually verify the success of our design of a transformer model based on rough set theory. Most point input-based models, such as the PointNet family, have only 92% OA results and perform even worse on mA. The transformer-based methods are generally better than the non-transformer-based methods, generally achieving a 93% OA. This is due to the transformer’s ability to learn global information. There are also a few methods, such as PointConv and Point Transformer, which rely on increasing the number of inputs or fusing voxels to improve accuracy. Compared to these, our model achieves the best performance in terms of OA and mA, while still having great potential for improvement.
In order to emphasize the advantages of the neighborhood granulation we employed, the experimental results for other conventional rough set granulation methods are shown in Table 2. Obviously, the overall performance of the rough set-based transformer model is significantly superior to the traditional transformer. It is especially reflected in the mA metric, which is attributed to the tolerance of the rough set approximation-guided representation to the uncertain information. It is worth clarifying that different granularization relations will be the most appropriate choice depending on the characteristics of different data. Since the attention of the point cloud has a strong influence on the neighboring points, the granulation function of the Gaussian kernel function obtains the best performance.
In Table 3, the degree of fit of the model to the data is evaluated, reflecting the accuracy of the model in classifying point cloud data. From the results, it can be observed that our method achieves the lowest error, which further validates the superiority of the method.
The disordered and unstructured point clouds make it challenging to quickly capture semantics information directly based on local models such as CNNs. The transformer-based models are competent because of its superior capability in capturing long-range dependencies. However, the non-local method is still inadequate for the complex point cloud data. Our rough set attention forms the concepts of features via the granulation of a cluster of tokens rather than an isolated feature vector, which is more objective than the recent transformer model. Then, it expresses the similarities and differences between concepts through a lower-approximation operator. The experimental results show that our method is more suitable for the uncertain characterization of point cloud classification.

4.2. Part Segmentation on ShapeNet

The model is also evaluated on ShapeNet dataset for part segmentation. ShapeNet consists of a total of 16,880 models, 14,006 of which are used for training and the rest for testing. In our experiment, we directly used the same train–test segmentation as PCT. 2048 points were sampled from each model as input, and only a few point sets had six parts with labels. In the segmentation task, we used mean Intersection-over-Union (mIoU) as an assessment metric. We selected representative networks using point input and self-attention as the main feature extraction for comparison. A training setup essentially identical to our classification task was used. We train our model using the SGD optimizer for 500 epochs with the batch size 16. The initial learning rate was set to 0.0001, with a cosine annealing schedule to adjust the learning rate at every epoch.
The experimental results are shown in Table 4. Similarly, the transformer-based models have obtained very good results. Models that use self-attention and other variants are primarily compared. In order to highlight the impact of our attention, we did not add any other modules and still obtained the best performance. Comparing the baseline models NPCT and SPCT for the same training case. The NPCT and SPCT underperformed us by 0.8 and 0.2, respectively. For the classical PointNet family, we obtained a boost of 0.9–2.3 mIoU. Our rough set-based self-attention achieves the best performance when using only the self-attention and transformer framework.
It is certain that our transformer yielded better performances because of the non-local feature extraction. Nevertheless, the existing transformer models consider the point as an isolated element, which can be described as a feature vector. Then, the dot product or addition relations between two points are employed to guide the generation of feature maps. On the contrary, our granular relation is able to describe the commonalities among multiple points. Then, we characterize their commonalities via approximate methods rather than specifying the dot product or addition function. This is the key reason as to why we have a better mIoU result.
Efficiency is a topic of concern. To this end, we add a comparative experiment between the traditional transformer model (PCT) and the rough set-based transformer in terms of efficiency. Herein, we display our time tests on the RTX2060 with a batch size of 16, shown in Table 5.
Clearly, the rough set-based transformer is more time-consuming than the traditional transformer (dot product attention) due to its intricate algorithm design and additional computations. Nevertheless, we have optimized it to reduce the time required, incorporating algorithm enhancements and developing CUDA parallel computing libraries. At present, the time required by our method is nearly equivalent to that of the traditional transformer. We have open-sourced the code and library files for this approach as part of our contribution. As a result, our approach offers a more comprehensive performance improvement, leading to a superior overall efficiency compared to the traditional transformer.

4.3. Visualization Analysis Experiments

To further analyze the improvements of our model, we visualized the weight matrix after its approximation operation. As shown in Figure 4, we performed one shallow-layer and one deep-layer visualization on the ModelNet40 and ShapeNet datasets. The rough set attention and dot product attention are employed for the comparison analysis.
The feature maps both on the shallow layer and the deep layer are all more obviously brighter than dot product attention. This demonstrates that rough set attention has a stronger global information extraction capability than dot product attention [38]. The traditional attention mechanism further suffers from information loss in the deep network. The disordered and unstructured point clouds result in uncertainty which the traditional attention is incapable of handling. On the contrary, rough set attention can deal with this uncertainty well by measuring the similarity in the view of approximation. It greatly reduces the effect via uncertain point cloud data.
To further validate the rough set-based transformer model’s ability to handle uncertainty information, we visualize the classification effect of a set of point cloud data with random unordered inputs. The results are shown in Figure 5, whereby the input comprises 1024 point cloud data, totaling 40 groups of different types of labels. The distance between the point clouds represents the degree of model recognition: the more obvious dispersion indicates a better recognition effect. The upper part shows the traditional transformer method; it can be seen that our method obviously has a better classification effect on uncertainty features than the traditional method.
Occlusion is also a common challenge that point cloud data often encounter in real-world scenarios. To validate our effective handling of such uncertainty features, we devised three levels (10%, 30%, 50%) of feature masking to assess the model’s learning performance, and the results of the comparative experiments with the traditional transformer are presented in Figure 6. Upon analyzing the experimental results, it becomes evident that as occlusion intensifies, the performance of the dot product attention-based transformer significantly deteriorates, whereas RST displays robustness in handling occluded features. This is attributed to the rough set-based attention mechanism’s ability to tolerate occluded tokens, resulting in a more objective guidance.
After experimental comparative analysis, we verify that the lower-approximation operator is representational, and also prove that it is advantageous compared to existing operators such as multiplicative operators. The reason for this analysis can be attributed to the fact that the approximation operator is able to characterize the uncertain and inconsistent points or problems inside the point cloud, and has tolerance for these uncertain relations, so it is superior compared to the multiplicative relations.

5. Conclusions

Recently, transformer models have exhibited excellent performance in addressing the challenge of point cloud learning. Nonetheless, the disorder and lack of structure in point clouds introduce uncertainty, which needs to be addressed within the traditional transformer framework. In this study, we introduce a novel rough set attention mechanism and extensively investigate the rough set transformer (RST) network. It is able to accurately classify and segment the point cloud data collected using LiDAR. RST has a more objective relationship guidance ability for the uncertain point cloud. All the experimental results have shown that our method yielded a better performance than the most advanced networks available.

Author Contributions

X.S.: Conceptualization, methodology, software, data curation, writing—original draft, writing—review and editing. K.Z.: Conceptualization, project administration, funding acquisition, writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Major Science and Technology Projects in Yunnan Province (202202AD080013, 202302AG050009), Yunnan Fundamental Research Projects (202301AV070003), Talent Training Fund of Kunming University of Science and Technology (KKZ3202303009).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, Y.; Zuo, Y.; Du, Z.; Song, X.; Luo, T.; Hong, X.; Wu, J. MInet: A Novel Network Model for Point Cloud Processing by Integrating Multi-Modal Information. Sensors 2023, 23, 6327. [Google Scholar] [CrossRef] [PubMed]
  2. Villa, F.; Severini, F.; Madonini, F.; Zappa, F. SPADs and SiPMs arrays for long-range high-speed light detection and ranging(LiDAR). Sensors 2021, 21, 3839. [Google Scholar] [CrossRef] [PubMed]
  3. Xie, Z.; Chen, J.; Peng, B. Point clouds learning with attention-based graph convolution networks. Neurocomputing 2020, 402, 245–255. [Google Scholar] [CrossRef]
  4. Li, Y.; Bu, R.; Sun, M.; Wu, W.; Di, X.; Chen, B. Pointcnn: Convolution on x-transformed points. In Proceedings of the Annual Conference on Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018. [Google Scholar]
  5. Atzmon, M.; Maron, H.; Lipman, Y.; Feldman, Y.A. Point Convolutional Neural Networks by Extension Operators. ACM Trans. Graph. 2018, 37, 1–12. [Google Scholar] [CrossRef]
  6. Wu, W.; Qi, Z.; Li, F. PointConv: Deep Convolutional Networks on 3D Point Clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 9621–9630. [Google Scholar]
  7. Pattaraintakorn, P.; Cercone, N. Integrating rough set theory and medical applications. Appl. Math. Lett. 2008, 21, 400–403. [Google Scholar] [CrossRef]
  8. Zhao, F.F.; Pang, B.; Mi, J.S. A new approach to generalized neighborhood system-based rough sets via convex structures and convex matroids. Inf. Sci. 2022, 612, 1187–1205. [Google Scholar] [CrossRef]
  9. Wu, Z. 3D ShapeNets: A Deep Representation for Volumetric Shapes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 8–10 June 2015; pp. 1912–1920. [Google Scholar]
  10. Yi, L. A scalable active framework for region annotation in 3D shape collections. ACM Trans. Graph. 2016, 35, 1–12. [Google Scholar] [CrossRef]
  11. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  12. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. In Proceedings of the Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5099–5108. [Google Scholar]
  13. Peyghambarzadeh, S.M.M.; Azizmalayeri, F.; Khotanlou, H.; Salarpour, A. Point-PlaneNet: Plane kernel based convolutional neural network for point clouds analysis. Digit. Signal Process. 2020, 98, 102633. [Google Scholar] [CrossRef]
  14. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic graph cnn for learning on point clouds. ACM Trans. Graph. 2020, 38, 1–12. [Google Scholar] [CrossRef]
  15. Wijaya, K.T.; Paek, D.; Kong, S. Advanced feature learning on point clouds using multi-resolution features and learnable pooling. arXiv 2022, arXiv:2205.09962. [Google Scholar]
  16. Boulch, A. ConvPoint: Continuous convolutions for point cloud processing. Comput. Graph. 2020, 88, 24–34. [Google Scholar] [CrossRef]
  17. Zhao, H. Point transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 16259–16268. [Google Scholar]
  18. Guo, M.H.; Cai, J.X.; Liu, Z.N.; Mu, T.J.; Martin, R.R.; Hu, S.M. Pct: Point cloud transformer. Comput. Vis. Media 2021, 7, 187–199. [Google Scholar] [CrossRef]
  19. Han, X.; He, Z.; Chen, J.; Xiao, G. 3CROSSNet: Cross-level cross-scale cross-attention network for point cloud representation. IEEE Robot. Autom. Lett. 2022, 7, 3718–3725. [Google Scholar] [CrossRef]
  20. Xinyi, L.; Yinlong, L.; Yan, X.; Venkatnarayanan, L.; Hu, C.; Feihu, Z.; Uwe, S.; Alois, K. Fast and deterministic (3+1)DOF point set registration with gravity prior. ISPRS J. Photogramm. Remote. Sens. 2023, 199, 118–132. [Google Scholar]
  21. Xia, Y.; Xu, Y.; Li, S.; Wang, R.; Du, J.; Cremers, D.; Stilla, U. SOE-Net: A Self-Attention and Orientation Encoding Network for Point Cloud based Place Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Kuala Lumpur, Malaysia, 18–20 December 2021; pp. 11343–11352. [Google Scholar]
  22. Wu, X.; Tian, Z.; Wen, X.; Peng, B.; Liu, X.; Yu, K.; Zhao, H. Towards Large-scale 3D Representation Learning with Multi-dataset Point Prompt Training. arXiv 2023, arXiv:2308.09718. [Google Scholar]
  23. Zhu, H.; Yang, H.; Wu, X.; Huang, D.; Zhang, S.; He, X.; He, T.; Zhao, H.; Shen, C.; Qiao, Y.; et al. PonderV2: Pave the Way for 3D Foundation Model with A Universal Pre-training Paradigm. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver Convention Center, Vancouver, BC, Canada, 18–22 June 2023. [Google Scholar]
  24. Yu, D.; Hu, Q.; Bao, W. Combining multiple neural networks for classification based on rough set reduction. Int. Conf. Neural Netw. Signal Process. 2003, 1, 543–548. [Google Scholar]
  25. Hu, Q.; Yu, D.; Liu, J.; Wu, C. Neighborhood rough set based heterogeneous feature subset selection. Inf. Sci. 2008, 178, 3577–3594. [Google Scholar] [CrossRef]
  26. Liu, X. Point2Sequence: Learning the Shape Representation of 3D Point Clouds with an Attention-based Sequence to Sequence Network. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; pp. 8778–8785. [Google Scholar]
  27. Lee, D.; Lee, J.; Lee, J.; Lee, H.; Lee, M.; Woo, S.; Lee, S. Regularization Strategy for Point Cloud via Rigidly Mixed Sample. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Kuala Lumpur, Malaysia, 18–20 December 2021; pp. 15895–15904. [Google Scholar]
  28. Yang, D.; Gao, W. PointManifold: Using Manifold Learning for Point Cloud Classification. In Proceedings of the IEEE/CVF International Conference on Image Processing, New York, NY, USA, 7–20 February 2020; pp. 1–5. [Google Scholar]
  29. Sun, C.; Zheng, Z.; Wang, X.; Xu, M.; Yang, Y. Self-supervised Point Cloud Representation Learning via Separating Mixed Shapes. IEEE Trans. Multimed. 2022, 25, 6207–6218. [Google Scholar] [CrossRef]
  30. Yan, X.; Zheng, C.; Li, Z.; Wang, S.; Cui, S. PointASNL: Robust Point Clouds Processing Using Nonlocal Neural Networks with Adaptive Sampling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5588–5597. [Google Scholar]
  31. Yu, J.; Zhang, C.; Wang, H.; Zhang, D.; Song, Y.; Xiang, T.; Liu, D.; Cai, W. 3D Medical Point Transformer: Introducing Convolution to Attention Networks for Medical Point Cloud Analysis. arXiv 2021, arXiv:2112.04863. [Google Scholar]
  32. Cheng, Z.; Wan, H.; Shen, X.; Wu, Z. PatchFormer: An Efficient Point Transformer with Patch Attention. In Proceedings of the IEEE/CVF International Conference on Computer Vision, New Orleans, LA, USA, 19–24 June 2022; pp. 11789–11798. [Google Scholar]
  33. Huang, Z.; Zhao, Z.; Li, B.; Han, J. LCPFormer: Towards Effective 3D Point Cloud Analysis via Local Context Propagation in Transformers. arXiv 2022, arXiv:2210.12755. [Google Scholar] [CrossRef]
  34. Liu, Y. Relationshape convolutional neural network for point cloud analysis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8887–8896. [Google Scholar]
  35. Xie, X.; Zheng, J.; Gao, Z. Attentional ShapeContextNet for Point Cloud Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–26 June 2018; pp. 4606–4615. [Google Scholar]
  36. Li, Y.; Wang, J.; Zhang, J.; Chen, H. DT-Net: Dynamic Transformation Network for Point Cloud Completion. arXiv 2021, arXiv:2109.06385. [Google Scholar]
  37. Qiu, S.; Anwar, S.; Barnes, N. Geometric back-projection network for point cloud classification. IEEE Trans. Multimed. 2021, 24, 1943–1955. [Google Scholar] [CrossRef]
  38. Raghu, M.; Unterthiner, T.; Kornblith, S.; Zhang, C.; Dosovitskiy, A. Do vision transformers see like convolutional neural networks? In Proceedings of the Conference on Neural Information Processing Systems, Online, 6–14 December 2021; pp. 12116–12128. [Google Scholar]
Figure 1. The relation representation of point cloud data. (a) The dot product or addition relations between two isolated feature vectors in classical transformer. (b) The approximation relations after granulation in the view of rough set theory.
Figure 1. The relation representation of point cloud data. (a) The dot product or addition relations between two isolated feature vectors in classical transformer. (b) The approximation relations after granulation in the view of rough set theory.
Sensors 23 09042 g001
Figure 2. The granulation operation and approximation operation. The relationship function guides the commonality between each feature, and the generated granulation matrix contains all the information about each feature with other features. The mutual approximation between concepts leads to the degree to which information grains are necessarily related to each other.
Figure 2. The granulation operation and approximation operation. The relationship function guides the commonality between each feature, and the generated granulation matrix contains all the information about each feature with other features. The mutual approximation between concepts leads to the degree to which information grains are necessarily related to each other.
Sensors 23 09042 g002
Figure 3. Architecture of RST. The encoder mainly comprises an Input Embedding module and multiple multi-headed attention modules. The decoder mainly comprises multiple linear layers for classification and segmentation.
Figure 3. Architecture of RST. The encoder mainly comprises an Input Embedding module and multiple multi-headed attention modules. The decoder mainly comprises multiple linear layers for classification and segmentation.
Sensors 23 09042 g003
Figure 4. Attention visualization on the ModelNet40 and ShapeNet dataset. (a,b,e,f) Heat maps based on the rough set. (c,d,g,h) Heat maps based on the traditional transformer.
Figure 4. Attention visualization on the ModelNet40 and ShapeNet dataset. (a,b,e,f) Heat maps based on the rough set. (c,d,g,h) Heat maps based on the traditional transformer.
Sensors 23 09042 g004
Figure 5. Visualization of classification results. The upper part shows the point cloud classification effect of common transformer models, and the lower part shows the effect of a rough set-based transformer model. The better the dispersion effect, the better the classification result.
Figure 5. Visualization of classification results. The upper part shows the point cloud classification effect of common transformer models, and the lower part shows the effect of a rough set-based transformer model. The better the dispersion effect, the better the classification result.
Sensors 23 09042 g005
Figure 6. Comparison experiments between traditional transformer and RST with different occlusion levels.
Figure 6. Comparison experiments between traditional transformer and RST with different occlusion levels.
Sensors 23 09042 g006
Table 1. Comparison on the ModelNet40 classification dataset. * represents the baseline.
Table 1. Comparison on the ModelNet40 classification dataset. * represents the baseline.
MethodInputInput SizeOA (%)mA (%)
PointNet [11]P1024 × 389.286.2
PointNet++ [12]P,N5120 × 691.9
PlaneNet [13]P1024 × 392.190.5
PointConv [5]P,N1024 × 692.588.1
PointCNN [6]P1024 × 392.288.1
DGCNN [14]P2048 × 693.590.7
Point2Seq [26]P1024 × 392.290.4
RSMix [27]P1024 × 393.5
Manifold [28]P2048 × 693.090.4
PointStack [15]P1024 × 393.389.6
DGCNN+MD [29]P1024 × 393.389.99
OGNet+MD [29]P1024 × 393.3990.71
PointASNL [30]P,N1024 × 693.31
Add-attentionP1024 × 392.488.0
Bmm-attentionP1024 × 392.889.0
PCT [18] *P1024 × 393.290.0
3DMedPT [31]P1024 × 393.4
3CROSSNet [19]P1024 × 393.5
PatchFormer [32]P,N1024 × 693.6
PT [17]P,N1024 × 693.790.6
LCPFormer [33]P1024 × 393.690.7
RST (ours)P1024 × 393.790.8
Table 2. Comparative experiments with other granulation relations.
Table 2. Comparative experiments with other granulation relations.
Granulation RelationsOA (%)mA (%)
Dominant Relationship93.090.0
Euclidean Norm93.290.2
Multiquadric Kernel93.690.4
Gaussian Kernel93.890.8
Table 3. Normal estimation average cosine distance error.
Table 3. Normal estimation average cosine distance error.
MethodPointsError (%)
PointNet [11]1k0.47
PointNet++ [12]1k0.29
PCNN [5]1k0.19
RS-CNN [34]1k0.15
PCT [18]1k0.13
RST (ours)1k0.11
Table 4. Comparison on the ShapeNet part segmentation dataset. * represents the baseline.
Table 4. Comparison on the ShapeNet part segmentation dataset. * represents the baseline.
MethodmIoU (%)
PointNet [11]83.7
3DMedPT [31]84.3
ShapeContextNet [35]84.6
PointNet++ [12]85.1
P2Sequence [26]85.1
DGCNN [14]85.2
DT-Net [36]85.6
PointConv [5]85.7
3CROSSNet [19]85.9
CAA [37]85.9
PT [17]85.9
NPCT [18] *85.2
SPCT [18] *85.8
PointCNN [6]86.1
RS-CNN [34]86.2
RST (ours)86.5
Table 5. Comparison experiments between PCT and RST with efficiency.
Table 5. Comparison experiments between PCT and RST with efficiency.
MethodEpochs (Convergence)Average Training TimeAverage Testing Time
PCT25079.13 s/epoch8.58 s
RST (ours)30079.58 s/epoch8.67 s
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sun, X.; Zeng, K. RST: Rough Set Transformer for Point Cloud Learning. Sensors 2023, 23, 9042. https://doi.org/10.3390/s23229042

AMA Style

Sun X, Zeng K. RST: Rough Set Transformer for Point Cloud Learning. Sensors. 2023; 23(22):9042. https://doi.org/10.3390/s23229042

Chicago/Turabian Style

Sun, Xinwei, and Kai Zeng. 2023. "RST: Rough Set Transformer for Point Cloud Learning" Sensors 23, no. 22: 9042. https://doi.org/10.3390/s23229042

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop