Next Article in Journal
Evaluation of Emission Characteristics and Microstructure of Particulate Matters from Excavation and Restoration Work on Asphalt Concrete Pavement
Next Article in Special Issue
Unsupervised Deep Embedded Clustering for High-Dimensional Visual Features of Fashion Images
Previous Article in Journal
Photomechanical Solid Polymers: Model for Pressure and Strain Induced by Photoisomerization and Photo-Orientation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Density Peaks Clustering Algorithm with Isolation Kernel and K-Induction

1
School of Cyber Security and Computer, Hebei University, Baoding 071002, China
2
Hebei Machine Vision Engineering Research Center, Hebei University, Baoding 071000, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(1), 322; https://doi.org/10.3390/app13010322
Submission received: 29 November 2022 / Revised: 17 December 2022 / Accepted: 21 December 2022 / Published: 27 December 2022

Abstract

:
Density peaks clustering (DPC) algorithm can process data of any shape and is simple and intuitive. However, the distance between any two high-dimensional points tends to be consistent, which makes it difficult to distinguish the density peaks and easily produces “bad label” delivery. To surmount the above-mentioned defects, this paper put forward a novel density peaks clustering algorithm with isolation kernel and K-induction (IKDC). The IKDC uses an optimized isolation kernel instead of the traditional distance. The optimized isolation kernel solves the problem of converging the distance between the high-dimensional samples by increasing the similarity of two samples in a sparse domain and decreasing the similarity of two samples in a dense domain. In addition, the IKDC introduces three-way clustering, uses core domains to represent dense regions of clusters, and uses boundary domains to represent sparse regions of clusters, where points in the boundary domains may belong to one or more clusters. At the same time as determining the core domains, the improved KNN and average similarity are proposed to assign as many as possible to the core domains. The K-induction is proposed to assign the leftover points to the boundary domain of the optimal cluster. To confirm the practicability and validity of IKDC, we test on 10 synthetic and 8 real datasets. The comparison with other algorithms showed that the IKDC was superior to other algorithms in multiple clustering indicators.

1. Introduction

As the volume and dimensionality of data continue to increase, the focus has moved to how to analyze and process this data effectively. As a powerful data analysis tool [1,2], clustering has been widely applied in pattern recognition [3,4], intrusion detection [5,6], medicine [7,8], sensors [9,10], and many other fields.
Clustering can generally be categorized by partition [11,12], hierarchy [13,14], density [15,16,17,18], grid [19,20], model [21], and graph [22]. k-means [11] is a partition-based clustering algorithm. Although this algorithm can handle complex data well and has low complexity, it has many disadvantages; for example, it depends heavily on the input parameter k and is sensitive to outliers. To find the optimal parameter k, Dinh et al. [23] used the silhouette coefficient to evaluate the quality of different clusters to find the optimal solution. Density-based clustering can also avoid the above problems well. It finds the cluster center and noise by calculating the density of each sample point and avoids the influence of setting the number of clusters and noise in advance on the results. Some of the more classic ones are density-based spatial clustering of applications with noise (DBSCAN) [24] and DPC [25].
DBSCAN [24] can discover arbitrarily shaped clusters in the noise space with just two parameters, making the clustering results almost independent of the traversal order of the nodes. However, DBSCAN strongly depends on the distance and is not suitable for datasets with small density differences. However, DBSCAN has a strong dependence on the parameter and is not suitable for datasets with small differences in density. DBSCAN is only effective for spherical data and has a poor impact on aspheric data. Campello et al. [26] proposed the hierarchical density-based spatial clustering of applications with noise (HDBSCAN). It combined DBSCAN with hierarchical clustering, redefined the distance between samples, and solved the problem that DBSCAN relies heavily on input parameters. However, this algorithm does not work well with boundary points.
DPC was proposed by Alex et al. [25]. It eliminates the drawbacks of DBSCAN and HDBSCAN, which are only valid for spherical data, can find density peak points with fewer parameters, and can efficiently assign points and remove noise. Meanwhile, DPC can automatically find the cluster centers and realize efficient clustering of data of any shape. However, there are still some drawbacks with DPC, mainly in two aspects: (1) poor applicability in high-dimensional spaces. With the advent of massive data, the dimensionality of the data is growing exponentially. When the dimensionality of data is high, it can be proven by the theorem of large numbers that the sum of independent random variables converges to a fixed value when the size of the variables is large, at which time the distances between high-dimensional samples tend to be equal. This means that the traditional distance formula becomes meaningless when calculating the distance between high-dimensional samples, which leads to a poor clustering effect; (2) error message delivery. The leftover points are directly allocated to the cluster in which the closest and denser point is located, without consideration of the information of the neighboring points. If a point is incorrectly assigned during this process, this error will be passed on to subsequent points, causing points that should not belong to a cluster to be assigned to a cluster, thus resulting in clustering failure; (3) deviation from human semantic interpretation. DPC belongs to hard clustering and does not take human semantic interpretation into account.
Aiming at the poor practicability of DPC in high-dimensional data, Xu et al. [27] introduced the grid granularity framework into DPC to address the poor clustering of DPC on large-scale datasets. Du et al. [28] introduced Principal Component Analysis (PCA) into the KNN-DPC model to perform dimensionality reduction operations on high-dimensional data before clustering. Although the above two algorithms have good feasibility and can obtain satisfactory results on relatively high-dimensional datasets, when the dataset presents vertical stripes, the clustering can be reduced to poor results. Hu et al. [29] used PCA, autoencoder (AE), and t-distributed stochastic neighbor embedding (t-SNE) to automatically extract features from high-dimensional data and then used Quasi-Monte Carlo (QMC) to estimate the distribution of low-dimensional samples to obtain clustering results. Mohamed et al. [30] first convert high-dimensional data into a two-dimensional space using t-SNE and cluster low-dimensional data using mutual nearest neighbors. The above two algorithms both use dimensionality reduction technology to process high-dimensional data, which can easily cause information loss. Wang et al. [31] provided feasible channels for data of different dimensions to achieve dimensionality reduction of complex data.
Aiming at the “bad label” delivery in the DPC, Yu et al. [32] improved the validity of DPC by collecting information on the dynamics of nearest neighbors. Li et al. [33] put all data points into polar coordinates, calculated the distance and cosine values between any two points, and finally output the results with a sparse matrix. It can be efficient to minimize the running time of DPC. Liu et al. [34] proposed a two-step clustering using shared nearest-neighbor information. Sun et al. [35] redefined local density by taking into account information about nearest neighbors. Chen et al. [36] grouped samples with k of the same nearest neighbors into a cluster. Seyedi et al. [37] established a similarity fusion matrix to dynamically update the label information. All the above algorithms consider only the K nearest neighbors around the leftover points and lack consideration of global information. Yu et al. [38] proposed the use of evidence theory to collect and fuse neighbor information of the leftover samples, which can effectively solve the error message delivery. However, it is difficult to solve the conflict problem, and the fusion accuracy is not enough.
To surmount the above-mentioned defects, this paper put forward a novel density peaks clustering algorithm with isolation kernel and K-induction. The following are the major contributions:
  • The isolation kernel is introduced and optimized. In this paper, the traditional distance is replaced by an optimized isolation kernel, which increases the similarity of two samples in a sparse domain and decreases the similarity of two samples in a dense domain. It solves the problem of poor applicability of high-dimensional data, further improves the accuracy of the algorithm, and solves the failure of traditional distance in high-dimensional spaces;
  • Three-way clustering is introduced, and the interval sets of the core domain and boundary domain are used to represent the clustering result. First, the KNN based on the optimized isolation kernel is exploited to find the K neighbors of the core domain center, and the initial core domain is obtained. Then, the range of the initial core domain is continuously expanded through the average similarity to obtain the final core domain;
  • The K-induction similarity is proposed. According to the K-induction similarity, each leftover point is allocated to the boundary domain of the cluster where the neighbor point with the largest K-induction similarity is located, which effectively avoids the domino effect and better solves the “bad label” delivery.
The rest of the paper is as follows: The DPC, the isolation kernel, and the three-way clustering are detailed in Section 2. The IKDC is further elaborated upon in Section 3. The preliminary results are displayed in Section 4 to support the efficiency of the IKDC. The implications and future work are mentioned in Section 5.

2. Preliminaries

The main symbols used in this section are listed in Table 1.

2.1. Density Peaks Clustering

The DPC finds the peak density points by calculating the local density ρ i . The DPC uses them as cluster centers, and then the leftover points are allocated directly to the cluster where the denser and closest points are located. For any point x i , the local density of x i is ρ i , as shown in Equation (1).
ρ i = j = 1 n χ d i j d c
where n is the number of points, d i j is the distance between x i and x j , d c is the cutoff distance, and χ ( x ) is the logical function when x < 0 , χ ( x ) = 1 , otherwise χ ( x ) = 0 .
δ i is the shortest distance between x i and a point denser than it, as shown in Equation (2). If no point with a higher density is found, the local density is made to be the farthest distance in the dataset.
δ i = min i : ρ i < ρ j ( d i j )
According to the above definition, by constructing a decision diagram of δ i relative to ρ i , the points are classified into three different types, i.e., density peaks, normal points, and noise. The density peaks are chosen as the cluster centers, and then the DPC allocates the leftover points to denser and the closest clusters.

2.2. Isolation Kernel

The isolation kernel is a special kernel function that uses the data isolation mechanism to design a similarity measurement method. It can increase the similarity between points in low-density areas and reduce the similarity between points in high-density areas. Different from traditional distances or kernels, isolation kernel [39,40] has no closed expressions and is stemmed directly from a dataset without learning [41]. The similarity between the two points is measured based on the partitions created in the data space.
A key requirement for an isolation kernel is a spatial partitioning mechanism that isolates a point from other points in the sample set. There are many kinds of spatial segmentation mechanisms, the commonly used ones are isolation forest [42,43], Voronoi Diagram [44,45] and isolating hyperspheres [46,47]. The definition of isolation kernel is as follows [39,40]: Suppose H ψ X D represents the set of all admissible partitions of H derived from a given finite dataset X D , where each H is from a random subset X X D , and each point in X has an equal probability of being chosen from X D , and | X | = ψ . The role of the isolation partition θ ( θ H ) is to isolate the point z X from other points in X in a given random subset X . The union of θ covers the entire dataset.
Given a data space X D , the isolation kernel about X D between any two points, x , y X D is defined as the density distribution expectation that all H, x, and y belong to the same isolation partition θ , as shown in Equation (3).
K ψ x , y X D = E H ψ X D [ I ( x , y θ θ H ) ]
where I ( B ) is the indicator function; if B is true, it outputs 1, otherwise, it outputs 0.
In fact, the isolation kernel K ψ is computed through a partition H m H ψ ( X D ) ( m = 1 , 2 , , t ) , as shown in Equation (4).
K ψ ( x , y | X D ) = 1 t m = 1 t θ H m i I ( x θ ) I ( y θ )
where K ψ ( x , y ) [ 0 , 1 ] .

2.3. Three-Way Clustering

Given a data space X D = { x 1 , , x i , , x n } , where x i = { x i 1 , x i 2 , , x i D } , each object x i has D-dimensional attributes. The traditional cluster result is shown as C = { C 1 , , C r , , C k } , where k is the number of clusters, which does not conform to the semantic interpretation of human thinking. Three-way clustering (3w) [48,49,50] introduces the idea of a three-way decision [51,52] and divides the global data space into C k P , C k B , and C k N , which are, respectively, the core domain, boundary domain, and noise domain of the cluster, as shown in Equation (5). The samples in the core domains definitely belong to this cluster, the samples in the boundary domains may or may not belong to this cluster, and the samples in the noise domain definitely do not belong to this cluster. Three-way clustering uses two disjoint sets ( C r P C r B ) to represent a cluster, which is in accordance with human cognition.
C r P C r B C r N = X D

3. The Model of IKDC

As the dimensionality of the data becomes higher and higher, the distances between points tend to be equal, and the traditional Euclidean distance metric fails in high-dimensional space. In addition, DPC will assign the leftover points to the closest and denser clusters. If a point is incorrectly assigned, this “bad label” will be passed to the next leftover point that is assigned, which will have a domino effect. To surmount the above-mentioned defects, this paper will make improvements to the DPC in two directions: (1) introducing a new distance based on isolation kernel; (2) defining the K-induction similarity to assign the leftover points and assigning them to the boundary domain of the cluster with the highest K-induced similarity. In this section, the basic details of the IKDC will be given, and its time complexity will be theoretically analyzed.

3.1. Determining Cluster Centers and Noise Points

When the dimensionality of data is high, it can be proved by the theorem of large numbers that the sum of independent random variables converges to a fixed value when the size of the variables is large, and the distances between any high-dimensional samples tend to be equal. At this time, density peak points cannot be found using the traditional distance formula, resulting in poor clustering results. Therefore, to solve the above problem, this paper introduces the isolation kernel and improves it. The isolation kernel can increase the similarity of samples in the sparse domain and reduce the similarity of samples in the dense domain, effectively solving the problem of consistent similarity between high-dimensional samples. From this, it can be seen that compared with the distance calculation metric used by the DPC, the isolation kernel is more suitable for calculating the distance between high-dimensional samples. In this paper, the space segmentation mechanism of the isolation forest is used to construct the isolation kernel, and function K ψ is used to replace the distance calculation metric in the DPC. The optimizing distance based on the isolation kernel is further obtained, as shown in Equation (6).
Definition 1 
(distance based on isolation kernel). Given a data space X D , the distance d ˜ i j based on the isolation kernel between points x i and x j is given by Equation (6).
d ˜ i j = 1 K ψ ( x i , x j | D ) + 1 t
where t is the number of total trees in the isolation forest.
IKDC uses the distance based on the isolation kernel instead of the traditional distance and then follows the idea of DPC to find cluster centers and noise. The cluster centers and noise are put into the core domains and noise domain Ω , respectively, and the noise domain is removed from the global dataset to put the effect of noise on the clustering results to the minimum.
To overcome the problem that the KNN algorithm fails in high-dimensional space, the improved KNN is obtained by using the distance based on the isolation kernel instead of the distance formula in the KNN algorithm.

3.2. Using Improved KNN and the Average Distance to Determine the Core Domains

After determining the cluster centers and noise points, IKDC first removes the noise points from X D , uses the improved KNN algorithm to determine the initial core domains in the set X D Ω , denoted as q j ( j [ 1 , K ] ) , and then continuously expands the range of the core domain through the average distance to obtain the final core domain.
Definition 2 
(the K-nearest neighbors). Given a data space X D , for any point p in the dataset, its K-nearest neighbor representation is shown in Equation (7).
K p = { q j | j [ 1 , K ] }
where K is the number of nearest neighbors.
Definition 3 
(the average similarity). Given a data space X D and the point p, the average similarity is shown in Equation (8).
d p a v g = 1 K 1 K d ˜ p q j
where d ˜ p q j ( j [ 1 , K ] ) is the distance based on the isolation kernel between point p and its j-th nearest neighbor.
Based on the initial core domain, the final core domain is obtained by continuously expanding the core domain range through the average similarity, as shown in Equation (8). More details are shown in Algorithm 1.
Algorithm 1 Determining the core domain based on the average similarity
Input: cluster center c = { c 1 , , c r , , c k } , data space X D Ω
Output: core domain { c 1 p , c 2 p , , c k p }
      1:
forr = 1, r <= k do
      2:
   find K-nearest neighbors of c r K N N c r ;
      3:
    C r P = [ c r ; K N N c r ] ;
      4:
   initialize Q = K N N c r ;
      5:
   while Q is not empty do
      6:
       select the head point p;
      7:
       find its K-nearest neighbors K N N p = { q 1 , , q j , , q K } ;
      8:
        d p a v g = 1 k 1 K d p q j ;
      9:
       for each point q j in K N N p  do
      10:
          if  d p q j < = d p a v g  then
      11:
             C r P = C r P q j ;
      12:
            add q j to the end of Q;
      13:
          end if
      14:
       end for
      15:
       remove p from the head of Q;
      16:
     end while
      17:
end for
      18:
return { c 1 p , c 2 p , , c k p }

3.3. Using K-Induction to Assign Remaining Points

After the final core domains of the cluster are determined by Algorithm 1, K-induction is used to assign the other remaining points to the boundary region of the determined cluster. The main idea is that, for a point p, the K neighbors of its surrounding can be regarded as K pieces of information, and by fusing this information, p is derived as to which cluster it belongs to. The K-induction expands the search range by integrating neighbor information of its neighbor, which effectively avoids the transfer of bad labels. It can not only avoid the limitation caused by only considering neighbor information, but also reduce the time consumed when searching for global information. Therefore, this paper first calculates the K-induction between the leftover point p and the surrounding K-neighbor and, secondly allocates the leftover points to the cluster where the point with the maximum K-induction is located. At this time, two situations may occur: (1) If there is only one neighbor q j with the maximum K-induction, it allocates the point p to the boundary domains of the cluster where the neighbor with the largest K-induction is located. If q j has not been assigned, it continues to iterate until the most suitable cluster is found; (2) If there are two or more neighbor points with the maximum K-induction, then it assigns the points to the boundary domains of the clusters where these neighbor points are located. If these neighbor points have not been assigned, then it continues to iterate. More details are shown in Algorithm 2.
Algorithm 2 Assign the remaining points to the boundary domain based on K-induction
Input: the set of R remaining points X D Ω C 1 P , C 2 P , , C k P
Output: boundary domain C 1 B , C 2 B , , C k B
      1:
for each remaining point p i  do
      2:   
calculate the neighbor similarity between point p i and the j-th K-nearest-neighbor in the direction of q j using Equations (9)–(12);
      3:
end for
      4:
construct a neighbor similarity matrix S = s p i , q j R × K ;
      5:
while ∃ the maximum s = p i , q j in S do
      6:
    m , n arg max i 1 . R , j 1 , K s p i , q j ;
      7:
    C n B = C n B p m ;
      8:
   for each unassigned point e satisfying p m e , p m K N N e  do
      9:       
calculate the neighbor similarity between point e and the m-th K-nearest-neighbor in the direction of q m using Equations (9)–(12);
    10:  
     update the related values in the matrix S;
    11:
   end for
    12:
end while
    13:
for each fringe point p i  do
    14:     
calculate the neighbor similarity between point p i and the j-th K-nearest-neighbor in the direction of q j using Equations (9)–(12);
    15:
      n arg max i 1 . R , j 1 , K s p i , q j ;
    16:
      C n B = C n B p i ;
    17:
     if the point p is still not assigned then
    18:
         find the assigned point q which is nearest to p;
    19:
         assign p to the boundary domains of the same cluster as q is located;
    20:
   end if
    21:
end for
    22:
return { c 1 p , c 2 p , , c k p }
Definition 4 
(the neighbor difference). Given a data space X D and the point p, the neighbor difference between the K-neighbors of the point p and the j-th K-nearest neighbor q j is shown in Equation (9).
dif p q j = 1 K K p K q j
Definition 5 
(the neighbor weight). Given a data space X D and the point p, the weight of the point p and the j-th K-nearest neighbor q j is shown in Equation (10).
ω p q j = exp α d i f p q j j [ 1 , K ] exp α d i f p q j
where α is an input parameter in ( 0 , 1 ) .
Definition 6 
(the weighted difference). Given a data space X D and the point p, the weighted difference between the point p and the j-th K-nearest neighbor in the direction of q j is shown in Equation (11).
d p ω q j = 1 K ω p q j K p K q j
whererepresents the dot product of two vectors.
Definition 7 
(the K-induction similarity). Given a data space X D and the point p, the K-induction similarity of point p with the j-th K-nearest neighbor q j is shown in Equation (12).
s p , q j = exp d p ω q j
Given the original data distribution and core domain range as shown in Figure 1, the detailed steps for calculating the K-induction similarity of unassigned points are shown in Figure 2, where α = 0.05 , K = 4 .
The K-induction similarity of unassigned point 1 with its four neighbors (2, 3, 4, 5) and the K-induction similarity of neighboring points 2, 3, 4 and 5 with their neighbors are shown in Figure 2. The neighbor difference between unassigned points 1 and 2, 3, 4 and 5 is obtained by Equation (9), and based on this, the neighbor weight is calculated using Equation (10), as shown in Figure 2. Then, the weighted difference of unassigned point 1 from its neighboring points is obtained using Equation (11), as in Figure 2. Finally, the K-induction similarity between unassigned point 1 and its neighboring points is calculated using Equation (12), as in Figure 2.

3.4. The Computation Complexity Analysis of IKDC

The computation complexity of the IKDC is discussed below. The major difference between the IKDC and the DPC lies in Algorithms 1 and 2. Therefore, this paper limits discussion to these two algorithms. In Algorithm 1, for each data point, the distance based on the isolation kernel d ˜ i j is calculated, and the core centers are determined, resulting in O ( n 2 ) complexity. The K-nearest neighbors for each core center are found, resulting in O( K n ) complexity. Algorithm 2 involves finding K-nearest neighbors and calculating K-induction for each of the remaining points. Noticing that K-nearest neighbors have been obtained in Algorithm 1, this algorithm has a computation complexity of O( K n ). In summary, due to K being much smaller than n, IKDC has a computation complexity of O ( n 2 ).

4. Experiment Analysis

To validate IKDC, we carried out experiments on 10 synthetic datasets and 8 real datasets. These datasets for this paper are from http://archive.ics.uci.edu/ml/index.php, accessed on 29 November 2022.
The details of the datasets are shown in Table 2 and Table 3. This paper uses Accuracy (ACC), Normalized Mutual Information (NMI), and Adjusted Rand Index (ARI) [53] to evaluate the performance of the algorithm. The maximum value of the above three indicators is 1. The closer the value of the indicators is to 1, the stronger the clustering performance.
To reflect the experimental results of the algorithm more objectively, in this paper, the variable parameters of different datasets are set within the allowable range, and the most suitable parameters are obtained by repeating the experiment many times. The experiment is repeated 100 times independently according to this method, and the average results are recorded.

4.1. Experiments on Synthetic Datasets

In this subsection, the effectiveness of IKDC on the 10 synthetic datasets in Table 2 will be demonstrated. For ACC, NMI, and ARI, IKDC is compared with DPC [25], DBSCAN [24], KNN-DPC [28], SNN-DPC [34], and FCM [54]. The experimental results of the six algorithms on different datasets are presented in Table 4 and Table 5. Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7 can more intensively display the experimental results of the six different algorithms on the Aggregation, Flame, and Jain datasets, where different clusters are distinguished by color, In Figure 7, the solid dots of the same color represent the core domain, and the dots of the same color and different shapes represent the boundary domain of this cluster. Since DIM512 is a high-dimensional dataset, no graph is formed to show its clustering results in this paper.
As shown in Figure 3a, Figure 4a, Figure 5a, Figure 6a and Figure 7a, the clustering effect of IKDC is improved over the other algorithms on the Aggregation dataset, with DBSCAN having the worst effect. IKDC was 0.024, 0.034, and 0.038 higher than DBSCAN in ACC, NMI, and ARI values, respectively. As shown in Figure 3b, Figure 4b, Figure 5b, Figure 6b and Figure 7b, for the Flame dataset, IKDC, DPC, DBSCAN, and KNN-DPC all perform well, and FCM has the worst clustering result. As shown in Figure 3c, Figure 4c, Figure 5c, Figure 6c and Figure 7c, for dataset Jain, the clustering results of both IKDC and SNN-DPC are superior. As shown in Table 4, on the Spiral dataset, IKDC, DPC, DBSCAN, KNN-DPC, and SNN-DPC all have good clustering effects, and only FCM has the worst clustering effect. Compared with FCM, IKDC improved the ACC, NMI, and ARI values by 0.66, 1, and 1.006, respectively. This shows that after introducing the fuzzy domain, IKDC is more effective.
From Table 5, it can be concluded that IKDC outperforms the other five algorithms in clustering on the DIM512 dataset with higher dimensionality. In comparison with FCM with the worst clustering effect, the ACC, NMI, and ARI values of IKDC are improved by 0.223, 0.296, and 0.314. Therefore, IKDC has a better clustering effect in high-dimensional space.
From Table 4 and Table 5, it can be concluded that IKDC outperforms the other five clustering algorithms for the remaining five datasets, especially when the datasets are unevenly distributed and of high dimensionality.
To sum up, by comprehensively considering the performance of the six algorithms in Table 4 and Table 5 on the three clustering indicators and Figure 3, Figure 4, Figure 5, Figure 6, Figure 7 and Figure 8, it can be concluded that IKDC outperforms the other five algorithms on the 10 synthetic datasets, which verifies that IKDC has a superior clustering effect.

4.2. Experiments on Real Datasets

In this subsection, the validity of IKDC on the eight real datasets in Table 3 will be shown. For ACC, NMI, and ARI, IKDC is compared with DPC, DBSCAN, KNN-DPC, SNN-DPC, and FCM. Table 6 and Table 7 show the comparison results of the six algorithms.
As shown in Table 6 and Table 7, IKDC has markedly more clustering effects for the Ionosphere, Wine, Waveform, Waveform (noise), Libras, WDBC, Spectrometer, and Seeds datasets. On Ionosphere and WDBC datasets, IKDC has improved values in ACC, NMI, and ARI compared to SNN-DPC and FCM. Overall, the clustering performance of IKDC outperforms the other five algorithms in high-dimensional space. On the Libras dataset, although the clustering effect of IKDC is the same as that of DPC, IKDC introduces three-way clustering, which is more consistent with human cognition, and IKDC outperforms the other five algorithms. On the other datasets, IKDC significantly outperforms the other five algorithms.
To sum up, by comprehensively considering the performance of the six algorithms on the three indicators, it is demonstrated that the IKDC outperforms the other five clustering algorithms on the eight real datasets, which verifies that the clustering effect of IKDC is more excellent.

4.3. The Influence of Own Parameters on the Experimental Results

In this subsection, to verify the effect of its own parameters on the clustering effect, we conducted experiments with the Wine dataset as an example, and the experimental results of the control parameter ψ and parameter t are shown in Figure 8. When ψ = len(wine)/5, as the number of isolation trees increases, the number of wrong points becomes less and less. When the number of isolation trees is the same, as ψ increases, the experimental results are better. Therefore, the experiments show that when the number of ψ and t is within the maximum limit, the clustering effect is better.

5. Discussion

The limitations of this paper mainly lie in two points: (1) the need to manually set parameter K in the nearest neighbor; and (2) the calculation of K-induction is more complicated. To ameliorate the above limitations, in the future, we can proceed in the following two orientations: (1) finding the most suitable K through an adaptive method; and (2) a deep neural network can be utilized to assign the leftover points.

6. Conclusions

The proposed IKDC makes two improvements to the DPC. The first improvement is to introduce an isolation kernel to replace the distance calculation metric in the DPC, which effectively solves the “curse of dimensionality” problem and avoids the failure of traditional distance formulas for high-dimensional samples. The second improvement is to use the K-induction to assign the leftover points, which can effectively avoid the transmission of the error message and prevent the domino effect. Through experiments on several high-dimensional datasets, this paper compares the clustering performance of the IKDC with five representative algorithms (including DPC, DBSCAN, KNN-DPC, SNN-DPC, and FCM) for ACC, NMI, and ARI. The IKDC algorithm works better for high-dimensional samples obtained by comparing the experimental results.

Author Contributions

Conceptualization, S.Z. and K.L.; methodology, S.Z.; software, S.Z.; validation, S.Z. and K.L.; formal analysis, S.Z.; investigation, S.Z.; resources, S.Z.; data curation, S.Z.; writing—original draft preparation, S.Z.; writing—review and editing, S.Z. and K.L.; visualization, S.Z. and K.L.; supervision, K.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data used to support the findings of this study are included at http://archive.ics.uci.edu/ml/index.php, accessed on 29 November 2022.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Macia, L. Using Clustering as a Tool: Mixed Methods in Qualitative Data Analysis. Qual. Rep. 2015, 20, 1083–1094. [Google Scholar] [CrossRef]
  2. Assent, I. Clustering high dimensional data. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2012, 2, 340–350. [Google Scholar] [CrossRef]
  3. Jiang, G.; Wang, H.; Peng, J.; Fu, X. Parallelism Network with Partial-aware and Cross-correlated Transformer for Vehicle Re-identification. In Proceedings of the ICMR ’22: International Conference on Multimedia Retrieval, Newark, NJ, USA, 27–30 June 2022. [Google Scholar]
  4. Hou, J.; Pelillo, M.; Yuan, H. Hypergraph Matching via Game-Theoretic Hypergraph Clustering. Pattern Recognit. 2022, 125, 108526. [Google Scholar] [CrossRef]
  5. Otair, M.; Ibrahim, O.; Abualigah, L.; Altalhi, M.; Sumari, P. An enhanced Grey Wolf Optimizer based Particle Swarm Optimizer for intrusion detection system in wireless sensor networks. Wirel. Netw. 2022, 28, 721–744. [Google Scholar] [CrossRef]
  6. Sheikhan, M.; Bostani, H. Hybrid and Modified OPFs for Intrusion Detection Systems and Large-Scale Problems; Academic Press: Pittsburg, PA, USA, 2022; pp. 109–136. [Google Scholar]
  7. Quinodoz, M.; Peter, V.; Cisarova, K.; Royer-Bertrand, B.; Stenson, P.; Cooper, D.; Unger, S.; Superti-Furga, A.; Rivolta, C. Analysis of missense variants in the human genome reveals widespread gene-specific clustering and improves prediction of pathogenicity. Am. J. Hum. Genet. 2022, 109, 457–470. [Google Scholar] [CrossRef]
  8. Wen, J.; Varol, E.; Sotiras, A.; Yang, Z.; Chand, G.B.; Erus, G.; Shou, H.; Abdulkadir, A.; Hwang, G.; Dwyer, D.B.; et al. Multi-scale semi-supervised clustering of brain images: Deriving disease subtypes. Med. Image Anal. 2022, 75, 102304. [Google Scholar] [CrossRef]
  9. Mohan, P.; Subramani, N.; Alotaibi, Y.; Alghamdi, S.; Khalaf, O.; Ulaganathan, S. Improved metaheuristics-based clustering with multihop routing protocol for underwater wireless sensor networks. Sensors 2022, 22, 1618. [Google Scholar] [CrossRef]
  10. Sharma, R.; Vashisht, V.; Singh, U. eeFFA/DE-a fuzzy-based clustering algorithm using hybrid technique for wireless sensor networks. Int. J. Adv. Intell. Paradig. 2022, 21, 129–157. [Google Scholar] [CrossRef]
  11. Yang, Y.; Cai, J.; Yang, H.; Li, Y.; Zhao, X. ISBFK-means: A new clustering algorithm based on influence space. Expert Syst. Appl. 2022, 201, 117018. [Google Scholar] [CrossRef]
  12. Zhang, E.; Li, H.; Huang, Y.; Hong, S.; Zhao, L.; Ji, C. Practical multi-party private collaborative k-means clustering. Neurocomputing 2022, 467, 256–265. [Google Scholar] [CrossRef]
  13. Grabowski, J.; Smoliński, A. The application of hierarchical clustering to analyzing ashes from the combustion of wood pellets mixed with waste materials. Environ. Pollut. 2021, 276, 116766. [Google Scholar] [CrossRef] [PubMed]
  14. Wu, C.; Peng, Q.; Lee, J.; Leibnitz, K.; Xia, Y. Effective hierarchical clustering based on structural similarities in nearest neighbor graphs. Environ. Pollut. 2021, 276, 107295. [Google Scholar] [CrossRef]
  15. Xia, D.; Bai, Y.; Zheng, Y.; Hu, Y.; Li, Y.; Li, H. A parallel SP-DBSCAN algorithm on spark for waiting spot recommendation. Multimed. Tools Appl. 2022, 81, 4015–4038. [Google Scholar] [CrossRef]
  16. Sun, L.; Qin, X.; Ding, W.; Xu, J.; Zhang, S. Density peaks clustering based on k-nearest neighbors and self-recommendation. Int. J. Mach. Learn. Cybern. 2021, 12, 1913–1938. [Google Scholar] [CrossRef]
  17. Zhou, Z.; Si, G.; Sun, H.; Qu, K.; Hou, W. A robust clustering algorithm based on the identification of core points and KNN kernel density estimation. Expert Syst. Appl. 2022, 195, 116573. [Google Scholar] [CrossRef]
  18. Yang, M.; Chang-Chien, S.; Nataliani, Y. Unsupervised fuzzy model-based Gaussian clustering. Inf. Sci. 2019, 481, 1–23. [Google Scholar] [CrossRef]
  19. Li, Y.; Xi, Y.; Lu, Y. Improving the Grid-based Clustering by Identifying Cluster Center Nodes and Boundary Nodes Adaptively. In Proceedings of the 10th International Conference on Pattern Recognition Applications and Methods—ICPRAM, Vienna, Austria, 4 February 2021. [Google Scholar]
  20. Cheng, M.; Ma, T.; Ma, L.; Yuan, J.; Yan, Q. Adaptive grid-based forest-like clustering algorithm. Neurocomputing 2022, 481, 168–181. [Google Scholar] [CrossRef]
  21. García-Escudero, L.; Mayo-Iscar, A.; Riani, M. Constrained parsimonious model-based clustering. Stat. Comput. 2022, 32, 2. [Google Scholar] [CrossRef]
  22. Jiang, G.; Peng, J.; Wang, H.; Mi, Z.; Fu, X. Tensorial Multi-View Clustering via Low-Rank Constrained High-Order Graph Learning. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 5307–5318. [Google Scholar] [CrossRef]
  23. Dinh, D.T.; Fujinami, T.; Huynh, V.N. Estimating the Optimal Number of Clusters in Categorical Data Clustering by Silhouette Coefficient. In Communications in Computer and Information Science; Chen, J., Huynh, V., Nguyen, G.N., Tang, X., Eds.; Springer: Singapore, 2019; pp. 1–17. [Google Scholar]
  24. Ester, M.; Kriegel, H.; Sander, J.; Xu, X. A density-based algorithm for discovering clusters in large spatial databases with noise. Knowl. Discov. Data Min. 1996, 1, 226–231. [Google Scholar]
  25. Rodriguez, A.; Laio, A. Clustering by fast search and find of density peaks. Science 2014, 344, 1492–1496. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Campello, R.J.; Moulavi, D.; Sander, J. Density-based Clustering Based on Hierarchical Density Estimates. In Advances in Knowledge Discovery and Data Mining, Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining; Pei, J., Tseng, V.S., Cao, L., Motoda, H., Xu, G., Eds.; Springer: Berlin, Germany, 2013; pp. 160–172. [Google Scholar]
  27. Xu, J.; Wang, G.; Deng, W. DenPEHC: Density peak based efficient hierarchical clusterin. Inf. Sci. 2016, 373, 200–218. [Google Scholar] [CrossRef]
  28. Du, M.; Ding, S.; Jia, H. Study on density peaks clustering based on k-nearest neighbors and principal component analysis. Knowl.-Based Syst. 2016, 99, 135–145. [Google Scholar] [CrossRef]
  29. Hu, Z.; Wei, X.; Han, X.; Kou, G.; Zhang, H.; Liu, X.; Bai, Y. Density Peaks Clustering Based on Feature Reduction and Quasi-Monte Carlo. Sci. Program. 2022, 2022, 8046620. [Google Scholar] [CrossRef]
  30. Abbas, M.; El-Zoghabi, A.; Shoukry, A. DenMune: Density peak based clustering using mutual nearest neighbor. Pattern Recognit. 2021, 109, 107589. [Google Scholar] [CrossRef]
  31. Wang, H.; Wang, Y.; Zhang, Z.; Fu, X.; Zhuo, L.; Xu, Y.; Wang, M. Kernelized Multiview Subspace Analysis By Self-Weighted Learning. IEEE Trans. Multimed. 2020, 23, 3828–3840. [Google Scholar] [CrossRef]
  32. Yu, D.; Liu, G.; Guo, M.; Liu, X.; Yao, S. Density peaks clustering based on weighted local density sequence and nearest neighbor assignment. IEEE Access 2019, 7, 34301–34317. [Google Scholar] [CrossRef]
  33. Li, C.; Ding, S.; Xu, X.; Du, S.; Shi, T. Fast density peaks clustering algorithm in polar coordinate system. Appl. Intell. 2022, 52, 14478–14490. [Google Scholar] [CrossRef]
  34. Liu, R.; Wang, H.; Yu, X. Shared-nearest-neighbor-based clustering by fast search and find of density peaks. Inf. Sci. 2018, 450, 200–226. [Google Scholar] [CrossRef]
  35. Sun, L.; Qin, X.; Ding, W.; Xu, J. Nearest neighbors-based adaptive density peaks clustering with optimized allocation strategy. Neurocomputing 2022, 473, 159–181. [Google Scholar] [CrossRef]
  36. Chen, M.; Li, L.; Wang, B.; Cheng, J.; Pan, L.; Chen, X. Effectively clustering by finding density backbone based-on kNN. Pattern Recognit. 2016, 60, 486–498. [Google Scholar] [CrossRef]
  37. Seyedi, S.; Lotfi, A.; Moradi, P.; Qader, N. Dynamic graph-based label propagation for density peaks clustering. Expert Syst. Appl. 2019, 115, 314–328. [Google Scholar] [CrossRef]
  38. Yu, H.; Chen, L.Y.; Yao, J.T. A three-way density peak clustering method based on evidence theory. Wiley Knowl.-Based Syst. 2021, 211, 106532. [Google Scholar] [CrossRef]
  39. Qin, X.; Ting, K.; Zhu, Y.; Lee, V. Nearest-neighbour-induced isolation similarity and its impact on density-based clustering. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence, Hilton Hawaiian Village, Honolulu, HI, USA, 27 January–1 February 2019. [Google Scholar]
  40. Radovanovic, M.; Nanopoulos, A.; Ivanovic, M. Hubs in space: Popular nearest neighbors in high-dimensional data. J. Mach. Learn. Res. 2010, 11, 2487–2531. [Google Scholar]
  41. Ting, K.; Washio, T.; Zhu, Y.; Xu, Y. Breaking the curse of dimensionality with Isolation Kernel. arXiv 2021, arXiv:2109.14198. [Google Scholar]
  42. Rajeev, H.; Devi, U. Detection of Credit Card Fraud Using Isolation Forest Algorithm; Springer: Singapore, 2022; pp. 23–34. [Google Scholar]
  43. Tokovarov, M.; Karczmarek, P. A probabilistic generalization of isolation forest. Inf. Sci. 2022, 584, 433–449. [Google Scholar] [CrossRef]
  44. Adhinugraha, K.; Taniar, D.; Phan, T.; Beare, R. Predicting travel time within catchment area using Time Travel Voronoi Di-agram (TTVD) and crowdsource map features. Inf. Process. Manag. 2022, 59, 102922. [Google Scholar] [CrossRef]
  45. Abdallah, W.; Mnasri, S.; Val, T. Distributed approach for the indoor deployment of wireless connected objects by the hy-bridization of the Voronoi diagram and the Genetic Algorithm. arXiv 2022, arXiv:2202.13735. [Google Scholar]
  46. Choudhury, J.; Ky, P.; Ren, Y.; Shi, C. Hypersphere for Branching Node for the Family of Isolation Forest Algorithms. In Proceedings of the 2021 IEEE International Conference on Smart Computing (SMARTCOMP), Irvine, CA, USA, 23 August–27 August 2021. [Google Scholar]
  47. Ting, K.; Xu, B.; Washio, T.; Zhou, Z. Isolation distributional kernel: A new tool for kernel based anomaly detection. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Virtual, 6 July–10 July 2020. [Google Scholar]
  48. Yu, H. Three-way decisions and three-way clustering. In Proceedings of the International Joint Conference on Rough Sets, Quy Nhon, Vietnam, 20–24 August 2018. [Google Scholar]
  49. Wang, P.; Yao, Y. CE3: A three-way clustering method based on mathematical morphology. Knowl.-Based Syst. 2018, 155, 54–65. [Google Scholar] [CrossRef]
  50. Yu, H.; Chen, L.; Yao, J.; Wang, X. A three-way clustering method based on an improved DBSCAN algorithm. Phys. A Stat. Mech. Its Appl. 2019, 535, 122289. [Google Scholar] [CrossRef]
  51. Yao, Y. Three-way decision: An interpretation of rules in rough set theory. In Proceedings of the International Conference on Rough Sets and Knowledge Technology, Gold Coast, Australia, 14–16 July 2009. [Google Scholar]
  52. Yao, Y. The geometry of three-way decision. Appl. Intell. 2021, 51, 6298–6325. [Google Scholar] [CrossRef]
  53. Vinh, N.; Epps, J.; Bailey, J. Information theoretic measures for clusterings comparison: Variants, properties, normalization and correction for chance. J. Mach. Learn. Res. 2010, 11, 2837–2854. [Google Scholar]
  54. Bezdek, J.; Ehrlich, R.; Full, W. FCM: The fuzzy c-means clustering algorithm. Comput. Geosci. 1984, 10, 191–203. [Google Scholar] [CrossRef]
Figure 1. Raw data distribution and core domain range.
Figure 1. Raw data distribution and core domain range.
Applsci 13 00322 g001
Figure 2. The detailed steps of K-induction similarity.
Figure 2. The detailed steps of K-induction similarity.
Applsci 13 00322 g002
Figure 3. Clustering effect of the DBSCAN algorithm on Aggregation, Flame, and Jain.
Figure 3. Clustering effect of the DBSCAN algorithm on Aggregation, Flame, and Jain.
Applsci 13 00322 g003
Figure 4. Clustering effect of the KNN-DPC algorithm on Aggregation, Flame and Jain.
Figure 4. Clustering effect of the KNN-DPC algorithm on Aggregation, Flame and Jain.
Applsci 13 00322 g004
Figure 5. Clustering effect of the SNN-DPC algorithm on Aggregation, Flame, and Jain.
Figure 5. Clustering effect of the SNN-DPC algorithm on Aggregation, Flame, and Jain.
Applsci 13 00322 g005
Figure 6. Clustering effect of the FCM algorithm on Aggregation, Flame, and Jain.
Figure 6. Clustering effect of the FCM algorithm on Aggregation, Flame, and Jain.
Applsci 13 00322 g006
Figure 7. Clustering effect of the IKDC algorithm on Aggregation, Flame, and Jain.
Figure 7. Clustering effect of the IKDC algorithm on Aggregation, Flame, and Jain.
Applsci 13 00322 g007
Figure 8. Influence of own parameters on experimental results.
Figure 8. Influence of own parameters on experimental results.
Applsci 13 00322 g008
Table 1. Symbols and descriptions.
Table 1. Symbols and descriptions.
SymbolDescription
X D global sample points
X a subset of sample points in the universe
x i i-th sample point
ρ i local density of sample point i
δ i distance from x i to the closest point with higher density than it
θ isolation partition
C r P the core domain of the r-th cluster
C r B the boundary domain of the r-th cluster
C r N the noise domain of the r-th cluster
Table 2. Synthetic datasets.
Table 2. Synthetic datasets.
DatasetInstancesDimensionsClusters
Aggregation78827
Flame24022
Jain37322
Pathbased30023
R15600215
Spiral31223
D313100231
DIM512102451216
S25000215
Compound39926
Table 3. Real datasets.
Table 3. Real datasets.
DatasetInstancesDimensionsClusters
Wine178133
WDBC569302
Seeds21073
Libras3609015
Ionosphere351342
Waveform5000213
Waveform (noise)5000403
Spectrometer53110248
Table 4. ACC, NMI, and ARI of six algorithms on different synthetic datasets.
Table 4. ACC, NMI, and ARI of six algorithms on different synthetic datasets.
AlgorithmACCNMIARIAlgorithmACCNMIARI
Aggregation                  Spiral                  
IKDC1.0001.0001.000IKDC1.0001.0001.000
DPC0.9950.9920.990DPC1.0001.0001.000
DBSCAN0.9730.9580.958DBSCAN1.0001.0001.000
KNN-DPC0.9970.9920.996KNN-DPC1.0001.0001.000
SNN-DPC0.9780.9550.959SNN-DPC1.0001.0001.000
FCM0.7780.8250.684FCM0.3400.000-0.006
Flame                  D31                  
IKDC1.0001.0001.000IKDC0.9700.9620.953
DPC1.0001.0001.000DPC0.9680.9580.936
DBSCAN1.0001.0001.000DBSCAN0.9680.9570.935
KNN-DPC1.0001.0001.000KNN-DPC0.9700.9600.940
SNN-DPC0.9980.8990.950SNN-DPC0.9740.9630.974
FCM0.8500.4420.488FCM0.8910.8620.936
Jain                  DIM512                  
IKDC1.0001.0001.000IKDC0.9660.9510.956
DPC0.9810.9760.970DPC0.9440.9400.935
DBSCAN0.9280.8950.890DBSCAN0.8510.7740.749
KNN-DPC0.9700.9600.940KNN-DPC0.9180.8970.890
SNN-DPC1.0001.0001.000SNN-DPC0.9390.8960.926
FCM0.7780.8310.707FCM0.7430.6550.6428
Table 5. ACC, NMI, and ARI of six algorithms on different synthetic datasets.
Table 5. ACC, NMI, and ARI of six algorithms on different synthetic datasets.
AlgorithmACCNMIARIAlgorithmACCNMIARI
Pathbased                  S2                  
IKDC0.9800.9200.916IKDC0.9660.9510.956
DPC0.7530.5550.472DPC0.9440.9400.935
DBSCAN0.8230.7310.613DBSCAN0.8510.7740.749
KNN-DPC0.7600.5610.561KNN-DPC0.9180.8970.890
SNN-DPC0.9770.9010.929SNN-DPC0.9390.8960.926
FCM0.7470.5500.465FCM0.7410.6900.694
R15                  Compound                  
IKDC1.0001.0001.000IKDC0.8850.9130.899
DPC0.9970.9940.993DPC0.8320.8730.833
DBSCAN0.9930.9890.986DBSCAN0.8400.8280.844
KNN-DPC0.9970.9940.993KNN-DPC0.8700.5520.809
SNN-DPC0.9970.9940.993SNN-DPC0.8570.8530.835
FCM0.9970.9650.993FCM0.5010.6190.406
Table 6. ACC, NMI, and ARI of six algorithms on different real datasets.
Table 6. ACC, NMI, and ARI of six algorithms on different real datasets.
AlgorithmACCNMIARIAlgorithmACCNMIARI
Ionosphere                  Waveform                  
IKDC1.0001.0001.000IKDC0.9800.9840.955
DPC1.0001.0001.000DPC0.9680.9580.936
DBSCAN1.0001.0001.000DBSCAN0.5240.1520.135
KNN-DPC1.0001.0001.000KNN-DPC0.6350.2180.223
SNN-DPC0.9980.8990.950SNN-DPC0.5980.3260.311
FCM0.8500.4420.488FCM0.6280.3740.353
Wine                  Waveform (noise)                  
IKDC0.9520.8730.852IKDC0.9830.9790.959
DPC0.8820.7100.627DPC0.9690.9580.936
DBSCAN0.9100.7530.741DBSCAN0.9680.9570.935
KNN-DPC0.9040.7430.727KNN-DPC0.9700.9600.940
SNN-DPC0.9660.8780.899SNN-DPC0.9740.9630.974
FCM0.9490.8500.834FCM0.8910.8620.936
Table 7. ACC, NMI, and ARI of different clustering algorithms on different real datasets.
Table 7. ACC, NMI, and ARI of different clustering algorithms on different real datasets.
AlgorithmACCNMIARIAlgorithmACCNMIARI
Libras                  Spectrometer                  
IKDC0.5030.5430.401IKDC0.9790.9750.975
DPC0.4200.5140.345DPC0.9680.9580.936
DBSCAN0.4580.6260.309DBSCAN0.9680.9570.935
KNN-DPC0.4970.6340.361KNN-DPC0.9700.9600.940
SNN-DPC0.4940.6610.393SNN-DPC0.9740.9630.974
FCM0.1830.2160.069FCM0.8910.8620.936
WDBC                  Seeds                  
IKDC0.9170.7200.751IKDC0.9340.7630.830
DPC0.8300.3730.373DPC0.9100.7160.742
DBSCAN0.6310.5300.530DBSCAN0.6240.4230.487
KNN-DPC0.8700.7660.436KNN-DPC0.9140.7340.766
SNN-DPC0.8740.5350.650SNN-DPC0.9240.7540.789
FCM0.9280.6090.631FCM0.9000.6910.727
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, S.; Li, K. A Novel Density Peaks Clustering Algorithm with Isolation Kernel and K-Induction. Appl. Sci. 2023, 13, 322. https://doi.org/10.3390/app13010322

AMA Style

Zhang S, Li K. A Novel Density Peaks Clustering Algorithm with Isolation Kernel and K-Induction. Applied Sciences. 2023; 13(1):322. https://doi.org/10.3390/app13010322

Chicago/Turabian Style

Zhang, Shichen, and Kai Li. 2023. "A Novel Density Peaks Clustering Algorithm with Isolation Kernel and K-Induction" Applied Sciences 13, no. 1: 322. https://doi.org/10.3390/app13010322

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop