Next Article in Journal
IH-TCGAN: Time-Series Conditional Generative Adversarial Network with Improved Hausdorff Distance for Synthesizing Intention Recognition Data
Next Article in Special Issue
Infrared Image Caption Based on Object-Oriented Attention
Previous Article in Journal
Novel Schemes of No-Slip Boundary Conditions for the Discrete Unified Gas Kinetic Scheme Based on the Moment Constraints
Previous Article in Special Issue
An Ensemble and Multi-View Clustering Method Based on Kolmogorov Complexity
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Density Spatial Clustering Method Fusing Chameleon Swarm Algorithm

1
School of Computer Science and Technology, Changchun University of Science and Technology, Changchun 130022, China
2
School of Information, Guangdong University of Finance & Economics, Guangzhou 510320, China
3
College of Information Science and Technology, Jinan University, Guangzhou 510632, China
4
College of Information Engineering, Yangzhou University, Yangzhou 225127, China
5
School of Economics and Management, Shenyang University of Chemical Technology, Shenyang 110142, China
*
Authors to whom correspondence should be addressed.
Entropy 2023, 25(5), 782; https://doi.org/10.3390/e25050782
Submission received: 13 March 2023 / Revised: 7 May 2023 / Accepted: 9 May 2023 / Published: 11 May 2023
(This article belongs to the Special Issue Pattern Recognition and Data Clustering in Information Theory)

Abstract

:
The density-based spatial clustering of application with noise (DBSCAN) algorithm is able to cluster arbitrarily structured datasets. However, the clustering result of this algorithm is exceptionally sensitive to the neighborhood radius (Eps) and noise points, and it is hard to obtain the best result quickly and accurately with it. To solve the above problems, we propose an adaptive DBSCAN method based on the chameleon swarm algorithm (CSA-DBSCAN). First, we take the clustering evaluation index of the DBSCNA algorithm as the objective function and use the chameleon swarm algorithm (CSA) to iteratively optimize the evaluation index value of the DBSCAN algorithm to obtain the best Eps value and clustering result. Then, we introduce the theory of deviation in the data point spatial distance of the nearest neighbor search mechanism to assign the identified noise points, which solves the problem of over-identification of the algorithm noise points. Finally, we construct color image superpixel information to improve the CSA-DBSCAN algorithm’s performance regarding image segmentation. The simulation results of synthetic datasets, real-world datasets, and color images show that the CSA-DBSCAN algorithm can quickly find accurate clustering results and segment color images effectively. The CSA-DBSCAN algorithm has certain clustering effectiveness and practicality.

1. Introduction

With the continuous development of society and rapid changes in Internet technology, big data has come into people’s view and developed rapidly, which has a very broad prospect for social progress [1]. Data mining techniques can extract various correlations and validity information from massive data to enhance the value of the data [2]. Clustering is one of the key techniques in data mining [3]. Clustering algorithms can categorize data with unknown relationships based on the similarity between the data and then mine the valuable information hidden in the data. Clustering algorithms have promising applications in the fields of face recognition [4], image segmentation [5], stock prediction [6], and medical image processing [7].
Clustering algorithms can generally be used in partition-based [8], hierarchy-based [9], density-based [10], grid-based [11], model-based [12], and graph-based [13] ways to perform classification. The k-means algorithm is the most traditional method of partitioning clusters [14]. Affinity propagation (AP) clustering is a method based on the transfer of information between data points [15]. DBSCAN is a classical density-based clustering method that can identify arbitrarily shaped classes with noisy points in spatial data [16]. The DBSCAN method seeks to uncover potentially meaningful messages in a given dataset. This method can effectively handle anomalous data by connecting neighboring regions with high densities in a given spatial database. In addition, the DBSCAN clustering algorithm has the benefits of higher efficiency and wider applicability as it can effectively handle noisy points without inputting the number of clustering centers.
Swarm intelligence is a computational method that simulates the smart behavior of biological groups; this method has the advantages of fast convergence, high solution accuracy, and robustness and is suitable for solving large-scale complicated problems and optimization issues. Typical swarm intelligence algorithms include particle swarm optimization [17], the cuckoo search algorithm [18], and the bird swarm algorithm [19]. In recent years, researchers have proposed a chameleon swarm algorithm (CSA) based on the feeding behavior of chameleons under different conditions [20].
The DBSCAN algorithm parameter Eps is exceptionally sensitive to clustering results and is prone to the problem of the over-identification of noise points during the clustering process. To solve the above problems, we use the CSA method to perform an adaptive search for the clustering parameter Eps of DBSCAN and obtain the best clustering parameter value by continuous iterative optimization. The initial clustering results can be obtained by inputting the optimized parameters into DBSCAN. To prevent the algorithm from over-identifying boundary points as anomalies, we use the nearest neighbor mechanism to assign the noise points of the optimization algorithm to the nearest category points, forming the final CSA-DBSCAN algorithm.
Although the DBSCAN method has a very good clustering effect, it still has some drawbacks. To address the challenges DBSCAN faces when handling large natural datasets that are fuzzy, Smiti et al. [21] proposed the “Soft DBSCAN” method using the fuzzy membership function of fuzzy c-mean theory. This method can cluster arbitrary shape data and has a better processing effect on noise points. Furthermore, DBSCAN performs neighborhood queries for every object in the dataset, resulting in low efficiency; to address this problem, Severino et al. [22] proposed a most-related region search strategy for the DBSCAN method. The approach checks the neighborhoods of a subset of objects in the dataset to improve the clustering efficiency. All of the above algorithms significantly improve the performance of the DBSCAN algorithm but do not address one key problem, which is that the parameters of the DBSCAN algorithm are exceptionally sensitive to the clustering results. Compared with the above literature, the CSA-DBSCAN algorithm can use the CSA algorithm with its excellent global search capability to perform an adaptive search on the Eps parameter in the DBSCAN method and can accurately find the optimal position without setting a specific search range.
The parameter problem of the DBSCAN algorithm has always been a hot topic for scholars at home and abroad. Bryant et al. [23] proposed a heuristic approach for RNN-DBSCAN. The algorithm utilizes the core observe and observe achievability definitions based on inverse nearest neighbors, which reduces the complexity of the problem (i.e., compared to the two parameters Eps and MinPts, RNN-DBSCAN needs to use only one parameter k). Hou et al. [24] proposed a parameter-free clustering method for DSets-DBSCAN. The algorithm first equalizes the histogram of pairwise similarity matrices used for the input data to obtain the dominant set (dset) and finally uses the dset to automatically determine the parameters to extend the clusters. All the above studies reduced the influence of parameters on clustering results. However, the best values of clustering parameters could not be found quickly and accurately. Compared with the above literature, the use of the CSA method to deal with the parameter problem of DBSCAN can greatly reduce the time consumption of the tuning process.
Since the swarm intelligence algorithm has a good global search capability, many scholars use this method to adaptively search for the parameters of the density clustering algorithm [25,26,27,28]. Zhu et al. [29] proposed the K-DBSCAN algorithm. The algorithm achieves a better clustering property by applying the proposed new harmony search method to DBSCAN to optimize the parameters “Eps” and “MinPts” and by using a multi-objective co-optimization function. Wang et al. [30] proposed a parameter-adaptive DBSCAN algorithm for bird swarm optimization. The algorithm finds the exact resulting clusters corresponding to the optimal Eps parameters of the DBSCAN algorithm by continuously adjusting the search range. All the above algorithms solve the parameter setting problem in different ways, but they all lack the mechanism to deal with the identified anomalies. The CSA-DBSCAN algorithm uses the nearest neighbor mechanism and is able to solve the problem of the over-identification of noise points and the inability to obtain an effective assignment of anomalies.
To further measure the compactness relationship between noise and boundary points, we introduce the theory of deviation of spatial data in the nearest neighbor mechanism. Finally, an advanced adaptive CSA-DBSCAN method is developed and applied to color image segmentation. We test the CSA-DBSCAN algorithm on synthetic datasets and real-world datasets and find that the improved method has better clustering accuracy and a better clustering effect than other algorithms for data with arbitrarily complex structures. To improve the computational efficiency of DBSCAN technology in image processing, we use the simple linear iterative clustering (SLIC) method to construct superpixel information [31] and then used an improved method to segment the superpixel data according to the characteristics of the image color space to realize image segmentation by the adaptive CSA-DBSCAN method. The practicality and effectiveness of the improved method are confirmed by the comprehensive evaluation of the clustering effect and image segmentation performance. Through comparison and analysis with other advanced clustering algorithms, the superior performance of the improved algorithm in adaptive optimization and image segmentation is verified. The main contributions to this article are presented as follows:
1.
The CSA-DBSCAN method is proposed to realize the parameter adaption of the clustering process and reduce the parameter tuning complexity.
2.
The nearest neighbor mechanism is used to assign the noise points identified by the CSA-DBSCAN method to the nearby clusters, which addresses the issue of assigning the noise points of the DBSCAN algorithm.
3.
Deviation theory is introduced in the nearest neighbor mechanism to further measure the compactness between data points.
4.
The construction of color image superpixel information using the SLIC algorithm increases the clustering performance of the CSA-DBSCAN algorithm in color image segmentation.
The rest of this article is organized as follows. Section 2 explain the theory of DBSCAN and the chameleon swarm method (CSA) in depth. Section 3 analyzes the newly presented adaptive DBSCAN method based on the CSA method (CSA-DBSCAN). Section 4 experiments with and summarizes the performance of CSA-DBSCAN in various datasets and evaluates it against other methods. Finally, conclusions and future progress are given in Section 5.

2. Research Background

In this section, we give two basic approaches related to our proposed algorithm. The first is the DBSCAN method, and the second is the chameleon swarm optimization algorithm.

2.1. Density-Based Spatial Clustering of Application with Noise

Esterl et al. [16] proposed a popular and highly efficient DBSCAN method. The method can handle sample sets of any shape and size, and the clustering results are independent of the input order of data objects, with good interference resistance. The core principle of the method is that for each target cluster, there are a series of sets containing a specified point threshold within a given neighborhood radius. The clustering efficacy of the method depends entirely on two parameters: the minimum number of objects in the search radius (MinPts) and Eps. The number of MinPts generally takes the value of 4 and is not very sensitive to the results of the process, which facilitates the parameterization of the method. The parameter Eps is very sensitive to the clustering performance, which causes some difficulty when setting the Eps parameter.
Depending on the aforementioned two parameters, we can classify the points in the sample into three types.
  • Core point: If the quantity of sample points in the Eps neighborhood is greater than MinPts, the points of this type are called core points.
  • Border point: The points within the radius Eps are smaller than MinPts. The boundary and some core points within the cluster are mutually contained within the Eps.
  • Outlier: A single site that does not satisfy the core point condition or the bounding point condition.
Since the DBSCAN method marks three types of points by progressively searching for them, there are three descriptions of the point-to-point density relationship in the search process. Suppose that given a dataset S = { y 1 , y 2 , , y n } , the relationship between data points is described as follows:
  • Direct density reachable: If y j is within the range of the Eps of y i , and y i is a core point, y j is said to be reachable by the direct density of y i .
  • Density reachable: For points y i and y j , if there exists a sequence p 1 , p 2 , , p n , where p 1 = y i , p n = y j , and p n + 1 is directly density reachable by p i , we say y i is density reachable by y j .
  • Density connected: For y i and y j , y i and y j are called connected by density if the existence of y k makes both y i and y j realizable by the density y k .

2.2. Chameleon Swarm Algorithm

In 2021, inspired by the foraging behavior of chameleons, Braik [20] proposed a novel chameleon swarm optimization algorithm (CSA). This method focuses on modeling the dynamic behavior of chameleons as they search for food near forests, dunes, and marshes. This method of math modeling mimics the process of chameleons searching for food, in which they rotate their eyes nearly 360 degrees for prey localization and capture prey with their sticky tongue that fires at high speed.
In nature, chameleons live in a variety of places, such as lowlands, rainforests, deserts, semi-deserts, and even mountains. Chameleons are very good climbers and have excellent eyesight, being able to see up to 32 feet in front of them. Chameleons usually feast upon insects such as sulfur worms, praying beetles, grasshoppers, and crickets. Some big individuals even use lizards and baby birds to improve their diet. The individual first uses their eyes to locate their target and then uses their tongue to capture it to eat and seek optimal for the solving way. On the flip side, chameleons have some natural predators, such as snakes, vultures, and lemurs. Chameleons can adapt their color to their surroundings, which is very beneficial when hunting or escaping from natural predators. The CSA method is particularly unique in its advantages based on the unique capabilities of the chameleon.
  • Initialization: This method, as with other optimization algorithms, is initialized randomly in the search range.
    y t i = [ y t , 1 i , y t , 2 i , , y t , d i ]
    y i = l j + ( u j l j ) × r
    where t is the numeration of iterations, i = { 1 , 2 , , n } , d denotes the dimension, and y t , d i is the location of the individual i in dimension d. In Equation (2),  r denotes the creation of random numbers in the range [ 0 , 1 ] , y i denotes the initial level of the ith individual, and u j and l j are the upper and lower bounds of dimension j, respectively.
  • Search for prey: At this point, the chameleon begins to search for prey in order to solve the problem of needing food.
    y i , j t + 1 = y i , j t + r 2 ( P i , j t G i t ) p 1 + r 1 ( G j t y i , j t ) p 2 r i P p y i , j t + μ ( l b j + r 3 ( u i l i ) ) s g n ( r a n d 0.5 ) r i < P p
    where y i , j t + 1 donates the position of chameleon i at step ( t + 1 ) in dimension j, G j t is the current optimal individual position, P i , j t is the best position of chameleon i so far, p 1 and p 2 are two parameters controlling the exploration ability, r 1 , r 2 , and  r 3 are all random values within [ 0 , 1 ] , and  r i looks different from the first three but is also a random quantity within [ 0 , 1 ] . P p = 0.1 indicates the probability that the chameleon perceives its prey. s g n ( r a n d 0.5 ) is 1 or −1, which mainly affects the direction of exploration and development. μ is calculated as follows:
    μ = γ e ( α t / T ) β
    where γ , α , and  β are 1, 3.5, and 3, respectively, t is the current number of iterations, and T is the maximum number of iterations.
  • Prey location: The eyes of chameleons move independently, which allows them to skillfully explore spatially and find prey. These eyes can view in two different directions and rotate and focus simultaneously, allowing them to see a sweeping view of their environments.
To simulate this process, the authors performed four setups: translating the location of the original chameleon to its gravity center, locating the matrix of rotations that recognizes the location of the prey, updating the chameleon’s position using the matrix of rotations at the gravity center, and moving the chameleon back to its original position.
After the above process, we started the step that involved updating the chameleon’s position. This stage simulates the position update that occurs when the chameleon locates its prey by eye rotation.
y t + 1 i = y r t i + y ¯ t i
where y ¯ t i is the average location of the individual in each dimension before position rotation, y t + 1 i is the coordinates after rotation, and  y r t i is the coordinates of the center of rotation, specified as follows:
y r t i = m × y c t i ,
where y r t i is the centering coordinate.
y c t i = y t i y ¯ t i
m = R ( θ , V z 1 , z 2 )
where m denotes the rotation matrix of the chameleon position rotation, V z 1 and z 2 represent two quadrature variants in the spatial coordinates, and R denotes the rotation matrix on each axis.
θ = r s g n ( r a n d 0.5 ) × 180
where θ represents the rotation angle of the chameleon’s eyes, and r is a random value within [ 0 , 1 ] , so that θ can be limited to [ 180 , 180 ] .
  • Hunting prey: The algorithm determines that the chameleon close to the prey is the best-positioned chameleon. This chameleon attacks its prey with its tongue. Therefore, its position will be slightly updated, since its tongue can extend to twice its original length. The speed of the chameleon’s tongue as it moves toward its prey is defined as:
v t + 1 i , j = ω v t i , j + r 1 ( G t j y t i , j ) c 1 + r 2 ( P t j y t i , j ) c 2
Many of these parameters were previously described. c 1 = c 2 = 1.75 controls the effect of G and P on the velocity of the tongue. ω is updated as follows:
ω = ( 1 t / T ) ( ρ ( t / T ) )
where ρ = 1 is the parameter that controls the exploitation capacity. When the chameleon’s tongue is projected towards the prey, the acceleration of the tongue bouncing up is a. The position of the tongue indicates the position of the chameleon, and the formula is as follows:
v t + 1 i , j = y t i , j + ( ( v t i , j ) 2 ( v t 1 i , j ) 2 ) / ( 2 a )
a = 2590 × ( 1 e l o g ( t ) )
At this point, the CSA algorithm has completed the merit-seeking foraging mechanism of the chameleon. Next, we will further illustrate the algorithm fusion process of CSA-optimized DBSCAN.

3. Materials and Methods

In this section, we show the technical details of the present DBSCAN method based on chameleon optimization (CSA-DBSCAN). Firstly, we focus on the theoretical model and optimization details of the adaptive parameter search system. Then, we obtain the best processing results corresponding to the best parameter values. To further evaluate the potential relationship between the noise and boundary points, we use the nearest neighbor approach to assign the noise points to the nearest boundary points. Finally, we introduce the theory of deviation in the process of measuring data similarity to evaluate the compactness between data, which provides a more reliable reference basis for the assignment of noise points.

3.1. Adaptive Parameter Seeking Strategy

DBSCAN is an elegant method that allows the clustering of data with arbitrary shapes and constructions. However, when testing the superiority of the clustering capability of a dataset, two parameters need to be jointly adopted to achieve better clustering results. This adds, to some extent, to the complexity and time consumption of clustering tuning parameters. In the clustering process, we usually choose the value of MinPts as a positive integer from 1 to 15. However, the value of Eps is exceptionally sensitive to the clustering outcomes. When using a dataset, the value of Eps might be taken to be as precise as 0.01. This poses a challenge and difficulty in the selection of parameters for the DBSCAN method.
To overcome this limitation, we used the chameleon optimization algorithm with its excellent search power to solve the parameter-setting task by finding the best arguments for every dataset. To this end, we proposed an adaptability mechanism for density clustering parameters based on chameleon optimization. First, we set the value of the argument Eps as the position of the initial chameleon population. We first initialized a certain amount of Eps populations as candidate resolutions to further search for the location of the optimality parameter. Since there is only one parameter to be searched for, we set the dimension as 1. We set the size of the population as n and the current quantity of iterations as t. Then, the candidate solutions for the Eps parameter in the one-dimensional space can be expressed by Equation (14).
E t i = [ E t , 1 1 , E t , 1 2 , , E t , 1 n ]
In Equation (14), E denotes the candidacy solution of the n Eps of the chameleon groups. To facilitate the search for the best parameter, we set the cluster evaluation index value as the fitness value to be optimized by the CSA method. The chameleon optimization technique has a very sharp global search performance, which can balance the accuracy of the search procedure by adjusting the position of the most optimal individual in all directions. After running the continuous iterative process of the method and updating the size and position of the fitness value, we very quickly obtained the optimal population position corresponding to the Eps parameter and thus obtained the optimized clustering effect.
In data mining clustering tasks, different clustering evaluation metrics are available to further measure the ability of an algorithm to deal with a given problem. To facilitate testing the clustering performance of the proposed method, we used the obtained adjusted Rand index (ARI) as an adaptation function to verify the clustering performance. In addition, to test the clustering effect of the CSA-DBSCAN method in image segmentation, we used the most popular boundary displacement error (BDE) [32] as the fitness function to measure the method’s effectiveness in image segmentation. Therefore, we set a total of two fitness functions that solve the different problems of the algorithm.
Suppose a denotes the number of identical samples in the real label A and the clustered label B, and b denotes the number of different samples in A and B and the total number of n. To observe the clustering performance more carefully, the external evaluation index ARI can be used as the fitness function. ARI is a range of values (0,1), and the closer the value is to 1, the better the clustering effect. Therefore, in this subsection, we reset the ARI index fitness function, which is set as follows:
F i t A R I = m a x ( R I ) E [ R I ] ( R I E [ R I ] ) m a x ( R I ) E [ R I ] × 100 %
R I = a + b C n 2
To verify the effectiveness of the CSA-DBSCAN algorithm in practical applications, we used real color images to verify the ability of the CSA-DBSCAN method to perform image segmentation. The fitness function for image segmentation is set as follows.
F i t B D E = i N 2 d ( p i , B 1 ) / N 2 + i N 1 d ( p i , B 2 ) / N 1 2
d ( p i , B 2 ) = m i n p B 2 p i p
where d is the length from the boundary point p i in B 1 to the boundary set B 2 , and  N 1 and N 2 are the number of boundary sets B 1 and B 2 .
The BDE is usually greater than 1; the lower the metric value, the better the image segmentation effect. Therefore, this section directly uses the BDE as the fitness function for the adaptive search.
The CSA-DBSCAN algorithm allows different fitness functions to be used according to their actual application environment. Different fitness functions can solve the issue of parameter setting for the DBSCAN method in different problems to avoid the influence of artificial parameter settings on experimental results. Therefore, this method improves the precision of clustering and the clustering effect. This adaptive CSA-DBSCAN method also has a good clustering effect in color image segmentation.
This subsection achieved the cross-fertilization of disciplines by fusing swarm intelligence theory and clustering methods and solved the problem related to the adaptive selection of the clustering parameter. The final parameter-seeking system based on chameleon optimization was thus obtained.

3.2. Noise Point Allocation Mechanism

The nearest neighbor mechanism is an extension of the k-nearest neighbor (KNN) method [33], which is a prominent and simple intelligent classification approach. It is a theoretically mature method. It is both one of the easiest machine learning methods and the most basic of the instance-based teaching methods, as well as one of the finest text categorization methods.
The underlying principle is that if the k most resembling examples of an instance in the spatial characteristic belong to a certain class, then that example also falls into that class. The chosen neighbors are all instances that have been properly categorized. This method supposes that all examples respond to points in the N-dimensional Euclidean spatial aspect. By calculating the separation between one point and every other point, the k nearest sites to that point are taken out, and then the point that belongs to the largest proportion of the categories within these k points is counted; that point then belongs to that category.
While obtaining the optimal resulting clusters for the best parameter values, we also saw the noise points identified by the method. To further explore the hidden relationship between noise points and boundary points, we used the nearest neighbor mechanism to assign noise points to the nearest clusters. The nearest neighbor mechanism evolved from the k-nearest neighbor method. The result of DBSCAN after an adaptive search has fewer noisy points, and the distance to the nearest boundary point is much smaller than the distance to other points. In this work, we unified the assignment of anomalies to the nearest boundary points based on the nearest neighbor feature to better measure the potential relationship between noise points and boundary points.
First, we calculated the separation matrix among the data sites and then sorted the separation matrix from smallest to largest. Then, we extracted the size and the original position of the distance matrix separately for storage. To obtain the data characteristics of the noise points, we extracted all the positions corresponding to the distance matrix with category points, and finally, we obtained the distance matrix and the corresponding positions of the noise points. We based our judgment on the first column of the noise point position (i.e., the point closest to the boundary). Assigning it to the nearest category completed the assignment of noise points.
In the noise point assignment strategy, the distance was calculated in such a way that we usually used the Euclidean distance. Suppose there is a dataset S = { x 1 , , x i , , x N } , and N is the number of data points. Then, in d dimensions, x i and x j are x i = { x i 1 , x i 2 , , x i d } and x j = { x j 1 , x j 2 , , x j d } , respectively. The Euclidean distance is as follows:
d i j = k = 1 d ( x i k x j k ) 2
In Equation (19), d i j denotes the distance between different data points. This calculation only measures the distance between data points and does not describe the compactness of the relationship between data points. To solve this problem, we introduced deviation theory in the nearest neighbor assignment mechanism and used this theory to measure the compactness between data points in an old way. The deviation theory is calculated as shown below.
d e i = d i j m e a n ( d i j ) N
where m e a n ( d i j ) denotes the average of the data point distances, N represents the overall quantity of data points, d i j m e a n ( d i j ) represents the degree of deviation of the data points from the average, and  d e i denotes the value of the attribute used to describe the compactness of the current data point obtained after obtaining the distance between data point i and the other data points.
Equation (20) can better help the nearest neighbor assignment of anomalies. Using d e i instead of the Euclidean distance can further measure the compactness between anomalies and boundary points, which helps in anomaly assignment. Finally, using adaptive parameter optimization, anomaly assignment strategy, and deviation theory, the CSA-DBSCAN method was proposed.
In Figure 1, we give the anomaly assignment process for the Complex8 dataset. We give the convergence curve of the CSA-DBSCAN algorithm (see Figure 1a). In Figure 1a, the CSA-DBSCAN algorithm searches for the optimal parameter positions when the number of iterations is 3. To facilitate understanding, we give a simulation graph of the anomaly assignment process of the CSA-DBSCAN algorithm. In Figure 1b, the three points in white indicate the anomalies identified by the DBSCAN algorithm after iteration to obtain the best parameters. The red arrows indicate that the anomalies are assigned to the nearest category points after the compactness degree calculation of the fusion deviation theory. The yellow solid line, black dashed line, and blue dotted line indicate the degree of compactness (also called nearest deviation distance) between the anomalies and other category points, respectively. The CSA-DBSCAN algorithm assigned the anomalies to the category points with the nearest deviation distance by comparing them with the deviation distance of other data points from different categories. The final clustering results were obtained by assigning the anomalies (see Figure 1c).
In Algorithm 1, we give the specific flow of the CSA-DBSCAN method. From the algorithm flow, we know that the time complexity of updating the position of the CSA-DBSCAN algorithm population when searching for prey with the number of iterations I is O ( I n ) , the time complexity of updating the position of eye rotation is O ( I n ) , and the time complexity of updating the speed and position of the tongue attack on prey is O ( I n ) . Since the time complexity of the DBSCAN algorithm is O ( n l o g n ) , the time complexity of the CSA-DBSCAN algorithm can be calculated as O ( I n ) + O ( I n ) + O ( I n ) + O ( I n l o g n ) O ( I n l o g n ) .
Algorithm 1 The proposed CSA-DBSCAN method.
Input: 
Dataset S, density number M i n P t s , problem dimension d, maximum iteration number I, chameleons number n, search bounds u and l, position update probability P p , and random numbers r 1 , r 2 , r 3 , and  r i
Output: 
The optimal Eps, best cluster result
 1:
The DBSCAN evaluation index is used as the object function to evaluate the adaptation value
 2:
Use the value of Eps as the location of the chameleon
 3:
while  t < I  do
 4:
    Using Equations (4), (11), and (13) to define parameters μ , ω , and a
 5:
    for  i = 1 : n  do
 6:
        for  j = 1 : d  do
 7:
           if  r i P p  then
 8:
               According to Equation (3), y i , j t + 1 = y i , j t + r 2 ( P i , j t G i t ) p 1 + r 1 ( G j t y i , j t ) p 2 ;
 9:
           else
10:
               According to Equation (3), y i , j t + 1 = y i , j t + μ ( l b j + r 3 ( u i l i ) ) s g n ( r a n d 0.5 ) ;
11:
           end if
12:
        end for
13:
    end for
14:
    for  i = 1 : n  do
15:
         y t + 1 i = y r t i + y ¯ t i
16:
    end for
17:
    Determine the speed of the chameleon when attacking the prey at the optimal Eps position according to Equation (10)
18:
    Determine the optimal chameleon location according to Equation (12)
19:
    Update the new Eps position in the CSA-DBSCAN algorithm
20:
     t = t + 1
21:
end while
22:
Assign noise points using the nearest neighbor mechanism and deviation theory
23:
Obtain the best clustering result

4. Results and Discussion

In this paper, we tested the clustering performance of the improved method using synthetic and real-world datasets and applied it to a clustering task on the Berkeley image segmentation dataset (BSDS500) [34] to further validate the utility of the CSA-DBSCAN algorithm. We demonstrated the effectiveness of the new adaptive method for outlier assignment on datasets with more complex density structures. We compared the clustering results of the proposed algorithm with the classical k-means [14], AP [15], DBSCAN [16], DPC [10], DPCSA [35], AmDPC [36], and DeDPC [37] algorithms for analysis. Among them, DPCSA, AmDPC and DeDPC are the most advanced clustering algorithms proposed in recent years. In this experiment, we compared and analyzed the values of the normalized mutual information (NMI), adjusted Rand index (ARI), and F-measure (FM) evaluation metrics of the above four mentioned algorithms, respectively, to further evaluate the clustering ability of these different algorithms. As shown in Table 1, we used twelve datasets to measure the efficiency of the method. As also shown in Table 1, the former six were synthetic datasets, and the latter six were real-world datasets. All datasets are described in detail below.

4.1. Evaluation of Clustering Effectiveness

The experimental environment was simulated using a computer with an Intel Pentium 2.9 GHz, 8.00 GB memory, and 500 GB hard drive running a Windows 10 operating system and using the MATLAB2019a programming language.
A detailed description of the clustering performance evaluation metrics is given before we derive the experimental results. We employed the three most valid cluster performance metrics (NMI, ARI, ACC, and FM). These metrics take values between −1 and 1, with values nearer to 1 indicating the better performance of the clusters.
NMI is a very valuable external evaluation metric of clustering. Suppose the true label of a dataset is A, and the clustering result label is B. Then, the unique values in A are extracted to form vector E, and the unique values in B are extracted to form vector F. The NMI of A and B are as follows:
N M I ( A , B ) = I ( E , F ) H ( E ) H ( F )
I ( E , F ) = e E f F l o g ( p ( e , f ) p ( e ) p ( f ) )
H ( E ) = 1 n p ( e i ) log 2 p ( e i )
H ( F ) = 1 n p ( f i ) log 2 p ( f i )
where p ( e ) indicates the probability for e in A, and p ( f ) denotes the proportion of f in B. P ( e , f ) indicates the joint probability of e and f.
The Rand index (RI) uses a paired approach to detect true negatives (TN), true positives (TP), false negatives (FN) and false positives (FP). In clustering methods, the metric values of the RI tend to have high values, and it is difficult to distinguish the effectiveness of the methods. Therefore, we used the ARI to calculate the algorithm performance instead of RI.
R I = T P + T N T P + T N + F P + F N
A R I = R I E [ R I ] m a x ( R I ) E [ R I ]
ACC is a classical evaluation index often used to measure clustering accuracy. Assuming that r i and s i represent cluster labels and real labels, respectively, in a dataset of n samples, ACC can be defined as:
A C C = i = 1 n δ ( s i , m a p ( r i ) ) n
δ ( x , y ) = 1 i f   x = y 0 o t h e r w i s e
where m a p ( r i ) is a permutation mapping function that matches the cluster label and the real label.
Assume that the real label of the dataset is A and the label of the clustering is B. a 1 denotes the quantity of samples that belong to the same clusters in both A and B. a 2 indicates the quantity of samples that belong to the same clusters in A and not in B. a 3 indicates the quantity of samples that belong to different clusters in A but belong to the same clusters in B. FM was calculated as shown below.
F M = a 1 a 1 + a 2 a 1 a 2 + a 3

4.2. Performance Evaluation on Synthetic Datasets

We validated the clustering performance of the CSA-DBSCAN algorithm on synthetic datasets. The external evaluation index (ARI) was chosen as the fitness function ( F i t A R I ) for this experiment. We selected Donutcurves, Target, Pearl, Complex8, and Complex9 datasets, all of which have complex density structures, to observe the ability of the CSA-DBSCAN method to handle complex data.
In the experiments, we kept the other parameters of the adaptive search constant and set the number of iterations to 100 to verify the adaptive effect of the improved method. The Twocirclesnoise and Target datasets for this experiment are datasets with anomalies as a separate class. Using this dataset allows further testing of the CSA-DBSCAN algorithm’s anomaly assignment performance. In Table 2, we give the evaluation metric values for the clustering of the eight state-of-the-art algorithms for the six synthetic datasets. The evaluation metric values for the different clustering methods correspond to the clustering results in Figure 2, Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7. In this experiment, the best clustering results were obtained for all algorithms. Therefore, we analyzed the clustering accuracy and effectiveness of the algorithms in combination with the tabular and visual clustering results.
In Table 2, Par indicates the parameter settings of the method on different datasets. The visual clustering results of the method in Table 2 correspond to Figure 2, Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7. In Table 2, the parameter of the k-means method is the quantity of clusters (k), the parameter of the AP method is the value of the matrix diagonal (preference), the parameters of the DBSCAN method are the quantities of neighborhoods (MinPts) and the domain radius (Eps), and the parameters of the DPC method are the average number of neighboring points, represented as a percentage (p). The DPCSA method is a parameterless clustering algorithm; the AmDPC model’s parameters are the number of neighbors ( N i ), peak split lines ( C 1 , C 2 ) and average number of neighbors, represented as a percentage (p). The parameters of the DeDPC algorithm are the number of data point neighbors (N) and the average number of neighbors, represented as a percentage (p). The CSA-DBSCAN method can find the best Eps parameter adaptively, so the parameter of this method is MinPts. For the synthetic datasets, we uniformly set the search range of the CSA-DBSCAN method parameter to [0, 20]. Since the value of MinPts is a positive integer within 15, the CSA-DBSCAN algorithm nicely overcomes the drawback inherent to parameter setting in the DBSCAN approach.
In Table 2, we have marked the best clustering evaluation metric values for the different methods in bold. As can be seen in Table 2, the CSA-DBSCAN method obtained the best cluster evaluation metric values on the six synthetic datasets. Table 2 shows all the complex structured datasets with their uneven density distributions. It can be seen from the table that the k-means, AP, DPC, and DPCSA algorithms do not work well with this type of dataset, and AmDPC and DeDPC do not achieve the best metric values for some of the datasets, though the values they achieve are still high. The DBSCAN algorithm is very close to 1 for most of the clustering metric values but cannot reach 1. The main reason is that the algorithm incorrectly identifies the boundary points as anomalies in the clustering process. The outlier assignment mechanism of the CSA-DBSCAN algorithm solves the problem of incorrectly identified outliers in the DBSCAN method very well. The algorithm can adaptively find the best parameters and clustering data for any complex structure with good clustering accuracy and clustering performance.
To better analyze the clustering performance, in Figure 2, Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7 and Table 2, all the algorithms correspond to the best clustering results. As shown in Figure 2, the Donutcurves dataset is a complex dataset of ring and semi-ring structures. We can see that the k-means, AP, DPC, DPCSA, AmDPC, and DeDPC methods do not accurately cluster the Donutcuvers dataset. The DBSCAN method clustered the data accurately but identified one of the category points as noisy, resulting in poor clustering accuracy. The CSA-DBSCAN approach obtained better clustering results in this complex structured dataset. In Figure 3, the Target dataset is a closed-loop dataset with an inhomogeneous density structure. We find that both the DBSCAN and CSA-DBSCAN algorithms can handle this type of dataset well. However, neither the AP nor the k-means algorithm can handle this type of dataset well. In Figure 4, we can see that both the k-means and AP algorithms do not handle the torus data and the central category points well. The DBSCAN approach uniformly misidentifies the central category points and the points outside the torus as noisy points. The CSA-DBSCAN method not only identifies the central category points but also assigns the points next to the torus to the corresponding clusters. It is possible to state that the CSA-DBSCAN method has good clustering accuracy and clustering efficiency for closed-loop datasets with irregular density compared to other algorithms.
In Figure 5, Pearl is a semi-annular multi-density dataset. We can see that the k-means, AP, DPC, and DeDPC algorithms are not able to handle the half-loop data, but both the DBSCAN and CSA-DBSCAN algorithms can handle this dataset well. However, the DBSCAN approach incorrectly identifies the boundary points as noisy points, while the DBSCAN, DPCSA, AmDPC, and CSA-DBSCAN algorithms obtain the best clustering results perfectly. In Figure 6, Complex8 is a dataset with a complex structure and an uneven density distribution, and the processing of this dataset can particularly detect the clustering effect of the method. In Figure 6, we can see that the k-means, AP, DPC, DPCSA, AmDPC, and DeDPC algorithms have poor clustering results and cannot identify complete classes for this dataset. It is encouraging that the CSA-DBSCAN method can obtain the optimal clustering effects more accurately than the DBSCAN algorithm. In Figure 7, the Complex9 dataset is a complex dataset with annular, semi-annular, and long arc structures. We can see that the k-means, AP, DPC, DPCSA, AmDPC, and DeDPC methods cannot handle this class of data. The DBSCAN and CSA-DBSCAN methods can handle this dataset perfectly, and the CSA-DBSCAN algorithm has better clustering accuracy and performance.
To further verify the utility of the proposed method, we further tested the effectiveness of the algorithm on real-world datasets.

4.3. Performance Evaluation on Real-World Datasets

In this subsection, we describe the clustering properties of different algorithms on real-world datasets. In Table 3, we give the values of clustering evaluation metrics for different methods on six high-dimensional datasets. The parameters of the algorithm tested on the real-world dataset are the same as those in Table 2. To facilitate comparison and analysis of the algorithm performance, we give in the table the values of the best clustering evaluation metrics obtained after repeated training and testing. Since real-world datasets have characteristics such as high dimensionality and nonlinearity, we uniformly set the search range of the CSA-DBSCAN method to [0, 20]. The search space of the proposed method is one parameter, so the search dimension is set to 1.
From Table 3, we can see that the CSA-DBSCAN algorithm for the Seeds dataset has higher values for all metrics than the other state-of-the-art clustering algorithms. From the Thyroid dataset, we can see that the NMI, ARI, and FM clustering metric values for the CSA-DBSCAN and DBSCAN methods are higher than the other algorithms, and the ACC metric values are slightly lower than the k-means method. In the Ionosphere dataset, the NMI, ARI, and FM clustering metrics of the CSA-DBSCAN algorithm are higher than the other algorithms, and the ACC metrics are slightly lower than those of the AmDPC method. In the Glass dataset, both the DeDPC and AmDPC algorithms have only one metric value higher than the other algorithms, respectively. In the Vehicle dataset, the ACC and FM metrics of the DeDPC algorithm are higher than the other algorithms, and the NMI and ARI metrics of the CSA-DBSCAN algorithm are higher than the other algorithms. In the Vehicle data set, the CSA-DBSCAN and DeDPC algorithms perform similarly. From the Iris dataset, we can see that the DPCSA algorithm achieves the best clustering evaluation metric values. The CSA-DBSCAN and AmDPC algorithms have the same and slightly lower clustering evaluation metric values than the DPCSA algorithm and are higher than the other clustering algorithms. Additionally, we can see that the CSA-DBSCAN algorithm only needs to input one parameter to obtain the best clustering result adaptively. Therefore, the parameters of the CSA-DBSCAN algorithm are simple and efficient. The CSA-DBSCAN method can redistribute these identified noisy points, which gives the new method better clustering capabilities and effectiveness in real-world datasets.
From Table 3, we can find that the clustering metric values of the DBSCAN algorithm differ from the clustering metric values of the CSA-DBSCAN algorithm. The main reason is that the data distribution in the high-dimensional dataset is discrete, and the DBSCAN method sometimes incorrectly identifies some boundary points as anomalies. The CSA-DBSCAN method can redistribute these identified noisy points, which makes the method better regarding clustering accuracy and effectiveness in the real dataset. In summary, the CSA-DBSCAN algorithm can handle datasets with a complex structure of inhomogeneous density.
The CSA-DBSCAN algorithm also has its unique advantages in practical applications. For complex image data, such as image segmentation and face recognition, the CSA-DBSCAN method can further explore the potential relationship between boundary points and anomalies, providing more reference values for image processing.

4.4. Performance Test of Outlier Processing after Fusion Deviation Theory

In this subsection, we will further explore the impact of introducing deviation theory to measure the compactness between data points on the clustering performance. We will select three representative synthetic datasets and two real-world datasets from Table 1 to fully measure the impact of deviation theory on the performance of outlier assignments. The nearest neighbor assignment mechanism of the CSA-DBSCAN algorithm is an extension of the KNN algorithm [33]. Before the use of deviation theory, the CSA-DBSCAN algorithm mainly relied on the KNN algorithm for the assignment of anomalies. Therefore, this assignment mechanism is influenced by the parameter k. Therefore, we will further discuss the effect of introducing the deviation theory and the variation of parameter value k on the clustering performance. We will control the other parameters of the CSA-DBSCAN algorithm in Table 2 and Table 3 to keep them constant and observe the effect of introducing deviation theory and different k values on the clustering results.
In Figure 8, Figure 9 and Figure 10, we present the results of the anomaly assignment for the three synthetic datasets Twocirclesnoise, Complex8, and Complex9. Figure 8a shows the results of the CSA-DBSCAN algorithm after iteration to obtain the optimal parameters, where the white points are the identified anomalies. Figure 8b,c shows the results of anomaly assignment for different parameter values k using the KNN method, respectively. After introducing the deviation theory to measure the compactness between data points, we directly assign the anomalies to the clustering results of the nearest category points according to the compactness relationship between the anomalies and other data points (see Figure 8d). Figure 9a and Figure 10a represent the anomaly identification results after iterations of the Complex8 and Complex9 datasets, respectively. In Figure 9b,c, we can see that some anomalies are assigned to other categories. In Figure 9d, we can see that the anomalies are well assigned to the corresponding categories. In Figure 10b, we can see that one outlier is incorrectly assigned to other classes along the direction of the red arrow. Both Figure 10c and Figure 10d display the accurate assignment of the outliers.
In Table 4, we give the values of the clustering evaluation metrics for the Twocirclesnoise, Complex8, Complex9, Seeds, and Vehicle datasets under different allocation strategies, respectively. Figure 8, Figure 9 and Figure 10 show the visual clustering results in Table 4. Seeds and Vehicle are real-world datasets. For easier observation, we bold the clustering results after introducing the deviation theory. From the table, we can see that different k values have an impact on the clustering performance, and the results of anomaly assignment using deviation theory have a significant improvement on the clustering performance. For the Seeds dataset, the clustering index value after introducing deviation theory is significantly higher than the KNN assignment method. In the Vehicle dataset, for NMI and ARI metric values, the method of assignation using the introduction of deviation theory always outperforms the method of assigning in other cases, and the variation in the values of ACC and FM is not significant. Therefore, overall, the Vehicle dataset has better performance for the assignment method with the introduction of deviation theory.
It can be seen that the KNN algorithm is used to directly assign outliers, and different parameter values k have a large impact on the clustering results. The deviation theory can measure the compactness between data points by processing the distance property in the data points. We assign anomalies directly to the nearest boundary points based on the compactness relationship between anomalies and category points. The assignment method with the introduction of deviation theory effectively solves the influence of the parameter k on the clustering performance and enables the CSA-DBSCAN algorithm to obtain more accurate clustering results. The assignment method with the introduction of deviation theory does not require any input parameters to complete the effective assignment of outliers. This allocation method greatly improves the effectiveness and stability of the CSA-DBSCAN algorithm’s outlier allocation.

4.5. Comparative Analysis of Algorithm Running Time

In this subsection, we compare and examine the algorithms’ running times in further detail. Before analyzing the algorithms’ running time, we provide the time complexity of the various methods (see Table 5). In Table 5, n represents the number of data points in the dataset, I represents the number of iterations, and K represents the number of clusters determined by the k-means method. According to Table 5, the DBSCAN algorithm has the lowest time complexity, followed by the k-means algorithm. The time complexity of the DPC, DPCSA, AmDPC, and DeDPC algorithms is similar and reasonable. Both the AP and CSA-DBSCAN algorithms are affected by the number of iterations during operation, so they have a higher time complexity. Theoretically, the time complexity of the AP algorithm is higher than that of the CSA-DBSCAN algorithm. However, the analysis of the runtime flow of the AP algorithm shows that the AP algorithm is only locally iterative during the runtime, while the CSA-DBSCAN algorithm is globally iterative during the runtime. In this case, it is difficult to further determine the magnitude of the time consumption of the AP and CSA-DBSCAN algorithms.
To further appreciate the time required to execute the various methods, we provide detailed running times for the 12 datasets in Table 2 and Table 3 (see Table 6). Each dataset in Table 6 corresponds to a different ordinal number. Table 6’s runtime parameter settings are identical to those in Table 2 and Table 3. As seen in Table 6, the DBSCAN algorithm has the shortest total runtime, while the k-means technique has the second shortest. The CSA-DBSCAN algorithm has the longest runtime compared to the other algorithms in the 12 datasets. Among them, the runtime consumption of the Complex8 and Complex9 datasets is 55.3762 and 86.5821 s, respectively, which far exceeds the runtime of the other algorithms. The Complex8 and Complex9 datasets have particularly large data volumes, and as a result, their runtime increases significantly. The CSA-DBSCAN algorithm is a fusion of CSA optimization and DBSCAN algorithm, and the overall running process is affected by the number of iterations, so the time the CSA-DBSCAN algorithm spends is higher than the other algorithms.
Table 6 includes a running time comparison (see Figure 11) to help visualize the time cost of the various algorithms. The CSA-DBSCAN, AP, and AmDPC algorithms have the longest overall runtimes, and runtimes greater than one second can be seen in Figure 11. We provide line comparison graphs of the CSA-DBSCAN, AP, and AmDPC algorithms to further observe the running time trends on different datasets (see Figure 11). In Figure 11, the horizontal coordinates represent the ordinal numbers corresponding to the different datasets in Table 6. We redesigned the scale of the vertical coordinates because of the excessive differences in the running times of the different algorithms. As a result, the time cost of different algorithms can all be compared and analyzed more intuitively. In Figure 11, we represent the algorithms with lower running times as bar charts. As can be seen in Figure 11, the runtime of the line graph is overall greater than one second, and the runtime of the bar graph is overall less than one second.
Figure 11 shows that the CSA-DBSCAN algorithm has the longest runtime, followed by the AP algorithm. The AmDPC algorithm has a shorter runtime than the CSA-DBSCAN and AP algorithms but a longer runtime than the others. The DBSCAN method has the shortest running time, followed by the k-means algorithm, while the DPC, DPCSA, and DeDPC algorithms have times that are comparable. The first six datasets in Figure 11 are synthetic, while the latter six are real-world datasets. From the description of the data attributes in Table 1, we can see that the amount of data in the synthetic dataset is significantly higher than that in the real-world dataset, and the running times of all eight algorithms are higher in the synthetic dataset than in the real-world dataset. In the real-world dataset, Ionosphere and Glass are both particularly high-dimensional datasets, but their runtime consumption is still lower than that of the synthetic dataset among these eight state-of-the-art algorithms.
Therefore, we may conclude that changing the dimensionality has little effect on the running time consumption of the algorithms. The amount of data in different datasets has a significant impact on an algorithm’s running time. Because the CSA-DBSCAN algorithm is a systematic fusion of CSA optimization and DBSCAN algorithm in an iterative process, its running time consumption is significant when compared to other algorithms.

4.6. Robustness Analysis of the CSA-DBSCAN Algorithm

In this subsection, we tested and analyzed the CSA-DBSCAN algorithm’s robustness. The CSA-DBSCAN algorithm is an adaptive method that fuses swarm intelligence optimization and a density clustering method. The method is divided into two stages: chameleon swarm optimization and density clustering. The process of optimizing the parameters of the DBSCAN algorithm using the chameleon swarm method is primarily influenced by two parameters: the number of iterations I and the number of populations n. We uniformly set these two parameters to fixed values of 60 and 10, respectively, before running the optimization algorithm. The DBSCAN algorithm is mainly influenced by the MinPts parameter. This parameter is usually set to a positive integer between 1 and 15. Therefore, we further verified the robustness of the CSA-DBSCAN method by analyzing the effect of different parameters on the clustering performance.
We separately tested the effects of the three CSA-DBSCAN algorithm parameters I, n, and MinPts on the clustering results and investigated their robustness. We began with the number of iterations I, which influences the CSA-DBSCAN algorithm’s parameter search quality. We set different values of I for testing: I ∈ [50, 500] (I is a positive integer), and then performed tests on three synthetic datasets and three real-world datasets, whose clustering results are shown in Table 7. The values in Table 7 are the means and standard deviations of the CSA-DBSCAN algorithm’s 10 experimental results for parameter I ranging from 50 to 500. We highlighted the mean and standard deviation of the better evaluation metrics’ results. To ensure experiment reliability, we kept the population size parameters n and MinPts constant to test the effect of the CSA-DBSCAN algorithm parameter I on clustering performance.
As expressed in Table 7, the mean values of the assessment metrics for the datasets Twocirclesnoise, Pearl, Complex8, and Glass are greater and have a standard deviation of 0, which suggests that the number of iterations does not affect the clustering outcomes. The mean values and low standard deviations of the evaluation measures for the Vehicle and Iris datasets are comparable to those in Table 3. Therefore, the clustering performance of the CSA-DBSCAN algorithm is sound for the Vehicle and Iris datasets. Generally, the variation of the clustering performance of the CSA-DBSCAN algorithm with the number of iterations I is robust. The main reason for the variation in the clustering results for the Vehicle and Iris datasets is that the optimization process is not optimal at 50 iterations. In the experimental process, we only need to set the number of iterations to be greater than 50 to achieve the optimal optimization of the algorithm.
In Table 8, we give a range of values for the number of chameleon populations, n. For the experiments, we set n to be between 5 and 55 and divided it into ten equal parts. Table 8 shows the mean and standard deviation of the clustering index values for ten experiments with the number of iterations I and MinPts held constant. The better values are highlighted in bold. From Table 8, we can see that the mean values of the indicators for the six datasets are the same as those in Table 2 and Table 3 and have zero standard deviation. Thus, it can be seen that the changes in the number of chameleon populations n and the number of iterations I hardly affect the clustering results. Therefore, the number of iterations I and the number of chameleon populations n are robust in the clustering process.
As shown in Table 9, we maintained the number of iterations I and the number of populations n of the optimization process as constant values to see how the DBSCAN algorithm parameter MinPts affects the clustering results. We set MinPts from 1 to 10 and split them into 10 equal portions. By computing the mean and standard deviation of the clustering metrics for the six datasets at 10 different MinPts, we were able to determine the sensitivity of the DBSCAN and CSA-DBSCAN algorithms to parameter changes. For easy observation, we bolded the improved mean and standard deviation in Table 9. From Table 9, we can see that the mean values of the CSA-DBSCAN algorithm are always higher than the DBSCAN algorithm for the six datasets. For the Twocirclesnoise, Pearl, and Complex8 datasets, the standard deviations of the CSA-DBSCAN algorithm are always lower than those of the DBSCAN method. This indicates that the CSA-DBSCAN algorithm is more robust with the variation of MinPts in these three datasets. The standard deviations of a few measures in the DBSCAN algorithm are marginally lower than those in the CSA-DBSCAN algorithm in the Glass, Vehicle, and Iris datasets but are otherwise extremely close. This indicates that the clustering results of the CSA-DBSCAN algorithm are better and more robust in these six datasets. This shows that when the MinPts values are changed, the CSA-DBSCAN algorithm is more robust than the DBSCAN method. In summary, the CSA-DBSCAN algorithm considerably enhances the DBSCAN algorithm’s soundness.
To examine the significance of the differences in performance between the CSA-DBSCAN and DBSCAN algorithms with differentMinPts settings in Table 9, we conduct statistical tests, as shown in Table 10. In Table 10, we use p-values to determine if there is a statistical difference in algorithm performance based on NMl, ARI, ACC, and FM evaluation criteria. As shown in Table 10, the p-values for the Pearl, Glass, and Iris datasets are less than 0.05, and in some cases, less than 0.01 for most measures, demonstrating that a performance difference between the CSA-DBSCAN and DBSCAN algorithms exists and is extremely statistically significant. Only the p-value for the NMI metric in the Twocirclesnoise and Complex8 datasets was less than 0.05, indicating a statistically significant difference between the NMI values in these two datasets. In the Vehicle dataset, the p-values for the ARI and FM measures are less than 0.01, showing a highly significant difference between these two metrics. In Table 10, we highlight the values of the indicators with significant differences. Combining the mean and standard deviation of the evaluated metric values in Table 9, we can see that the CSA-DBSCAN algorithm outperforms the DBSCAN method in general. Therefore, for the MinPts parameter, the CSA-DBSCAN algorithm is overall more robust than DBSCAN and has a statistically significant difference.
In conclusion, as the number of iterations I, the number of populations n, and the value of MinPts are modified, the CSA-DBSCAN algorithm can still produce superior clustering results. The clustering performance of the CSA-DBSCAN algorithm has a certain robustness and effectiveness.

4.7. Color Image Segmentation Clustering Applications

In this subsection, we selected four real images from the Berkeley segmentation dataset (BSDS500) to test the clustering performance of the CSA-DBSCAN algorithm’s image segmentation. In this experiment, the image segmentation evaluation metric F i t B D E was chosen as the fitness function to adaptively search for the optimal parameter values. In Figure 12, the original image is shown on the leftmost side, and the image segmentation results of eight state-of-the-art clustering methods, k-means, AP, DBSCAN, DPC, DPCSA, AmDPC, DeDPC, and CSA-DBSCAN, are shown from left to right. In order to reduce the computational effort, an effective superpixel SLIC algorithm was used in this paper to pre-segment the image into multiple superpixel information uniformly before image processing. Finally, this information is fed into the different algorithms mentioned above for testing.
To evaluate the effectiveness of the CSA-DBSCAN algorithm for practical applications in image segmentation, we give the boundary displacement error (BDE) [32] and probabilistic Rand index (PRI) metrics [38] for different image segmentation methods. PRI is calculated as the ratio of the number of pixels with identical algorithmic segmentation and multiple manual segmentation labels to the whole number of pixels; BDE is a measure of the average displacement error of the boundary pixels in both segmentation results. In Table 11, we give the segmentation metric metrics for the four real images in Figure 12 in order from top to bottom. The “↓” indicates that the smaller the BDE value is, the better the image segmentation is. The “↑” indicates that the larger the PRI value, the better the image segmentation effect, and the PRI value is in the range of (0, 1). The evaluation index values in Table 11 and Table 12 correspond to the clustering results in Figure 12. In the table, we bolded the best indicator values for a clearer view of the data.
From Table 11 and Table 12 and Figure 12, we can see that the first three image k-means, AP, DPC, DPCSA, AmDPC, and DeDPC algorithms counting from top to bottom do not segment the images as well as the DBSCAN and CSA-DBSCAN algorithms. Both the DBSCAN and CSA-DBSCAN methods segment the images well and accurately describe the overall image contour. In the last image, the AmDPC algorithm PRI metric value reaches its highest value, and the CSA-DBSCAN algorithm PRI value is very close compared to AmDPC. However, the BDE metric value of the CSA-DBSCAN algorithm is much higher than that of the AmDPC algorithm. Overall, the CSA-DBSCAN algorithm outperforms the AmDPC algorithm for the segmentation of the fourth image. As can be seen in Table 11 and Table 12, the BDE and PRI metric values of the DBSCAN method are slightly lower than those of the CSA-DBSCAN method. This indicates that the CSA-DBSCAN method has better image segmentation accuracy compared to the DBSCN algorithm.
Therefore, compared with other mainstream clustering algorithms, the CSA-DBSCAN algorithm has better clustering accuracy and effectiveness in the application of image segmentation. In summary, the CSA-DBSCAN algorithm has a certain effectiveness and practicality in adaptive search parameters and image segmentation.

5. Discussion

This research focuses on the fusion strategy of the swarm intelligence optimization adaptive clustering algorithm. Firstly, the Eps parameters of DBSCAN are utilized as individual candidate solutions for chameleons, and the optimal Eps clustering parameter values are obtained adaptively by the optimization mechanism of the CSA algorithm; then, the identified anomalies are found. Finally, the nearest neighbor mechanism of deviation theory is used to reacquire the compactness relationship between the data and assign the anomalies to the proximity cluster classes to obtain the final clustering results. The CSA-DBSCAN algorithm takes advantage of the CSA optimization and DBSCAN clustering performance to form a new adaptive clustering fusion algorithm. Compared with other algorithms, the CSA-DBSCAN algorithm can adaptively obtain the best clustering parameter values and clustering results, reducing the complexity of clustering tuning, enabling the clustering of arbitrarily structured complex datasets and the identification and assignment of outliers.
To boost computational efficiency in image processing, the CSA-DBSCAN algorithm employs the SLIC method to create superpixel information on the original image. The method then uses an updated way to segment the superpixel data based on the image color space features to achieve image segmentation using the adaptive CSA-DBSCAN method. The approach of pre-processing color photos with SLIC significantly reduces the algorithm’s computing time consumption. The CSA-DBSCAN algorithm is a single-parameter adaptive clustering algorithm. To discover the optimum clustering result quickly and reliably, the technique requires only one integer number parameter. When compared to existing algorithms, the modified approach performs exceptionally well in adaptive search and image segmentation.
The CSA-DBSCAN algorithm can handle color images well, but its performance for vector image data in space is unknown. For vector image data clustering, the DBSCAN algorithm’s Euclidean distance calculation is insufficient to tackle the associated special challenges. In the subsequent study, we will use the spatial clustering method with network path distances proposed by Yu [39], combined with the advantages of the CSA-DBSCAN algorithm, to cluster vector image data through a preprocessing step of mathematical morphology and vector-to-raster conversion.

6. Conclusions

In this research, we employ the most recent chameleon swarm optimization method for the adaptive optimization of the DBSCAN clustering algorithm parameter Eps, which solves the problem of Eps being particularly sensitive to clustering outcomes. To obtain the optimal clustering outcome, the CSA-DBSCAN algorithm requires only one parameter to be entered during operation. We employ the closest neighbor mechanism and deviation theory to assign noisy points to the nearest boundary points in the DBSCAN method to solve the problem of the over-identification of noisy points. The simulation results show that the improved CSA-DBSCAN method can handle complex structural data with arbitrary irregular density and has higher clustering accuracy and effect in high-dimensional datasets. The CSA-DBSCAN algorithm implements the fusion of the swarm intelligence algorithm, clustering algorithm, nearest neighbor mechanism, and deviation theory. Compared with other algorithms, this algorithm can find the best clustering results quickly and identify outliers more accurately, with better clustering precision and performance. The image processing experiment results suggest that the CSA-DBSCAN algorithm has a high practical value in the field of image segmentation.
However, the CSA-DBSCAN algorithm must also change the MinPts parameter, which requires considerable tuning time for unknown data. We will apply multi-objective optimization algorithms with stronger optimization capabilities in the future to achieve optimal clustering with the benefit of parameter-free clustering.

Author Contributions

Conceptualization, W.Z. and L.W.; methodology, W.Z.; software, W.Z.; validation, X.H., Y.W. and Y.Z.; formal analysis, Z.J.; investigation, Y.W.; resources, W.Z.; data curation, W.Z.; writing—original draft preparation, W.Z.; writing—review and editing, W.Z.; visualization, Y.W.; supervision, L.W.; project administration, L.W.; funding acquisition, L.W. All authors have read and agreed to the published version of the manuscript.

Funding

The research was funded by the National Social Science Foundation, grant number 22BTJ057.

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Not Applicable.

Data Availability Statement

The real-world datasets can be obtained from https://archive.ics.uci.edu/ml/index.php (accessed on on 14 July 2022). The BSD500 datasets can be obtained from https://www2.eecs.berkeley.edu/Research/Projects/CS/vision/grouping/resources.html (accessed on 25 June 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Philip Chen, C.L.; Zhang, C.Y. Data-intensive applications, challenges, techniques and technologies: A survey on Big Data. Inf. Sci. 2014, 275, 314–347. [Google Scholar] [CrossRef]
  2. Zamora, J.; Sublime, J. An Ensemble and Multi-View Clustering Method Based on Kolmogorov Complexity. Entropy 2023, 25, 371. [Google Scholar] [CrossRef]
  3. Li, M.; Bi, X.; Wang, L.; Xuming, H.; Wang, L.; Zhou, W. Text Similarity Measurement Method and Application of Online Medical Community Based on Density Peak Clustering. J. Organ. End User Comput. 2022, 34, 1–25. [Google Scholar] [CrossRef]
  4. Wang, Y.; Wang, D.; Zhang, X.; Pang, W.; Miao, C.; Tan, A.H.; Zhou, Y. McDPC: Multi-center density peak clustering. Neural. Comput. Appl. 2020, 32, 13465–13478. [Google Scholar] [CrossRef]
  5. Ji, S.; Zhu, H.; Wang, P.; Ling, X. Image clustering algorithm using superpixel segmentation and non-symmetric gaussian–cauchy mixture model. IET Image Process. 2020, 14, 4132–4143. [Google Scholar] [CrossRef]
  6. Li, M.; Bi, X.; Wang, L.; Han, X. A method of two-stage clustering learning based on improved DBSCAN and density peak algorithm. Comput. Commun. 2021, 167, 75–84. [Google Scholar] [CrossRef]
  7. Mújica-Vargas, D.; Matuz-Cruz, M.; García-Aquino, C.; Ramos-Palencia, C. Efficient System for Delimitation of Benign and Malignant Breast Masses. Entropy 2022, 24, 1775. [Google Scholar] [CrossRef]
  8. Namburu, A.; Samayamantula, S.K.; Edara, S.R. Generalised rough intuitionistic fuzzy c-means for magnetic resonance brain image segmentation. IET Image Process. 2017, 11, 777–785. [Google Scholar] [CrossRef]
  9. Guan, J.; Li, S.; He, X.; Zhu, J.; Chen, J. Fast hierarchical clustering of local density peaks via an association degree transfer method. Neurocomputing 2021, 455, 201–218. [Google Scholar] [CrossRef]
  10. Rodriguez, A.; Laio, A. Clustering by fast search and find of density peaks. Science 2014, 344, 1492–1496. [Google Scholar] [CrossRef]
  11. Du, M.; Wu, F. Grid-Based Clustering Using Boundary Detection. Entropy 2022, 24, 1606. [Google Scholar] [CrossRef] [PubMed]
  12. Zhang, Y.; Brady, M.; Smith, S. Segmentation of brain MR images through a hidden markov random field model and the expectation-maximization algorithm. IEEE Trans. Med. Imaging 2001, 20, 45–57. [Google Scholar] [CrossRef] [PubMed]
  13. Kumar, D.K.S.; D’Mello, D.A. DPISCAN: Distributed and parallel architecture with indexing for structural clustering of massive dynamic graphs. Int. J. Data Sci. Anal. 2022, 13, 199–233. [Google Scholar] [CrossRef]
  14. Hartigan, J.A.; Wong, M.A. A K-Means Clustering Algorithm. Appl. Stat. 1979, 28, 100–108. [Google Scholar] [CrossRef]
  15. Frey, B.J.; Dueck, D. Clustering by Passing Messages between Data Points. Science 2007, 315, 972–976. [Google Scholar] [CrossRef]
  16. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. A Density-Based Algorithm for Discovering Clusters in Large Spatial Databases with Noise. In Proceedings of the Second International Conference on Knowledge Discovery and Data Mining (KDD’96), Portland, OR, USA, 2–4 August 1996; AAAI Press: Washington, DC, USA, 1996; pp. 226–231. [Google Scholar]
  17. Zhang, S.; Chen, J. Optimization of energy-efficient dynamic task assignment for wireless sensor networks based on particle swarm algorithm. J. Intell. Fuzzy Syst. 2016, 20, 1–11. [Google Scholar] [CrossRef]
  18. Wang, G.G.; Deb, S.; Gandomi, A.H.; Zhang, Z.; Alavi, A.H. Chaotic cuckoo search. Soft Comput. 2016, 20, 3349–3362. [Google Scholar] [CrossRef]
  19. Meng, X.B.; Gao, X.Z.; Liu, L. A new bio-inspired optimisation algorithm: Bird swarm algorithm. J. Exp. Theor. Artif. Intell. 2015, 28, 673–687. [Google Scholar] [CrossRef]
  20. Braik, M.S. Chameleon Swarm Algorithm: A bio-inspired optimizer for solving engineering design problems. Expert Syst. Appl. 2021, 174, 114685. [Google Scholar] [CrossRef]
  21. Smiti, A.; Eloudi, Z. Soft DBSCAN: Improving DBSCAN clustering method using fuzzy set theory. In Proceedings of the 2013 6th International Conference on Human System Interactions (HSI), Sopot, Poland, 6–8 June 2013; pp. 380–385. [Google Scholar]
  22. Galán, S.F. Comparative evaluation of region query strategies for DBSCAN clustering. Inf. Sci. 2019, 502, 76–90. [Google Scholar] [CrossRef]
  23. Bryant, A.; Cios, K. RNN-DBSCAN: A density-based clustering algorithm using reverse nearest neighbor density estimates. IEEE Trans. Knowl. Data Eng. 2018, 30, 1109–1121. [Google Scholar] [CrossRef]
  24. Hou, J.; Gao, H.; Li, X. DSets-DBSCAN: A parameter-free clustering algorithm. IEEE Trans. Image Process. 2016, 25, 3182–3193. [Google Scholar] [CrossRef]
  25. Liu, Y.H.; Ma, Z.M.; Yu, F. Adaptive density peaks clustering based on k-nearest neighbors with aggregating strategy. Knowl. Based Syst. 2017, 133, 208–220. [Google Scholar] [CrossRef]
  26. Shi, B.; Han, L.; Yan, H. Adaptive clustering algorithm based on knn and density. Pattern Recognit. Lett. 2018, 104, 37–44. [Google Scholar] [CrossRef]
  27. Zhang, W.; Zhi, Y.; Qin, L. DBSCAN Clustering Algorithm Based on Improved Beetle Swarm Optimization. Stat. Decis. 2022, 10, 20–25. [Google Scholar]
  28. Jin, L.; He, Y.; Wang, H.; Huo, Z.; Xie, X.; Guo, B. Point Cloud Segmentation Algorithm Based on Adaptive Threshold DBSCAN for Roadside LiDAR. Automot. Eng. 2022, 44, 987–996. [Google Scholar]
  29. Zhu, Q.; Tang, X.; Elahi, A. Application of the novel harmony search optimization algorithm for DBSCAN clustering. Expert Syst. Appl. 2021, 178, 115054. [Google Scholar] [CrossRef]
  30. Wang, L.; Wang, H.; Han, X.; Zhou, W. A novel adaptive density-based spatial clustering of application with noise based on bird swarm optimization algorithm. Comput. Commun. 2021, 174, 205–214. [Google Scholar] [CrossRef]
  31. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Sabine, S. SLIC Superpixels Compared to State-of-the-Art Superpixel Methods. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2274–2282. [Google Scholar] [CrossRef]
  32. Wang, X.; Tang, Y.; Masnou, S.; Chen, L. A Global/Local Affinity Graph for Image Segmentation. IEEE Trans. Image Process. 2015, 24, 1399–1411. [Google Scholar] [CrossRef]
  33. Zhang, B.; Srihari, S.N. Fast k-nearest neighbor classification using cluster-based trees. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 525–528. [Google Scholar] [CrossRef] [PubMed]
  34. Arbeláez, P.; Maire, M.; Fowlkes, C.; Malik, J. Contour Detection and Hierarchical Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 898–916. [Google Scholar] [CrossRef] [PubMed]
  35. Yu, D.; Liu, G.; Guo, M.; Liu, X.; Yao, S. Density Peaks Clustering Based on Weighted Local Density Sequence and Nearest Neighbor Assignment. IEEE Access 2019, 7, 34301–34317. [Google Scholar] [CrossRef]
  36. Zhou, W.; Wang, L.; Han, X.; Parmar, M.; Li, M. A novel density deviation multi-peaks automatic clustering algorithm. Complex Intell. Syst. 2023, 9, 177–211. [Google Scholar] [CrossRef]
  37. Zhou, W.; Wang, L.; Han, X.; Li, M. A novel deviation density peaks clustering algorithm and its applications of medical image segmentation. IET Image Process. 2022, 16, 3790–3804. [Google Scholar] [CrossRef]
  38. Unnikrishnan, R.; Pantofaru, C.; Hebert, M. Toward Objective Evaluation of Image Segmentation Algorithms. IEEE Comput. Soc. 2007, 29, 929–944. [Google Scholar] [CrossRef] [PubMed]
  39. Yu, W. A mathematical morphology based method for hierarchical clustering analysis of spatial points on street networks. Appl. Soft Comput. 2019, 85, 105785. [Google Scholar] [CrossRef]
Figure 1. The anomaly assignment process on the Complex8 dataset.
Figure 1. The anomaly assignment process on the Complex8 dataset.
Entropy 25 00782 g001
Figure 2. Clustering results on Donutcurves dataset.
Figure 2. Clustering results on Donutcurves dataset.
Entropy 25 00782 g002
Figure 3. Clustering results on Target dataset.
Figure 3. Clustering results on Target dataset.
Entropy 25 00782 g003
Figure 4. Clustering results on Twocirclesnoise dataset.
Figure 4. Clustering results on Twocirclesnoise dataset.
Entropy 25 00782 g004
Figure 5. Clustering results on Pearl dataset.
Figure 5. Clustering results on Pearl dataset.
Entropy 25 00782 g005
Figure 6. Clustering results on Complex8 dataset.
Figure 6. Clustering results on Complex8 dataset.
Entropy 25 00782 g006
Figure 7. Clustering results on Complex9 dataset.
Figure 7. Clustering results on Complex9 dataset.
Entropy 25 00782 g007
Figure 8. Results of outlier assignment on Twocirclesnoise dataset.
Figure 8. Results of outlier assignment on Twocirclesnoise dataset.
Entropy 25 00782 g008
Figure 9. Results of outlier assignment on Complex8 dataset.
Figure 9. Results of outlier assignment on Complex8 dataset.
Entropy 25 00782 g009
Figure 10. Results of outlier assignment on Complex9 dataset.
Figure 10. Results of outlier assignment on Complex9 dataset.
Entropy 25 00782 g010
Figure 11. Comparison of the running time of different algorithms.
Figure 11. Comparison of the running time of different algorithms.
Entropy 25 00782 g011
Figure 12. Image segmentation results of different clustering algorithms.
Figure 12. Image segmentation results of different clustering algorithms.
Entropy 25 00782 g012
Table 1. Description of the experimental datasets.
Table 1. Description of the experimental datasets.
DatasetObjectsDimensionClusters
Donutcurves100024
Target77026
Twocirclesnoise61023
Pearl100022
Complex8255128
Complex9303129
Seeds21073
Thyroid21553
Ionosphere351302
Glass21496
Vehicle846184
Iris15043
Table 2. Description of clustering evaluation on synthetic datasets.
Table 2. Description of clustering evaluation on synthetic datasets.
DatasetsAlgorithmsParNMIARIACCFM
Donutcurvesk-means40.68210.15250.58400.6796
AP−6.41250.58020.40850.52900.6139
DBSCAN0.01/20.99770.99870.99900.9990
DPC30.94290.91300.88100.9353
DPCSA-0.94280.91280.87500.9351
AmDPC3/0.05/0.15/80.86600.55510.75000.8160
DeDPC11/50.89660.82400.78200.8702
CSA-DBSCAN21.00001.00001.00001.0000
Targetk-means60.75070.10480.70000.8306
AP−82.82590.69050.65570.65320.8055
DBSCAN0.3/21.00001.00001.00001.0000
DPC200.77670.70200.70520.8341
DPCSA-0.65250.62340.63510.7850
AmDPC5/1/0.39/0.30.97710.99240.98830.9961
DeDPC8/20.66980.65190.70780.8031
CSA-DBSCAN21.00001.00001.00001.0000
Twocirclesnoisek-means30.0039-0.22000.33770.4026
AP−97.43050.52050.18600.21970.4281
DBSCAN0.8/40.95740.98350.99180.9916
DPC80.43960.38140.62620.6619
DPCSA-0.35740.26630.69510.6456
AmDPC3/1.6/1.9/51.00001.00001.00001.0000
DeDPC1/21.00001.00001.00001.0000
CSA-DBSCAN31.00001.00001.00001.0000
Pearlk-means20.39200.45110.83600.7310
AP−158.35150.64540.60670.66200.7790
DBSCAN0.575/40.99480.99800.99900.9990
DPC80.78480.66820.68600.8173
DPCSA-1.00001.00001.00001.0000
AmDPC3/2/2/51.00001.00001.00001.0000
DeDPC17/20.75560.69330.74800.8325
CSA-DBSCAN41.00001.00001.00001.0000
Complex8k-means80.6129−0.04300.52140.4954
AP−6349.19390.64400.39360.46060.4891
DBSCAN15.58/30.99850.99960.99880.9997
DPC50.78120.51130.73270.7177
DPCSA-0.84050.74700.75890.7907
AmDPC3/50/38.6/160.85740.75280.80090.7955
DeDPC20/20.74580.62640.64520.6876
CSA-DBSCAN21.00001.00001.00001.0000
Complex9k-means90.66840.02700.49850.3145
AP−2972.51680.73800.28450.35830.4393
DBSCAN15.12/100.99420.99750.99700.9979
DPC30.71670.44620.57370.5627
DPCSA-0.71990.42210.45930.5188
AmDPC3/20/21/500.91320.89190.87460.9132
DeDPC1/20.72970.33890.64570.6213
CSA-DBSCAN81.00001.00001.00001.0000
Table 3. Description of clustering evaluation on real-world datasets.
Table 3. Description of clustering evaluation on real-world datasets.
DatasetsAlgorithmsParNMIARIACCFM
Seedsk-means30.69490.71660.89520.8106
AP−61.15250.71010.71030.89050.8068
DBSCAN0.92/40.48230.40010.59520.6439
DPC60.71880.71320.89050.8093
DPCSA-0.71510.72360.89520.8166
AmDPC10/1.5/3/100.71880.71320.89050.8093
DeDPC15/150.71880.71320.89050.8093
CSA-DBSCAN110.72590.73210.90000.8212
Thyroidk-means30.43540.58230.85580.8062
AP−15.03330.47280.08450.15350.3027
DBSCAN4.28/10.50130.69150.72090.8648
DPC60.37640.35350.78600.7828
DPCSA-0.35170.31850.79540.7750
AmDPC15/5.8/5/20.42560.41140.82330.7952
DeDPC10/150.43130.42400.82790.7978
CSA-DBSCAN10.50130.69150.72090.8648
Ionospherek-means20.12920.16810.70660.6004
AP−80.40750.12430.16340.70370.5983
DBSCAN1.4/10.36970.51120.60680.7526
DPC20.26690.25180.65810.6082
DPCSA-0.12240.19690.72650.6335
AmDPC3/1.3/0.87/20.32120.39270.81770.7553
DeDPC3/50.25860.26120.71230.6453
CSA-DBSCAN10.61540.37360.52890.7640
Glassk-means60.43530.23890.47660.5493
AP−46.67390.29690.17500.44390.5408
DBSCAN0.5/40.33270.17710.41590.4381
DPC10.30530.17070.45790.4784
DPCSA-0.29710.17720.45330.5384
AmDPC4/1.5/3/50.40080.19460.48600.5700
DeDPC18/280.45400.25100.50930.4938
CSA-DBSCAN10.52690.29300.46730.5570
Vehiclek-means40.11630.07830.34990.3422
AP−30.70220.18760.11650.34400.3085
DBSCAN0.55/100.21460.06020.40310.3993
DPC20.16030.08640.37230.4280
DPCSA-0.20210.10980.40540.4074
AmDPC5/0.77/0.71/0.80.18630.01160.37590.4333
DeDPC17/50.19530.00280.40540.4349
CSA-DBSCAN40.21600.12660.34100.3569
Irisk-means30.75820.89330.73020.8208
AP−34.93740.76120.66670.37050.7715
DBSCAN0.4/50.71960.70630.80670.7972
DPC20.80580.75920.90670.8407
DPCSA-0.88510.90380.96670.9355
AmDPC4/1/1/10.87050.88580.96000.9234
DeDPC5/20.80580.75920.90670.8407
CSA-DBSCAN50.87050.88580.96000.9234
Table 4. Clustering evaluation index values of different outlier allocation strategies.
Table 4. Clustering evaluation index values of different outlier allocation strategies.
DatasetsAllocation StrategyNMIARIACCFM
Twocirclesniosek = 60.98270.99340.99670.9967
k = 21.00001.00001.00001.0000
Deviation theory1.00001.00001.00001.0000
Complex8k = 1070.99860.99940.99960.9995
k = 1080.99730.99870.99920.9989
Deviation theory1.00001.00001.00001.0000
Complex9k = 800.99890.99970.99970.9998
k = 21.00001.00001.00001.0000
Deviation theory1.00001.00001.00001.0000
Seedsk = 30.71010.71030.89050.8068
k = 40.71140.72980.90000.8195
Deviation theory0.72590.73210.90000.8212
Vehiclek = 40.21490.11590.34280.3595
k = 140.18550.09700.34750.3649
Deviation theory0.21600.12660.34160.3569
Table 5. Time complexity comparison.
Table 5. Time complexity comparison.
Algorithmsk-MeansAPDBSCANDPCDPCSAAmDPCDeDPCCSA-DBSCAN
Time Complexity O ( I K n ) O ( I n 2 ) O ( n l o g n ) O ( n 2 ) O ( n 2 ) O ( n 2 ) O ( n 2 ) O ( I n l o g n )
Table 6. Running time for different algorithms (in s).
Table 6. Running time for different algorithms (in s).
No.Datasetsk-MeansAPDBSCANDPCDPCSAAmDPCDeDPCCSA-DBSCAN
1Donutcurves0.47994.37140.15110.84410.94280.89270.958211.9972
2Target0.44932.73220.27131.22560.66700.87410.95138.2078
3Twocirclesnoise0.36071.80380.24270.87850.55241.06530.83243.9055
4Pearl0.44384.48420.27990.92780.75041.13950.861712.6280
5Complex81.403232.71250.37981.47181.27322.18081.253355.3762
6Complex92.417249.04490.64961.76111.81372.34491.221386.5821
7Seeds0.32890.75860.15530.70710.32640.71680.79371.6091
8Thyroid0.32850.81310.29530.67460.29890.81530.78431.5749
9Ionosphere0.12921.01160.71090.77370.41640.93750.80963.4974
10Glass0.55610.75680.23750.81590.39571.30380.87641.7904
11Vehicle0.77943.29290.23370.86350.50841.31280.856711.3781
12Iris0.39830.68630.10590.69820.31150.69150.82731.0434
Table 7. Robustness analysis of the CSA-DBSCAN algorithm with different numbers of iterations I.
Table 7. Robustness analysis of the CSA-DBSCAN algorithm with different numbers of iterations I.
Datasets I ∈ [50, 500]NMIARIACCFM
Twocirclesnoise1.0000 ± 0.00001.0000 ± 0.00001.0000 ± 0.00001.0000 ± 0.0000
Pearl1.0000 ± 0.00001.0000 ± 0.00001.0000 ± 0.00001.0000 ± 0.0000
Complex81.0000 ± 0.00001.0000 ± 0.00001.0000 ± 0.00001.0000 ± 0.0000
Glass0.5269 ± 0.00000.2930 ± 0.00000.4673 ± 0.00000.5570 ± 0.0000
Vehicle0.1957 ± 0.05840.1133 ± 0.03800.3337 ± 0.02460.3722 ± 0.0435
Iris0.8486 ± 0.04610.7827 ± 0.21730.9013 ± 0.12370.8930 ± 0.0641
Table 8. Robustness analysis of the CSA-DBSCAN algorithm with different chameleon numbers n.
Table 8. Robustness analysis of the CSA-DBSCAN algorithm with different chameleon numbers n.
Datasets   n ∈ [5, 55]NMIARIACCFM
Twocirclesnoise1.0000 ± 0.00001.0000 ± 0.00001.0000 ± 0.00001.0000 ± 0.0000
Pearl1.0000 ± 0.00001.0000 ± 0.00001.0000 ± 0.00001.0000 ± 0.0000
Complex81.0000 ± 0.00001.0000 ± 0.00001.0000 ± 0.00001.0000 ± 0.0000
Glass0.5269 ± 0.00000.2930 ± 0.00000.4673 ± 0.00000.5570 ± 0.0000
Vehicle0.2160 ± 0.00000.1266 ± 0.00000.3416 ± 0.00000.3569 ± 0.0000
Iris0.8705 ± 0.00000.8858 ± 0.00000.9600 ± 0.00000.9234 ± 0.0000
Table 9. Robustness analysis of the CSA-DBSCAN algorithm with different values of MinPts.
Table 9. Robustness analysis of the CSA-DBSCAN algorithm with different values of MinPts.
Datasets MinPts ∈ [1, 10]AlgorithmsNMIARIACCFM
TwocirclesnioseDBSCAN0.8815 ± 0.15580.9005 ± 0.17160.9234 ± 0.14530.9434 ± 0.0997
CSA-DBSCAN0.9827 ± 0.01490.9902 ± 0.00850.9951 ± 0.00420.9951 ± 0.0042
PearlDBSCAN0.9903 ± 0.00760.9941 ± 0.00280.9970 ± 0.00140.9970 ± 0.0014
CSA-DBSCAN0.9990 ± 0.00320.9996 ± 0.00130.9998 ± 0.00060.9998 ± 0.0006
Complex8DBSCAN0.9644 ± 0.04880.9620 ± 0.06350.9545 ± 0.06560.9691 ± 0.0514
CSA-DBSCAN0.9840 ± 0.01810.9876 ± 0.01720.9784 ± 0.02420.9898 ± 0.0142
GlassDBSCAN0.2902 ± 0.11380.1194 ± 0.08020.3925 ± 0.03360.4231 ± 0.0193
CSA-DBSCAN0.3580 ± 0.08850.2248 ± 0.04470.4579± 0.03370.5474 ± 0.0111
VehicleDBSCAN0.1658 ± 0.04070.0561 ± 0.03750.3369 ± 0.05630.3529 ± 0.0220
CSA-DBSCAN0.1878± 0.04810.1022 ± 0.02020.3482 ± 0.01750.3724± 0.0252
IrisDBSCAN0.6710 ± 0.10030.5889 ± 0.10510.7427 ± 0.05630.7198 ± 0.0682
CSA-DBSCAN0.8023 ± 0.04500.7219± 0.19210.8493± 0.10540.8511 ± 0.0547
Table 10. The robustness statistics test results (p-value) with different MinPts values.
Table 10. The robustness statistics test results (p-value) with different MinPts values.
Datasets MinPts ∈ [1, 10]NMIARIACCFM
Twocirclesnoise0.03090.06300.07680.0658
Pearl0.00020.00020.00030.0002
Complex80.04370.06400.06340.0626
Glass0.00010.00000.00000.0000
Vehicle0.15690.00780.21470.0031
Iris0.00000.01480.00910.0000
Table 11. Image segmentation BDE index comparison.
Table 11. Image segmentation BDE index comparison.
No.k-MeansAPDBSCANDPCDPCSAAmDPCDeDPCCSA-DBSCAN
Image 115.766219.71505.093715.47328.51497.453010.73064.9178
Image 211.568014.21347.61859.94519.945111.265013.53907.5544
Image 39.563112.21076.773410.05327.165411.91579.54746.1981
Image 416.413616.583012.575418.816811.873112.051913.647211.7218
Table 12. Image segmentation PRI index comparison.
Table 12. Image segmentation PRI index comparison.
No.k-MeansAPDBSCANDPCDPCSAAmDPCDeDPCCSA-DBSCAN
Image 10.60630.54100.88750.64020.68510.67570.62270.8927
Image 20.84670.82050.90470.88140.88140.86570.83230.9051
Image 30.85510.81460.89340.85930.89450.82520.56060.8997
Image 40.74790.74030.78870.77300.78230.80830.76150.7990
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, W.; Wang, L.; Han, X.; Wang, Y.; Zhang, Y.; Jia, Z. Adaptive Density Spatial Clustering Method Fusing Chameleon Swarm Algorithm. Entropy 2023, 25, 782. https://doi.org/10.3390/e25050782

AMA Style

Zhou W, Wang L, Han X, Wang Y, Zhang Y, Jia Z. Adaptive Density Spatial Clustering Method Fusing Chameleon Swarm Algorithm. Entropy. 2023; 25(5):782. https://doi.org/10.3390/e25050782

Chicago/Turabian Style

Zhou, Wei, Limin Wang, Xuming Han, Yizhang Wang, Yufei Zhang, and Zhiyao Jia. 2023. "Adaptive Density Spatial Clustering Method Fusing Chameleon Swarm Algorithm" Entropy 25, no. 5: 782. https://doi.org/10.3390/e25050782

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop