Next Article in Journal
Discrete Time Hybrid Semi-Markov Models in Manpower Planning
Next Article in Special Issue
Special Issue “Statistical Data Modeling and Machine Learning with Applications”
Previous Article in Journal
Calibrating the CreditRisk+ Model at Different Time Scales and in Presence of Temporal Autocorrelation
Previous Article in Special Issue
A Cascade Deep Forest Model for Breast Cancer Subtype Classification Using Multi-Omics Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Self-Expressive Kernel Subspace Clustering Algorithm for Categorical Data with Embedded Feature Selection

1
Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
2
Shenzhen College of Advanced Technology, University of Chinese Academy of Sciences, Shenzhen 518055, China
3
Department of Computer Science, University of Sherbrooke, Sherbrooke, QC J1K 2R1, Canada
4
College of Computer and Cyber Security, Fujian Normal University, Fuzhou 350007, China
*
Author to whom correspondence should be addressed.
Mathematics 2021, 9(14), 1680; https://doi.org/10.3390/math9141680
Submission received: 18 June 2021 / Revised: 11 July 2021 / Accepted: 14 July 2021 / Published: 16 July 2021
(This article belongs to the Special Issue Statistical Data Modeling and Machine Learning with Applications)

Abstract

:
Kernel clustering of categorical data is a useful tool to process the separable datasets and has been employed in many disciplines. Despite recent efforts, existing methods for kernel clustering remain a significant challenge due to the assumption of feature independence and equal weights. In this study, we propose a self-expressive kernel subspace clustering algorithm for categorical data (SKSCC) using the self-expressive kernel density estimation (SKDE) scheme, as well as a new feature-weighted non-linear similarity measurement. In the SKSCC algorithm, we propose an effective non-linear optimization method to solve the clustering algorithm’s objective function, which not only considers the relationship between attributes in a non-linear space but also assigns a weight to each attribute in the algorithm to measure the degree of correlation. A series of experiments on some widely used synthetic and real-world datasets demonstrated the better effectiveness and efficiency of the proposed algorithm compared with other state-of-the-art methods, in terms of non-linear relationship exploration among attributes.

1. Introduction

One of the goals of clustering is to mine the internal structure and characteristics of unlabeled data, which is known as unsupervised learning [1,2]. Real-world applications, i.e., pattern recognition [3], text mining [4], image retrieval [5], and bioinformatics [6], generate unlabeled data. All of these data are not just numerical data but are increasingly categorical data, which are flooding into practical applications. Clustering analysis for categorical data has attracted a great deal of interest from the scientific community. One example is that political philosophy is often measured as liberal, moderate, or conservative. Another example is that breast cancer diagnoses based on a mammograms use the categories normal, benign, probably benign, suspicious, and malignant.
In the past few decades, various clustering algorithms have been proposed [7,8,9,10,11] for numerical data. However, the attributes of categorical data are discrete, and their attribute values come from a limited symbol set. Unlike continuous data, categorical data are unable to produce a mathematical calculation, such as the mean and standard deviation. As a result, algorithms suitable for continuous data cannot be directly used for categorical data. To deal with this disadvantage, researchers have developed some clustering algorithms for categorical data, such as ROCK [12], ScaLable Information Bottleneck (LIMBO) [13,14], MGR [15], DHCC [16], and k-modes type algorithms [17,18,19,20,21,22,23]. However, each of these algorithms has its own merits and disadvantages. Even state-of-the-art algorithms have their shortcomings, and they are not effective for all datasets. For instance, ROCK is a non-k-mode agglomerative hierarchical clustering method that uses the conventional Jaccard coefficient to compute the similarity of two samples. However, the Jaccard cannot measure the specific value of the difference; it can only obtain whether the result is the same or not. In addition, the time complexity of this algorithm is high, which is quadratic with the number of objects. LIMBO uses an agglomerative information bottleneck to measure the entities’ distance, but is not comprehensive enough to extract data clustering features. The MGR algorithm proposes a mean gain ratio to select cluster attributes. LIMBO and MGR are based on information theory, meaning that they can quickly take into account one related variable, but one only, while ignoring other important feature information. DHCC can analyze multiple correspondence, avoiding a one-to-one similarity calculation. However, this method is sensitive to strange objects and, compared with agglomerative approaches, DHCC is a divisive algorithm with less application. The conventional k-modes algorithm and its variants have been extensively used for categorical data clustering. The distance of the samples was measured by simple matching coefficient (SMC). However, these methods only consider the attributes’ mode, while ignoring the statistical information of the data itself. Meanwhile, they can be trapped into local optima and are sensitive to initial clusters and modes. Our numerical experiments even showed that the k-modes algorithm could not identify the optimal clustering results for some particular datasets, regardless of the selection of the initial centers.
To solve the k-modes type algorithms’ problems, Chen [24] proposed a probabilistic framework in which the kernel bandwidth was introduced with a soft feature selection scheme so that the cluster center equals to the smoothed frequency estimator for the categories. Feature selection is of great significance to data processing in the era of big data [25,26]. It often involves the process of selecting the most important features representing an object’s attributes and then building a learning model in tasks clustering. Feature selection can not only relieve the curse of dimensionality caused by too many attributes but can also retain relevant features, remove irrelevant features, reduce the difficulty of learning tasks, and look for the essential features. Based on evaluation criteria, embedded feature selection methods such as CART [27] not only overcome the low efficiency of the wrapper feature selection method [28,29,30] but also avoid the disconnection of the filter feature selection method. Algorithms that take a filter-method approach to feature selection, such as Chi-Square [31], information gain [32], gain ratio [33], support vector machine [34,35], ReliefF [36,37], and hybrid ReliefF [38,39], are used in many practical applications. The embedded feature selection approach uses a learning model, so that the feature selection process is automatically integrated with the learner training process. Although several clustering analysis methods employ feature selection [24,40], many of the current approaches have one or more of the following disadvantages: considering all features independently, considering all attributes’ importance equally, and lack of an optimization solution.
The kernel clustering method that increases the sample features’ optimization process uses the Mercer kernel to map the samples in the input space to the high-dimensional feature space and clusters in the feature space. The kernel clustering method is widely used and is considered superior to classical clustering algorithms in performance. It can distinguish, extract, and enlarge useful features through non-linear mapping, so as to achieve more accurate clustering. Kernel k-means algorithm [41] makes the sample linearly separable (or nearly linearly separable) in kernel space by the “kernel” method. Still, the kernel function is defined for continuous data. Thus it cannot be directly transposed to categorical data and the algorithm based on the assumption that the original features are equally important. Some recent self-expressiveness-based methods [42,43,44] use subspace self-expressiveness property related to regularization terms. They are also not suitable for categorical data, and they all involve a linear combination of attributes.
In this paper, we view the task of clustering categorical data from a kernel clustering approach and propose a non-linear clustering algorithm for categorical data. The algorithm, named self-expressive kernel subspace clustering for categorical data (SKSCC), is based on the kernel density estimation (KDE) and probability-based similarity measurement. SKSCC not only considers the relationship between attributes in non-linear space but also gives each attribute a feature weight to measure the correlation degree. KDE has been employed in the estimation of probability distribution for categorical data [24,45,46]. This work introduces the self-expressive kernel density estimation (SKDE) in which every attribute has its own bandwidth. It then proposes a new non-linear similarity measurement method for categorical data in which a weight is added for each attribute to determine the importance of the attribute. Therefore, the objective function of the derived clustering algorithm is non-linear. As is commonly accepted, non-linear equations and equalities are not easy to solve. Therefore, we propose an efficient non-linear optimization method to solve the objective function of the clustering algorithm.
In summary, the main contributions of our work are as follows:
  • We define the self-expressive kernel density estimation approach, in which the symbols can be expressed by probability that is proportional to the kernel bandwidth, and the cluster center is smoothed to the frequency estimator for the categories;
  • We propose a non-linear feature-weighted similarity measurement method that gives consideration to the relationship between the attributes;
  • We put forward a non-linear optimization method in kernel subspace. Furthermore, we present the SKSCC, an efficient self-expressive kernel subspace clustering algorithm for categorical data that uses feature selection to choose the important attributes;
  • A series of experiments on several synthesis and real-world datasets were conducted to compare the performance of the proposed algorithm. The experimental results show that the proposed algorithm outperforms other algorithms in terms of non-linear relationship exploration among attributes and improves the performance and efficiency of clustering.
The remainder of this paper is organized as follows: Section 2 describes related work. Section 3 introduces the KDE-based similarity for categorical data. In Section 4, the new clustering algorithm is elaborated. Experimental results are analyzed in Section 5. Section 6 presents our conclusions.

2. Related Work

The similarity measure of categorical data is the basis of categorical data analysis. A good clustering algorithm maximizes the similarity within clusters and minimizes the similarity between clusters. Although many researchers have proposed different methods to measure the similarity or dissimilarity of categorical data, none of them have been widely recognized. For numerical data, there are Euclidean distance, vector dot product, and other similar or different degrees of measurement objects. For categorical data, the mean and variance are not defined, and the vector dot product operation is meaningless.
In 1998, Huang [17] proposed the conventional k-modes algorithm, which is a non-weighted feature clustering approach. The k-modes algorithm can be formulated into a mathematical optimization model as follows:
min J W , Q   = l = 1 k i = 1 n w l i d X i , Q l
where w l i composes a partition matrix and l = 1 k w l i = 1 , w l i { 0 , 1 } , and Q l = { q l 1 , q l 2 , , q l m } is the cluster center. The algorithm adopted a simple method, called overlap measure (OM) [19], to measure the distance, as shown in Equations (2) and (3). The differences between symbols are just equal or unequal (equal is 1, unequal is 0), as shown in Equation (3).
d X , Y   = i = 1 D S x i , y i
where,
S ( x i , y i ) =   1 i f   x i = y i 0 i f   x i y i .
This measure method is easy to use and has great computational efficiency, since there are no involved parameters. However, its defined distances are not always reasonable in indicating the real dissimilarity because it ignores the valuable information about the relationship of the correlated attributes. There are some variants of k-mode algorithms, such as presented in [47,48]. All of these algorithms suppose that features are equally important for clustering analysis but have seen limited use in real-world practice.
In weighted features clustering algorithms, such as WKM [22], wk-modes [21], and SCC [24], features are weighted according to their importance to the clustering tasks. In these algorithms, the features are of different importance. They calculate the similarity between the two samples by supposing each dimension independently. The mathematical optimization model of these algorithms can be expressed as follows:
min J W , Q   = l = 1 k i = 1 n j = 1 m w l i λ l j β d X i , Q l
where W is also a partition matrix and l = 1 k w l i = 1 , w l i { 0 , 1 } , Λ = [ λ l j ] is a weight matrix, and β is an excitation parameter which is used to control the feature weight.
The algorithm also utilized the OM method to measure the distance, as Equations (2) and (3). These methods have the advantage of high clustering efficiency. In addition, feature weighting clustering algorithms assign uniform weight to all the intra-attribute distances measured on the feature, which is suitable for well-defined distances. However, the distance measure is not well-defined for categorical data, as evidenced by the OM distance measurement. To solve this problem, most existing methods focus on exploring appropriate distance measures and attribute weighted mechanisms, such as MWKM [23]. These methods are all linear algorithms, in that they are based on the assumption that features are independent of each other, so that the relationship between features is ignored, which means that a great deal of information between the features is lost.
At present, two methods are mainly used to explore the non-linear relationship between attributes: deep neural network (DNN) and the kernel method. As we all know, DNNs need a large amount of data to train. The larger the amount of data, the more accurate the result. The kernel method uses the Mercer kernel function to implicitly describe the non-linear relationship between attributes and has been widely studied and applied because of its simplicity of mathematical expression and the high efficiency of calculation. Chen et al. [24] proposed a soft subspace clustering approach based on probabilistic distance. Its mathematical optimization model can be expressed as follows:
min O B J Π , W   = k = 1 K x π k d = 1 D w k d θ D i s d x , π k
where W is the weight of the dth dimension for cluster k, x is the data sample and π k is the kth cluster. D i s d ( x , π k ) denotes the distance of sample x to the kth cluster on the dth dimension, which is computed by two discrete probabilities. This method also proposes to define a kernel density function κ ( X d o d l ; λ k ) , as shown in Equation (6), to estimate the probability, where λ k [ 0 , 1 ] is the bandwidth for every cluster.
κ X d o d l ; λ k   =   1 O d 1 O d λ k X d = o d l 1 O d λ k X d o d l
where | O d | represents the power of O d , which is the number of aggregates, and o d l denotes the lth category in O d , o d l O d .
Although this method considers the relationship between attributes in non-linear space, it does not distinguish the importance of attributes. This method also can be seen as one in which all attributes are independent of each other and all attributes in the same cluster use the same bandwidth.

3. KDE-Based Similarity for Categorical Data

In this section, we first propose a kernel density estimation (KDE) method for categorical attributes, by which each attribute has its own bandwidth. Then, the distance between categorical data objects can be expressed by a probabilistic data distribution. Moreover, a new similarity measure in the kernel subspace is defined to clustering.

3.1. Self-Expressive Kernel Density Estimation (SKDE)

Kernel density estimation method does not use the prior knowledge of the data distribution and does not attach any assumptions to data distribution. It is used to study the characteristics of data distribution from the data sample itself and is a non-parametric probability density estimation method. Unlike the kernel function seen in Equation (6), we define the kernel density function as follows:
X d o d l ; λ d   =   1 O d 1 O d λ d X d = o d l 1 O d λ d X d o d l
where O d represents the power of O d , which is the number of aggregates, and λ d represents the width of the dth attribute.
It can be simply expressed as follows:
X d o d l ; λ d   = 1 O d λ d +   1 λ d I X d = o d l
where, I ( · ) denotes the indicator function; I ( t r u e ) = 1 and I ( f a l s e ) = 0 .
According to the Equation (7), we can obtain:
o d l O d X d o d l ; λ d   = 1 O d 1 O d λ d +   O d 1 λ d O d = 1 .
The above equation shows that the kernel function we defined satisfies the basic properties of probability distribution.
We use p ^ ( o d l | λ d ) to express the kernel probability estimation of p ( o d l ) . According to the basic principle of the SKDE method, we have:
p ^ o d l | λ d = 1 N x D B X d o d l ; λ d = f o d l 1 O d 1 O d λ d + 1 f o d l λ d O d = λ d O d + 1 λ d f o d l
where D B is a sample set, f ( o d l ) is the frequency estimation of o d l .
In order to map categorical data to the high-dimensional space through the kernel function, a symbolic vectorization technique is used, as Definition 1 follows.
Definition 1.
We define a data object x i d as follows:
x i d = x i d 1 , , x i d l , , x i d | O d |
where x i d ( l ) denotes the probability of o d l O d with regard to x i d , denoted by: x i d ( l ) = P d ( o d l | x d ) , and satisfies the constraint condition: l = 1 | O d | x i d ( l ) = 1 .
x i d ( l ) can be estimated using the kernel function shown in Equation (8), as follows:
x i d l = P d o d l | x d = d e f o d l | x d ; λ d = 1 | O d | λ d + 1 λ d I X d = o d l .

3.2. Similarity Measurement Based on Kernel Subspace

The existing mainstream methods fail to consider the relationship between features. We formally define the non-linear similarity measurement in the kernel subspace as follows:
Definition 2.
The similarity measure of kernel subspace is given by:
s i m x i , x j = κ w x i , x j
where κ w ( x i , x j ) represents the weighted features’ kernel function, denoting the combination of two sample objects on each attribute.
According to Definition 2, the polynomial kernel function can be expressed as:
  • origin polynomial kernel function:
    κ w x i , x j = x i · x j + 1 p = d D x i d x j d + 1 p ,
  • weighted feature polynomial kernel function:
    κ w x i , x j = x i · x j + 1 p = d D w k d θ x i d x j d + 1 p .
We introduce a kernel function that originally acts on continuous data to project categorical data into the kernel space and a weight vector w k = { w k d | d = 1 , 2 , , D } for each cluster in the kernel space for original feature selection. The greater the dth dimension’s contribution to cluster, the more important it is. w k d meets the constraints:
k , d : w k d 0 k : d = 1 D w k d = 1 .
We introduce an index θ ( θ 0 ) for w k d to control the incentive intensity, and suppose θ as a known constant. The bigger the value of θ , the smoother the weight distribution.
This similarity measure not only uses the kernel method to “kernel” the categorical data, but also considers the relationship between features in the non-linear space. We also select features in the mapped kernel space, which distinguishes the importance of features to the cluster.

4. Proposed Clustering Algorithm

In cluster analysis, the cluster is defined as the sample set with the minimum compactness (or dispersion), in which the compactness is measured by the similarity between the sample and the cluster center. Combined with the defined non-linear similarity measurement formula of kernel subspace, the kernel subspace clustering optimization objective function of categorical data can be defined as follows:
J Π , W   = k = 1 K x i π k S i m x i , v k = k = 1 K x i π k κ w x i , v k
where, v k is the center of the cluster π k , denoted as a D dimension vector v k = ( v k 1 , , v k d , , v k D ) . Since a categorical attribute value is represented by a vector by Definition 1, so the dth dimension’s center of the cluster π k should also be represented by a vector. Each component v k d represents the dth dimension’s center, denoted as v k d =   < v k d ( 1 ) , , v k d ( l ) , , v k d ( | O d | ) , which meets the constraints l = 1 | O d | v k d ( l ) = 1 , and v k d ( l ) represents the probability of o d l O d in the dth dimension.
Therefore, we have:
v k d l = 1 | π k | x i π k o d l | x d ; λ d = 1 | O d | λ d + 1 λ d f k o d l
where f k ( o d l ) denotes the frequency estimation of o d l O d in the dth attribute.

4.1. Non-Linear Optimization in Kernel Subspace

In the process of calculation, the sum function is operated in the kernel function (such as the polynomial kernel subspace function mentioned above), which makes it difficult to solve w k d , which, in turn, greatly increases the difficulty of solving the objective function. Therefore, we propose an efficient optimization method for solving the kernel subspace clustering optimization objective function. The objective function is transformed into the form of the existing mainstream methods (such as WKM [22] method) in order to improve the computational efficiency. The optimization objective defined by Equation (14) is further analyzed. Theorem 1 shows that for all convex kernel functions, the maximum value of Equation (14) is equivalent to the maximum value of the function of Equation (16), given by:
J Π , W   = k = 1 K x i π k d = 1 D w k d θ κ d x i , v k
where κ d ( x i , v k ) represents the mapping function’s inner product of x i and v k in the dth dimension, that is, the kernel function in the dth dimension. For example, the polynomial kernel function can be expressed as follows:
κ d x i , v k   = x i d v k d + 1 p .
Theorem 1.
When θ 1 , for all convex kernel functions κ ( · , · ) , the maximum Equation (14) has the same solution as the maximum Equation (16).
Proof. 
We define z d as the two input objects’ combination in the dth dimension for similarity measurement in the kernel subspace. When the two input objects are the sample x i and the cluster center v k , z d represents the combination of x i and v k in the dth dimension. If we let
f ( d = 1 D w k d θ z d ) = κ d ( x i , v k ) ,
in which f ( · ) is the newly defined function, we can obtain f ( z d ) = κ d ( x i , v k ) . We use mathematical induction to prove
d = 1 D w k d θ f ( z d ) f ( d = 1 D w k d θ z d ) .
(1)
When D = 1 , 2 , the inequality clearly holds;
(2)
We suppose that the inequality clearly holds when D = n , then,
d = 1 n w k d θ f ( z d ) f ( d = 1 n w k d θ z d ) .
When D = n + 1 , let p n = d = 1 n w k d , then, we have:
d = 1 n + 1 w k d θ f z d = w k n + 1 θ f z n + 1 + d = 1 n w k d θ f z d = w k n + 1 θ f z n + 1 + p n θ d = 1 n w k d p n θ f z d w k n + 1 θ f z n + 1 + p n θ f d = 1 n w k d p n θ z d f w k n + 1 θ z n + 1 + p n θ d = 1 n w k d p n θ z d = f w k n + 1 θ z n + 1 + d = 1 n w k d θ z d = f d = 1 n + 1 w k d θ z d .
   □
We can thus obtain
d = 1 D w k d θ f ( z d ) f ( d = 1 D w k d θ z d ) .
In particular, when θ = 1 , the inequality is Jesson inequality. We acquire f ( d = 1 D w k d θ z d ) by stretching the lower bound d = 1 D w k d θ f ( z d ) to upper bound. Then, we adjust w k d to maximize d = 1 D w k d θ f ( z d ) . Through step-by-step iteration, we finally obtain the maximum of f ( d = 1 D w k d θ z d ) .
Combining Definition 1 and Theorem 1, the Gaussian kernel function [49] can be expressed as follows:
κ w x i , x j = exp d = 1 D w k d θ x i d x j d 2 2 σ 2 = f d = 1 D w k d θ z d
where z d = x i d x j d 2 2 σ 2 , · is the Euclidean norm, σ 2 is variance, and f x   = exp x .

4.2. SKSCC Clustering Algorithm

The Gaussian kernel function is the most widely used kernel function, because it has a better performance for large, as well as small samples and has fewer parameters than other kernel functions. This paper proposes the SKSCC that takes the Gaussian kernel function to be the objective function, as shown in Equation (16). We can now transfer the Equation (16) to Equation (19), as follows:
J Π , W   = k = 1 K x i π k d = 1 D w k d θ f z d f z d   = exp z d z d = l | O d | I x i d = o d l λ d | O d | 1 λ d f k o d l 2 2 σ 2
where σ 2 is defined as the global variance, and
σ 2 = 1 N D i = 1 N d = 1 D o O d [ I ( x i d = o ) f k ( o ) ] 2 ,
in which N is the number of sample set, and D is the dimension of the attributes.
Equation (19) is a non-linear optimization problem with constraints. Using Lagrange multipliers, the objective function can be transferred to Equation (20) as follows:
max J Π , W   = k = 1 K x i π k d = 1 D w k d θ f z d + k = 1 K ξ k 1 d = 1 D w k d f z d   = exp z d z d = l | O d | I x i d = o d l λ d | O d | 1 λ d f k o d l 2 2 σ 2 .
In this paper, we use the EM algorithm to optimize max J ( Π , W ) , In other words, the local optimal value of J can be obtained by the iterative method. According to this principle, we first set Π = Π ^ to maximize J ( Π ^ , W ) , and then obtain the value W, recorded as W ^ . Next, we set W = W ^ and then maximize J ( Π , W ^ ) to calculate Π , recorded as Π ^ . The two steps are calculation of W ^ and clustering, which are detailed as follows:
(1)
Weight Computing
We define K independent suboptimal objective functions, as follows:
J k w k , λ k = x i π k d = 1 D w k d θ f z d + ξ k 1 d = 1 D w k d f z d = exp z d z d = l | O d | I x i d = o d l λ d | O d | 1 λ d f k o d l 2 2 σ 2 .
Let J k w k d = 0 , then:
J k w k d = θ w k d θ 1 x i π k f z d ξ k = 0 .
Let J k ξ k = 0 , then:
J k ξ k = 1 d = 1 D w k d = 0 .
From Equations (22) and (23), we can obtain the representation of w k d as follows:
w k d = x i π k exp l | O d | I x i d = o d l λ d | O d | 1 λ d f k o d l 2 2 σ 2 1 1 θ d = 1 D x i π k exp l | O d | I x i d = o d l λ d | O d | 1 λ d f k o d l 2 2 σ 2 1 1 θ .
(2)
Clustering
Cluster can be generated by dividing x i into the cluster with the most similarity. The algorithm can be expressed as follows:
k = a r g max k κ w x i , v k   = a r g max k exp d = 1 D w k d θ z d z d = l | O d | I x i d = o d l λ d | O d | 1 λ d f k o d l 2 2 σ 2 .
In summary, the algorithm is outlined in Algorithm 1. According to the algorithmic structure, SKSCC can be viewed as an extension to the k-modes clustering algorithm, by adding step (3) to update the cluster and step (5) to compute the attribute weights, both of which are proportional to the kernel bandwidth that can be learned by the objects themselves. Therefore, as the k-modes algorithm, the SKSCC algorithm can also converge in a finite number of iterations. The time complexity of SKSCC is O ( K N D ) .
Algorithm 1 SKSCC clustering algorithm.
Input:
   The categorical dataset D B , the number of clusters K, incentive intensity θ ;
Output:
   Cluster Π and weight set W.
1: Initialization:
    iterations’ times t, t = 0;
    Set all W to 1 D , that’s W ( 0 ) = 1 D ;
    Calculate bandwidth λ d ; d = 1 , 2 , , D ;
    Calculate global variance σ 2 ;
    Randomly select k objects as the initial cluster center, generating initial datasets, denoted as Π ( 0 ) ;
2: repeat
3:  let W ^ = W ( t ) , divide all the samples into clusters using Equation (25), and then get Π ( t + 1 ) ;
4:  Update cluster center: v k d ;
5:  Update W: set Π ^ = Π ( t + 1 ) , update weight W using Equation (24), then get W ( t + 1 ) ;
6:   t = t + 1 ;
7: until The clustering set does not change, that is, Π ( t ) = Π ( t + 1 ) .
8: return  Π ( t ) and W ( t ) .

4.3. Optimization of Kernel Bandwidths

In light of the weight calculation formula Equation (24), the weights depend on the kernel bandwidths, which is the bandwidth optimization problem in the defined SKDE method. Here, we use the mean integrated squared error (MSE) method, which is a data-driven method for estimating optimal bandwidth. For the dth attribute, the kernel probability estimation’s MSE for o d l O d can be expressed as follows:
M S E o d l , λ d   = E o d l O d p ^ o d l | λ d p o d l 2 .
According to the definition of kernel function and the properties of expectation, the bandwidth λ d can be obtained. The objective function of the optimal estimation of bandwidth is as follows:
λ d = o d l O d E λ d | O d | + 1 λ d f o d l p o d l 2 = o d l O d 1 λ d 2 E f 2 o d l + 2 λ d 1 λ d | O d | + λ d 1 p o d l E f o d l + p 2 ( o d l ) 2 λ d | O d | p ( o d l ) + λ d 2 | O d | 2 .
Because of
f o d l   = 1 N x i D B I x i d = o d l
where N represents the number of samples.
Then, we have:
E f o d l   = E I X d = o d l = p o d l .
Due to V a r X = E X 2 E X 2 , I · 2 = I · ; then, we have:
V a r f o d l = 1 N V a r I x i d = o d l = 1 N p o d l p 2 o d l .
Therefore, we obtain:
λ d = 1 1 | O d | λ d 2 + 1 λ d 2 N λ d 2 σ d 2
where σ d 2 = 1 o d l O d p 2 ( o d l ) .
Let λ d λ d = 0 , then:
λ d λ d = 1 1 | O d | 2 λ d + 2 1 λ d 1 N λ d 2 σ d 2 = 0 .
Therefore, we have:
λ d = | O d | σ d 2 | O d | N + σ d 2 N σ d 2 N .
We use the frequency distribution of the training samples to estimate p ( o d l ) , and we calculate σ d 2 by the standard deviation of the training samples. Hence, we obtain
s d 2 = 1 o d l O d f 2 o d l .
The kernel bandwidth algorithm is outlined in Algorithm 2. Several properties of the kernel bandwidth’s optimal estimation are analyzed:
(1)
The larger the number of samples N, the smaller the bandwidth.
λ d = | O d | s d 2 | O d | N + σ d 2 N σ d 2 N = s d 2 N o d l O d f 2 o d l 1 | O d | + s d 2
The coefficient of N is o d l O d f 2 o d l   1 | O d | ; its values’ range is [ 0 , 1 ] . The larger the number of samples N, the smaller the bandwidths. When N , the bandwidth λ d 0 . This is consistent with the effect of bandwidth as the smoothing parameter of the kernel function.
(2)
The larger the data dispersion, the larger the bandwidth.
λ d = | O d | s d 2 | O d | N + σ d 2 N σ d 2 N = s d 2 N N | O d | N 1 s d 2
Let us calculate the derivative of λ d with respect to s d 2 as follows:
λ d s d 2 = N 1 1 | O d | N N | O d | N 1 s d 2 2 .
Because 1 1 | O d | > 0 , then λ d s d 2 > 0 ; so, λ d is the increasing function with respect to s d 2 in the range [0,1). The larger the data dispersion s d 2 , the larger the bandwidth λ d , that is to say, the larger the discreteness of an attribute, the larger the kernel bandwidth corresponding to the attribute. In particular, when an attribute categorical data are uniformly distributed, the corresponding kernel bandwidth takes the maximum value.
Algorithm 2 The kernel bandwidth calculation algorithm.
Input:
   The categorical dataset D B ;
Output:
    Λ = { λ d | d = 1 , 2 , , D } ;
1:   for d = 1 to D do
2:    Compute s d 2 using Equation (23);
3:    Compute λ d using Equation (22);
4:   end for

5. Experimental Analysis

Experiments were performed to verify the effectiveness of our proposed SKSCC on synthetic and real datasets. Comparative experiments were carried out on some current mainstream categorical clustering algorithms.

5.1. Experimental Setup

In practical applications, the Gaussian kernel function is the most widely used kernel function, because it is suitable for a variety of samples and has few parameters. Moreover, the mapping space provided by this type of kernel function isinfinitely dimensional, so that the data that are not separated in the original space can be directly mapped into linear separable points. Therefore, we chose the Gaussian kernel to mine the non-linear relationship between categorical attributes. The parameter defined as
σ 2 = 1 N D i = 1 N d = 1 D o O d I x i d = o λ d | O d | 1 λ d f o d l 2
which is the global variance, and is learned from the data themselves.
We chose three algorithms—k-mode [17], WKM [22], MWKM [23]—for our comparative experiments. WKM introduced attributes-weighting within the framework of the k-modes algorithm, which is a linear weighting. The MWKM algorithm weights the attributes through the frequency of the mode. All three methods are based on the principle of feature independence to calculate the sample similarity (or dissimilarity). These algorithms are selected for comparison with the non-linear similarity measurement SKSCC. The parameter β is set to 2 in WKM. The parameter β is set to 2 and T s = T v = 1 in MWKM.
Synthetic data can control the cluster structure of datasets through the number and size of clusters, which is conducive for analyzing the performance of the algorithm and its adaptability to various datasets. For this paper, we first tested on several synthetic datasets and then carried out experiments on many real datasets. Because the labels are all known, two external evaluation indices—accuracy and F-score [22]—were selected to evaluate the clustering performance of the new algorithm. The larger the value of the two indices, the better the clustering effect. F-score is defined as follows:
F s c o r e = k = 1 K n k N max 1 i K 2 × R c l a s s k , π i × P c l a s s k , π i R c l a s s k , π i + P c l a s s k , π i
where c l a s s k represents the kth real class in datasets, n k represents the sample number of c l a s s k , and P ( c l a s s k , π i ) and R ( c l a s s k , π i ) separately represent accuracy and recall compared real class c l a s s k and cluster π i of clustering results, that is,
P = T P T P + F P
R = T P T P + F N
where TP represents the number of predicting correct clusters as correct clusters; FN represents the number of predicting correct clusters as false clusters; FP represents the number of predicting false clusters as correct clusters.

5.2. Discussion of Parameters

In the kernel space, each attribute is automatically given a weight to measure its similarity, and the corresponding subspace is found through feature selection.
w k d θ = x i π k exp l | O d | I x i d = o d l λ d | O d | 1 λ d f k o d l 2 2 σ 2 θ 1 θ d = 1 D x i π k exp l | O d | I x i d = o d l λ d | O d | 1 λ d f k o d l 2 2 σ 2 θ 1 θ
where θ is the incentive intensity, and is the allocation parameter of control weight. Figure 1 shows the change in parameters for the weight of the three attributes in the Breastcancer dataset. Here, the discreteness of the three attributes is set to increase from attribute 1. There are four comments for θ .
(1)
When θ = 0 , w k d θ is the constant, that is, each attribute will be assigned an equal weight;
(2)
When θ = 1 , θ 1 θ , but all of the weights must meet the restriction d = 1 D w k d = 1 , so when θ 1 + , the attribute with the minimum deviation of the sample will be weighted, while the rest of the attributes will be given zero weight; when θ 1 , the importance of all attributes tends to be the same;
(3)
When 0 < θ < 1 , the more discrete the attribute, the greater its weight;
(4)
When θ < 0 and θ > 1 , the attribute weight is inversely proportional to the dispersion of data distribution. Considering Theorem 1, we should set θ > 1 , but when θ is too larger, the difference between attribute weights is reduced.

5.3. Analysis of Synthetic Data and Results

This study used MATLAB (Version 9.9.0.1495850 R2020b) to generate the synthetic data in the experiment. First, four multi-dimensional numerical datasets were generated by the MATLAB function m v n r n d ( · ) , in which the weight of attributes was controlled by setting the variance of attributes, and the correlation degree between attributes was controlled by adjusting the parameters of the covariance matrix. The synthesized numerical data were then discretized by equal width [40] and transformed into categorical data. The synthetic datasets that contain the correct category labels are presented in Table 1. Four datasets were used to verify the advantages of SKSCC compared with the current mainstream categorical clustering methods.
  • The covariance of attribute 1 and attribute 2 is set to −2 in DataSet1, which makes their attributes a negative correlation. The covariance of attribute 1 and attribute 4 is set to 2, which makes their attributes a positive correlation. The variances are set to be equal on each attribute;
  • DataSet2 and DataSet1 are set to the same clusters, but the number of attributes differs. Ten attributes are extracted to set their covariance. The variances are set to be equal on each attribute;
  • DataSet3 and DataSet2 are set to the same attributes, but the clusters are different. The variances are set to be equal in two clusters. Ten attributes are extracted to set their covariance;
  • DataSet4 is set to the most number of attributes and the clusters. Twenty attributes are extracted to set their covariance in seven clusters. All attributes are set to covariance in one cluster. A half clusters set the same variances, as well as other half clusters.
We implemented 100 runs on each algorithm and each dataset, and set θ = 1.5 . The average clustering accuracy reported in Table 2 reflects the overall performance of each clustering algorithm, and the stability of clustering performance of each algorithm can be judged according to the listed variance. The smaller the variance of clustering accuracy, the better the stability of clustering performance.
From Table 2, we can see that with the increase in the number of related attributes, the clustering accuracy of SKSCC is significantly higher than that of other algorithms. This is because SKSCC employs a “kernel” operation and take into consideration the relationship between attributes.

5.4. Analysis of Real-World Data and Results

In this part of the experiments, we set out to test and verify the performance of SKSCC in real-world datasets. We compared the SKSCC algorithm with three other algorithms: the original k-modes algorithm (k-mode), the weighting algorithm (WKM), and the mixed weighting algorithm (MWKM).

5.4.1. Real-World Datasets

To carry out the experiments, we obtained 10 datasets from the University of California Irvine (UCI) Machine Learning Repository [7]. Table 3 lists the details of these 10 datasets. The Breastcancer, Vote, Mushroom, and Adult+stretch datasets have the same clusters, but Mushroom dataset has the most samples, and Adult+stretch dataset has the least number of samples. The Balance and Splice datasets each have the same number of clusters (3), but the dimensionality of Splice is higher. The Soybeansmall and Car datasets each have the same number of clusters (4), but different attributes and samples. Dermatology and Zoo are multi-cluster datasets.

5.4.2. Comparison of Clustering Quality

Because the initial cluster centers can affect the algorithm results, we randomly selected 100 initial centers, and all of the algorithms used the same initial centers in each experiment. We implemented 100 runs on each algorithm and each dataset, and set θ = 1.5 . The average values and the errors for F-score and accuracy are presented in Table 4. The results showed that our proposed method, SKSCC, achieved the best performance in the comparative experiments on most of the datasets. Because the k-mode [17], WKM [22], and MWKM [23] algorithms are all based on the mode-type category theory, it is easy for them to descend to the clustering objective algorithm’s local minimum, causing them to lose applicability. However, WKM achieved good results on the Car and Splice datasets, while MWKM achieved high accuracy on the Dermatology dataset.
Figure 2 shows the distribution of clustering accuracy for all the algorithms when run 100 times. SKSCC has the best stability. The abscissa represents each algorithm’s running time, and the ordinate is the F-score value to express the results of each clustering. SKSCC has the smallest fluctuation among all the algorithms, although WKM has the best average F-score on the Splice and Car datasets, and MWKM has the best average F-score on the Dermatology dataset. The clustering results for the k-mode algorithms show significant contrast, because they consider only the module in the clustering process, which makes it easy to fall into the local optimum, and the initial cluster center is k randomly selected objects. This is reflected in the standard deviation of average precision. Because SKSCC quantizes the module, it avoids the above-mentioned problems and has more stable performance than the other algorithms.

5.4.3. Feature Weighting Results

Our SKSCC approach also has a feature selection effect. Using the Breastcancer dataset as an example, Figure 3 shows the attribute weights generated by the MWKM and SKSCC algorithms. It does not show the k-mode algorithm or WKM algorithm, because the former method is not weighted in its features and the latter method calculates the weights based on mode frequency, which is similar to MWKM algorithm. From Figure 4, we can see that for SKSCC, A1 and A9 acquire the largest and the smallest weights, respectively, of the benign class, but MWKM algorithm achieved the opposite results. To test the feature weighting method’s rationality for the SKSCC, we removed the A1 and A9 features from the original Breastcancer data in order to form two reduced datasets. The F-score values of the different clustering algorithms on the Breastcancer dataset with the original and reduced feature sets are shown in Figure 4. For all the algorithms, the reduced dataset with the A9 feature removed achieved the highest F-score values, while the reduced dataset with the A1 feature removed showed decreased F-score values. The results indicate that our SKSCC algorithm with non-linear similarity measurement does a better job, by considering the relationship of the attributes, than the other algorithms.

5.4.4. Time Consumption

This paper uses a logarithm of the average time of clustering to compare the actual average times. The ordinate represents the average time (in MS) of each algorithm running on the real-world dataset. It can be seen from Figure 5 that k-mode, WKM, and MWKM algorithms have high clustering efficiency, which is one of the advantages of the module-based clustering algorithms. Because only the module of the categorical attribute needs to be considered, the statistical information of the other categorical symbols can be ignored, which greatly reduces the algorithms’ clustering times.

6. Conclusions

Kernel clustering with categorical data is a vital direction in application research. In view of current problems, such as supposing all features independently, considering all attributes’ importance equally, and finding an optimization solution, this paper proposes a novel kernel clustering approach for categorical data, that is, a self-expressive kernel subspace clustering algorithm for categorical data (SKSCC). This paper first defines a kernel function for self-expression kernel density estimation (SKDE), in which each attribute has its own bandwidth and can be calculated by the data themselves. We also propose a novel non-linear similarity measurement method and an efficient non-linear optimization method (Theorem 1) to solve the objective function of the kernel clustering. Finally, the SKSCC algorithm is presented for categorical data. Our method not only considers the relationship between attributes in non-linear space but also gives each attribute a feature weight to measure the correlation degree in the algorithmic process. The experimental results indicate that the proposed algorithm outperforms the other algorithms on the synthetic and UCI datasets.
There are many directions that are of interest for future exploration. We will expand our approach to other kernel functions and test the performance on more datasets for various data. Our efforts will also be directed at combining our method with deep learning to estimate the parameters adaptively.

Author Contributions

Conceptualization, Q.J. and L.C.; methodology, H.C. and K.X.; software, H.C.; validation, H.C. and K.X.; formal analysis, H.C. and K.X.; investigation, Q.J. and L.C.; resources, Q.J. and L.C.; data curation, H.C.; writing—original draft preparation, H.C.; writing—review and editing, H.C.; visualization, H.C.; supervision, Q.J. and L.C.; project administration, Q.J. and L.C.; funding acquisition, Q.J. and L.C. All authors have read and agreed to the published version of the manuscript.

Funding

The work is supported by the Key-Area Research and Development Program of Guangdong Province Grant No. 2019B010137002, and the National Natural Science Foundation of China under Grant Nos. U1805263, 61672157.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank all the anonymous reviewers for their insightful comments and constructive suggestions that have obviously upgraded the quality of this manuscript.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal circumstances that could have appeared to influence the work reported in this manuscript.

References

  1. Tang, J.; Liu, H. An unsupervised feature selection framework for social media data. IEEE Trans. Knowl. Data Eng. 2014, 26, 2914–2927. [Google Scholar] [CrossRef] [Green Version]
  2. Alelyani, S.; Tang, J.; Liu, H. Feature selection for clustering: A review. Data Clust. Algorithms Appl. 2013, 29, 144. [Google Scholar]
  3. Han, J.; Kamber, M. Data Mining: Concepts and Techniques; Morgan Kaufmann: San Francisco, CA, USA, 2001. [Google Scholar]
  4. Bharti, K.K.; Singh, P.K. A survey on filter techniques for feature selection in text mining. In Proceedings of the Second International Conference on Soft Computing for Problem Solving (SocProS 2012), Jaipur, India, 28–30 December 2012; Springer: New Delhi, India, 2014; pp. 1545–1559. [Google Scholar]
  5. Yasmin, M.; Mohsin, S.; Sharif, M. Intelligent image retrieval techniques: A survey. J. Appl. Res. Technol. 2014, 12, 87–103. [Google Scholar] [CrossRef] [Green Version]
  6. Saeys, Y.; Inza, I.; Larranaga, P. A review of feature selection techniques in bioinformatics. Bioinformatics 2007, 23, 2507–2517. [Google Scholar] [CrossRef] [Green Version]
  7. Frank, A. UCI Machine Learning Repository. 2010. Available online: http://archive.ics.uci.edu/ml (accessed on 28 March 2021).
  8. Jain, A.K.; Murty, M.N.; Flynn, P.J. Data clustering: A review. ACM Comput. Surv. (CSUR) 1999, 31, 264–323. [Google Scholar] [CrossRef]
  9. Xu, R.; Wunsch, D. Survey of clustering algorithms. IEEE Trans. Neural Netw. 2005, 16, 645–678. [Google Scholar] [CrossRef] [Green Version]
  10. Jain, A.K. Data clustering: 50 years beyond k-mean. Pattern Recognit. Lett. 2010, 31, 651–666. [Google Scholar] [CrossRef]
  11. Wu, S.; Lin, J.; Zhang, Z.; Yang, Y. Hesitant fuzzy linguistic agglomerative hierarchical clustering algorithm and its application in judicial practice. Mathematics 2021, 9, 370. [Google Scholar] [CrossRef]
  12. Guha, S.; Rastogi, R.; Shim, K. ROCK: A robust clustering algorithm for categorical attributes. Inf. Syst. 2000, 25, 345–366. [Google Scholar] [CrossRef]
  13. Andritsos, P.; Tzerpos, V. Information-theoretic software clustering. IEEE Trans. Softw. Eng. 2005, 31, 150–165. [Google Scholar] [CrossRef]
  14. Andritsos, P.; Tsaparas, P.; Miller, R.J.; Sevcik, K.C. LIMBO: Scalable clustering of categorical data. In Proceedings of the International Conference on Extending Database Technology, Heraklion, Crete, Greece, 14–18 March 2004; Springer: Berlin/Heidelberg, Germany, 2004; pp. 123–146. [Google Scholar]
  15. Qin, H.; Ma, X.; Herawan, T.; Zain, J.M. MGR: An information theory based hierarchical divisive clustering algorithm for categorical data. Knowl.-Based Syst. 2014, 67, 401–411. [Google Scholar] [CrossRef] [Green Version]
  16. Xiong, T.; Wang, S.; Mayers, A.; Monga, E. DHCC: Divisive hierarchical clustering of categorical data. Data Min. Knowl. Discov. 2012, 24, 103–135. [Google Scholar] [CrossRef]
  17. Huang, Z. Extensions to the k-means algorithm for clustering large data sets with categorical values. Data Min. Knowl. Discov. 1998, 2, 283–304. [Google Scholar] [CrossRef]
  18. Huang, Z.; Ng, M.K. A fuzzy k-modes algorithm for clustering categorical data. IEEE Trans. Fuzzy Syst. 1999, 7, 446–452. [Google Scholar] [CrossRef] [Green Version]
  19. Ng, M.K.; Li, M.J.; Huang, J.Z.; He, Z. On the impact of dissimilarity measure in k-modes clustering algorithm. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 503–507. [Google Scholar] [CrossRef] [PubMed]
  20. Bai, L.; Liang, J.; Dang, C.; Cao, F. The impact of cluster representatives on the convergence of the k-modes type clustering. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 1509–1522. [Google Scholar] [CrossRef] [PubMed]
  21. Cao, F.; Liang, J.; Li, D.; Zhao, X. A weighting k-modes algorithm for subspace clustering of categorical data. Neurocomputing 2013, 108, 23–30. [Google Scholar] [CrossRef]
  22. Chan, E.Y.; Ching, W.K.; Ng, M.K.; Huang, J.Z. An optimization algorithm for clustering using weighted dissimilarity measures. Pattern Recognit. 2004, 37, 943–952. [Google Scholar] [CrossRef]
  23. Bai, L.; Liang, J.; Dang, C.; Cao, F. A novel attribute weighting algorithm for clustering high-dimensional categorical data. Pattern Recognit. 2011, 44, 2843–2861. [Google Scholar] [CrossRef]
  24. Chen, L.; Wang, S.; Wang, K.; Zhu, J. Soft subspace clustering of categorical data with probabilistic distance. Pattern Recognit. 2016, 51, 322–332. [Google Scholar] [CrossRef]
  25. Han, J.; Kamber, M.; Pei, J. Data mining concepts and techniques third edition. Morgan Kaufmann Ser. Data Manag. Syst. 2011, 5, 83–124. [Google Scholar]
  26. Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2003, 3, 1157–1182. [Google Scholar]
  27. Breiman, L.; Friedman, J.; Stone, C.J.; Olshen, R.A. Classification and Regression Trees; CRC Press: Boca Raton, FL, USA, 1984. [Google Scholar]
  28. Kohavi, R.; John, G.H. Wrappers for feature subset selection. Artif. Intell. 1997, 97, 273–324. [Google Scholar] [CrossRef] [Green Version]
  29. Pashaei, E.; Aydin, N. Binary black hole algorithm for feature selection and classification on biological data. Appl. Soft Comput. 2017, 56, 94–106. [Google Scholar] [CrossRef]
  30. Rasool, A.; Tao, R.; Kamyab, M.; Hayat, S. Gawa—A feature selection method for hybrid sentiment classification. IEEE Access 2020, 8, 191850–191861. [Google Scholar] [CrossRef]
  31. Liu, H.; Setiono, R. Chi2: Feature selection and discretization of numeric attributes. In Proceedings of the 7th IEEE International Conference on Tools with Artificial Intelligence, Herndon, VA, USA, 5–8 November 1995; pp. 388–391. [Google Scholar]
  32. Quinlan, J.R. Induction of decision trees. Mach. Learn. 1986, 1, 81–106. [Google Scholar] [CrossRef] [Green Version]
  33. Quinlan, J.R. C4. 5: Programs for Machine Learning; Elsevier: Amsterdam, The Netherlands, 2014. [Google Scholar]
  34. Kandaswamy, K.K.; Pugalenthi, G.; Hazrati, M.K.; Kalies, K.U.; Martinetz, T. BLProt: Prediction of bioluminescent proteins based on support vector machine and relieff feature selection. BMC Bioinform. 2011, 12, 345. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Shao, J.; Liu, X.; He, W. Kernel based data-adaptive support vector machines for multi-class classification. Mathematics 2021, 9, 936. [Google Scholar] [CrossRef]
  36. Robnik-Šikonja, M.; Kononenko, I. Theoretical and empirical analysis of ReliefF and RReliefF. Mach. Learn. 2003, 53, 23–69. [Google Scholar] [CrossRef] [Green Version]
  37. Le, T.T.; Urbanowicz, R.J.; Moore, J.H.; McKinney, B.A. Statistical inference Relief (STIR) feature selection. Bioinformatics 2019, 35, 1358–1365. [Google Scholar] [CrossRef]
  38. Huang, Z.; Yang, C.; Zhou, X.; Huang, T. A hybrid feature selection method based on binary state transition algorithm and ReliefF. IEEE J. Biomed. Health Inform. 2018, 23, 1888–1898. [Google Scholar] [CrossRef] [PubMed]
  39. Deng, Z.; Chung, F.L.; Wang, S. Robust relief-feature weighting, margin maximization, and fuzzy optimization. IEEE Trans. Fuzzy Syst. 2010, 18, 726–744. [Google Scholar] [CrossRef]
  40. Chen, L.F. A probabilistic framework for optimizing projected clusters with categorical attributes. Sci. China Inf. Sci. 2015, 58, 1–15. [Google Scholar] [CrossRef]
  41. Kong, R.; Zhang, G.; Shi, Z.; Guo, L. Kernel-based k-means clustering. Comput. Eng. 2004, 30, 12–14. [Google Scholar]
  42. Elhamifar, E.; Vidal, R. Sparse subspace clustering: Algorithm, theory, and applications. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2765–2781. [Google Scholar] [CrossRef] [Green Version]
  43. Ji, P.; Zhang, T.; Li, H.; Salzmann, M.; Reid, I. Deep subspace clustering networks. arXiv 2017, arXiv:1709.02508. [Google Scholar]
  44. You, C.; Li, C.G.; Robinson, D.P.; Vidal, R. Oracle based active set algorithm for scalable elastic net subspace clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 3928–3937. [Google Scholar]
  45. Chen, L.; Guo, G.; Wang, S.; Kong, X. Kernel learning method for distance-based classification of categorical data. In Proceedings of the 2014 14th UK Workshop on Computational Intelligence (UKCI), Bradford, UK, 8–10 September 2014; pp. 1–7. [Google Scholar]
  46. Ouyang, D.; Li, Q.; Racine, J. Cross-validation and the estimation of probability distributions with categorical data. J. Nonparametr. Stat. 2006, 18, 69–100. [Google Scholar] [CrossRef]
  47. Huang, Z. Clustering large data sets with mixed numeric and categorical values. In Proceedings of the 1st Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD), Singapore, 23–24 February 1997; pp. 21–34. [Google Scholar]
  48. Cheung, Y.M.; Jia, H. Categorical-and-numerical-attribute data clustering based on a unified similarity metric without knowing cluster number. Pattern Recognit. 2013, 46, 2228–2238. [Google Scholar] [CrossRef]
  49. Zhong, S.; Chen, D.; Xu, Q.; Chen, T. Optimizing the gaussian kernel function with the formulated kernel target alignment criterion for two-class pattern classification. Pattern Recognit. 2013, 46, 2045–2054. [Google Scholar] [CrossRef]
Figure 1. Analysis of weight with different θ .
Figure 1. Analysis of weight with different θ .
Mathematics 09 01680 g001
Figure 2. Comparison of F-score with different algorithms on different datasets.
Figure 2. Comparison of F-score with different algorithms on different datasets.
Mathematics 09 01680 g002aMathematics 09 01680 g002b
Figure 3. Weight distributions generated by two algorithms on Breastcancer dataset.
Figure 3. Weight distributions generated by two algorithms on Breastcancer dataset.
Mathematics 09 01680 g003
Figure 4. F-score values of the different clustering algorithms on the Breastcancer dataset with original and reduced feature sets.
Figure 4. F-score values of the different clustering algorithms on the Breastcancer dataset with original and reduced feature sets.
Mathematics 09 01680 g004
Figure 5. F-Score values of the different clustering algorithms on the Breastcancer dataset with original and reduced feature sets.
Figure 5. F-Score values of the different clustering algorithms on the Breastcancer dataset with original and reduced feature sets.
Mathematics 09 01680 g005
Table 1. Data categorized in four synthetic datasets.
Table 1. Data categorized in four synthetic datasets.
Attributes (D)Clusters (K)Samples (N)
Datasets1621000
Datasets22021000
Datasets32041000
Datasets44081000
Table 2. Comparison of F-score and Accuracy results of four algorithms performed on the four synthetic datasets.
Table 2. Comparison of F-score and Accuracy results of four algorithms performed on the four synthetic datasets.
IndexDatasetsK-Mode [17]WKM [22]MWKM [23]SKSCC
F-ScoreDatasets10.9823 ± 0.00000.9489 ± 0.00790.9738 ± 0.0018 1.0000 ± 0.0000
Datasets20.9762 ± 0.00150.9860 ± 0.00000.9860 ± 0.0000 0.9940 ± 0.0000
Datasets30.6346 ± 0.00110.5766 ± 0.00180.6311 ± 0.0009 0.6771 ± 0.0005
Datasets40.5268 ± 0.00080.3839 ± 0.00330.5367 ± 0.0010 0.6224 ± 0.0017
AccuracyDatasets10.9823 ± 0.00000.9589 ± 0.00380.9746 ± 0.0012 1.0000 ± 0.0000
Datasets20.9762 ± 0.00150.9860 ± 0.00000.9860 ± 0.0000 0.9939 ± 0.0000
Datasets30.6755 ±0.00160.6037 ± 0.00240.6644 ± 0.0009 0.7033 ± 0.0004
Datasets40.5863 ± 0.00130.5053 ± 0.01470.5848 ± 0.0014 0.6655 ± 0.0014
Table 3. Details of 10 DataSets from UCI.
Table 3. Details of 10 DataSets from UCI.
No.UCI DatasetsAttributes (D)Clusters (K)Samples (N)
1Breastcancer92699
2Vote162435
3Mushroom2128124
4Adult+stretch4220
5Balance43625
6Splice6033190
7Soybeansmall35447
8Car641728
9Dermatology336366
10Zoo157101
Table 4. Comparison of clustering results in terms of F-score and accuracy.
Table 4. Comparison of clustering results in terms of F-score and accuracy.
IndexDatasetsK-Mode [17]WKM [22]MWKM [23]SKSCC
F-ScoreBreastcancer0.8637 ± 0.00000.7683 ± 0.00050.8645 ± 0.0155 0.9660 ± 0.0000
Vote0.8610 ± 0.00000.8238 ± 0.00730.8698 ± 0.0000 0.8749 ± 0.0000
Mushroom0.7159 ± 0.01710.6645 ± 0.00340.7480 ± 0.0202 0.7901 ± 0.0193
Adult + stretch0.6691 ± 0.01350.6722 ± 0.01590.6876 ± 0.0163 0.7537 ± 0.0085
Balance0.4882 ± 0.00160.4782 ± 0.00220.4630 ± 0.0024 0.5672 ± 0.0017
Splice0.4155 ± 0.0000 0.5321 ± 0.0007 0.4313 ± 0.0000 0.5258 ± 0.0019
Soybeansmall0.8324 ± 0.01520.7336 ± 0.01570.8436 ± 0.0175 0.8641 ± 0.0146
Car0.4412 ± 0.0018 0.5006 ± 0.0057 0.4268 ± 0.00120.4738 ± 0.0028
Dermatology0.6476 ± 0.00830.5573 ± 0.0136 0.6685 ± 0.0088 0.6357 ± 0.0034
Zoo0.7273 ± 0.00900.6716 ± 0.01300.7417 ± 0.0074 0.7701 ± 0.0070
AccuracyBreastcancer0.8621 ± 0.00000.8284 ± 0.00000.8659 ± 0.0156 0.9659 ± 0.0000
Vote0.8625 ± 0.00000.8244 ± 0.00660.8681 ± 0.0000 0.8734 ± 0.0000
Mushroom0.7536 ± 0.01340.8481 ± 0.01570.7733 ± 0.01430.8194 ± 0.0131
Adult + stretch0.7150 ± 0.01600.7165 ± 0.01680.6910 ± 0.0159 0.8620 ± 0.0086
Balance0.5251 ± 0.00100.4629 ± 0.00330.4327 ± 0.0024 0.8722 ± 0.0321
Splice0.4237 ± 0.0000 0.6149 ± 0.0011 0.4314 ± 0.0000 0.5426 ± 0.0017
Soybeansmall0.8740 ± 0.0110 0.9423 ± 0.0039 0.8915 ± 0.01100.9085 ± 0.0083
Car0.4023 ± 0.0013 0.4550 ± 0.0095 0.3593 ± 0.00000.4251 ± 0.0038
Dermatology0.7085 ± 0.0076 0.9298 ± 0.0038 0.7367 ± 0.00630.6911 ± 0.0048
Zoo0.7937 ± 0.0066 0.8260 ± 0.0084 0.7895 ± 0.00730.8043 ± 0.0061
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chen, H.; Xu, K.; Chen, L.; Jiang, Q. Self-Expressive Kernel Subspace Clustering Algorithm for Categorical Data with Embedded Feature Selection. Mathematics 2021, 9, 1680. https://doi.org/10.3390/math9141680

AMA Style

Chen H, Xu K, Chen L, Jiang Q. Self-Expressive Kernel Subspace Clustering Algorithm for Categorical Data with Embedded Feature Selection. Mathematics. 2021; 9(14):1680. https://doi.org/10.3390/math9141680

Chicago/Turabian Style

Chen, Hui, Kunpeng Xu, Lifei Chen, and Qingshan Jiang. 2021. "Self-Expressive Kernel Subspace Clustering Algorithm for Categorical Data with Embedded Feature Selection" Mathematics 9, no. 14: 1680. https://doi.org/10.3390/math9141680

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop