Next Article in Journal
Decoupling Induction and Multi-Order Attention Drop-Out Gating Based Joint Motion Deblurring and Image Super-Resolution
Next Article in Special Issue
Knowledge-Based Scene Graph Generation with Visual Contextual Dependency
Previous Article in Journal
Collatz Attractors Are Space-Filling
Previous Article in Special Issue
Multi-View Graph Clustering by Adaptive Manifold Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Clustering Ensemble Framework with Integration of Data Characteristics and Structure Information: A Graph Neural Networks Approach

School of Computer and Information Technology, Shanxi University, Taiyuan 030006, China
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(11), 1834; https://doi.org/10.3390/math10111834
Submission received: 18 April 2022 / Revised: 24 May 2022 / Accepted: 24 May 2022 / Published: 26 May 2022
(This article belongs to the Special Issue Trustworthy Graph Neural Networks: Models and Applications)

Abstract

:
Clustering ensemble is a research hotspot of data mining that aggregates several base clustering results to generate a single output clustering with improved robustness and stability. However, the validity of the ensemble result is usually affected by unreliability in the generation and integration of base clusterings. In order to address this issue, we develop a clustering ensemble framework viewed from graph neural networks that generates an ensemble result by integrating data characteristics and structure information. In this framework, we extract structure information from base clustering results of the data set by using a coupling affinity measure After that, we combine structure information with data characteristics by using a graph neural network (GNN) to learn their joint embeddings in latent space. Then, we employ a Gaussian mixture model (GMM) to predict the final cluster assignment in the latent space. Finally, we construct the GNN and GMM as a unified optimization model to integrate the objectives of graph embedding and consensus clustering. Our framework can not only elegantly combine information in feature space and structure space, but can also achieve suitable representations for final cluster partitioning. Thus, it can produce an outstanding result. Experimental results on six synthetic benchmark data sets and six real world data sets show that the proposed framework yields a better performance compared to 12 reference algorithms that are developed based on either clustering ensemble architecture or a deep clustering strategy.

1. Introduction

Data clustering, also known as cluster analysis, is the process of partitioning a set of data objects into some clusters based on a similarity measure without any prior knowledge. The objects in each cluster are similar to each other, and different from data objects in other clusters [1]. All of these clusters are usually referred to as a clustering result. The core of data clustering is to discover the inherent structure from the unlabeled data. Thus, it is usually employed as an effective tool to understand raw data in the initial phase of processing, particularly for the problems where prior knowledge is absent or expensive to obtain. Driven by the demand of acquiring knowledge from various complicated data, clustering analysis has become a research hotspot. Over the past decades, many studies have focused on the solution of data clustering from diverse perspectives, such as clustering algorithms with different similarity criteria [2], extensions and modification for particular data types [3], identifying the optimal number of clusters [4], subspace clustering and multiview clustering [5,6], clustering results evaluation [7], and applications of data clustering [8].
Unlike supervised learning methods, data clustering is essentially an ill-posed problem [9] because the results of different clustering algorithms can be equally accepted without prior knowledge. In other words, any clustering result may be superior to others in a particular pattern distribution. This problem results in a dilemma: users need to select an appropriate clustering algorithm according to the inherent characteristics of data and domain knowledge. The clustering ensemble appears in this background, and it can generate an excellent and stable cluster assignment by combining multiple clustering results [10]. Clustering ensemble outperforms the single clustering algorithm in several aspects [11,12]: (i) the average performance of clustering ensemble on different data types and pattern distributions is superior to its optimal ensemble member; (ii) it can obtain a consistent result that cannot be achieved by any single clustering algorithm; (iii) it is more robust to noises, outliers, and sampling changes than single clustering algorithm; (iv) it outperforms the single clustering algorithm in terms of parallelizability and scalability; and (v) it can integrate multiple clustering results generated from distributed data sources or features. Generally, clustering ensemble comprises two main phases. In the first phase, a group of base clustering results is produced. In the second phase, a consensus function is designed to generate a final cluster assignment by combing base clustering results. The base clustering members can be produced in many ways, such as by using different clustering algorithms, employing one clustering algorithm with different parameters, and utilizing subsets of data or features [13]. The typical methods used to design a consensus function include a relabeling strategy [14], feature-based methods [15,16], pairwise-similarity-based algorithms [17,18], and graph-based approaches [19,20]. In general, three ensemble-information matrices [11] can be acquired from the base clustering results: the label-assignment matrix, pairwise similarity matrix, and binary cluster association matrix. Each ensemble strategy mentioned above utilizes one of these matrices to combine base clustering results. That is to say, the clustering ensemble results of these methods are derived either from a reorganized data description in feature space or from the relationships between data objects in structure space. In reality, the data characteristics that are expressed by features or attributes and the structure information manifested as relationships between objects are two presentations of the data from different perspectives. Therefore, both of them provide valuable guidance on producing the final ensemble result. However, the existing clustering ensemble algorithms, according to our knowledge, seldom consider combining these two data descriptions in the design of the consensus function.
Data characteristics and structure information focus on different aspects of data description. Thus, their combination can potentially improve thereliability of clustering ensemble. With this motivation, we develop a novel clustering ensemble framework to integrate data characteristics with structure information. To serve this purpose, our work is focused on addressing the following two key issues: (i) What structural information should be employed to construct the consensus function? Many local relationships are commonly used to describe structure information, such as the label-assignment information originally obtained from an ensemble, co-association matrix, or associations between data objects or those among clusters [13]. In general, the structural information represents the implicit relationships between data objects. It includes not only the local similarity between objects, but also a global structure implying the intrinsic pattern of the data. Therefore, using only the local structure to design the clustering consensus function is far from sufficient, and determining how to effectively extract and utilize suitable structure information is an important problem. (ii) How can we integrate data characteristics and structure information to produce the consensus clustering result? If only concentrating on data characteristics, the deep clustering algorithms [21,22,23] provide an effective data partitioning strategy using the powerful representation learning ability of deep learning. For example, an autoencoder (AE) [24] is a commonly used framework with multiple layers, in which, each layer captures specific latent information from data characteristics, while in a clustering ensemble context, one can acquire various forms of structure information from the data and its base clusterings. Therefore, what is the interaction relationship between data characteristics and structure information in the generation of the final clustering result? Furthermore, how can these two different types of information be combined elegantly, and how can they be represented in a suitable form for clustering ensemble? These problems constitute the other key issue of our work.
To address the first problem, we intend to capture the structural information of the data from a set of base clusterings. All of the base clusterings are produced on the same data; thus, some forms of relationships certainly exist between base clusterings, between clusters, and between objects. Viewed from this aspect, the intrinsic structure information can be captured from a set of base clustering results by explicating and coupling the above relationships.
We employ a variational graph autoencoder (VGAE) module [25] to cope with the second key issue by learning the joint representations of data characteristics and structure information suitable for the clustering objective. On this basis, we derive a joint optimization model that incorporates representation learning and consensus cluster partitioning into a unified framework.
In this study, we propose a novel framework, namely clustering ensemble viewed from graph embedding (CEvGE), to integrate data characteristics with structure information by employing the powerful representation ability of GNNs. Our work can be viewed as an improvement on a clustering ensemble model by learning appropriate representations in the latent space. In addition, it can be viewed as an improvement on a deep clustering method by imposing a global structure constraint. Our main contributions can be summarized as follows:
(i)
We develop a novel clustering ensemble framework viewed from GNNs that produces the ensemble result by exploiting both data characteristics and structure information.
(ii)
We reconstruct raw data as an object similarity graph, and integrate data characteristics and structure information elegantly by learning their joint representations in a variational GNN. We employ a Gaussian mixture model (GMM) to predict the consensus cluster assignment for the latent embeddings, which is a natural model for clustering.
(iii)
We construct a unified optimization model to integrate the objectives of joint embedding learning and final cluster assignment, in which, clustering can provide a correct guidance for embedding learning.
(iv)
We conduct extensive experiments on synthetic and real-world data sets. The results demonstrate the validity and superiority of our framework against reference algorithms.
The remainder of this paper is organized as follows. Section 2 introduces some closely related works. In Section 3, we discuss the construction and implementation of our CEvGE framework in detail. In Section 4, we compare the CEvGE with several state-of-the-art clustering ensemble models and deep clustering algorithms in a series of experiments Finally, we conclude this work in Section 5.

2. Related Work

In this section, we introduce works related to the AE and variational autoencoder (VAE), deep clustering, and graph embedding.

2.1. AE and VAE

The AE can be regarded as a nonlinear generalization of PCA used to reduce data dimensionality, in the form of a multilayer neural network with a small middle layer [24]. It consists of three basic elements: an adaptive, multilayer encoder network that transforms the high-dimensional data into low-dimensional codes; a decoder network with a similar structure that recovers data from the codes; and a loss function that measures the lost information caused by dimensionality reduction. Driven by the popularity of various deep learning models, especially the generative adversarial networks (GANs), the AE has been regarded as at the forefront of generative modeling. Various extended AE models have been developed successively. In general, these extended models fall into three categories [26]: (i) instantiation-based models trained by combining them with other learning schemes, such as the convolutional autoencoder (CAE) and extreme learning machine autoencoder (ELMAE), etc.; (ii) regularization-based models implemented by introducing regularization constraints into the loss function, including the sparse autoencoder, contractive autoencoder, and information theoretic-learning autoencoder, etc.; and (iii) variational inference-based models, such as the VAE and adversarial autoencoder (AAE).
Using a similar encoder–decoder structure, the VAE is in fact deeply rooted in the variational Bayesian methods [27], which map input data into a distribution in latent space rather than a fixed vector In the VAE, the decoder network can be viewed as a generative module that generate new samples similar to but not identical to input data. The necessary assumptions for the VAE are relatively weak, and it can be trained fast by backpropagation. Furthermore, the error introduced by approximation in the VAE is arguably small. These merits make the VAE a very popular generative model. To improve the performance of the VAE for data with a complex distribution, some works attempt to enhance the prior assumption for latent variables by introducing a Bayesian probabilistic mixture model. For example, Jiang et al. [28] approximated the prior distribution of latent variables by a mixture of Gaussians, and develped an improved VAE model, namely variational deep embedding (VaDE). Nalisnick et al. [29] also enhanced the prior description in the VAE using a GMM, and proposed a deep latent Gaussian mixture model (DLGMM) that employs multiple inference networks to generate variational posterior distribution. Another similar work is the DGG developed by Yang et al. [30], which constructs a single generative model with an isotropic Gaussian mixture prior description.

2.2. Deep Clustering with the Encoder–Decoder Schema

Recently, driven by the powerful representation ability of the deep neural network (DNN), some works attempted to improve the clustering performance by introducing the DNN to learn effective data representations for the clustering objective. Tian et al. constructed a two-stage deep clustering framework that utilized a DNN to acquire feature representations in subspace, and then divided clusters using these learned features. To guide the DNN to learn suitable representations for cluster assignment, some later works tried to incorporate a clustering objective into the deep learning framework. For example, Xie et al. employed an incomplete AE to learn data embeddings, and designed an assistant distribution to estimate cluster assignment for the learned embeddings. In that model, called deep embedded clustering (DEC), the cluster assignment and model training are achieved simultaneously in a self-learning schema. Guo et al. [31] introduced a complete autoencoder to improve DEC by overcoming the misguiding problem in embedding learning. In this way, the clustering loss can provide a positive guidance for embedding learning on maintaining the original distribution pattern. Similarly, another deep learning model proposed by Guo et al., namely deep convolutional embedded clustering (DCEC), utilized a convolutional AE and a single-layer classifier to learn the data representations and the cluster distributions, respectively. In this model, the DNN is trained by minimizing the reconstruction loss and the estimation error measured by relative entropy. In [32], the authors elected some representative data objects as landmarks and measured the similarity of landmarks between all data objects as the input of the autoencoder. They combined embedding learning and cluster assignment to further enhance the clustering performance. They also employed clustering loss to update the cluster centers and parameters of AE simultaneously. Bo et al. [33]. proposed a structural deep clustering network (SDCN) model to integrate the structural information into deep clustering, in which a delivery operator was designed to transfer the representations learned by autoencoder to the corresponding GCN layer. On this basis, they also developed a dual self-supervised mechanism to unify the two different deep neural architectures and guide the update of the whole model. Wang et al. [34] designed a dual-stacked autoencoder feature embedded clustering (DSAFEC) for human activity recognition (HAR) that uses dual-stacked autoencoder features (DSAF) to learn new representations for the original input and then predicts cluster assignment probabilities for the learned representations by employing a softmax regression.
In the above deep clustering algorithms, embeddings are learned effectively by the AE with a strict one-to-one correlation to input data. However, the pattern of learned embedding depends on the training data; as a result, the distribution in latent space is discontinuous. That is to say, the clustering performance would degrade due to the decoded data possibly losing some characteristics implied in the original data. To address this issue, some later works apply VAE as the embedding learning network for deep clustering, rather than AE. In VAE, the latent variables sampled from the learned distributions can effectively capture the statistics of original data. In addition, the decoder can generate data objects that help the clustering process acquire more information about the inherent distribution of original data. For example, Jiang et al. [28] developed an unsupervised generative deep clustering model called VaDE, in which, the latent embeddings are learned by a DNN and the cluster distributions are predicted by a GMM. In the optimization of VaDE, the SGVB estimator and the reparameterization trick are introduced. For the clustering problem of high-dimensional data, Li et al. [35] proposed a VAE-based model named the latent tree variational autoencoder (LTVAE)that produces several clustering results for high-dimensional data using distinct portions of features. They described latent embeddings in a tree structure, and used the location of each cluster in the tree to organize the data relationships. Hwang et al. [36] addressed the issue of the clustering complex and high-dimensional wafer maps in semiconductor manufacturing by proposing a variational deep clustering algorithm, namely one-step VAE+DPGMM. In this algorithm, a GMM is implemented into a VAE framework to extract more suitable features for the clustering environment, and a Dirichlet process is further applied in the variational autoencoder mixture framework for automated one-step clustering. Compared with conventional two-step clustering methods, the model can considerably increase the chance to distinguish small differences in wafer map patterns.
All of the deep clustering algorithms mentioned above introduce the powerful representational ability of DNNs for clustering tasks based on an AE or VAE framework. However, they all focus on improving the conventional clustering approaches rather than the clustering ensemble context. Additionally, they only explore data relationships in feature space and neglect the information in structure space, which is also an important description of the data pattern. In a clustering ensemble context, the overlooked structure information is just hidden in the set of base clustering results. This fact motivates us to introduce the powerful representation learning ability of DNNs to clustering ensemble, and to exploit data characteristics and structure information together to produce the consensus clustering result.

2.3. Graph Embedding

As a ubiquitous form of data organization, a graph provides a natural tool to express structure information of data, as it can record relationships between data objects effectively. To incorporate the non-Euclidean graph data into machine learning models, graph embedding is always used to encode the graph from high-dimensional, sparse space into a low-dimensional, compact, and continuous feature space, where characteristics in the original graph are preserved as much as possible. Various graph embedding models can be abstracted to an encoder–decoder framework, in which, the encoder maps each vertex to a low-dimensional embedding and the decoder reconstructs the graph from the learned embeddings [37]. In general, the objective of the encoder–decoder graph embedding model is optimized by minimizing a certain reconstruction loss. Graph embedding models can be divided into two categories: (i) shallow embedding models, which are largely inspired by classic matrix factorization techniques [38] or random walks [39], and always utilize an “embedding lookup” encoder function; and (ii) generalized encoder–decoder architectures, which construct highly complex encoders often using GNNs to map topology and vertex attributes together into a low-dimensional latent space [40,41]. In the latter category, some works aggregate local topology to learn graph embedding in a VAE framework. For example, the model in [25] constructs a VAE-based learning framework, where a graph convolutional network (GCN) and an inner product function are employed as the encoder and the decoder, respectively. To overcome the inferior embedding problem in real-world graph data, Pan et al. [42] proposed an adversarially regularized VGAE, in which, the embedding learned by the encoder is forced to match a prior distribution by using an adversarial training scheme. To solve the data corruption issue, Kang et al. [43] proposed a graph learning scheme to learn reliable graphs from the real-world noisy data by adaptively removing noise and errors in the raw data. The model can also be viewed as a robust version of manifold regularized robust principle component analysis (RPCA), where the quality of the graph plays a critical role.

3. The Proposed Framework

For a set of N data objects X = x i i = 1 N in the D-dimensional space, let Π = π t t = 1 T represent a set of T base clusterings for X , where π t = C γ t γ = 1 k t is the t th clustering member, k t denotes the number of clusters in π t , and C γ t is the γ th cluster. For the object x i , λ i t denotes its cluster label in the t th base clustering. The label set, including all of the different labels in π t , is denoted by Λ t , and the objects assigned with label λ i t in π t are marked as s t λ i t . The task of clustering ensemble is to produce a consensus partition π * = C k * k = 1 K for X by combining all of the base clusterings, where K is the number of clusters in π * , and C k * denotes the k th cluster in the consensus result.

3.1. Overview of the Framework

The integral construction of the CEvGE framework is illustrated in Figure 1. In the CEvGE, the structure information of data is firstly extracted from a set of base clusterings. With the help of the structure information, the data set can be reorganized as an an object similarity graph. Then, the data characteristics and structure information are integrated in a VGAE module viewed from graph embedding perspective. In the latent space, a GMM is applied to predict the consensus cluster assignment for the joint embedding. Finally, the VGAE module and the GMM are trained jointly in a unified optimization model, which integrates objectives of graph embedding and consensus cluster assignment.

3.2. Extraction of Structure Information

An affinity matrix is always constructed to describe the data structure, where each element represents relationships between two data objects evaluated by a certain similarity measure. Thus, an adequate similarity measure that can express the inherent data relationships comprehensively is very important when describing structure information. In the original data space before clustering, the data structure can be directly captured by calculating pairwise similarity. However, this form of structure information can only describe presentational relationships from a local perspective. In the clustering ensemble context, we can explore extra structure information from a set of base clusterings. On the one hand, data relationships are reflected in the inner-dependence within a base clustering, and, on the other hand, they are also implied in the inter-dependence among different base clusterings. As a result, the global structure information that reveals an intrinsic pattern of data can be acquired by uncovering the coupling relationships between base clusterings, between clusters, and between data objects. In this work, the coupling relationships are in terms of two elements: interactions between base clusterings and between data objects [17]. We propose capturing and incorporating them to tease out the implicit global relationships in the data, which will be used subsequently to construct the affinity matrix.

3.2.1. Coupling of Base Clusterings

All base clustering results are different partitionings on the same data; as a result, there must be some relationships among these base results. From intra and inner perspectives, the coupling of base clusterings comprises two components: the intra-coupling impresses the interaction between cluster labels within one base clustering, whereas the inner-coupling indicates the interaction between two base clusterings. According to this idea, we use the coupled clustering similarity for clusters (CCSC) between two cluster labels to describe the coupling relationship of base clusterings:
S i m t π λ i t , λ j t | Λ o o = 1 T = S i m t π I a λ i t , λ j t S i m t π I e λ i t , λ j t | Λ o o t ,
where S i m t π λ i t , λ j t | Λ o o = 1 T is the CCSC between two cluster labels λ i t and λ j t . S i m t π I a λ i t , λ j t is the intra-coupled clustering similarity (IaCS) between λ i t and λ j t , used to describe the intra-coupling of base clusterings by calculating the frequency of cluster labels within a base clustering. S i m t π I e λ i t , λ j t | Λ o o t is the inter-coupled relative similarity (IeRS) between λ i t and λ j t based on another base clustering π o , used to characterize the intra-coupling of base clusterings by comparing the co-occurrence of the cluster labels among different base clusterings, where Λ o is the label set of π o . The IaCS is defined as:
S i m t π I a λ i t , λ j t = s t λ i t s t λ j t s t λ i t + s t λ j t + s t λ i t s t λ j t .
The IeRS is defined as:
S i m t π I e λ i t , λ j t | Λ o o t = o = 1 , o t T ω o S i m t | o λ i t , λ j t | Λ o ,
where ω o 0 , 1 is the weight of the base clustering π o , o = 1 , o t T ω o = 1 , and S i m t | o λ i t , λ j t | Λ o is defined as:
S i m t | o λ i t , λ j t | Λ o = λ o Ω min s o λ o s t λ i t s t λ i t , s o λ o s t λ j t s t λ j t .
In Equation (4), Ω denotes the set L o s t λ i t L o s t λ j t , where L o s t λ i t is the label set of s t λ i t in π o .

3.2.2. Coupling of Data Objects

Similarly, the coupling relationships between objects can also be described from the following two perspectives. In terms of the intra-perspective, we use the intra-coupled object similarity (IaOS) to measure the similarity between two objects, which is defined as the average sum of the CCSC between the associated cluster labels ranging over all of the base clusterings. The IaOS between x i and x j can be calculated as
S i m I a O S x i , x j = 1 T t = 1 T S i m t π λ i t , λ j t | Λ o o = 1 T .
whereas, from the inter-perspective, we reflect the interaction between data objects by mining their neighbors’ correlation. Accordingly, we utilize the common neighbors between two objects to mesure the inter-coupled object similarity (IeOS), which is defined as
S i m I e O S x i , x j = 1 N x n X | x n N x i N x j .
In Equation (6) where S i m I e O S x i , x j denotes the IeOS between objects x i and x j . N x i represents the neighbor set of x i , and is defined as
N x i = x n | κ x i , x n θ .
κ ( · , · ) is a Gaussian kernel function with a threshold θ [ 0 , 1 ] , which is defined as
κ x i , x n = 1 α i exp φ ( x i , x n ) 2 2 ϑ 2 ,
where α i > 0 and ϑ > 0 are the normalizer and width of the kernel function, respectively, and φ ( · ) denotes a certain similarity measure for data objects.
The location of a data object in a clustering is determined by the cluster it belongs to. Accordingly, we can integrate the relationships from the inner-perspectives and inter-perspectives together through the associated clusters. In particular, we induce the coupled similarity of objects by specifying the similarity measure employed in Equation (8) to be IaOS, and we have
κ x i , x n I a O S = 1 α i exp S i m I a O S ( x i , x n ) 2 2 ϑ 2 .
Then, the IeOS can be converted to a new similarity measure for objects, namely coupled clustering and object similarity (CCOS). The CCOS between x i and x j can be calculated as
S i m C C O S x i , x j = 1 N x n X | x n N x i I a O S N x j I a O S ,
where the neighbor sets are defined as N x i I a O S = x n | κ I a O S x i , x n θ and N x j I a O S = { x n | κ I a O S x j , x n θ } , respectively. In this manner, both the intra-coupled and inter-coupled relationships between data objects are considered in the CCOS. In the meantime, the intra-coupled and inter-coupled interactions between base clusterings are also taken into account by the new similarity measure.

3.2.3. Structure Information Organized by a Affinity Matrix

According to the above discussion, a global structure description can be extracted from a set of base clusterings, i.e., CCOS. Compared with previous local similarity measurements, the CCOS expresses comprehensive relationships between data objects from multiple perspectives by exploiting base clusterings. Thus, we use the CCOS to construct an affinity matrix A N × N that organizes the structure information of the data. The element A i , j in the matrix is denoted as
A i , j = S i m C C O S x i , x j .

3.3. Clustering Ensemble Viewed from Graph Embedding Perspective

The affinity matrix of structure information can be seen as geometric constraints of the data. As a result, we can reorganize the input data X as an objects similarity graph G ( X , A ) , which includes both the characteristics and structure of the data. In this way, the clustering ensemble task on X is transformed into a clustering problem on the reorganized graph G ( X , A ) . To this end, we develop a joint model within the VGAE framework. In this model, data characteristics and structure information are integrated by learning their joint embeddings in latent space, and, subsequently, a GMM is utilized to predict the final cluster assignments for the latent embeddings.

3.3.1. Inference Model

The inference model is used to integrate data characteristics and structure information jointly by mapping the graph-reorganized data to latent embeddings, which are parameterized by a two-layer GCN as
q Z | X , A = i = 1 N q z i | X , A .
In Equation (12), Z = z i i = 1 N denotes the embedding of the data in latent space. For a certain data object x , the corresponding embedding can be acquired from the following distribution:
q ( z | x , a ) = N ( z | μ z , d i a g ( σ z 2 ) ) ,
where a denotes the associated row in the affinity matrix. The vectors μ z = GC N μ ( X , A ) and log σ z = GC N σ ( X , A ) are the mean and variance of the latent distribution, respectively. They are all learned by the GCN, whose structure is defined as
GCN ( X , A ) = G c o n v ( ReLU ( G c o n v ( A , X ; W 0 ) ) ; W 1 ) ,
where the function G c o n v ( · ) is a graph convolutional layer [39], and W 0 and W 1 are learnable weight matrices of the first and second layers, respectively. W 0 is shared between GC N μ ( X , A ) and GC N σ ( X , A ) .

3.3.2. Generative Model

In our framework, the generative model functions are used to reconstruct input data from the learned embeddings, as well as to predict the consensus cluster assignment. Specifically, we assume the input data are generated from a mixture of Gaussian distributions. Thus, we approximate the clustering ensemble result C k * k = 1 K by a GMM, and introduce a K-dimensional vector c to indicate the prior probabilities of each cluster in the consensus cluster assignment. The generative process can be modeled as follows:
  • Sample a cluster C k * Cat ( c ) from the consensus clustering result π * , where Cat ( c ) is the categorical distribution of π * parameterized by c .
  • Sample a vector z N z | μ c , k , d i a g σ c , k 2 from the picked cluster, where μ c , k and σ c , k 2 denote the mean and variance of the kth Gaussian component, respectively.
  • Sample a vector a from the reconstructed data X ˜ = x ˜ i i = 1 N . For binary data, the vector can be sampled from a multivariate Bernoulli distribution as a Ber μ x ˜ , where μ x ˜ is computed by μ x ˜ = g z ; ϕ , whereas, for real-value data, the vector can be sampled from a multivariate Gaussian distribution, where μ x ˜ and σ x 2 are learned by μ x ˜ ; log σ x ˜ 2 = g z ; ϕ . g z ; ϕ is a nonlinear function parameterized by ϕ . In our framework, the inner product decoder is employed.
g z ; ϕ = σ z i T z j .
According to the above generative process, the joint probability p a , z , C k * can be factorized as
p a , z , C k * = p a | z p z | C k * p C k * .
Since a and C k * are independently conditioned on z , we have
p a | z = Ber μ x ˜ o r N μ x ˜ , d i a g σ x ˜ 2 ,
p z | C k * = N z | μ c , k , d i a g σ c , k 2 ,
p C k * = C a t C k * | c .

3.4. Learning Algorithm

Our CEvGE framework can be tuned by maximizing the log-likelihood of the input data as
max W , ϕ , C * x log p ϕ x = max W , ϕ , C * x log z C * p ϕ x , z , C k * .
Using the Jensen’s inequality, we have
log p ϕ x L ELBO x = E q z , C * | x , a log p a , z , C k * q z , C k * | x , a ,
where L ELBO x denotes the ELBO of x , and q z , C k * | x , a is the variational approximation of the true posterior p z , C k * | x , a . By using a mean-field distribution, q ( z , C k * | x , a ) can be factorized as
q z , C k * | x , a = q z | x , a q C k * | x , a .
According to Equations (16) and (22), L ELBO x can be rewritten as
L ELBO x = E q z , C k * | x , a log p a | z p z | C k * p C k * q z | x , a q C k * | x , a = E q z , C k * | x , a [ log p a | z + log p z | C k * + log p C k * log q z | x , a log q C k * | x , a ] .
Substituting Equations (13), (17), (18) and (19), into Equation (23), and using the Monte Carlo SGVB estimator, we can further transform L ELBO x into
L ELBO x = 1 M m = 1 M d = 1 D x d log μ x ˜ ( m ) | d + 1 x d log 1 μ x ˜ ( m ) | d 1 2 k = 1 K q C k * | x , a r = 1 R log σ c 2 | r + σ z 2 | r σ c , k | r + μ z | r μ c | r 2 σ c , k 2 | r + k = 1 K q C k * | x , a log p C k * q C k * | x , a + 1 2 r = 1 R 1 + log σ z 2 | r ,
where M is the total number of samples in the SGVB estimator, D and R are the dimensionalities of the input data and latent embedding, respectively, x d is the d th feature of x , and the operators | i and | r denote the i th and r th component of the vector •, respectively. μ x ˜ ( m ) is calculated by μ x ˜ ( m ) = g z ( m ) ; ϕ , in which, z ( m ) is the m th Monte Carlo sample picked from q z | x , a . The reparameterization trick is used to employ gradient backpropagation on the stochastic layer, and z ( m ) is calculated as
z ( m ) = μ z + σ z * ρ m ,
In Equation (25), the learning rate ρ m is sampled from N 0 , I and the symbol * denotes the element-wise multiplication operator. μ z and σ z are learned by the GCN formulated in Equation (14).
In the proposed CEvGE framework, the posterior distribution q C k * | x , a that maximizes the ELBO must be found to obtain the final clustering result. L ELBO x can be rewritten by regrouping Equation (23) as
L ELBO x = E q z , C k * | x , a log p a , z , C k * q z , C k * | x , a = z C * q z | x , a q C k * | x , a log p x , a | z p z q z | x , a + log p C k * | z q C k * | x , a d z = z q z | x , a log p x , a | z p z q z | x , a d z z q z | x , a KL p C k * | z q C k * | x , a d z ,
where KL ( · ) is the Kullback–Leibler (KL) divergence function used for the divergence between two distributions, and p z = N z | 0 , I is the Gaussian prior distribution for the latent embedding. The first term in Equation (26) is independent of C k * , and the second term is non-negative according to the definition of KL divergence. Thus, L ELBO x can achieve the maximum value when KL p C k * | z q C k * | x , a 0 holds. Accordingly, the optimal cluster assignment q C k * | x , a can be estimated by
q C k * | x , a = p C k * | z = p C k * p z | C k * k = 1 K p C k * p z | C k * .
The latent embedding z will be an appropriate representation for clustering ensemble as the embedding learning and the cluster assignment are incorporated into the joint framework. The information loss caused by the mean-field assumption in Equation (22) can be preserved by the relationship between C k * and z implied in p C k * | z .
To further explore how our optimization model can work, we rewrite the ELBO in Equation (21) as
L ELBO x = E q z , C k * | x , a log p a , z , C k * q z , C k * | x , a = E q z , C k * | x , a log p a , z , C k * log q z , C k * | x , a = E q z , C k * | x , a log p a , z , C k * p z , C k * + log p z , C k * log q z , C k * | x , a = E q z , C k * | x , a log p a | z , C k * log q z , C k * | x , a p z , C k * = E q z , C k * | x , a log p a | z , C k * KL q z , C k * | x , a p z , C k * .
In Equation (28), the first term is obviously a reconstruction component, which helps our framework to explain the relationships between data objects well by employing latent embeddings and their cluster assignments, whereas the second term represents the divergence between the variational posterior distritution q z , C k * | x , a and the prior distribution p z , C k * modeled by a GMM. This divergence can be considered as a regularization term in our optimization model, which makes the learned embedding z lie on a Gaussian mixture manifold. Accordingly, we can draw the following conclusions: (i) the data characteristics and structure information of data are integrated elegantly in a generative graph embedding framework; (ii) the joint embeddings are learned with the guidance of the process of clustering, and, meanwhile, the prediction of cluster assignments is enhanced by the appropriate representations; (iii) the validity and reliability of the clustering ensemble result produced by the CEvGE framework can be improved.

3.5. Overall Implementation

The implementation of the developed CEvGE framework is formally summarized in Algorithm 1.
Algorithm 1 The implementation of CEvGE framework
Input : Data objects X , learning rate ρ , number of Monte Carlo samples in SGVB estimator M, epochs L.
Output : Consistent clustering result q C k * | x i , a i .
1:
Produce an ensemble of base clusterings for X ;
2:
From Equation (11), construct affinity matrix A to represent structure information for X
3:
Choose c U 0 , 1
4:
for   l = 1 , , L do
5:
    for  i = 1 , , N  do
6:
         μ z , i = GCN μ ( x i , a i ) ;
7:
        log σ z , i = GCN σ ( x i , a i ) ;
8:
        Sample C k * Cat C k * | c
9:
        Sample z i N ( z | μ c , k , d i a g ( σ c , k 2 ) )
10:
       Generate reconstructed a ˜ i = σ z i T z j
11:
       From Equation (24), compute L ELBO x i
12:
       Backpropagate gradients
13:
    end for
14:
end for
15:
From Equation (27), obtain the category assignment q ( C k * | x i , a i )
16:
return   q C k * | x i , a i

4. Experiments and Analysis

In this section, extensive experiments are conducted on several synthetic and real data sets to evaluate the validity and superiority of the proposed algorithm.

4.1. Data Sets and Evaluation Metrics

The distributions of synthetic data sets used in the experiments are shown in Figure 2. Several widely known real data sets, namely, Iris, Breast, KDD’99, MNIST, STL-10, and HHAR, were also employed for testing. Iris, Breast, and KDD’99 are data sets consisting of different attributes; MNIST and STL-10 are image data sets; and HHAR is a sensor signal data set. Since KDD’99 and MNIST are very large, implementing complex algorithms on the entire data set is very difficult. Thus, subsets sampled from the large data sets were utilized in the experiments. For KDD’99, we drew 4000 samples from each class of network-connected records; for MNIST, we drew 2000 samples from each class. Preprocessing is also required for some real data sets to simplify computing. For instance, the images in MNIST were vectorized, and features in STL-10 were subtracted using the pretrained ResNet-50 [44]. The details of all data sets we used are listed in Table 1.
In the experiments, several external and internal indexes were used to evaluate the clustering results. External indexes measure the similarity between the clustering result and the true partition on a data set, whereas internal indexes assess the performance of algorithms based on the inherent feature and measurement of a data set itself. The external indexes employed in the experiments included clustering accuracy (CA), adjusted rand index (ARI), and normalized mutual information (NMI). For any object x n in a data set X with N data objects, we used l n and c n to denote the real label and the cluster label of the object, respectively. The CA can be calculated as
C A = n max ( c n | l n ) N
Given two partitions of data set X , namely the true partition P = p 1 , p 2 , , p k ^ and the cluster assignment on the data C = c 1 , c 2 , , c k , respectively, the interaction relationships between these two partitions can be described by a contingency table (Table 2), where n i j is defined as n i j = c i p j . Based on the contingency table, the ARI of the clustering result C can be calculated by Equation (30).
A R I = i j n i j 2 i b i 2 j d j 2 / N 2 1 2 [ i b i 2 + j d j 2 ] i b i 2 j d j 2 / N 2
The NMI between two partitions can be measured as
N M I = 2 i j n i j log n i j N b i d j i b i log b i N j d j log d j N .
The definitions of CA, ARI, and NMI indicate that, the closer a cluster partition result is to the true partition, the higher its CA, ARI, and NMI values become.
Two internal indexes were also applied to compare the clustering results of different algorithms. They were Davies–Bouldin index (DBI) and Dunn validity index (DVI). The DBI measures the average similarity of each cluster between another closest cluster, which is calculated as
D B I = 1 k max 1 i j k i a v g ( c i ) + a v g ( c j ) d i s t c t r ( c i , c j ) ,
where d i s t c t r ( c i , c j ) denotes the squared Euclidean distance between the centroids of the clusters c i and c j , and a v g ( c i ) represents the average distance between each object and the center in cluster c i . A small DBI value means good inner-cluster cohesiveness and inter-cluster separation. DVI is a metric that measures the stability of a cluster assignment, and is defined as
D V I = min 1 i k min 1 j k , i j d i s t min ( c i , c j ) max 1 r k d i a m ( c r ) ,
where d i a m ( c r ) denotes the diameter of the cluster c r . The higher a clustering result’s DVI value, the more stable its structure.

4.2. Competitive Algorithms

To evaluate the performance of our CEvGE model comprehensively, four types of clustering ensemble algorithms were employed in the comparison:
  • One relabeling-based approach that employs a voting strategy was chosen, i.e., active density peak clustering ensemble algorithm (A1) proposed in [14].
  • Two feature-based algorithms were selected. They derive the final clustering result using the nominal information obtained from basic clusterings. They are the multiple k-means clustering ensemble algorithm (A2) [15] and the extended fuzzy k-means-based algorithm (A3) [16].
  • Two pairwise similarity approaches. One is the coupled clustering ensemble (CCE) algorithm (A4) developed by Wang et al. [17], which extracts coupling relationships from base clusterings similar to our framework, and the other is the reliability-based weighted fuzzy clustering ensemble framework (A5) designed by Bagherinia et al. [45].
  • Two graph-based approaches were chosen. They are the hypergraph partitioning algorithm (A6) [11], and the ultra-scalable ensemble clustering algorithm (A7) [20], respectively.
Our framework can be considered as a modified deep clustering model that incorporates structure information in the prediction of cluster assignment. As a result, in the experiments, we also compared it with five deep clustering methods: DEC (A8) [22], DSAFEC (A9) [34], IDEC (A10) [31], VaDE (A11) [28], and one-step VAE+DPGMM (A12) [36].

4.3. Experimental Setting

In the experiments, parameters in all reference algorithms were set according to their authors’ recommendations. In addition, some settings were predefined as follows:
  • For all of the algorithms in the experiments, base clusterings are produced by k-means. On each given data set, we set the value of K to be the true number of categories, and conducted the k-means T times independently with random initializations. On each data set, each algorithm was executed with T = 10 , 20 , 30 , 40 , 50 , and the best result was selected for comparison.
  • For the reference deep clustering algorithms, all of the layers are fully connected, and ReLU is utilized as the activation function. Adam optimizer was employed to improve the computational efficiency, and the size of mini-batch was 100. The learning rate was initialized to be 0.02, which decreased every 10 epochs with a decay factor of 0.9. On all of the synthetic data sets and two of the real data sets (Iris and Breast), the network structures of the encoder and decoder were set as D 30 30 2 and 2 30 30 D , respectively, where D is the dimensionality of the input data. On the rest of the real data sets, the network structures of the encoder and decoder were set as D 500 500 2000 10 and 10 2000 500 500 D , respectively. The AE network pretraining method that was used in DEC was also adopted by other AE/VAE-based algorithms to prevent the model falling into saddle points or local minima at the beginning of training.
  • The kernel width of the Gaussian kernel function used in the CEvGE framework was set within the interval θ [ 0.1 , 2 ] for different data sets with a step size of 0.1. We also sampled the hyperparameter θ from the interval [0.1, 0.9] with a step size of 0.1.
Under these settings, we chose the optimal parameters of each algorithm for further analysis.

4.4. Experimental Results

4.4.1. Performance analysis

First, we ran all of the algorithms 50 times on both of the synthetic and real data sets, and used the average results to evaluate their clustering performances. The values of external and internal indexes for different algorithms are presented in Table 3, Table 4, Table 5 and Table 6. From these tables, we can find that the CEvGE generally outperforms other algorithms in terms of external and internal indexes on most data sets. Subsequently, by further comparing and analyzing, we can draw the following conclusions:
(i) In Table 3 and Table 4, the contrastive clustering ensemble algorithms do not work well on most of the synthetic data sets. This is mainly because these ensemble algorithms that produce the final clustering results totally depend on the base clusterings. However, each data set is composed of several linearly inseparable categories; consequently, the base clusterings generated by k-means contain some unreliable partitions. These poor base results have a negative impact on the performances of the contrastive ensemble algorithms. The deep clustering algorithms intend to find good representations for the clustering objective. However, they learn latent embeddings or distributions for each data object independently, and overlook the relationships between data objects, which is fundamental to capturing the data pattern. As a result, they do not yield perfect outcomes. In comparison, in the CEvGE framework, comprehensive structure information is extracted from base clusterings and incorporated with data characteristics in a VGAE module. The learned latent embeddings not only facilitate the prediction of cluster assignments, but also preserve structural relationships existing in the original data space. With the assistance of structural information, the capacity of handling data with complex distributions is improved in our framework.
(ii) In Table 5 and Table 6, the clustering performances of our framework on the real data sets are also superior or close to the best results of other approaches. It is worth noting that the results of A11 and A12 on the KDD’99 are very close to the best result achieved by the CEvGE. The main reason is that A11 and A12 are both designed within a VAE framework, which gives them a generative ability. Therefore, the clustering result can be enhanced to some extent when the input data objects are insufficient for expressing the true pattern of the original data. In addition, the GMM used as their classifier is a tractable parametric model that can smoothly approximate an arbitrarily shaped distribution. Unlike A11 and A12, the CEvGE extracts and incorporates the structure information that provides assistance for the depiction of the cluster formation of the data. Thus, it outperforms other deep clustering algorithms. On image data sets MNIST and STL-10, the deep clustering algorithms can obtain suitable representations for the clustering objective aided by the powerful learning capability of DNN, and they outperform the reference ensemble algorithms. In the CEvGE, the embeddings learned by a GNN integrate both the characteristics from the original data space and the structure information, taking full advantage of the base clusterings; simultaneously, the process of embedding learning is guided by the prediction of consensus clustering in a unified optimization model. Thus, the CEvGE achieves a superior clustering performance.
(iii) It is worth noting that the algorithm A4, which extracts coupling relationships from base clusterings in the same way as our framework, produces unsatisfactory results on both synthetic and real data sets. The major reason for this is that the coupling relationships extracted from the poor base clusterings are inadequate in reflecting the intrinsic structure of the data.
To eliminate the adverse impact on the evaluation of the results caused by randomness, we conducted a pairwise Student’s t-test (CEvGE vs. each contrastive algorithm) for external (CA) and internal (DBI) indexes. The p-values of all pairs of the test result on both synthetic and real world data sets are recorded in Table 7 and Table 8. It can be found that, in most test results, the p-values are smaller than 0.05, and only a few p-values are slightly bigger than 0.5. That is to say, the superior performance of the proposed model is statistically significant in general.
To test the generative ability of the CEvGE framework, we replaced the VAE module in the framework with an AE, and marked this framework as CEvGE/AE. Then, we randomly sampled various proportions (10% to 100%) of objects from KDD’99 and MNIST to construct two new data sets. The clustering results (ARI) of CEvGE and CEvGE/AE performed on these sampled data sets are illustrated in Figure 3. The results show that the performances of these two algorithms reveal a downward trend as the sample size decreases. This finding is largely due to the fact that randomly sampling destroys the intrinsic structure of each category. In particular, under small sample sizes, the sampled objects cannot reflect the original category distributions. It is worth noting that the CEvGE keeps its ARI at a constantly high level, as the sample size is larger than 60%, and it outperforms the comparative algorithm under all of the different sample sizes. The only difference between the proposed algorithm and CEvGE/AE is the generative capacity. Thus, we can conclude that our framework can create some new samples similar to, but not identical to, the input data, and that the generative capacity plays a positive role in the prediction of clustering assignment.

4.4.2. Hyperparameter Analysis

In this section, the influence of the hyperparameter θ on the clustering performance of the CEvGE framework is analyzed. θ is used to control the size of the neighbor set during structure information extraction, and it is directly related to the construction of the affinity matrix. A smaller θ takes more objects into a neighbor set, and consequently involves considerable information that assists the clustering. However, inconsistent neighbors also tend to be included, which may lead to additional computations and may misguide cluster partitioning. In the experiments, we ran the CEvGE 50 times on all of the real data sets with different values of θ . The average results are presented in Table 9. The CEvGE achieves the best result when hyperparameter θ takes a certain value, and its performance degrades if θ is too large or too small. The reason for this is that taking excessive or insufficient objects into the neighborhood may lead to a performance degradation on the structure information extraction. Furthermore, our model generally achieves relatively sound results regardless of the value of θ , as long as the structure information is introduced. This finding indicates that the structure information can enhance the model’s ability to recognize the cluster pattern.

4.4.3. Ablation Study

To investigate the behavior of the proposed CEvGE framework, we conducted several ablation studies. By removing or substituting certain components, we derived four models from CEvGE: (i) CEvGE_del_GE, in which, the graph embedding is removed, and the consensus clustering assignment is obtained by spectral clustering on the object-similarity graph; (ii) CEvGE_del_GR, which constructs structure information using pairwise similarity in the original data space rather than extracting global relationships from base clusterings; (iii) CEvGE_del_DC, in which, the ensemble result is produced purely relying on base clusterings, and the data charateristics are replaced by an all-one matrix; and (iv) CEvGE/AE, which replaces the VAE with an AE. In the ablation experiments, we compared these four models with the CEvGE on KDD’99 and MNIST data sets, and report the results in Table 10 and Table 11, respectively.
(i) Effectiveness of embedding. We first discuss the role of joint embedding learning in our framework by comparing it with the CEvGE_del_GE. Without the embedding learning, the CEvGE_del_GE suffers from a certain performance degradation. These results strongly demonstrate that our joint embedding learning mechanism is capable of acquiring suitable embeddings for the clustering task, consequently enhancing the clustering performance.
(ii) Effectiveness of global structure information. Next, we aim to explore the impact of global structure relationships on clustering assignment. From the results, we can recognize that the structure information extracted from base clusterings helps our framework to capture the intrinsic pattern of the data more accurately compared with the local relationship in the CEvGE_del_GR.
(iii) Integration of data characteristics and structure information. From Table 10 and Table 11, the integration of data characteristics can clearly improve the clustering performance, while the CEvGE_del_DC achieves relatively poor results by being totally reliant on base clusterings when producing the final clustering. It verifies that the integration of two types of information in our framework can eliminate the adverse impacts caused by unreliability in base clusterings.
(iv) Effect of generative model. We also show the effect of generative ability in our CEvGE by replacing the VAE network with an AE. It is obvious that the CEvGE/AE is inferior to the CEvGE in both the aspects of external and internal indexes. The results demonstrate that the generative ability in our framework is beneficial to the preservation of the inherent distribution of original data in latent embedding when training data are insufficient.

5. Conclusions and Future Work

Data characteristics and structure information play different roles in describing the intrinsic pattern of the data. A GNN-based clustering ensemble framework, namely CEvGE, was developed to utilize these two types of information effectively for the production of a consensus clustering result. In this framework, the structure information of data was extracted from base clusterings and used to reorganize the data as an object similarity graph. Subsequently, the data characteristics and structure information were combined elegantly in the form of graph embedding. Lastly, a joint optimization model was constructed to unify the objectives of graph embedding and cluster partitioning, which makes the final clustering result that is produced by appropriate data representations enhanced by structure information. In the experimental analysis, we compared our framework with several state-of-the-art clustering ensemble and deep clustering algorithms on both synthetic and real data sets. The results reflect the validity of the proposed framework.
This work mainly focuses on a general issue of how to enhance clustering ensemble by integrating data characteristics and structure information. It provides a novel research perspective for the clustering ensemble problem. Meanwhile, some specific problems are yet to be solved. For example, the scalability issue of the proposed framework cannot be overlooked, which potentially impedes its usage on large scale data. Concerning this issue, we plan to improve the computational efficiency by optimizing its calculation mode and execution mechanism in the coming research. In addition, extending the availability of the framework for different data types and application contexts is another focus of our future plan.

Author Contributions

H.-Y.D.: conceptualization, methodology, writing—review and editing, project administration. W.-J.W.: conceptualization, methodology, review, project administration. All authors have read and agreed to the published version of the manuscript.

Funding

The funding sources were the National Natural Science Foundation of China (No. 61902227, 62076154, 62022052, U21A20513), and the Technology Research Development Projects of Shanxi (No. 201901D211192).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors are very grateful to the anonymous reviewers and editor. Their many helpful and constructive comments and suggestions helped us to significantly improve this work. We also wish to thank the authors of the compared algorithms for sharing their codes.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Rodriguez, A.; Laio, A. Clustering by fast search and find of density peak. Science 2014, 344, 1492–1496. [Google Scholar] [CrossRef] [Green Version]
  2. Cai, Z.; Yang, X.; Huang, T.; Zhu, W. A new similarity combining reconstruction coefficient with pairwise distance for agglomerative clustering. Inf. Sci. 2020, 508, 173–182. [Google Scholar] [CrossRef]
  3. Yang, C.; Lee, B.; Lin, Y. Effect of money supply, population, and rent on real estate: A clustering analysis in Taiwan. Mathematics 2022, 10, 1155. [Google Scholar] [CrossRef]
  4. Hancer, E.; Karaboga, D. A comprehensive survey of traditional, merge-split and evolutionary approaches proposed for determination of cluster number. Swarm Evol. Comput. 2017, 32, 49–67. [Google Scholar] [CrossRef]
  5. Lin, Z.; Kang, Z.; Zhang, L.; Tian, L. Multi-view attributed graph clustering. IEEE Trans. Knowl. Data Eng. 2021. [Google Scholar] [CrossRef]
  6. Kang, Z.; Lin, Z.; Zhu, X.; Xu, W. Structured graph learning for scalable subspace clustering: From single view to multiview. IEEE Trans. Cybern. 2021. [Google Scholar] [CrossRef]
  7. Lipor, J.; Balzano, L. Clustering quality metrics for subspace clustering. Pattern Recognit. 2020, 104, 107328. [Google Scholar] [CrossRef]
  8. Sourav, S.B.; Boon, S.S.; Bharill, N. Clustering and summarizing protein-protein interaction networks: A Survey. IEEE Trans. Knowl. Data Eng. 2016, 28, 638–658. [Google Scholar]
  9. Saxena, A.; Prasad, M.; Gupta, A.; Bharill, N. A review of clustering techniques and developments. Neurocomputing 2017, 2676, 664–681. [Google Scholar] [CrossRef] [Green Version]
  10. Sandes, N.C.; Coelho, A.L. Clustering ensembles: A hedonic game theoretical approach. Pattern Recognit. 2018, 81, 95–111. [Google Scholar] [CrossRef]
  11. Boongoen, T.; Iam-On, N. Cluster ensembles: A survey of approaches with recent extensions and applications. Comput. Sci. Rev. 2018, 28, 1–25. [Google Scholar] [CrossRef]
  12. Zhang, M. Weighted clustering ensemble: A review. Pattern Recognit. 2022, 124, 108428. [Google Scholar] [CrossRef]
  13. Golalipour, K.; Akbari, E.; Hamidi, S.; Lee, M.; Enayatifar, R. From clustering to clustering ensemble selection: A review. Eng. Appl. Artif. Intell. 2021, 104, 104388. [Google Scholar] [CrossRef]
  14. Shi, Y.; Yu, Z.; Cao, W.; Chen, C.; Wong, H. Fast and effective active clustering ensemble based on density peak. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 3593–3607. [Google Scholar] [CrossRef] [PubMed]
  15. Bai, L.; Liang, J.; Cao, F. A multiple k-means clustering ensemble algorithm to find nonlinearly separable clusters. Inf. Fusion 2020, 61, 36–47. [Google Scholar] [CrossRef]
  16. Khan, I.; Luo, Z.; Shaikh, A.; Hedjam, R. Ensemble clustering using extended fuzzy k-means for cancer data analysis. Expert Syst. Appl. 2021, 172, 114622. [Google Scholar] [CrossRef]
  17. Wang, C.; Chi, C.; She, Z.; Cao, L.; Stantic, B. Coupled clustering ensemble by exploring data interdependence. ACM Trans. Knowl. Discov. Data 2018, 12, 63:1–63:38. [Google Scholar] [CrossRef]
  18. Huang, D.; Wang, C.; Peng, H.; Lai, J.; Kwoh, C. Enhanced ensemble clustering via fast propagation of cluster-wise similarities. IEEE Trans. Syst. Man Cybern. Syst. 2021, 51, 508–520. [Google Scholar] [CrossRef]
  19. Bai, L.; Liang, J.; Du, H.; Guo, Y. An information-theoretical framework for cluster ensemble. IEEE Trans. Knowl. Data Eng. 2019, 31, 1464–1477. [Google Scholar] [CrossRef]
  20. Huang, D.; Wang, C.; Wu, J.; Lai, J.; Kwoh, C. Ultra-scalable spectral clustering and ensemble clustering. IEEE Trans. Knowl. Data Eng. 2020, 32, 1212–1226. [Google Scholar] [CrossRef] [Green Version]
  21. Tian, F.; Gao, B.; Cui, Q.; Chen, E.; Liu, T. Learning deep representations for graph clustering. In Proceedings of the AAAI Conference on Artificial Intelligence, Quebec City, QC, Canada, 27–31 July 2014. [Google Scholar]
  22. Xie, J.; Girshick, R.; Farhadi, A. Unsupervised deep embedding for clustering analysis. In Proceedings of the 33rd International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016. [Google Scholar]
  23. Guo, X.; Liu, X.; Zhu, E.; Yin, J. Deep Clustering with Convolutional Autoencoders. In Proceedings of the 25th International Conference on Neural Information Processing, Montreal, QC, Canada, 3–8 December 2018. [Google Scholar]
  24. Hinton, G.; Salakhutdinov, R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Kipf, T.N.; Welling, M. Variational graph auto-encoders. In Proceedings of the NeurIPS Workshop Bayesian Deep Learning, Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  26. Dong, G.; Liao, G.; Liu, H.; Kuang, G. A review of the autoencoder and its variants: A comparative perspective from target recognition in synthetic-aperture radar images. IEEE Geosci. Remote Sens. 2018, 6, 44–68. [Google Scholar] [CrossRef]
  27. Kingma, D.; Welling, M. Auto-encoding variational Bayes. In Proceedings of the 2nd International Conference on Learning Representations, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  28. Jiang, Z.; Zheng, Y.; Tan, H.; Tang, B.; Zhou, H. Variational deep embedding: An unsupervised and generative approach to clustering. In Proceedings of the 26th International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017. [Google Scholar]
  29. Nalisnick, E.; Hertel, L.; Smyth, P. Approximate inference for deep latent Gaussian mixtures. In Proceedings of the NeurIPS Workshop Bayesian Deep Learning, Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  30. Yang, L.; Cheng, N.; Li, J.; Fang, J. Deep clustering by gaussian mixture variational autoencoders with graph embedding. In Proceedings of the 2nd International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  31. Guo, X.; Gao, L.; Liu, X.; Yin, J. Improved deep embedded clustering with local structure preservation. In Proceedings of the 26th International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017. [Google Scholar]
  32. Li, X.; Zhao, X.; Chu, D.; Zhou, Z. An autoencoder-based spectral clustering algorithm. Soft Comput. 2020, 24, 1661–1671. [Google Scholar] [CrossRef]
  33. Bo, D.; Wang, X.; Shi, C.; Zhu, M.; Lu, E.; Cui, P. Structural Deep Clustering Network. In Proceedings of the 29th International World Wide Web Conferences, Taipei, Taiwan, China, 20–24 April 2020. [Google Scholar]
  34. Wang, T.; Ng, W.W.Y.; Li, J.; Wu, Q.; Zhang, S.; Nugent, C.; Shewell, C. A Deep Clustering via Automatic Feature Embedded Learning for Human Activity Recognition. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 210–223. [Google Scholar] [CrossRef]
  35. Li, X.; Chen, Z.; Poon, L.K.M.; Zhang, N. Learning latent superstructures in variational autoencoders for deep multidimensional clustering. In Proceedings of the 7th International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  36. Hwang, J.; Kim, H. Variational deep clustering of wafer map patterns. IEEE Trans. Semiconduct. Manufact. 2020, 33, 466–475. [Google Scholar] [CrossRef]
  37. Hamilton, W.L.; Ying, R.; Leskovec, J. Representation learning on graphs: Methods and applications. arXiv 2018, arXiv:1709.05584. [Google Scholar]
  38. Cao, S.; Lu, W.; Xu, Q. Grarep: Learning graph representations with global structural information. In Proceedings of the 21st ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Sydney, Australia, 10–13 August 2015. [Google Scholar]
  39. Perozzi, B.; Al-Rfou, R.; Skiena, S. Deepwalk: Online learning of social representations. In Proceedings of the 23rd International Conference on World Wide Web, Seoul, Korea, 7–11 April 2014. [Google Scholar]
  40. Fan, S.; Wang, X.; Shi, C.; Kuang, K.; Liu, N.; Wang, B. Debiased graph neural networks with agnostic label selection bias. IEEE Trans. Neural Netw. Learn. Syst. 2022. [Google Scholar] [CrossRef] [PubMed]
  41. Jin, D.; Liu, Z.; Li, W.; He, D.; Zhang, W. Graph convolutional networks meet markov random fields: Semi-supervised community detection in attribute networks. In Proceedings of the 34th AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020. [Google Scholar]
  42. Pan, S.; Hu, R.; Long, G.; Jiang, J.; Yao, L.; Zhang, C. Adversarially regularized graph autoencoder for graph embedding. In Proceedings of the 27th International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018. [Google Scholar]
  43. Kang, Z.; Pan, H.; Hoi, S.C.H.; Xu, Z. Robust graph learning from noisy data. IEEE Trans. Cybern. 2019, 50, 1833–1843. [Google Scholar] [CrossRef] [Green Version]
  44. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 29th IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 28–30 June 2016. [Google Scholar]
  45. Bagherinia, A.; Minaei-Bidgoli, B.; Hosseinzadeh, M.; Parvin, H. Reliability-based fuzzy clustering ensemble. Fuzzy Sets Syst. 2021, 413, 1–28. [Google Scholar] [CrossRef]
Figure 1. Integral construction of the proposed CEvGE framework. The CEvGE consists of three components: (i) the structure extraction component is used to explore the global structure of data from a set of base clusterings; (ii) the inference model is realized by a GCN module to encode the data characteristics and information jointly; (iii) the generative model reconstructs input data from latent space and produces the final cluster assignment by a GMM. The whole framework can be optimized jointly by maximizing the ELBO, which is calculated and backpropagated to the latent embedding.
Figure 1. Integral construction of the proposed CEvGE framework. The CEvGE consists of three components: (i) the structure extraction component is used to explore the global structure of data from a set of base clusterings; (ii) the inference model is realized by a GCN module to encode the data characteristics and information jointly; (iii) the generative model reconstructs input data from latent space and produces the final cluster assignment by a GMM. The whole framework can be optimized jointly by maximizing the ELBO, which is calculated and backpropagated to the latent embedding.
Mathematics 10 01834 g001
Figure 2. Distributions of synthetic data sets. (a) Rings. (b) Zigzag. (c) Parabolic. (d) Complex. (e) Atom. (f) Chainlink.
Figure 2. Distributions of synthetic data sets. (a) Rings. (b) Zigzag. (c) Parabolic. (d) Complex. (e) Atom. (f) Chainlink.
Mathematics 10 01834 g002aMathematics 10 01834 g002b
Figure 3. Test of generative ability on randomly sampled data sets. (a) ARI metrics on KDD’99. (b) ARI metrics on MNIST.
Figure 3. Test of generative ability on randomly sampled data sets. (a) ARI metrics on KDD’99. (b) ARI metrics on MNIST.
Mathematics 10 01834 g003
Table 1. Details of data sets: number of data samples (N), number of dimensions (D), and number of clusters (k).
Table 1. Details of data sets: number of data samples (N), number of dimensions (D), and number of clusters (k).
Data SetNDk
Synthetic dataRings150023
Zigzag160223
Parabolic100022
Complex303129
Atom80032
Chainlink100032
Real dataIris15043
Breast569302
KDD’9920,000415
MNIST20,00078610
STL-1013,000204810
HHAR10,2995616
Table 2. The contingency table for two partitions on the same data set.
Table 2. The contingency table for two partitions on the same data set.
C P p 1 p 2 p k ^ Sums
c 1 n 11 n 12 n 1 k ^ b 1
c 2 n 21 n 22 n 2 k ^ b 2
c k n k 1 n k 2 n k k ^ b k
S u m s d 1 d 2 d k ^
Table 3. External indexes of different algorithms on synthetic data sets.
Table 3. External indexes of different algorithms on synthetic data sets.
AlgorithmRingZigzagParabolicComplexAtomChainlink
CAARINMICAARINMICAARINMICAARINMICAARINMICAARINMI
A10.52370.31280.28860.51020.37510.40010.57720.30260.42200.41680.26930.23500.58060.22580.31180.41520.19960.1797
A20.78760.63710.70240.70430.56610.60420.69920.58930.61580.64390.51060.60500.58170.43440.38390.68810.55080.4673
A30.43310.15630.48320.41570.26680.31200.50630.21170.23520.37120.16090.35540.60070.41160.38600.57880.37350.2961
A40.73020.63880.47350.67380.45070.48610.73290.58140.51030.60720.52280.53910.67360.46110.37700.58220.41040.3847
A50.62440.23680.28550.60770.30420.37110.64420.35320.29940.47230.21470.30630.63080.41050.35070.52560.31140.2831
A60.26430.00260.02570.42050.38340.14470.21630.00350.01050.37400.19660.36560.2895−0.00130.00300.3381−0.00130.0020
A70.79660.61130.48090.74410.57440.51750.79610.53060.50080.73920.61700.62880.85290.65810.51040.80490.68870.6231
A80.81350.38950.34170.74630.47620.52660.81310.44270.45590.70540.51600.59880.73820.26550.36520.71430.21790.4261
A90.85020.55770.43700.77040.61390.50120.83570.62420.65080.72450.44680.53190.78820.47030.43350.67240.54720.4337
A100.83820.46180.38860.76170.51160.55410.82430.48220.49060.75880.48820.61790.82570.37330.42310.75500.40080.4238
A110.83570.53260.50630.74030.52830.57200.82320.67070.62770.76350.52660.67920.82960.59310.41790.75880.57750.4561
A120.85140.50670.47250.78530.49250.54470.81960.66820.59830.78090.55710.56690.85550.60080.46340.76470.55380.4829
CEvGE0.95480.78380.65530.87660.75540.79340.92210.75210.81540.83630.73340.75900.91350.82780.72460.79460.72500.6533
Table 4. Internal indexes of different algorithms on synthetic data sets.
Table 4. Internal indexes of different algorithms on synthetic data sets.
AlgorithmRingZigzagParabolicComplexAtomChainlink
DBIDVIDBIDVIDBIDVIDBIDVIDBIDVIDBIDVI
A11.35090.15531.71140.18630.96170.14751.05630.16131.71320.12051.30080.1582
A20.57620.16071.25580.17460.50140.14801.22630.18461.26840.12651.14350.1488
A31.74220.15883.52010.17032.13660.13384.30280.17723.41050.12302.33710.1461
A41.03520.15171.22400.16711.21450.14231.15730.17201.58010.12441.36390.1428
A51.47620.17061.20480.17290.99200.14411.30460.18092.04490.13072.16170.1554
A62.38520.15193.07520.17041.52660.13642.78140.16631.57290.12151.42700.1403
A70.78810.16830.85480.16651.40850.15010.99520.17360.82410.12370.84800.1481
A80.87250.16501.45220.16420.81530.14611.70030.17131.28310.12090.90570.1442
A90.52270.16520.81640.17631.12040.15061.38820.18161.51080.09760.93720.1519
A100.81240.17101.26130.18261.10770.14221.50910.17941.40580.12551.16830.1471
A110.65370.16830.85740.18551.46020.14031.23310.19841.37160.12940.90740.1552
A120.75210.17620.78090.18861.21530.14761.33270.21571.28830.13211.12100.1486
CEvGE0.43080.18530.55830.19030.51480.15410.86770.21040.84720.13810.57160.1662
TRUE0.39520.24770.53060.19740.49730.16520.84730.22430.80380.14210.53090.1758
Table 5. External indexes of different algorithms on real data sets.
Table 5. External indexes of different algorithms on real data sets.
AlgorithmIrisBreastKDD’99MNISTSTL-10HHAR
CAARINMICAARINMICAARINMICAARINMICAARINMIACARINMI
A10.82330.64030.62250.85710.59450.43080.82690.78810.76640.82400.76380.72050.81550.76290.58280.78940.65080.6043
A20.93870.74330.77630.88540.77120.72600.83070.88010.82440.90010.85490.75630.78420.73680.65720.85950.74600.6889
A30.78630.65810.58220.74280.70870.73850.76030.67720.84470.82930.73250.69040.73010.60520.70060.81660.71330.6527
A40.82450.68290.62170.71920.67730.62400.81530.82710.82550.82530.78740.76600.80020.78410.68250.81030.73510.6423
A50.80140.63100.62910.77530.68940.67420.72490.60070.55090.80610.65810.57190.68420.47260.44100.76330.53990.4937
A60.54800.10260.16090.4752-0.00070.00070.5829-0.00050.35020.60040.21140.47720.72410.58420.52700.82530.71920.6035
A70.91360.73380.68280.90620.78620.66300.85790.85910.74320.89630.86450.78860.84150.72080.71940.85500.66780.6877
A80.86370.62550.65070.78250.68640.55840.88460.85250.85750.85910.82690.83720.85540.78850.73710.81060.75420.6558
A90.89720.72600.71320.82430.71100.62560.85140.79630.81440.82260.79880.81730.83410.78070.72580.92880.85110.7094
A100.85280.65410.62350.77920.72350.57710.81150.85390.85200.83320.85200.86720.81190.82040.70040.80950.78260.6894
A110.88420.72260.70040.80340.74430.65520.92720.90330.86610.88960.90140.86770.84370.83280.75320.85460.82100.6782
A120.87790.73480.72140.82230.67130.62380.91030.88070.86350.87220.88350.86420.88710.76040.76040.83320.70960.7033
CEvGE0.93350.75670.75670.91760.77080.71410.95270.92240.90240.95620.92030.88340.91740.85420.78870.92080.85640.7246
Table 6. Internal indexes of different algorithms on real data sets.
Table 6. Internal indexes of different algorithms on real data sets.
AlgorithmIrisBreastKDD’99MNISTSTL-10HHAR
DBIDVIDBIDVIDBIDVIDBIDVIDBIDVIDBIDVI
A10.71250.15690.74430.17020.82080.13710.84530.09861.10620.14761.05130.1538
A20.61830.15470.77080.18100.77420.13200.86640.11721.08550.15531.17960.1829
A30.63700.15560.61390.14630.82260.14290.82510.12280.99430.13620.92740.1772
A40.91570.16450.72830.13910.84160.12381.12760.11591.37010.12471.22600.1578
A50.78650.15480.89340.13570.81350.11091.03520.11791.38560.13152.18690.1680
A63.55160.09614.21940.11462.24050.11073.73410.08424.55610.12043.74780.1477
A70.45760.20350.58420.18670.67130.15230.78750.11551.11470.15140.81570.1692
A81.18270.12511.53770.16820.98740.12051.26330.10411.55140.13680.97050.1618
A90.73480.17660.77310.15261.13690.10371.30040.09341.23380.13180.83200.2073
A101.04920.13161.44830.16641.07680.12411.31550.09481.42360.13290.96430.1671
A110.85190.16491.18940.17500.96630.14461.08620.11591.21070.12471.08830.1523
A120.88480.15730.94420.17440.89670.14650.94770.11711.22480.12590.89320.1596
CEvGE0.43220.23070.59410.20670.55680.16890.70050.14050.84030.15940.70690.2009
TRUE0.41530.27630.56180.21570.52240.17640.65360.17450.81740.17410.68420.2136
Table 7. The p-values of Student’s t-test of CEvGE vs. contrastive algorithms on synthetic data sets (p > 0.05 are underlined).
Table 7. The p-values of Student’s t-test of CEvGE vs. contrastive algorithms on synthetic data sets (p > 0.05 are underlined).
PairsRingZigzagParabolicComplexAtomChainlink
CADBICADBICADBICADBICADBICADBI
CEvGE vs. A10.01270.01500.00650.01140.03270.02950.01300.01430.02250.02070.01440.0186
CEvGE vs. A20.04330.03520.03150.02280.05310.04300.02010.01770.01740.02110.00490.0041
CEvGE vs. A30.00870.01050.01660.01470.03040.02660.01390.02450.01570.01880.00740.0043
CEvGE vs. A40.00680.00740.01100.01260.01980.01830.00780.00620.01460.01440.01300.0114
CEvGE vs. A50.02770.02190.00780.01830.02110.03270.00760.00850.02180.00920.02090.0231
CEvGE vs. A60.00620.00550.00740.00630.01360.01900.00350.00320.01060.01260.03020.0269
CEvGE vs. A70.01330.01730.02310.01960.04270.03280.04470.03160.04190.05240.05170.0389
CEvGE vs. A80.02640.02230.01840.02620.00620.00550.00810.00430.00540.00750.02750.0239
CEvGE vs. A90.01780.00790.01460.02030.02750.02470.01880.02170.01580.00940.00750.0042
CEvGE vs. A100.01860.01640.01180.01050.00310.00540.00490.00580.00680.00540.01330.0097
CEvGE vs. A110.00830.00880.01000.01650.01090.00930.00620.00500.00660.00590.00930.0124
CEvGE vs. A120.00750.00910.00770.00490.00610.00400.00230.00380.00620.00440.00420.0040
Table 8. The p-values of Student’s t-test of CEvGE vs. contrastive algorithms on real data sets (p > 0.05 are underlined).
Table 8. The p-values of Student’s t-test of CEvGE vs. contrastive algorithms on real data sets (p > 0.05 are underlined).
PairsIrisBreastKDD’99MNISTSTL-10HHAR
CADBICADBICADBICADBICADBICADBI
CEvGE vs. A10.01720.00860.02730.01890.02500.03090.01150.00820.02410.01860.01780.0153
CEvGE vs. A20.00540.00770.01660.01400.01330.01520.04080.05180.00770.00820.01880.0209
CEvGE vs. A30.01790.01560.02060.02170.01580.01740.00270.00550.02880.05220.02540.0213
CEvGE vs. A40.02450.03000.02120.02260.01550.01350.00480.00550.01390.01460.02760.0331
CEvGE vs. A50.01090.01220.01140.01520.02600.04280.01320.01460.01510.01320.01840.0189
CEvGE vs. A60.01650.01470.05270.03530.00710.01050.00840.00770.02440.02070.01760.0160
CEvGE vs. A70.02550.02300.03770.04150.00170.00090.00610.00390.04060.02790.03180.0341
CEvGE vs. A80.00840.01240.01830.01970.00140.00310.01030.01280.01300.01470.05300.0317
CEvGE vs. A90.01260.01080.00720.00420.02780.03250.00620.00700.01450.01280.04650.0387
CEvGE vs. A100.00750.00910.00630.00460.0021000500.01330.01090.01940.01730.01550.0097
CEvGE vs. A110.01030.00950.00190.00270.01020.00670.02040.01730.01440.01400.01270.0109
CEvGE vs. A120.01380.01210.00300.00230.01310.01050.00550.00370.00830.01000.00930.0120
Table 9. Ensemble results of the CEvGE algorithm with different values of hyperparameter θ .
Table 9. Ensemble results of the CEvGE algorithm with different values of hyperparameter θ .
θ IrisBreastKDD’99MNISTSTL-10HHAR
CAARINMICAARINMICAARINMICAARINMICAARINMIACARINMI
0.10.87200.77020.72910.89950.75530.67440.93520.91420.87740.92600.91720.88190.88930.84410.77600.88690.83860.7185
0.20.89020.77870.73320.90310.75970.70290.93910.91880.89030.94340.92030.88210.89640.85080.78040.89420.84770.7207
0.30.92730.78340.73860.91760.77080.71410.94560.92030.89250.95620.91930.88340.91030.85420.78870.90360.85050.7246
0.40.93350.76920.75670.90730.76420.70840.95270.92240.90240.95190.91570.88150.91570.85190.77950.91660.85640.7210
0.50.91100.76370.75280.90140.75750.69550.95060.91940.88860.94870.91380.88080.91740.85110.77360.92080.85330.7202
0.60.88530.76230.74020.88600.75320.68020.93770.91720.88170.94150.91300.87620.89220.84870.77240.91810.85270.7163
0.70.88180.76020.73370.87250.75260.67750.93080.91660.87550.93760.91210.87550.88710.84460.77060.89940.85070.7152
0.80.87340.75960.73160.87060.75090.67330.92270.91440.86810.93320.90960.87310.88370.84380.75550.89230.83860.7084
0.90.86300.75590.72230.86800.75040.66720.92030.91380.86790.91880.90650.87040.86590.83950.75370.87050.83310.6941
Table 10. Ablation studies on KDD’99 data set.
Table 10. Ablation studies on KDD’99 data set.
ModelCAARINMIDBIDVI
CEvGE0.95270.92240.90240.55680.1689
CEvGE_del_GE0.93550.90820.87610.64800.1566
CEvGE_del_GR0.93070.90530.86930.93520.1477
CEvGE_del_DC0.85470.83030.83100.82040.1273
CEvGE/AE0.92510.89520.85260.77350.1482
Table 11. Ablation studies on MNIST data set.
Table 11. Ablation studies on MNIST data set.
ModelCAARINMIDBIDVI
CEvGE0.95620.92030.88340.70050.1405
CEvGE_del_GE0.91940.88520.86630.83170.1288
CEvGE_del_GR0.89810.90350.87060.99240.1214
CEvGE_del_DC0.84260.79880.79301.10360.1164
CEvGE/AE0.92330.88360.87240.82650.1236
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Du, H.-Y.; Wang, W.-J. A Clustering Ensemble Framework with Integration of Data Characteristics and Structure Information: A Graph Neural Networks Approach. Mathematics 2022, 10, 1834. https://doi.org/10.3390/math10111834

AMA Style

Du H-Y, Wang W-J. A Clustering Ensemble Framework with Integration of Data Characteristics and Structure Information: A Graph Neural Networks Approach. Mathematics. 2022; 10(11):1834. https://doi.org/10.3390/math10111834

Chicago/Turabian Style

Du, Hang-Yuan, and Wen-Jian Wang. 2022. "A Clustering Ensemble Framework with Integration of Data Characteristics and Structure Information: A Graph Neural Networks Approach" Mathematics 10, no. 11: 1834. https://doi.org/10.3390/math10111834

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop