Next Article in Journal
Digital Twin for Lyophilization by Process Modeling in Manufacturing of Biologics
Next Article in Special Issue
Snapse: A Visual Tool for Spiking Neural P Systems
Previous Article in Journal
CFD Hydrodynamics Investigations for Optimum Biomass Gasifier Design
Previous Article in Special Issue
A Representation of Membrane Computing with a Clustering Algorithm on the Graphical Processing Unit
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Consensus Fuzzy K-Modes Clustering Using Coupling DNA-Chain-Hypergraph P System for Categorical Data

Business School, Academy of Management Science, Shandong Normal University, Jinan 250014, China
*
Author to whom correspondence should be addressed.
Processes 2020, 8(10), 1326; https://doi.org/10.3390/pr8101326
Submission received: 24 September 2020 / Revised: 14 October 2020 / Accepted: 15 October 2020 / Published: 21 October 2020
(This article belongs to the Special Issue Modeling, Simulation and Design of Membrane Computing System)

Abstract

:
In this paper, a data clustering method named consensus fuzzy k-modes clustering is proposed to improve the performance of the clustering for the categorical data. At the same time, the coupling DNA-chain-hypergraph P system is constructed to realize the process of the clustering. This P system can prevent the clustering algorithm falling into the local optimum and realize the clustering process in implicit parallelism. The consensus fuzzy k-modes algorithm can combine the advantages of the fuzzy k-modes algorithm, weight fuzzy k-modes algorithm and genetic fuzzy k-modes algorithm. The fuzzy k-modes algorithm can realize the soft partition which is closer to reality, but treats all the variables equally. The weight fuzzy k-modes algorithm introduced the weight vector which strengthens the basic k-modes clustering by associating higher weights with features useful in analysis. These two methods are only improvements the k-modes algorithm itself. So, the genetic k-modes algorithm is proposed which used the genetic operations in the clustering process. In this paper, we examine these three kinds of k-modes algorithms and further introduce DNA genetic optimization operations in the final consensus process. Finally, we conduct experiments on the seven UCI datasets and compare the clustering results with another four categorical clustering algorithms. The experiment results and statistical test results show that our method can get better clustering results than the compared clustering algorithms, respectively.

1. Introduction

Data clustering has recently attracted more attentions in practical applications. However, most studies are about numerical data. In this method, the distance between the clustering center and the data objects are calculated by the standard distance metrics. However, there are many classification datasets that do not have a natural order or distance between the parts. For example, in the real world, each classification attribute of blood type has a unique classification value, such as [ A , B , O o r AB ] . Therefore, research into categorical data is a difficult and challenging task, which attracts many data mining researchers.
In 1998, Huang [1] proposed the k-modes algorithm for the categorical data clustering. The k-modes algorithm calculated the distance between the object and the cluster center by the Hamming distance instead of Euclidean distance which is used in the k-means algorithm. Then, Huang [2] proposed the fuzzy k-modes algorithm (FKM). This algorithm was an extended version of the k-modes algorithm. Thereafter, many algorithms have been proposed for the clustering of categorical data [3]. These methods were mostly based on numerical data clustering algorithms, such as ROCK [4], CACTUS [5], COOLCAT [6], LIMBO [7], wk-modes [8], MOGA [9], NSGA-FMC [10], SBC [11], MOFC [12], and so on. The research content is mainly divided into two categories. One is the method of the similarity measure about the categorical data. Hamming distance is the main method for calculating the distance between the data point and the clustering center which measures the distance two different categorical values at 1, otherwise is 0. This method has been used in many other clustering algorithms, such as the k-prototypes algorithm [1] and Squeezer [13]. In addition to the Hamming distance, there are many other distance metrics [14], such as Ahmad’s distance metric, the Jaccard metric, the association-based distance metric, and the context-based distance metric. Another direction is utilizing the optimization algorithm to improve the clustering performance. The WFKM algorithm [2] strengthened the basic k-modes clustering by associating higher weights with features useful in analysis. The GFKM algorithm [15] integrated the genetic algorithm into the fuzzy k-modes algorithm, aiming to find the global optimal solution. The IWFKM algorithm [16] replaced the Hamming distance with the frequency probability-based distance metric, which has been proved to improve the clustering results. Apart from this method, the multi-objective clustering algorithm was also considered to improve the performance of the categorical algorithm [17]. Rough set theory was also introduced into the K-modes algorithm, which is used to calculate the density of each candidate modes to characterize the distribution around it [18]. Some researchers developed cluster weighted estimators of marginal proportion, which remain unchanged under the amount of information, and derive the similarity, one sample ratio, goodness of fit and independent chi square test for clustering data [19]. Some research also introduced other evolutionary algorithms into K-modes algorithm. For example, the firefly algorithm was used to generate initial clusters [20]. Some scholars started from the data themselves, and studied the possibility of reducing the dimension of representation based on spatial structure while maintaining the same representation capability [21].
Pǎun [22] proposed membrane computing (also called P system) in 1998. This aims to abstract computational models from the structure and function of biological cells and the cooperation between organs and tissues. Up to now, membrane computing mainly includes three basic models: cell-like p system, tissue-like p system and neuron-like p system. In the process of calculation, each cell is regarded as an independent unit, and each cell operates independently and does not interfere with each other. The whole membrane system operates in a highly parallel mode. According to the research content about the P system, it can be divided into theoretical study and application study. For the theoretical research, researchers use the direct membrane to solve problems [23]. In this aspect, some new P system models are proposed which can improve the computation power with the min cells or spikes [24,25]. Many other variants of membrane systems were also proposed in [26,27,28,29]. In application research, the direct membrane algorithm was used to solve some practical problems by the researchers [30]. Some researchers also used the coupled membrane system to realize the clustering process [31,32]. These membrane systems are all improvements based on cell-like P systems, tissue-like P systems and neuron-like systems. For these kinds of system, they are all designed based on simple structures and can not solve the problems with complicated structure. For example, the traditional P system cannot store multivariate data with complex relationships. Therefore, our team previously proposed the concept of P system with the simple complex structure [33] and chain structure. For instance, Liu and Xue established the new P system based on the simple complex structure in [33], and Luan and Liu designed the chain P system [34], while Yan and Xue proposed the chain-hypergraph P system [35].
Based on the above analysis, we proposed a novel hybrid DNA-chain-hypergraph P system to implement the consensus clustering (DCHP-FCC). The DCHP-FCC system contains three reaction chain membrane subsystems and one consensus subsystem. Three different base clustering algorithms are used in the three subsystems, respectively. This operation combines the different advantages of the three algorithms, and the DNA genetic algorithm implements the consensus clustering process in the consensus system. The experiment on seven UCI datasets is conducted. The experimental results show that our proposed method outperforms the results of the state-of-the-art methods.
This work makes the following contributions:
(1) A novel DCHP system is designed which combines the advantage of the chain structure and hypergraph topology structure. Three reaction chain membrane subsystems and one consensus subsystem are designed to generate the basic partitions and integrate basic partitions, respectively.
(2) A revised k-means which is optimized by the DNA genetic algorithm is used for a basic partition integration strategy which can optimize the initial clustering center and obtain the global optimal solutions.
(3) Simulation is performed using well-known datasets in the UCI machine learning repository to verify the clustering quality of the DCHP-FCC.
The rest of this paper is organized as follows. Section 2 introduces the basic concepts of the k-modes algorithm, consensus clustering and chain and hypergraph structure. The coupling DCHP system is illustrated in Section 3. Experiments and results are analyzed in Section 4. Section 5 summarizes conclusions and future research directions.

2. Basic Concepts

2.1. Three Basic Fuzzy K-Modes Clustering Algorithms

The fuzzy k-modes algorithm (FKM) was proposed by Huang and Ng [1], and is one of the most popular clustering algorithms for categorical data. This type of method has improved the k-modes algorithm by the corresponding membership degree value in different clustering.
Definition 1.
Definiting 4-tuple S = ( U , A , V , f ) is an information system, where U represents non-empty finite set of objects, A refers to the non-empty finite set of attributes, V = a A V a , V a records the domain of attribute a , and f : U × A V is a total function such that f ( u , a ) V a , for every ( u , a ) U × A .
Let X be the dataset which has n categorical objects. Each object x i ( 1 i n ) has p attributes, so that x i = { x i 1 , x i 2 , , x i p } . The objective of the FKM is characterized as follows:
F F K M ( U , Z , X ) = j = 1 k i = 1 n u j i α d ( x i , z j )
subject to:
0 u j i 1 ,    1 i n , 1 j k ,
j = 1 k u j i = 1 ,    1 i n
0 < i = 1 n u j i < n ,    1 j k
where, α is weight component, U = ( u j i ) is a k × n matrix which records the fuzzy membership degree, Z = { z 1 , , z k } is the set of the clustering centers. X = ( X 1 , X 2 , , X n ) is the data matrix, where X i is the i th point. d ( x i , z j ) calculates the distance between the object x i and the clustering center z j . The distance is calculated by simple matching dissimilarity measures or Hamming distance which is showed as follows:
d ( x i , z j ) = l = 1 m δ ( x i l , z j l )
and
δ ( x i l , z j l ) = { 0 , i f x i l = z j l 1 , i f x i l z j l
Based on the proposed scheme, a weight vector is added in the conventional fuzzy k-modes algorithm [2]: W = [ w 1 , w 2 , , w p ] , where w l represents the weight for the l th variable, l = 1 , 2 , , p . So, the objective function of the WFK-modes (WFKM) algorithm is:
F α , β ( U , Z , X , W ) = j = 1 n i = 1 n u j i α d j i W
where, β is the power of the attribute weight w l which is a measure of emphasis on weights. Similar to the FKM algorithm, d j i W = d W ( Z j , X i ) = l = 1 p δ W ( z j l , x i l ) , where z j i is the l th term of Z j and x i l refer to the l th term of X i .
δ W ( x i l , z j l ) = { 0 , i f x i l = z j l w l β , i f x i l z j l
In addition to improving the algorithm itself, some optimization algorithm is also used to optimize the FKM algorithm. For example, the genetic fuzzy k-modes algorithm integrated the genetic algorithm and the conventional fuzzy k-modes algorithm, which is called GFKM [36]. This algorithm has five basic steps: (1) string representation, (2) population initialization, (3) selection process, (4) crossover process, and (5) mutation process. The fuzzy k-modes algorithm which is optimized by the genetic algorithm can obtain the globally optimal solution and speed up the process of the convergence.

2.2. Consensus Clustering

Consistent clustering is a framework for clustering multiple algorithms or the same algorithm under different parameters to obtain better results. As shown in Figure 1, let X = { x 1 , x 2 , , x n } represents the dataset, and arbitrary cluster algorithms are used p times to get p different basic partitions π 1 , π 2 , , π p (BPs) [37].
The number of clusters K 1 , K 2 , , K p in different partition is arbitrary. Each cluster result can transfer to the corresponding binary valued vector representation, and the binary value 1 indicates the sample i belonging to the cluster, otherwise 0. Then, the consensus function needs to aggregate the BPs and obtains the final clustering result π . In this paper, the revised k-means algorithm which optimized by the DNA genetic algorithm is used as the consensus function. The DNA genetic algorithm is used to optimize the initial clustering center of the k-means algorithm, and the clustering quality can be evaluated by some common evaluation indicators, such as normalized mutual information, accuracy and F_measure, etc. The consensus clustering result is typically better than that obtained by the best BP. The following notations can be used to illustrate it:
π i p a r t i t i o n i : ( k i , C 1 i C k i i ) k i n u m b e r o f c l u s t e r s i n p a r t i t i o n π i C j i = { s l : s l c l u s t e r j o f p a r t i t i o n π i }    l i s t o f s a m p l e s i n t h e j t h c l u s t e r o f p a r t i t i o n π i X j i : X j i ( k ) = { 1    i f s k C j i , k = 1 , , n 0    o t h e r w i s e    b i n a r y v a l u e d v e c t o r r e p r e s e n t a t i o n o f c l u s t e r C j i

2.3. Chain and Hypergraph Structure

A simple complex S is a set of non-empty simplex s 1 , s 2 , , s p . If s 1 s 2 , s 1 represent the vertex or face of the simplex s 2 . Each complex also should be oriented. Therefore, a S -chain is a simplicial complex with p dimensional simplices, which defined in [33]. All simplexes which have the same directions can combine into a chain domain.
A hypergraph H = ( v , e ) represents a graph whose edge contains an arbitrary number of vertices [38,39,40]. v is vertices sets and e is hyper-edges sets. A hyper-edge can contain more than two vertices and can be formally represented by a nonempty subset of v . As shown in Figure 1. A hypergraph can better represent the complex information and the local group information than the traditional graph. In the traditional graph, there is only one edge between the two vertices if they are similar. However, in the hypergraph, we can construct a hyper-edge to connect more than two vertices. Therefore, the local group information and complex relationship hidden in the data can be captured by the hypergraph model [41].
We also can represent the hypergraph H = ( v , e ) in an accessible matrix:
H = { 1 , v e 0 , o t h e r w i s e
where, H = 1 if the hyper-edge e contains the vertex v . So, the Figure 2 can be expressed as:
[ e 1 e 2 e 3 e 4 v 1 1 0 0 0 v 2 1 1 0 0 v 3 1 1 1 0 v 4 0 0 0 1 v 5 0 0 1 0 v 6 0 0 1 0 v 7 0 0 0 0 ]

3. Coupling DNA-Chain-Hypergraph P System for Consensus Clustering (DCHP-FCC)

In this section, the coupling DCHP system is proposed. Firstly, the membrane structure of the DCHP system is introduced. Then, the different operations in the subsystem and consensus system are introduced, respectively. The flowchart of the proposed DCHP-FCC algorithm is shown in Figure 3.

3.1. Membrane Structure of the DCHP System

The DCHP system has two main membrane structures—the chain membrane and hyper-membrane structure. The basic framework of the chain membrane structure is shown in Figure 4 and the structure of the hyper-membrane structure is shown in Figure 5.
Definition 2.
Based on the operation of the chain structure, the unit membrane can combine into chains. The chain membrane has two directions (i.e., “+” or “−”). As shown in Figure 4, the chain σ 1 is “+”: σ 11 σ 12 σ 13 σ 1 s and s is the length of the chain σ 1 . As in σ 2 , the chain σ 2 is “−”: σ 21 σ 22 σ 23 σ 2 s . Membranes σ 11 , σ 12 , , σ 1 s , σ 21 , , σ 2 s are all unit membranes. There is a channel between the adjacent unit membrane. τ is the skin membrane, and it is also the max membrane in the system. σ 11 , σ 12 , , σ 1 s , σ 21 , , σ 2 s are all called children membrane of the skin membrane τ , so σ 11 , σ 12 , , σ 1 s , σ 21 , , σ 2 s τ . A membrane without any other membranes inside it is an elementary membrane, so σ 11 , σ 12 , , σ 1 s , σ 21 , , σ 2 s are also represent elementary membrane.
Definition 3.
Based on the topology structure of the hypergraph, the hyper-membrane structure is designed as a membrane with two or more upper membranes. In this system, it also has a similar definition to the chain membrane. For example, a membrane without any upper membrane is as skin membrane, and a membrane without any children membranes is an elementary membrane. For the two membranes m 1 and m 2 , m 1 is the upper membrane of the m 2 if m 2 m 1 . If there is no membrane, m 3 satisfies m 2 m 3 m 1 , and the membrane m 2 is the correspondingly lower membrane of m 1 As shown in Figure 5, membrane 1 is a skin membrane, and membranes 2, 3, 6, 8, 9, 10, 11 are all elementary membranes. In particular, membrane 8 is a hyper-membrane which has two upper membranes, 4 and 7.
According to the basic membrane structure in Figure 4 and Figure 5, a novel P system is designed for the consensus clustering process. The membrane structure of the DCHP system is shown in Figure 6. The DCHP system of degree m > 0 is defined as:
= ( O , μ , ω 1 , ω 2 , , ω m , s u b s y s i , c o n s y s , i 0 )
where:
  • O is the finite set of objects;
  • μ represents the structure of the membrane. It includes the structure of the chain membrane, hyper-membrane and consensus membrane;
  • ω 1 , ω 2 , , ω m are objects in O , which represent the initial multisets objects in m membranes at the beginning of the calculation; we denote the number of chain membrane is m 1 , the number of hyper-membrane is m 2 and the number of membrane in consensus system is m 3 . m 1 + m 2 + m 3 = m . λ means the membrane has no object.
  • s u b s y s i is the subsystem which is used to generate the basic partition of clustering. In this system, three subsystems execute three kind of clustering algorithm, respectively.
  • c o n s y s is the consensus clustering membrane, which is used to generate the final clustering result.
  • i 0 is the output membrane of the system .

3.2. The Consensus Clustering Realized with the DCHP System

To implement the consensus clustering of the fuzzy k-modes algorithm, we propose three kinds of subsystem (i.e., reaction chain-hyper P system, consensus system, and global DCHP system). As shown in Figure 6, three classic algorithms (FKM [1], WFKM [2,16] and GFKM [36]) are simultaneously implemented in the three subsystems. The fuzzy k-modes algorithm generates a fuzzy partition matrix for the categorical data, and gives confidence to objects in different clusters. We call it soft partition, which is closer to reality than hard partition. This method is equal to all variables that determine cluster membership. However, the situation in the real world is usually different. The WFKM algorithm introduces a weight vector in the traditional FKM algorithm. This modification associates higher weights with the features which are instrumental in recognizing the clustering pattern of the data. These two methods are only improvements of the K-mode algorithm itself, and there are still some drawbacks in the convergence speed and the global optimization of the algorithm. In order to speed up the convergence process, the GFKM algorithm is proposed which used a one-step crossover process, mutation process and selection process in the clustering process.

3.2.1. Reaction Chain-Hypergraph P System in Subsystem

Initially, objects are randomly generated in the reaction chain hypergraph P system. The object represents a set of cluster centers. Suppose the dataset X = { X 1 , X 2 , , X n } R n × d has K clusters, where n is the number of data points and d is the dimension. The initial cluster centers K i ( i = 1 , 2 , 3 ) are randomly selected out of n data as the initial cluster centers and are denoted as Z = { z 1 , z 2 , , z k i } , where z i = { z i 1 , z i 2 , , z i d } R d . Z = { z 1 , z 2 , , z k i } is the initial string in the reaction chain hypergraph P system. Different reaction chain hypergraph P systems in the same subsystem conduct the same k-modes algorithm. Different subsystems conduct different K-modes algorithms with different parameters in parallel.
Then, each reaction chain-hyper P system generates a clustering result with different cluster number K i . The clustering result π i is transferred to the corresponding subsystem. These results are also called basic partitions (BPs). Next, the object co-occurrence strategy is used for these BPs, and the details of this method can be seen in [42]. The BPs are transformed into a binary dataset X ( 2 ) = { x l ( 2 ) | l = 1 , 2 , , n } : x l ( 2 ) = < x l , 1 2 , , x l , i 2 , , x l , p 2 > , w i t h x l , i ( 2 ) = < x l , i 1 2 , , x l , i j 2 , , x l , i K i 2 > , and
x l , i j ( 2 ) = { 1 , i f x l b e l o n g s t o c l u s t e r j i n B P π i 0 , o t h e r w i s e

3.2.2. Local Communication Membrane System

Afterwards, the transformed binary dataset X ( 2 ) is transferred to the consensus system by the communication rule:
( p , a k / q , λ )
where a k is the basic partition result in the subsystem and λ means there is no object in the membrane. p is the subsystem, and q is the consensus system.

3.2.3. Consensus System

The revised k-means algorithm which is optimized by the DNA genetic algorithm is used as the consensus function. The DNA genetic algorithm is used to optimize the initial clustering center of the k-means algorithm. When the dataset X ( 2 ) appears in the consensus system, it is transferred to binary. So, we can use these values as the initial population. Then, the specific selection, crossover and mutation process of DNA genetic algorithm are described as follows:
(1) Selection operation
The optimal individuals with the first 10% fitness are directly inherited to the next generation, and the rest of the individuals are selected according to the roulette selection strategy.
(2) Crossover and mutation operation
One-point crossover and adaptive mutation operation are used in this step, respectively. The crossover probability is set as P c . The mutation probability can be adjusted according to the fitness value of the individual. The specific mutation probability is updated as follows:
P m = ( f m a x f ) / ( f m a x f a v g )
where, P m is the mutation probability, f m a x refers to the maximum fitness in every generation, f a v g is the average fitness in every generation, and f is the fitness of the individual. If the fitness is equal to the maximum fitness, the mutation probability is 0. This operation can guarantee that the optimal individual does not change by the mutation operation.
(3) Fitness function
The absolute-deviation criterion in the following is used to measure the clustering quality:
F = k = 1 K x i C k d i s t ( x i , o k )
where, x i is data point i in the dataset X ( 2 ) , o k is the cluster center of the C k . d i s t ( x i , o k ) calculates the Euclidean distance between the data point x i and the corresponding cluster center o k .
The computation process can stop when the predefined maximum iteration is reached or the difference between two adjacent iterations is less than the given threshold ε . Then, the DCHP system is stopped, and the final results are output in the membrane i 0 .

4. Experiments and Discussions

4.1. Data Sets and Parameter Settings

Seven datasets which were collected from the UCI Machine Learning Repository [43] are used in this section. The data type of all seven datasets is categorical data. These datasets are used in the comparison algorithms. Although the comparison algorithms used several other data sets in addition to these seven datasets, considering that these datasets have poor results in the comparison algorithms, they are not used as experimental data in this paper. Table 1 shows the detail information of these datasets. To generate the BPs, three different clustering algorithms are used. The number of clusters K i ( i = 1 , 2 , 3 ) in the process of BPs generation is set to [ K a , n ] , and K a is the actual number of clusters of the dataset. The cluster number of the consensus clustering process is set as K a . All experiments are simulated by the MATLAB R2014b running on a Windows 7 platform of the 64-bit edition. The PC has an Intel Core i7-4770 3.4 GHZ CPU and 8 GB of RAM.

4.2. Evaluation Metric

To evaluation the performance of the DCHP-FCC algorithm, three external clustering evaluation metrics are selected [44], i.e., Adjusted Rank Index (ARI), Clustering Accuracy (ACC), and F_measure (F).
The ARI is defined as follows:
A R I ( T , C ) = 2 ( a d b c ) ( a + b ) ( b + d ) + ( a + c ) ( c + d )
where, T represents the pre-determined clustering label, and the C represents the label of the clustering result. a , b , c and d refer to: (1) in the same class as T and C , (2) in the same class as T but not in the same class as C , (3) in the same class as C but not in the same class as T , (4) in a different class C and T , respectively.
The ACC can be calculated by:
A C C = k = 1 n δ ( l k , m a p ( g k ) ) n
where, n is the number of data points, g k represents the clustering result labels which are obtained by the algorithm, and l k is the true class label of the data x i . m a p ( · ) is the mapping function that maps the clustering labels obtained by the algorithm to the real clustering labels. When l k = m a p ( g k ) , the function value of δ ( l k , m a p ( g k ) ) is equal to 1, and otherwise it is equal to 0.
The F_measure is defined as:
F _ m e a s u r e ( k , l ) = 2 × ( P ( k , l ) × R ( k , l ) ) ( P ( k , l ) + R ( k , l ) )
where, P ( k , l ) = s k l / s k and R ( k , l ) = s k l / s l . P ( k , l ) refers to the precision of cluster k with respect to class l , and R ( k , l ) represents the recall of cluster k with respect to class l . s k l refers to data points which belong to both cluster k and l , s k represents the number of data points in k , and s l is the number of data points in l .

4.3. Experiment Results and Analysis

In this section, we firstly select the number of BPs. According to the analysis in previous research [37], we can select the number of BPs as about 50, and the clustering effect gradually stabilizes. In this paper, we also need to determine the number of BPs, but we do not use the same BP generation strategy as in the other methods. In this paper, BPs need to be generated in each time. At the same time, considering the structure of the DCHP system and the characteristics of the three basic clustering algorithms, we need to guarantee that the BPs generated by the different algorithms are the same. So, the final number of BPs is a multiple of 3. Taking into account that the number of BPs is basically maintained at about 50, the clustering effect is the best, so we conduct preliminary experiments on the number of BPs with 30, 60 and 90, respectively. Experiments are run 30 times, and the mean and variance of the experimental results are recorded in each case. The experimental results in Table 2 show that when the number of BPs is 60, the experimental effect is the best.
At the same time, the boxplots for the 30 times are also shown in Figure 7. In Figure 7, (a) is the result of the NMI metric, (b) is the result of the ACC metric, and (c) is the result of the F_measure metric. The red line is the center of the box. The top line is the third quartile and the bottom line is the first quartile of the box. The upper and lower limits of the whiskers represent the maximum and the minimum values, respectively. The symbol ‘+’ represents the outlier. Note that the better clustering result, the higher and more compact the boxes are. We can see from Figure 7, when the BPs is 60, the best clustering results are obtained. Therefore, the number of BPs in subsequent experiments is determined to be 60.
Next, the DCHP-FCC algorithm is compared with three basic clustering algorithms and one improved algorithm, IWFKM, which was proposed in [16]. In order to maintain the originality of the benchmark algorithms, this paper obtains the results of these algorithms through their own algorithms and parameters, which are shown in Table 3. The mutation probability in the GFKM algorithm is also set as 0.01. For comparison, the maximum number of iterations of the comparison algorithm is 100.
Every experiment is also run 30 times, and the mean and variance of the experimental results are recorded in each case. The experimental results are shown in Table 4. As we can see from Table 4, the performance of the DCHP-FCC is much better than the compared algorithms on Soybean-small, Spect heart, Voting, Zoo and Mushroom datasets, and partly better than the compared algorithms on the Tic-tac-toe and Breast cancer datasets. For the Tic-tac-toe and Breast cancer datasets, the DCHP-FCC algorithm can obtain the best ARI value, and the GFKM algorithm attains the best ACC and F_measure values. Even though the DCHP-FCC algorithm does not perform best on the Tic-tac-toe and Breast cancer datasets, the metric value is only 0.01 lower than the best value. So, the DCHP-FCC algorithm is more effective in dealing with the categorical clustering than the compared clustering algorithms.

4.4. Significance Testing

In this section, the hypothetical tests on the values of seven UCI datasets are computed between the DCHP-FCC algorithm and other four comparison clustering algorithms. The results are shown in Table 5, Table 6 and Table 7. The significance level is set as ρ < 0.05 . According to the results in Table 5, Table 6 and Table 7, the symbol ‘+’ represents when the difference between the DCHP-FCC algorithm and the compared algorithm is significant, and the symbol ‘−’ represents when the difference between the DCHP-FCC algorithm and the compared algorithm is not significant. We can see from Table 5, Table 6 and Table 7 that the results of the hypothetical test are almost ‘+’. This proves that there is a clear difference between the algorithm in this paper and the comparison algorithms.

5. Conclusions

In this paper, we propose a novel P system (DCHP) with a hybrid structure which combines the advantage of the chain structure and hypergraph topology structure for the consensus fuzzy k-modes clustering. The DCHP system has three subsystems and one consensus system. The subsystems are used to generate three kinds of basic partitions, respectively, and the consensus system is used to realize the consensus clustering process with evolution operations. Then, the DCHP-FCC algorithm is compared with four k-modes clustering algorithms on seven UCI datasets and uses three performance validation indices: ARI, ACC and F_measure. The experimental results show that the DCHP-FCC algorithm can get better clustering results than the other compared algorithms.
There are several ways to continue this study in the future. Firstly, we can consider using different basic clustering algorithms in the consensus clustering or design a new P system structure instead of DCHP system. Secondly, the algorithm which can identify the optimal number of clusters should be investigated for categorical data in the consensus clustering process. In addition, the other clustering method can be used in the consensus clustering process.

Author Contributions

Conceptualization, Z.J. and X.L.; Methodology, Z.J. and X.L.; Software, Z.J.; Validation, Z.J.; Formal Analysis, Z.J.; Writing—Original Draft Preparation, Z.J.; Writing—Review & Editing, Z.J. and X.L.; Supervision, X.L.; Project Administration, X.L.; Funding Acquisition, X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the National Natural Science Foundation of China (Nos. 61876101, 61802234 and 61806114), the Social Science Fund Project of Shandong (16BGLJ06, 11CGLJ22), China Postdoctoral Science Foundation Funded Project (2017M612339, 2018M642695), Natural Science Foundation of the Shandong Provincial (ZR2019QF007), China Postdoctoral Special Funding Project (2019T120607) and Youth Fund for Humanities and Social Sciences, Ministry of Education (19YJCZH244).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Huang, Z. Extensions to the k-means algorithm for clustering large data sets with categorical values. Data Min. Knowl. Discov. 1998, 2, 283–304. [Google Scholar] [CrossRef]
  2. Saha, A.; Das, S. Categorical fuzzy k-modes clustering with automated feature weight learning. Neurocomputing 2015, 166, 422–435. [Google Scholar] [CrossRef]
  3. Liu, C.; Wang, X.; Huang, Y.; Liu, Y.; Li, R.; Li, Y.; Liu, J. A Moving Shape-based Robust Fuzzy K-modes Clustering Algorithm for Electricity Profiles. Electr. Power Syst. Res. 2020, 187, 106425. [Google Scholar] [CrossRef]
  4. Guha, S.; Rastogi, R.; Shim, K. ROCK: A robust clustering algorithm for categorical attributes. Inf. Syst. 2000, 25, 345–366. [Google Scholar] [CrossRef]
  5. Ganti, V.; Gehrke, J.; Ramakrishnan, R. CACTUS-clustering categorical data using summaries. In Proceedings of the 5th ACM SIGKDD Conference, San Diego, CA, USA, 15–18 August 1999; pp. 7–83. [Google Scholar]
  6. Barbara, D.; Li, Y.; Couto, J. COOLCAT: An entropy-based algorithm for categorical clustering. In Proceedings of the 11th ACM Conference on Information and Knowledge Management (CIKM ‘02), Mclean, VA, USA, 4–9 November 2002; pp. 582–589. [Google Scholar]
  7. Andritsos, P.; Tsaparas, P.; Miller, R.J.; Sevcik, K.C. LIMBO: A scalable algorithm to cluster categorical data. In Proceedings of the 9th International Conference on Extending Database Technology (EDBT), Heraklion, Greece, 14–18 March 2004; pp. 123–146. [Google Scholar]
  8. Cao, F.; Liang, J.; Li, D.; Zhao, X. A weighting k-modes algorithm for subspace clustering of categorical data. Neurocomputing 2013, 108, 113–122. [Google Scholar] [CrossRef]
  9. Mukhopadhyay, A.; Maulik, U.; Bandyopadyay, S. Multiobjective genetic algorithm-based fuzzy clustering of categorical attributes. IEEE Trans. Evol. Comput. 2009, 13, 991–1005. [Google Scholar] [CrossRef]
  10. Yang, C.L.; Kuo, R.J.; Chien, C.H.; Quyen, N.T.P. Non-dominated sorting genetic algorithm using fuzzy membership chromosome for categorical data clustering. Appl. Soft Comput. 2015, 30, 113–122. [Google Scholar] [CrossRef]
  11. Qian, Y.; Li, F.; Liang, J.; Liu, B.; Dang, C. Space structure and clustering of categorical data. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 1–13. [Google Scholar] [CrossRef]
  12. Zhu, S.; Xu, L. Many-objective fuzzy centroids clustering algorithm for categorical data. Expert Syst. Appl. 2018, 96, 230–248. [Google Scholar] [CrossRef]
  13. He, Z.; Xu, X.; Deng, S. Squeezer: An efficient algorithm for clustering categorical data. J. Comput. Sci. Technol. 2002, 17, 611–624. [Google Scholar] [CrossRef]
  14. Jia, H.; Cheung, Y.; Liu, J. A new distance metric for unsupervised learning of categorical data. IEEE Trans. Neural Netw. Learn. Syst. 2015, 27, 1065–1079. [Google Scholar] [CrossRef] [PubMed]
  15. Shang, R.; Tian, P.; Wen, A.; Liu, W.; Jiao, L. An intuitionistic fuzzy possibilistic C-means clustering based on genetic algorithm. In Proceedings of the IEEE Congress on Evolutionary Computation, Vancouver, BC, Canada, 24–19 July 2016. [Google Scholar] [CrossRef]
  16. Kuo, R.J.; Nguyen, T.P.Q. Genetic intuitionistic weighted fuzzy k-modes algorithm for categorical data. Neurocomputing 2019, 330, 116–126. [Google Scholar] [CrossRef]
  17. Zhou, Z.; Zhu, S. Kernel-based multiobjective clustering algorithm with automatic attribute weighting. Soft Comput. 2017, 22, 3685–3709. [Google Scholar] [CrossRef]
  18. Naouali, S.; Salem, S.B.; Chtourou, Z. Uncertainty mode selection in categorical clustering using the Rough Set Theory. Expert Syst. Appl. 2020, 159, 113555. [Google Scholar] [CrossRef]
  19. Gregg, M.; Datta, S.; Lorenz, D. Variance estimation in tests of clustered categorical data with informative cluster size. Stat. Methods Med. Res. 2020, 29, 3396–3408. [Google Scholar] [CrossRef]
  20. Yuvaraj, N.; Suresh Ghana Dhas, C. High-performance link-based cluster ensemble approach for categorical data clustering. J. Supercomput. 2020, 76, 4556–4579. [Google Scholar] [CrossRef]
  21. Zheng, Q.; Diao, X.; Cao, J.; Liu, Y.; Li, H.; Yao, J.; Chang, C.; Lv, G. From whole to part: Reference-based representation for clustering categorical data. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 927–937. [Google Scholar] [CrossRef]
  22. Pǎun, G. Computing with Membranes. J. Comput. Syst. Sci. 2000, 61, 108–143. [Google Scholar] [CrossRef] [Green Version]
  23. Pan, L.; Alhazov, A.; Su, H.; Song, B. Local synchronization on asynchronous tissue P systems with Symport/Antiport Rules. IEEE Trans. NanoBioence 2020, 19, 315–320. [Google Scholar] [CrossRef]
  24. Peng, H.; Li, B.; Wang, J.; Song, X.; Mario, J. Spiking neural P systems with inhibitory rules. Knowl. Based Syst. 2019, 188, 105064. [Google Scholar] [CrossRef]
  25. Wu, T.; Pan, L. The computation power of spiking neural P systems with polarizations adopting sequential mode induced by minimum spike number. Neurocomputing 2020, 401, 392–404. [Google Scholar] [CrossRef]
  26. Peng, H.; Wang, J.; Shi, P. A novel image thresholding method based on membrane computing and fuzzy entropy. J. Intell. Fuzzy Syst. Appl. Eng. Technol. 2013, 24, 229–237. [Google Scholar] [CrossRef]
  27. Tu, M.; Wang, J.; Peng, H. Fault diagnosis model of power systems based on adaptive fuzzy spiking neural P systems. Chin. J. Electron. 2016, 23, 87–92. [Google Scholar]
  28. Wang, J.; Shi, P.; Peng, H.; Pérez-Jiménez, M.J.; Wang, T. Weighted fuzzy spiking neural P systems. IEEE Trans. Fuzzy Syst. 2013, 21, 209–220. [Google Scholar] [CrossRef]
  29. Song, B.; Zhang, C.; Pan, L. Tissue-like P systems with evolutional symport/antiport rules. Inf. Sci. 2017, 378, 177–193. [Google Scholar] [CrossRef]
  30. Rong, H.; Yi, K.; Zhang, G.; Dong, J.; Huang, Z. Automatic Implementation of Fuzzy Reasoning Spiking Neural P Systems for Diagnosing Faults in Complex Power Systems. Complexity 2019, 2019, 1–16. [Google Scholar] [CrossRef]
  31. Jiang, Z.; Liu, X. Novel coupled DP system for fuzzy C-means clustering and image segmentation. Appl. Intell. 2020, 50, 1–16. [Google Scholar] [CrossRef]
  32. Liu, X.; Wang, L.; Qu, J.; Wang, N. A Complex Chained P System Based on Evolutionary Mechanism for Image Segmentation. Comput. Intell. Neurosci. 2020, 2020, 1–19. [Google Scholar] [CrossRef]
  33. Liu, X.; Xue, A. Communication P systems on simplicial complexes with applications in cluster analysis. Discret. Dyn. Nat. Soc. 2012, 2012. [Google Scholar] [CrossRef]
  34. Luan, J.; Liu, X.Y. Logic Operation in Spiking Neural P System with Chain Structure. In Frontier and Future Development of Information Technology in Medicine and Education; Springer: Dordrecht, The Netherlands, 2013; pp. 11–20. [Google Scholar]
  35. Yan, S.; Wang, Y.; Kong, D.T.; Hu, J.Y.; Qu, J.H.; Liu, X.Y.; Xue, J. Hybrid Chain-Hypergraph P Systems for Multiobjective Ensemble Clustering. IEEE Access 2019, 7, 143511–143523. [Google Scholar] [CrossRef]
  36. Gan, G.; Wu, J.; Yang, Z. A genetic fuzzy k-Modes algorithm for clustering categorical data. Expert Syst. Appl. 2009, 36, 1615–1620. [Google Scholar] [CrossRef]
  37. Zhao, Y.; Zhang, W.; Sun, M.; Liu, X. An Improved Consensus Clustering Algorithm based on Cell-Like P Systems with Multi-Catalysts. IEEE Access 2020, 8, 154502–154517. [Google Scholar] [CrossRef]
  38. Piergiulio, C.; Violeta, L. Graphs and Hypergraphs. In Applications of Hyperstructure Theory; Springer: Boston, MA, USA, 2003. [Google Scholar]
  39. Ha, T.W.; Seo, J.H.; Kim, M.H. Efficient Searching of Subhypergraph Isomorphism in Hypergraph Databases. In Proceedings of the IEEE International Conference on Big Data & Smart Computing, Shanghai, China, 15–18 January 2018; pp. 739–742. [Google Scholar]
  40. Zhou, D.; Huang, J.; Bernhard, S. Learning with Hypergraphs: Clustering, Classification, and Embedding. In Advances in Neural Information Processing Systems 19, Proceedings of the Twentieth Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 4–7 December 2006; MIT Press: Cambridge, MA, USA, 2006; Volume 19, pp. 1601–1608. [Google Scholar]
  41. Wang, X.; Liu, J.; Cheng, Y.; Liu, A.; Chen, E. Dual Hypergraph Regularized PCA for Biclustering of Tumor Gene Expression Data. IEEE Trans. Knowl. Data Eng. 2018, 31, 2292–2303. [Google Scholar] [CrossRef]
  42. Wu, J.; Liu, H.; Xiong, H.; Cao, J.; Chen, J. K-means-based consensus clustering: A unified view. IEEE Trans. Knowl. Data Eng. 2015, 27, 155–169. [Google Scholar] [CrossRef]
  43. Dua, D.; Graff, C. UCI Machine Learning Repository. Available online: http://archive.ics.uci.edu/ml (accessed on 21 October 2020).
  44. Shang, R.; Zhang, W.; Li, F.; Jiang, L.; Stolkin, R. Multi-objective artificial immune algorithm for fuzzy clustering based on multiple kernels. Swarm Evol. Comput. 2019, 50, 100485. [Google Scholar] [CrossRef]
Figure 1. Consensus clustering framework.
Figure 1. Consensus clustering framework.
Processes 08 01326 g001
Figure 2. An example of the hypergraph which has vertices v = { v 1 , v 2 , v 3 , v 4 , v 5 , v 6 , v 7 } , and hyper-edges e = { e 1 , e 2 , e 3 , e 4 } = { { v 1 , v 2 , v 3 } , { v 2 , v 3 } , { v 3 , v 5 , v 6 } } .
Figure 2. An example of the hypergraph which has vertices v = { v 1 , v 2 , v 3 , v 4 , v 5 , v 6 , v 7 } , and hyper-edges e = { e 1 , e 2 , e 3 , e 4 } = { { v 1 , v 2 , v 3 } , { v 2 , v 3 } , { v 3 , v 5 , v 6 } } .
Processes 08 01326 g002
Figure 3. Flow chart of the proposed DNA-chain-hypergraph P System for consensus clustering (DCHP-FCC) algorithm.
Figure 3. Flow chart of the proposed DNA-chain-hypergraph P System for consensus clustering (DCHP-FCC) algorithm.
Processes 08 01326 g003
Figure 4. Chain membrane structure.
Figure 4. Chain membrane structure.
Processes 08 01326 g004
Figure 5. Hyper-membrane structure of the P system.
Figure 5. Hyper-membrane structure of the P system.
Processes 08 01326 g005
Figure 6. The membrane structure of the DCHP system.
Figure 6. The membrane structure of the DCHP system.
Processes 08 01326 g006
Figure 7. Boxplots of Adjusted Rank Index (ARI), Clustering Accuracy (ACC), and F_measure by DCHP-FCC algorithm on Soybean-small dataset with basic partitions (BPs) equal to 30, 60 and 90, respectively.
Figure 7. Boxplots of Adjusted Rank Index (ARI), Clustering Accuracy (ACC), and F_measure by DCHP-FCC algorithm on Soybean-small dataset with basic partitions (BPs) equal to 30, 60 and 90, respectively.
Processes 08 01326 g007
Table 1. Benchmark datasets.
Table 1. Benchmark datasets.
Dataset# of Instance# of Attributes# of Classes
Soybean-small47354
Spect Heart267222
Tic-tac-toe95892
Voting435162
Breast cancer28692
Zoo101177
Mushroom8124222
Table 2. The preliminary experiments on the Soybean-small dataset.
Table 2. The preliminary experiments on the Soybean-small dataset.
UCI Datasets DCHP-FCC (BPs = 90)DCHP-FCC (BPs = 60)DCHP-FCC (BPs = 30)
ARIACCFARIACCFARIACCF
Soybean-smallMean0.93380.91490.90520.96760.95670.95050.92280.89500.8853
Std.0.00690.01120.01400.00440.00780.01020.00860.01450.0180
Table 3. Parameter setting of the proposed algorithm and the comparison algorithms.
Table 3. Parameter setting of the proposed algorithm and the comparison algorithms.
AlgorithmParametersDescriptionSetting
FKM K 1 The number of clustersRandom select in [ K a , n ]
WFKM/IWFKM K 2 The number of clustersRandom select in [ K a , n ]
GFKM K 3 The number of clustersRandom select in [ K a , n ]
Pop_sizePopulation size100
Max_iterMaximum number of generations100
PmMutation parameter0.01
Consensus clustering K a The number of clustersReal number of clusters
ε Iteration stopping criteria e 5
PcCrossover parameter0.7
Pm1Mutation parameter0.01
Table 4. The comparison of ARI, ACC and F-measure (F) benchmark for four clustering algorithms on seven datasets (mean ± std., BPs = 60).
Table 4. The comparison of ARI, ACC and F-measure (F) benchmark for four clustering algorithms on seven datasets (mean ± std., BPs = 60).
UCI Datasets DCHP-FCCFKMWFKMGFKMIWFKM
Soybean-smallARI0.97 ± 0.0040.86 ± 0.0060.88 ± 0.0100.89 ± 0.0090.87 ± 0.006
ACC0.96 ± 0.0080.83 ± 0.0090.86 ± 0.0150.86 ± 0.0130.85 ± 0.010
F0.95 ± 0.0100.80 ± 0.0120.84 ± 0.0210.84 ± 0.0170.83 ± 0.015
Spect HeartARI0.75 ± 0.0020.50 ± 4 × 10−50.50 ± 2 × 10−50.51 ± 0.0010.50 ± 4 × 10−5
ACC0.79 ± 5 × 10−320.79 ± 5 × 10−320.79 ± 5 × 10−320.79 ± 5 × 10−320.79 ± 5 × 10−32
F0.74 ± 0.0020.63 ± 0.0000.63 ± 0.0000.63 ± 0.0030.63 ± 0.000
Tic-tac-toeARI0.52 ± 0.0000.51 ± 0.0000.50 ± 3 × 10−50.52 ± 0.0010.51 ± 5 × 10−5
ACC0.65 ± 2 × 10−310.65 ± 4 × 10−50.65 ± 2 × 10−310.66 ± 0.0000.65 ± 2 × 10−31
F0.59 ± 0.0010.57 ± 0.0010.55 ± 0.0010.60 ± 0.0020.56 ± 0.001
VotingARI0.78 ± 0.0000.75 ± 0.0000.75 ± 3 × 10−60.75 ± 0.0000.75 ± 2 × 10−6
ACC0.88 ± 0.0000.86 ± 0.0000.85 ± 1 × 10−60.86 ± 0.0000.85 ± 1 × 10−6
F0.88 ± 0.0000.86 ± 0.0000.85 ± 1 × 10−60.86 ± 0.0000.85 ± 1 × 10−6
Breast cancerARI0.51 ± 1 × 10−40.50 ± 6 × 10−60.50 ± 3 × 10−50.51 ± 7 × 10−40.50 ± 6 × 10−6
ACC0.70 ± 5 × 10−320.70 ± 5 × 10−320.70 ± 5 × 10−320.71 ± 5 × 10−50.70 ± 5 × 10−32
F0.58 ± 8 × 10−40.56 ± 3 × 10−40.55 ± 7 × 10−40.59 ± 0.0020.54 ± 4 × 10−4
ZooARI0.90 ± 0.0000.87 ± 0.0010.87 ± 0.0020.89 ± 0.0020.86 ± 0.001
ACC0.87 ± 0.0000.84 ± 0.0020.83 ± 0.0030.84 ± 0.0020.83 ± 0.003
F0.78 ± 0.0040.74 ± 0.0050.73 ± 0.0080.77 ± 0.0040.72 ± 0.006
MushroomARI0.81 ± 1 × 10−310.67 ± 0.0140.70 ± 0.0140.68 ± 0.0160.67 ± 0.018
ACC0.89 ± 2 × 10−310.76 ± 0.0170.80 ± 0.0140.74 ± 0.0150.76 ± 0.017
F0.89 ± 2 × 10−310.76 ± 0.0150.79 ± 0.0140.77 ± 0.0160.76 ± 0.019
Table 5. The p-value produced by the T-test in terms of ARI.
Table 5. The p-value produced by the T-test in terms of ARI.
UCI DatasetsDCHP-FCC vs.
FKMWFKMGFKMIWFKM
Soybean-small0.9 × 10−8(+)5.0 × 10−5(+)3.0 × 10−4(+)6.6 × 10−6(+)
Spect Heart5.7 × 10−27(+)1.0 × 10−23(+)5.1 × 10−22(+)1.1 × 10−23(+)
Tic-tac-toe7.0 × 10−14(+)1.8 × 10−14(+)3.5 × 10−10(+)8.5 × 10−5(+)
Voting6.1 × 10−9(+)1.8 × 10−12(+)1.5 × 10−6(+)6.9 × 10−13(+)
Breast cancer6.6 × 10−4(+)0.0077(+)0.3373(−)3.1 × 104(+)
Zoo0.0092(+)0.0033(+)0.3104(−)0.0020(+)
Mushroom3.4 × 10−7(+)2.6 × 10−5(+)5.1 × 10−6(+)4.9 × 106(+)
Table 6. The p-value produced by the T-test in terms of ACC.
Table 6. The p-value produced by the T-test in terms of ACC.
UCI DatasetsDCHP-FCC vs.
FKMWFKMGFKMIWFKM
Soybean-small2.9 × 10−7(+)1.1 × 10−4(+)8.4 × 10−4(+)2.2 × 10−5(+)
Spect Heart----
Tic-tac-toe----
Voting7.2 × 10−9(+)1.2 × 10−12(+)1.8 × 10−6(+)4.3 × 10−13(+)
Breast cancer----
Zoo5.5 × 10−4(+)0.0027(+)0.0060(+)4.8 × 105(+)
Mushroom0.0414(+)0.5516(−)0.1215(−)7.9 × 106(+)
Table 7. The p-value produced by the T-test in terms of F_measure.
Table 7. The p-value produced by the T-test in terms of F_measure.
UCI DatasetsDCHP-FCC vs.
FKMWFKMGFKMIWFKM
Soybean-small4.3 × 10−7(+)2.0 × 10−4(+)8.7 × 10−4(+)1.1 × 10−4(+)
Spect Heart1.9 × 10−14(+)3.7 × 10−14(+)1.2 × 10−9(+)8.1 × 10−13(+)
Tic-tac-toe0.0540(−)6.0 × 10−6(+)0.5474(−)1.1 × 10−4(+)
Voting6.8 × 10−9(+)1.1 × 10−12(+)1.6 × 10−6(+)3.9 × 10−13(+)
Breast cancer9.7 × 10−4(+)1.9 × 10−4(+)0.3499(−)1.7 × 106(+)
Zoo0.0374(+)0.0280(+)0.3151(−)0.0023(+)
Mushroom2.3 × 10−6(+)8.8 × 10−5(+)1.0 × 10−5(+)6.6 × 106(+)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jiang, Z.; Liu, X. A Novel Consensus Fuzzy K-Modes Clustering Using Coupling DNA-Chain-Hypergraph P System for Categorical Data. Processes 2020, 8, 1326. https://doi.org/10.3390/pr8101326

AMA Style

Jiang Z, Liu X. A Novel Consensus Fuzzy K-Modes Clustering Using Coupling DNA-Chain-Hypergraph P System for Categorical Data. Processes. 2020; 8(10):1326. https://doi.org/10.3390/pr8101326

Chicago/Turabian Style

Jiang, Zhenni, and Xiyu Liu. 2020. "A Novel Consensus Fuzzy K-Modes Clustering Using Coupling DNA-Chain-Hypergraph P System for Categorical Data" Processes 8, no. 10: 1326. https://doi.org/10.3390/pr8101326

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop