Next Article in Journal
Quantum Randomness is Chimeric
Previous Article in Journal
Arbitrary-Order Finite-Time Corrections for the Kramers–Moyal Operator
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Unified Formulation of k-Means, Fuzzy c-Means and Gaussian Mixture Model by the Kolmogorov–Nagumo Average

1
Department of Computer and Information Science, Seikei University, 3-3-1 Kichijoji-Kitamachi, Musashino-shi, Tokyo 180-8633, Japan
2
The Institute of Statistical Mathematics, 10-3 Midori-cho, Tachikawa, Tokyo 190-8562, Japan
*
Author to whom correspondence should be addressed.
Entropy 2021, 23(5), 518; https://doi.org/10.3390/e23050518
Submission received: 22 March 2021 / Revised: 17 April 2021 / Accepted: 19 April 2021 / Published: 24 April 2021

Abstract

:
Clustering is a major unsupervised learning algorithm and is widely applied in data mining and statistical data analyses. Typical examples include k-means, fuzzy c-means, and Gaussian mixture models, which are categorized into hard, soft, and model-based clusterings, respectively. We propose a new clustering, called Pareto clustering, based on the Kolmogorov–Nagumo average, which is defined by a survival function of the Pareto distribution. The proposed algorithm incorporates all the aforementioned clusterings plus maximum-entropy clustering. We introduce a probabilistic framework for the proposed method, in which the underlying distribution to give consistency is discussed. We build the minorize-maximization algorithm to estimate the parameters in Pareto clustering. We compare the performance with existing methods in simulation studies and in benchmark dataset analyses to demonstrate its highly practical utilities.

1. Introduction

In data analysis or data mining, there are two fundamental types of methodologies: clustering and classification [1]. Clustering, which is categorized as an exploratory paradigm, detects the underlying structure behind the data and grasps the rough image before proceeding to more intensive and comprehensive data analysis [2,3]. On the other hand, classification predicts unknown class labels of test data based on models constructed by training data with known class labels. The former is called supervised learning, while the latter is called unsupervised learning in pattern recognition [4].
Clustering algorithms fall roughly into three categories: hierarchical, partitioning, and mixture model-based algorithms [5]. In hierarchical clustering, each observation is considered as one cluster in the initial setting. Then clusters are merged recursively based on a similarity matrix defined beforehand. The resultant clusters are expressed as a dendrogram. The partition algorithm starts with a fixed number of clusters and searches for the cluster centers to minimize an objective function such as the squared distances between the centers and observations. It finds the centers simultaneously. The model-based algorithm assumes a mixture of probability distributions, which generates the observations and assigns the distributions to one of the mixture components. A Gaussian-mixture distribution-based approach is widely used in this context.
In this paper, we propose a new clustering, called Pareto clustering in the framework of quasilinear modeling [6,7,8]. It combines the cluster components by the Kolmogorov–Nagumo average [9] in a flexible way. We consider a generalized energy function as an objective function to estimate cluster parameters, which is an extension of the energy function proposed by [10]. The objective function consists of a survival function of the Pareto distribution, which is widely used in extreme value theory [11]. We investigate the consistency of the parameters, resulting in the underlying probability distribution of the generalized energy function. We find that k-means [12,13] and fuzzy c-means [14] have the underlying probability distributions with singular points at the cluster centers. This fact shows a clear difference from the model-based clustering such as a Gaussian-mixture modeling. Moreover, we show that the quasilinear modeling based on the Kolmogorov–Nagumo average connects k-means, fuzzy c-means, and a Gaussian-mixture modeling using the hyperparameters of the generalized energy function. See [15,16] for the discussion of the relation between k-means and fuzzy c-means.
The paper is organized as follows. In Section 2, we introduce the generalized energy function as the objective function of the Pareto clustering and discuss the consistency of the parameters. Moreover, we show that k-means, fuzzy c-means, and a Gaussian-mixture are all derived from the generalized energy function as special cases. This fact leads to the fact that the parameters can be estimated in a unified manner by the the minorize-maximization (MM) algorithm [17], where the monotone decrease of generalized energy function is guaranteed. In Section 3, we demonstrate the performance of the Pareto clustering based on simulation studies and benchmark datasets and show its practical utilities. We summarize the results of the Pareto clustering and discuss the extensions and applications in various scientific fields.

2. Materials and Methods

2.1. Generalized Energy Function

Let T be a non-negative random variable with a probability density function f ( t ) . The survival function of T is defined as
S ( t ) = P ( T > t ) , t 0 .
Then for d-dimensional random variables x 1 , , x n , we define a generalized energy function to be minimized with respect to a parameter μ for clustering
L S ( μ ) = 1 τ i = 1 n S 1 1 K k = 1 K S ( τ x i μ k 2 ) ,
where μ = ( μ 1 , , μ K ) is a set of centers and τ > 0 is the shape parameter. If we take S ( t ) = exp ( t ) , the function corresponds to the energy function proposed by [10], where τ can be interpreted as the temperature in physics. The formulation in (2) is called the Kolmogorov–Nagumo average [9,18] and is widely applied to bioinformatics, ecology, fisheries, etc. [6,8,19].
In Equation (2), we express an average of probabilities that x i belongs to the kth cluster over all K clusters using 1 / K k = 1 K S ( x i μ k 2 ) , where x i μ k 2 is the energy of x i associated with μ k . Hence we view S 1 1 / K k = 1 K S ( x i μ k 2 ) as the Kolmogorov–Nagumo average of the energy of x i with the probabilistic meanings. In effect, we take summation of the Kolmogorov-Nagumo average over the observations { x 1 , . . . , x n } .
Remark 1.
The generalized energy function (2) has a relation with the Archimedean copula defined by
1 S k = 1 K S 1 ( 1 u k )
for { u k } k = 1 K in ( 0 , 1 ) , cf. [20] for an introductory discussion. In principle, the generalized energy function is a function from a vector of K cluster energy functions to a integrated energy function. The Archimedean copula is that of K marginal cumulative distribution functions to the joint cumulative distribution function. In this way, the generalized energy function expresses an interactive relation for cluster energy functions analogous with the Archimedean copula expressing the correlation among variables.
We consider an estimator of the generalized energy function as
μ ^ = argmin μ L S ( μ ) .
If we assume that x i ( i = 1 , , n ) is distributed according to a probability density function p ( x , μ * ) , the expected generalized energy function is given by
L S ( μ ) = 1 τ S 1 1 K k = 1 K S ( τ x μ k 2 ) p ( x , μ * ) d x .
Here we define a function for a set of cluster centers as
E μ ( x ) = 1 K k = 1 K S ( τ x μ k 2 ) .
Thus, we note that
E μ ( x ) d x = v d E ( T d 2 ) ,
where v d is a volume constant 2 π d / 2 / { τ 1 / 2 d Γ ( d / 2 ) } because E ( T ) = 0 S ( t ) d t (Appendix A). This property is a key idea in the following discussion.
Lemma 1.
Assume that the survival function S ( t ) in (1) is convex in t. We define a function G of ( μ * , μ ) as
G ( μ * , μ ) = S 1 ( E μ ( x ) ) f ( S 1 ( E μ * ( x ) ) ) d x ,
Then for any μ and μ *
G ( μ * , μ ) G ( μ * , μ * )
with equality if and only if μ = μ * .
Proof. 
We observe that the function S 1 ( t ) is a decreasing function of t given as
S 1 ( t ) t = 1 f ( S 1 ( t ) )
because S ( S 1 ( t ) ) = t and ( / t ) S ( t ) = f ( t ) . Similarly,
2 S 1 ( t ) t 2 = f ( S 1 ( t ) ) { f ( S 1 ( t ) ) } 3 ,
which is positive for all t 0 because ( 2 / t 2 ) S ( t ) = f ( t ) > 0 from the convexity assumption for S ( t ) . Therefore, S 1 ( t ) is also convex in t ( 0 , 1 ) . This leads to
G ( μ * , μ ) G ( μ * , μ * ) = { S 1 ( E μ ( x ) ) S 1 ( E μ * ) } f ( S 1 ( E μ * ( x ) ) ) d x
{ E μ ( x ) E μ * ( x ) } S 1 ( E μ * ( x ) ) t f ( S 1 ( E μ * ( x ) ) d x
= { E μ ( x ) E μ * ( x ) } d x
= 0 .
Here Equality in (13) holds if and only if μ = μ * from the convexity for S 1 . The Equality (14) is shown by
f ( S 1 ( t ) ) S 1 ( t ) t = 1
for any t 0 as seen in (10). Equality (15) holds due to (7).    □
Theorem 1.
If the p ( x , μ * ) has a form such as
p ( x , μ * ) = Z ( μ * ) f ( S 1 ( E μ * ( x ) ) ) ,
where Z ( μ * ) > 0 is a normalizing constant. Then we have
L S ( μ ) L S ( μ * ) .
Proof. 
Note that
L S ( μ ) L S ( μ * ) = Z ( μ * ) τ { G ( μ * , μ ) G ( μ * , μ * ) } ,
which concludes (18) from Lemma 1.    □
We note that μ ^ is asymptotically consistent to true parameter μ * if the probability density function has the form in (17).

2.1.1. Pareto Distribution

Let us consider a generalized Pareto distribution, where the survival function and its inverse function are defined by
S ( t ) = ( 1 + β t ) 1 β and S 1 ( t ) = t β 1 β ,
where β > 0 denotes the shape parameter. Then the generalized energy function is
L τ , β ( μ ) = 1 τ β i = 1 n k = 1 K 1 K { 1 + τ β x i μ k 2 } 1 β β 1 .
If we consider β 0 , then
lim β 0 L τ , β ( μ ) = 1 τ i = 1 n log k = 1 K 1 K exp τ x i μ k 2 ,
which is reduced to the energy function proposed by [10]. Hence, we can understand that Rose’s clustering (maximum-entropy clustering) is generated by a survival distribution function of an exponential distribution. Then we have
lim τ L τ , β ( μ ) = lim τ 1 τ β i = 1 n k = 1 K 1 K { 1 + τ β x i μ k 2 } 1 β β
= i = 1 n k = 1 K 1 K { x i μ k 2 } 1 β β
The gradient with respect to μ k is given by
2 β i = 1 n 1 K { x i μ k 2 } 1 β = 1 K π { x i μ 2 } 1 β 1 + β ( x i μ k ) ,
which exactly leads to the estimation equations of fuzzy c-means if we take β = m 1 [14]. Furthermore, we have
lim τ , β 0 L τ , β ( μ ) = i = 1 n min 1 k K x i μ k 2 ,
which is the loss function of k-means. The corresponding survival function is lim β 0 ( 1 + β t ) 1 β = exp ( t ) . Note that the loss function is directly derived from (2) as
lim τ 1 τ i = 1 n S 1 1 K k = 1 K S ( τ x i μ k 2 ) = i = 1 n min 1 k K x i μ k 2 .
In addition, we have
lim τ 0 1 τ i = 1 n S 1 1 K k = 1 K S ( τ x i μ k 2 ) = i = 1 n 1 K k = 1 K x i μ k 2
because S ( 0 ) = 1 .

2.1.2. Fréchet Distribution

Next, we consider Fréchet distribution with the survival function defined as
S ( t ) = 1 exp ( t γ ) .
where γ < 0 is the shape parameter. The generalized energy function is given by
L γ , τ ( μ ) = i = 1 n 1 τ γ log 1 K k = 1 K exp ( τ γ x i μ k 2 γ ) 1 γ .
We find that
lim τ 0 L γ , τ ( μ ) = i = 1 n lim τ 0 1 τ γ log 1 K k = 1 K exp ( τ γ x i μ k 2 γ ) 1 γ = i = 1 n min 1 k K x i μ k 2 γ 1 γ = i = 1 n min 1 k K x i μ k 2 .
Hence, this energy function is reduced to that of the K-means algorithm as shown in the Pareto distribution case. The estimating equation is given by
μ k L γ , τ ( μ ) = i = 1 n ω k ( x i , τ , γ ) ( μ k x i ) = 0 ,
where
ω k ( x i , τ , γ ) = 1 τ γ log 1 K = 1 K exp ( τ γ x i μ 2 γ ) 1 γ 1 × exp ( τ γ x i μ k 2 γ ) = 1 K exp ( τ γ x i μ 2 γ ) x i μ k 2 γ 2 .
When we assume the unbiasedness for the estimating function in (30), that is
E { ω k ( X , τ , γ ) ( μ k X ) } = 0 ,
the underlying distribution has a density function proportional to
1 τ γ log M ( x , μ ) γ 1 γ M ( x , μ ) .
where
M ( x , μ ) = 1 K = 1 K exp ( τ γ x μ 2 γ ) .
We confirm that
ω k ( x i , τ , γ ) = 1 if   x i μ k 2 = min 1 K x i μ 2 0 otherwise
as τ goes to 0. Then we consider the limit of τ to , which provides
lim τ L γ , τ ( μ ) = i = 1 n lim τ 1 τ γ log 1 K k = 1 K exp ( τ γ x i μ k 2 γ ) 1 γ = i = 1 n 1 K k = 1 K x i μ k 2 γ 1 γ
which is equal to (22). This also leads to Fuzzy c-means if we take as γ = 1 / ( 1 m ) [14].

2.2. Estimation of Variances and Mixing Proportions in Clusters

In stead of the Euclidean distance x i μ k 2 , we consider x i μ k k 1 2 = ( x i μ k ) k 1 ( x i μ k ) to incorporate the variance structure around μ k . Bezdek et al. [14] considered a common variance structure k = i = 1 n ( x i x ¯ ) ( x i x ¯ ) , where x ¯ = 1 / n i = 1 n x i for k = 1 , , K . On the other hand, we estimate distinct k for each μ k .
For this purpose, we modify the generalized energy function in (2) to allow for a variances 1 , , K and mixing proportions π 1 , , π K ( k = 1 K π k = 1 and π k 0 for k = 1 , , K ) as
L S ( θ ) = 1 τ i = 1 n S 1 k = 1 K π k | k | 1 2 S ( τ x i μ k k 1 2 ) ,
where θ = ( μ k , k , π k ) k = 1 K . We assume that S ( t ) is convex so that the domain of S 1 ( t ) can be extended from [ 0 , 1 ] to [ 0 , ) to allow for | k | 1 2 . The estimator of this modified generalized energy function is given as
θ ^ = argmin θ L S ( θ ) ,
The expected generalized energy function is given by
L S ( θ ) = 1 τ S 1 k = 1 K π k | k | 1 2 S ( τ x μ k k 1 2 ) p ( x , θ * ) d x ,
where p ( x , θ * ) is the underlying probability density function.
For a cumulative distribution function F ( t ) = 1 S ( t ) we have L S ( θ ) = L F ( θ ) if and only if k = 1 K π k | k | 1 / 2 = 1 . On the other hand, it always holds that L S ( μ ) = L F ( μ ) for the original generalized energy function in (2).
Similarly to (6), we define
E θ ( x ) = k = 1 K π k | k | 1 2 S ( τ x μ k k 1 2 ) ,
and we notice that E θ ( x ) d x is also independent of μ as
E θ ( x ) d x = v d E ( T d 2 ) .
Lemma 2.
Assume that the survival function S ( t ) in (1) is convex in t. We define a function G of ( μ * , μ ) as
G ( θ * , θ ) = S 1 ( E θ ( x ) ) f ( S 1 ( E θ * ( x ) ) ) d x .
Then for any θ and θ *
G ( θ * , θ ) G ( θ * , θ * )
with equality if and only if θ = θ * .
Proof. 
It is obvious from Lemma 1 and the fact that E θ ( x ) d x is independent of μ .    □
From Lemma 2, we can easily show the following theorem regarding L S ( θ ) .
Theorem 2.
If the p ( x , θ * ) has a form such as
p ( x , θ * ) = Z ( θ * ) f ( S 1 ( E θ * ( x ) ) ) ,
where Z ( θ * ) > 0 is a normalizing constant. Then we have
L S ( θ ) L S ( θ * ) .
For the Pareto distribution, we have from (37)
L τ , β ( θ ) = 1 τ β i = 1 n k = 1 K π k | k | 1 2 { 1 + τ β x i μ k k 1 2 } 1 β β 1
= 1 τ i = 1 n ϕ k = 1 K π k w ( x i , μ k , k ) ,
where
w ( x i , μ k , k ) = | k | 1 2 { 1 + τ β x i μ k k 1 2 } 1 β
ϕ ( t ) = t β 1 β .
From (44), the underlying probability density function is
p τ , β ( θ * ) = Z τ , β ( θ * ) k = 1 K π k * w ( x i , μ k * , k * ) 1 + β ,
where Z τ , β ( θ * ) is a normalizing constant. When β 0 , we have
lim β 0 L τ , β ( θ ) = 1 τ i = 1 n log k = 1 K π k | k | 1 2 exp ( τ x i μ k k 1 2 ) ,
which is the negative log likelihood function of the normal mixture distributions apart from a constant term ( 2 π ) d / 2 when τ = 1 / 2 .
Similarly, we have the estimation equation of fuzzy c-means allowing for the Mahalanobis distance when τ . Moreover, we have k-means with the use of the Mahalanobis distance when τ and β 0 . For the other extreme cases, we observe that both lim β L τ , β ( θ ) and lim τ 0 L τ , β ( θ ) diverge or converge to 0 depending on the values of π k and k ( k = 1 , , K ). Hence we choose large values for τ and small values for β in the subsequent data analysis.

2.3. Estimating Algorithm

The direct optimization of L τ , β ( θ ) in (46) is difficult due to the mixture structure. Thus, we employ the idea of expectation and maximization (EM) algorithm [21] and the minorize-maximization (MM) algorithm [17] similar to [19]. Our proposed clustering method (Pareto clustering) is as follows in Algorithm 1.
Algorithm 1: Pareto clustering
1. Set initial values ( μ k ( 0 ) , k ( 0 ) , π k ( 0 ) ) for k = 1 , , K .
2. Repeat the following steps for t = 0 , , T 1 and k = 1 , , K until convergence.
3.
q k ( t ) ( x i ) = π k ( t ) w ( x i , μ k ( t ) , k ( t ) ) = 1 K π k ( t ) w ( x i , μ k ( t ) , k ( t ) )
μ k ( t + 1 ) = i = 1 n q k ( t ) ( x i ) 1 + β x i i = 1 n q k ( t ) ( x i ) 1 + β
k ( t + 1 ) = τ ( 2 d β ) i = 1 n q k ( t ) ( x i ) 1 + β ( x i μ k ( t + 1 ) ) ( x i μ k ( t + 1 ) ) i = 1 n q k ( t ) ( x i ) 1 + β
π k ( t + 1 ) = i = 1 n q k ( t ) ( x i ) 1 + β w ( x i , μ k ( t + 1 ) , k ( t + 1 ) ) β 1 1 + β = 1 K i = 1 n q ( t ) ( x i ) 1 + β w ( x i , μ ( t + 1 ) , ( t + 1 ) ) β 1 1 + β
4. Output ( μ ^ k , ^ k , π ^ k ) = ( μ k ( T ) , k ( T ) , π k ( T ) ) for k = 1 , , K .
The initial values ( μ k ( 0 ) , k ( 0 ) , π k ( 0 ) ) are determined by the hierarchical clustering in a similar way to the algorithm by [22]. The derivation of the estimating algorithm is as follows. First, we have
L τ , β ( θ ) = 1 τ i = 1 n ϕ k = 1 K π k w ( x i , μ k , k )
= 1 τ i = 1 n ϕ k = 1 K q k ( x i ) π k w ( x i , μ k , k ) q k ( x i )
1 τ i = 1 n k = 1 K q k ( x i ) ϕ π k w ( x i , μ k , k ) q k ( x i )
where q k ( x i ) is a positive weight such as k = 1 K q k ( x i ) = 1 and ϕ ( t ) is the convex function defined in (49). The equality holds if and only if
π 1 w ( x i , μ 1 , 1 ) q 1 ( x i ) = = π K w ( x i , μ K , K ) q K ( x i ) ,
which is equivalent to
q k ( x i ) = π k w ( x i , μ k , k ) k = 1 K π k w ( x i , μ k , k ) , ( k = 1 , , K ) .
Based on q k ( t ) ( x i ) in (52), we define
Q ( θ | θ ( t ) ) = 1 τ i = 1 n k = 1 K q k ( t ) ( x i ) π k w ( x i , μ k , k ) q k ( t ) ( x i ) β 1 / β
= 1 τ β i = 1 n k = 1 K { q k ( t ) ( x i ) } 1 + β π k β | k | β 2 { 1 + τ β ( x i μ k ) k 1 ( x i μ k ) } q k ( t ) ( x i )
= 1 τ β i = 1 n k = 1 K { q k ( t ) ( x i ) } 1 + β π k β { | V k 1 | β d β 2 + τ β ( x i μ k ) V k 1 ( x i μ k ) } q k ( t ) ( x i ) ,
where V k 1 = | k | β / 2 k 1 . Then we have
μ k Q ( θ | θ ( t ) ) = 2 π k β V k 1 i = 1 n { q k ( t ) ( x i ) } 1 + β ( x i μ k )
= 0
which means that
μ k ( t + 1 ) = i = 1 n { q k ( t ) ( x i ) } 1 + β x i i = 1 n { q k ( t ) ( x i ) } 1 + β .
Similarly, we have
V k 1 Q ( θ | θ ( t ) ) | μ k = μ k ( t + 1 ) = 1 τ β i = 1 n { q k ( t ) ( x i ) } 1 + β π k β β d β 2 | V k 1 | β d β 2 V k + τ β ( x i μ k ( t + 1 ) ) ( x i μ k ( t + 1 ) )
= 1 τ β i = 1 n { q k ( t ) ( x i ) } 1 + β π k β β d β 2 k + τ β ( x i μ k ( t + 1 ) ) ( x i μ k ( t + 1 ) )
= 0 ,
which means that
k ( t + 1 ) = τ ( 2 d β ) i = 1 n { q k ( t ) ( x i ) } 1 + β ( x i μ k ( t + 1 ) ) ( x i μ k ( t + 1 ) ) i = 1 n { q k ( t ) ( x i ) } 1 + β .
Next we consider
R ( θ | θ ( t ) ) = Q ( θ | θ ( t ) ) + λ k = 1 K π k 1 ,
where λ is a Lagrange multiplier. Then
π k R ( θ | θ ( t ) ) | μ k = μ k ( t + 1 ) , k = k ( t + 1 ) = 1 τ π k β 1 i = 1 n { q k ( t ) ( x i ) } 1 + β w ( x i , μ k ( t + 1 ) , k ( t + 1 ) ) β + λ
= 0 ,
which means
π k ( t + 1 ) = { i = 1 n { q k ( t ) ( x i ) } 1 + β w ( x i , μ k ( t + 1 ) , k ( t + 1 ) ) β } 1 1 + β = 1 K { i = 1 n { q ( t ) ( x i ) } 1 + β w ( x i , μ ( t + 1 ) , ( t + 1 ) ) β } 1 1 + β .
Remark 2.
The generalized energy function in (46) is monotonically decreasing in the estimating algorithm. That is, we have
L τ , β ( θ ( t + 1 ) ) Q ( θ ( t + 1 ) | θ ( t ) ) Q ( θ ( t ) | θ ( t ) ) = L τ , β ( θ ( t ) ) .
See Appendix B for more details.
Remark 3.
The estimating algorithm of fuzzy c-means by [14] is given as
u i k ( t ) = = 1 K x i μ k ( t ) 2 x i μ ( t ) 2 1 m 1 m ,
μ k ( t ) = i = 1 n u i k ( t ) x i i = 1 n u i k ( t ) ,
where { u i k ( t ) } 1 / m is called the membership function of x i in cluster k at the iteration step t. These are special cases of (52) and (53) with τ , k = I , π k = 1 / K and β = m 1 . Hence we observe that the original algorithm of fuzzy c-means can be interpreted as the EM algorithm.
Remark 4.
In analogy with the membership function of fuzzy c-means by [14], we define q k ( t ) ( x i ) in (52) as a membership function of x i in cluster k at the iteration step t in Pareto clustering. Hence we estimate cluster C k as
C k = { x i | q k ( T ) ( x i ) q ( T ) ( x i ) , = 1 , , K , i = 1 , , n } ,
where k = 1 K C k = { x 1 , , x n } .
Remark 5.
In high-dimensional setting p 1 , we consider the ridge regularization for k ( t + 1 ) as in [23]
k ( t + 1 ) ( α ) = α k ( t + 1 ) + ( 1 α ) σ ^ k 2 I ,
where α = 0.95 and σ k 2 is the scalar variance estimated to be the maximum value of the diagonal elements of k ( t + 1 ) . Moreover, we take β 1 to make k ( t + 1 ) positive definite.

2.4. Evaluation of Clustering Methods

We compare the performances of k-means, fuzzy c-means, Gaussian mixture modeling (Gaussian), partitioning around medoids (PAM), and Pareto clustering. To implement these methods, we use the kmeans function in the stat package [24], the cmeans function in the e1071 package [14], the Mclust in the mclust package [22] and the pam function in the cluster package [25] in the statistical software R, where the default settings are used for each function. In Pareto clustering, τ = 0.5 and β = 1 are used as the default settings. We assume that the number of clusters K is known and compare the performances as in [26].

2.4.1. Metrics

Cluster C k ( k = 1 , , K ) estimated by a clustering method is evaluated by a predefined reference class set D ( = 1 , , L ) such as
Precision ( C k , D ) = | C k D | | C k | Recall ( C k , D ) = | C k D | | D | ,
where Recall ( C k , D ) = Precision ( D , C k ) . Precision ( C k , D ) counts data points in cluster C k belonging to class . Hence max Precision ( C k , D ) represents the purity of cluster C k regarding the classes. By taking the weighted average, we have
Purity = k = 1 K | C k | n max Precision ( C k , D ) ,
where n is the sample size. Recall ( C k , D ) counts data points in a class set D estimated to be in cluster C k . Precision and recall correspond to the positive predictive value and sensitivity, respectively [27].
A metric combining precision and recall is proposed by [28] such as
F-value = k = 1 K | D k | n max F ( D k , C ) ,
where
F ( D k , C ) = 2 × Recall ( D k , C ) × Precision ( D k , C ) Recall ( D k , C ) + Precision ( D k , C ) ,
which is the harmonic mean of Precision ( D k , C ) and Recall ( D k , C ) , and is called the F-measure [29].
The cluster level similarity between the estimated center μ ^ = ( μ ^ 1 , , μ ^ K ) and the reference value (ground truth) of center μ * = ( μ 1 * , , μ K * ) is the centroid index (CI) proposed by [26] as
CI ( μ ^ , μ * ) = max ( CI ( μ ^ , μ * ) , CI ( μ * , μ ^ ) ) ,
where
CI ( μ ^ , μ * ) = k = 1 K orphan ( μ k * ) orphan ( μ k * ) = 1 q k 0 otherwise q = argmin 1 k K μ ^ μ k * 2 .
Here, q indicates the index of the element of the reference center μ * that is the nearest to μ ^ . The function orphan ( μ k * ) indicates whether μ k * is an isolated element (orphan) or not, which is n o t nearest to any elements of μ ^ . Hence CI ( μ ^ , μ * ) indicates the dissimilarity between μ ^ and μ * . Due to the asymmetry of CI ( μ ^ , μ * ) with respect to μ ^ and μ * , we take the maximum of CI ( μ ^ , μ * ) and CI ( μ * , μ ^ ) . Hence CI ( μ ^ , μ * ) measures how many clusters are differently located among μ ^ and μ * .
Another metric to measure the similarity between μ ^ and μ * is defined as the mean squared error (MSE) over the number of clusters K, which is given as
MSE = 1 K k = 1 K | | μ ^ k μ k * | | 2 .
Differently from Purity and F-value, MSE can be calculated based on only estimated and reference centers μ ^ and μ * . This property is useful in a situation where the reference class sets D 1 , , D K are difficult to determine but μ * is easily identified. We use MSE in the simulation studies to evaluate the accuracy of μ ^ and Purity and F-value in the analysis of benchmark datasets.

2.4.2. Simulation Studies

We generate samples according to the density function p τ , β ( θ * ) in (50) using the Metropolis-Hastings algorithm [30,31] as
x Z τ , β ( θ * ) k = 1 K π k * w ( x i , μ k * , k * ) 1 + β ,
where μ 1 * = ( 0 , 0 ) , μ 2 * = ( 5 , 5 ) , μ 3 * = ( 5 , 5 ) , π 1 * = 0.5 , π 2 * = 0.2 , π 3 * = 0.3 and
1 * = 2 0.5 0.5 1 , 2 * = 1 0 0 1 , 3 * = 1 0 0 1 .
Figure 1 illustrates the perspective plots and contour plots for ( τ , β ) = ( 0.5 , 1 ) , ( 0.5 , 0 ) , (10, 1). The shape of p τ , β ( θ * ) varies according to the values of τ and β . The Gaussian mixture distribution corresponds with τ = 0.5 and β = 0 in panel (b). When β = 1 , the variance of each component increases and the contours connect with each other. On the other hand, for a large value of τ = 10 , the distribution shows high peaks around the centers. This indicates that p τ , β ( θ * ) , including fuzzy c-means when τ , has a quite different shape from he Gaussian mixture distribution. Other versions of the shapes are also illustrated in Appendix C. The performance of each method is evaluated by MSE based on 100 simulated samples with sample size n = 3000 .

2.4.3. Benchmark Data Analysis

The performance of our proposed method is evaluated using benchmark datasets prepared by [32]. It includes a variety of datasets with low and high cluster overlap, various sample sizes, low and high dimensionalities and unbalanced cluster sizes. Hence, these datasets are suitable for clarifying the statistical performance of the clustering methods. In this setting, we compare the performance of k-means, fuzzy c-means, Gaussian, PAM, and Pareto clustering as well as the variants of Pareto clustering with several values of ( τ , β ) as explained in Table S1. The characteristics of the benchmark datasets such as the sample sizes, the number of clusters, and dimensionality are summarized in Table S2.

3. Results

Figure 2 illustrates the results of MSE in the simulation studies. Pareto clustering provides the best performance in panel (a), where the samples are generated by the underlying distribution p 0.5 , 1 ( θ * ) of Pareto clustering. The shape of the distribution is similar to Gaussian mixture; however, the variance of each component becomes larger and the contour lines are connected to each other as in panel (a) of Figure 1. On the other hand, in panel (c), the variance of each component becomes smaller and contour lines are completely separated. In the both cases, the performances of the Gaussian mixture are clearly degenerated. In the case of panel (b) in which the data are generated from the Gaussian mixture, the performances are comparable to each other, suggesting that k-means, fuzzy c-means, PAM, and Pareto clustering are robust to the underlying distributions to some extent.
In the benchmark data analysis, metrics of Purity, F-value, and CI are evaluated in Tables S3–S5, where variance k and mixing proportion π k ( k = 1 , , K ) in Pareto clustering are estimated. For the two-dimensional shape datasets such as Flame, Compound, D31, Aggregation, Jain, Pathbased, and Spiral in the upper rows of Table S3, existing methods such as k-means, fuzzy c-means, PAM, and Gaussian mixture outperform our proposed methods. In high-dimensional data with d = 1024 (Dim1024), k-means and a Gaussian mixture do not work well. Other methods achieved the maximum value (1) of Purity. For datasets with a large number of clusters, D31 ( K = 31 ) and A3 ( K = 50 ), PAM performs best. For datasets with large sample size of n 5000 and a moderate number of clusters K = 8 , 15 , our proposed method performs best. As for the effect of τ , it barely affects the performance of our proposed method. On the other hand, β slightly affects the performance, resulting that the intermediates among Gaussian mixture, Pareto clustering, k-means, and fuzzy c-means, such as GP, GPKF 1 , GPKF 10 , and GPKF 100 , show relatively good performances as a whole. We observe similar tendencies regarding the F-value (Table S4).
As for the CI, the values are relatively small for all methods, suggesting that cluster locations are properly estimated. However, some methods do not work for some datasets: Gaussian mixture for A3 (CI = 18), Birch1 (CI = 34) and Birch2 (CI = 49); k-means for D31 (CI = 7), Dim1024 (CI = 4), A3 (CI = 6), Birch1 (CI = 12) and Birch2 (CI = 23); and fuzzy c-means for D31 (CI = 5), A3 (CI = 7), Birch1 (CI = 18) and Birch2 (CI = 25). On the other hand, PAM and our proposed methods show stable results. The results, where k and π k are not estimated and fix k = I and π k = 1 / K in Pareto clustering, are shown in Tables S6–S8.

4. Discussion

We propose a new clustering method based on the generalized energy function derived from the Kolmogorov–Nagumo average. The survival function used in the generalized energy function plays an important role to ensure the minimum consistency of the parameters, which is shown in Lemma 1 using the property of divergence G ( μ * , μ ) . We consider two examples of the survival function based on the Pareto and Fréchet distributions and show a connection among k-means, fuzzy c-means, and Gaussian mixture, leading to new methods that are intermediates among them. For the underlying distribution of our method in (50), we observe that k-means and fuzzy c-means do not have probabilistic interpretations because the corresponding underlying distributions become singular. We also propose an estimating algorithm for cluster locations, variances, and mixing proportions using the MM algorithm.
Simulation studies and benchmark data analysis show that intermediates among k-means, fuzzy c-means, and the Gaussian mixture perform well. This observation suggests that our proposed method has a wide range of applications in which k-means, fuzzy c-means, and the Gaussian mixture are used. For example, simultaneous deep learning and clustering [33] in which a deep neural network and k-means are jointly used, image segmentation using fuzzy c-means in a deep neural network [34], an application of fuzzy c-means in classification problems [35] and a parallel computation for large datasets by fuzzy c-means [36] can be investigated in the framework of the generalized energy function by the Pareto distribution.
As for the tuning parameters τ and β , we consider an approach using the leave-one-out cross validation in the Supplementary Materials in order to improve the clustering performance. The objective function in the leave-one-out cross validation is derived from an anchor loss as in [37] to estimate the optimal values of τ and β properly. The benchmark data analysis suggests that the performance is insensitive to the values of τ but is sensitive to the values of β . Hence, this approach should be useful to determine the optimal value of β .
Banerjee et al. [38] has proposed a clustering method based on Bregman divergences and clarified the relationship between the exponential families and the corresponding Bregman divergences. They separately consider hard and soft clustering; the former corresponds to k-means style clustering and the latter corresponds to mixture model clustering. In our proposed model, the tuning parameters τ and β bridge the gap between them and the performances are investigated by simulation studies and benchmark datasets. The extension of our method by replacing the squared distance x i μ k k 1 2 with Bregman divergence should improve its practical flexibility and utility. When β or γ divergence is used, the clustering method should be robust to contamination in observations as suggested by [39,40].
It is well known that MM algorithm and EM algorithm converge to a local optimum and the resultant clusters are sensitive to initial values [41]. One way to circumvent this difficulty is to prepare several sets of initial values and select the best one among them such as the global k-means algorithm [42]. Another approach is to combine MM algorithm and genetic algorithm (GA) to expand thoroughly the search space for the optimal solution [41,43]. The both approach can be incorporated into the Pareto clustering to make it robust to the initial values and to escape from local optimal solutions.

Supplementary Materials

The following are available at https://www.mdpi.com/article/10.3390/e23050518/s1, A: Notations of methods and characteristics of bench-mark datasets, Figure S1: Summary of clustering methods, Table S2: Sample sizes, the number of clusters and dimensions of benchmark datasets. B: Results of benchmark data with k I in Pareto clustering, Table S3: The result of Purity ( k = I ), Table S4: The result of F values ( k I ), Table S5: The result of Centroid index ( k I ). C: Results of benchmark data with k = I in Pareto clustering, Table S6: The result of Purity ( k = I ), Table S7: The result of F values ( k = I ), Table S8: The result of Centroid index ( k = I ). D: Tuning of parameters τ and β . E: R code of the Pareto clustering.

Author Contributions

Conceptualization, S.E.; methodology, S.E. and O.K.; formal analysis, O.K.; writing—original draft preparation, O.K.; All authors have read and agreed to the published version of the manuscript.

Funding

Financial support was provided by the Japan Society for the Promotion of Science KAKENHI Grant Number JP18K11190 and and JP18H03211.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Benchmark datasets are available at http://cs.uef.fi/sipu/datasets/.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Derivation of the Volume Constant vd

We define
E μ ( x ) = 1 K k = 1 K S ( τ x μ k 2 ) .
Then we have
S ( τ x μ k 2 ) d x = τ 1 2 S ( y y ) d y , ( y = τ 1 2 ( x μ k ) )
= τ 1 2 0 S ( r 2 ) 2 π d 2 Γ ( d 2 ) r d 1 d r , ( from   polar   coordinate   system )
= π d 2 Γ ( d 2 ) τ 1 2 0 S ( t ) t d 2 2 d t , ( t = r 2 )
= π d 2 Γ ( d 2 ) τ 1 2 2 d t d 2 S ( t ) 0 + 2 d 0 t d 2 f ( t ) d t , ( S ( t ) = f ( t ) )
= 2 π d 2 Γ ( d 2 ) τ 1 2 d E [ T d 2 ]
The last equality holds under a condition that lim t t d 2 S ( t ) = 0 . Hence we have
E μ ( x ) d x = 2 π d 2 Γ ( d 2 ) τ 1 2 d E [ T d 2 ] .
When d = 2 , we have
E μ ( x ) d x = π τ 1 2 E [ T ] .

Appendix B. Monotone Decrease of the Generalized Energy Function

The Q function at iteration t in the estimating algorithm is defined as
Q ( θ | θ ( t ) ) = 1 τ β i = 1 n k = 1 K { q k ( t ) ( x i ) } 1 + β π k β | k | β 2 { 1 + τ β ( x i μ k ) k 1 ( x i μ k ) } q k ( t ) ( x i ) ,
where
q k ( t ) ( x i ) = π k ( t ) w ( x i , μ k ( t ) , k ( t ) ) k = 1 K π k ( t ) w ( x i , μ k ( t ) , k ( t ) ) , ( k = 1 , , K ) .
The estimate of μ k is given as
μ k ( t + 1 ) = i = 1 n { q k ( t ) ( x i ) } 1 + β x i i = 1 n { q k ( t ) ( x i ) } 1 + β = i = 1 n v k ( t ) ( x i ) x i , ( k = 1 , , K ) ,
where
v k ( t ) ( x i ) = { q k ( t ) ( x i ) } 1 + β i = 1 n { q k ( t ) ( x i ) } 1 + β .
Here we observe that
i = 1 n v k ( t ) ( x i ) ( x i μ k ) k 1 ( x i μ k )
= i = 1 n v k ( t ) ( x i ) ( x i μ k ( t + 1 ) ) k 1 ( x i μ k ( t + 1 ) ) + i = 1 n v k ( t ) ( x i ) ( μ k ( t + 1 ) μ k ) k 1 ( μ k ( t + 1 ) μ k )
Hence we have
Q ( θ | θ ( t ) ) Q ( θ | θ ( t ) ) | μ k = μ k ( t + 1 )
= 1 τ β π k β | k | β 2 i = 1 n { q k ( t ) ( x i ) } 1 + β { 1 + τ β ( x i μ k ) k 1 ( x i μ k ) }
1 τ β π k β | k | β 2 i = 1 n { q k ( t ) ( x i ) } 1 + β { 1 + τ β ( x i μ k ( t + 1 ) ) k 1 ( x i μ k ( t + 1 ) ) }
= π k β | k | β 2 c q ( t ) i = 1 n v k ( t ) ( x i ) { ( x i μ k ) k 1 ( x i μ k ) ( x i μ k ( t + 1 ) ) k 1 ( x i μ k ( t + 1 ) ) }
= π k β | k | β 2 c q ( t ) i = 1 n v k ( t ) ( x i ) ( μ k ( t + 1 ) μ k ) k 1 ( μ k ( t + 1 ) μ k ) 0
where
c q ( t ) = i = 1 n { q k ( t ) ( x i ) } 1 + β .
The estimate of k is given by
k ( t + 1 ) = τ ( 2 d β ) i = 1 n v k ( t ) ( x i ) ( x i μ k ( t + 1 ) ) ( x i μ k ( t + 1 ) ) ,
where d β < 2 . Hence we have
Q ( θ | θ ( t ) ) | μ k = μ k ( t + 1 ) Q ( θ | θ ( t ) ) | μ k = μ k ( t + 1 ) , k = k ( t + 1 )
= 1 τ β π k β | k | β 2 i = 1 n { q k ( t ) ( x i ) } 1 + β { 1 + τ β ( x i μ k ( t + 1 ) ) k 1 ( x i μ k ( t + 1 ) ) }
1 τ β π k β | k ( t + 1 ) | β 2 i = 1 n { q k ( t ) ( x i ) } 1 + β { 1 + τ β ( x i μ k ( t + 1 ) ) k ( t + 1 ) 1 ( x i μ k ( t + 1 ) ) }
= 1 τ β π k β | k | β 2 c q ( t ) i = 1 n v k ( t ) ( x i ) { 1 + τ β ( x i μ k ( t + 1 ) ) k 1 ( x i μ k ( t + 1 ) ) }
1 τ β π k β | k ( t + 1 ) | β 2 c q ( t ) i = 1 n v k ( t ) ( x i ) { 1 + τ β ( x i μ k ( t + 1 ) ) k ( t + 1 ) 1 ( x i μ k ( t + 1 ) ) }
= 1 τ β π k β | k | β 2 c q ( t ) 1 + τ β trace k 1 i = 1 n v k ( t ) ( x i ) ( x i μ k ( t + 1 ) ) ( x i μ k ( t + 1 ) )
1 τ β π k β | k ( t + 1 ) | β 2 c q ( t ) 1 + τ β trace k ( t + 1 ) 1 i = 1 n v k ( t ) ( x i ) ( x i μ k ( t + 1 ) ) ( x i μ k ( t + 1 ) )
= 1 τ β π k β | k | β 2 c q ( t ) 1 + τ β trace k 1 1 τ ( 2 d β ) k ( t + 1 ) 1 τ β π k β | k ( t + 1 ) | β 2 c q ( t ) 1 + d β 2 d β
Here we notice that
Q ( θ | θ ( t ) ) | μ k = μ k ( t + 1 ) Q ( θ | θ ( t ) ) | μ k = μ k ( t + 1 ) , k = k ( t + 1 ) 0
| k ( t + 1 ) | β 2 2 2 d β | k | β 2 [ 1 + β 2 d β trace ( k 1 k ( t + 1 ) ) ] 1
2 2 d β | k 1 k ( t + 1 ) | β 2 1 + β 2 d β trace ( k 1 k ( t + 1 ) ) ,
where trace { k 1 k ( t + 1 ) } = trace { k 1 / 2 k ( t + 1 ) k 1 / 2 } = λ 1 + + λ d 0 with λ j being the non-negative eigen value of k 1 / 2 k ( t + 1 ) k 1 / 2 . Here we have
| k 1 k ( t + 1 ) | β 2 1 + β 2 d β trace ( k 1 k ( t + 1 ) )
= ( λ 1 λ d ) β 2 1 + β 2 d β ( λ 1 + + λ d )
λ 1 + + λ d p d β 2 1 + β 2 d β ( λ 1 + + λ d ) , ( β > 0 )
= Λ d β 2 p d β 2 1 + β 2 d β Λ ,
where Λ = λ 1 + + λ d 0 . The last term attains the minimum at Λ = d , leading to
2 2 d β | k 1 k ( t + 1 ) | β 2 1 + β 2 d β trace ( k 1 k ( t + 1 ) ) .
As for π k we define
R ( θ | θ ( t ) ) = Q ( θ | θ ( t ) ) + λ k = 1 K π k 1
and we have
R ( θ | θ ( t ) ) π k | μ k = μ k ( t + 1 ) , k = k ( t + 1 ) , π k = π k ( t + 1 ) = 0
2 R ( θ | θ ( t ) ) π k 2 | μ k = μ k ( t + 1 ) , k = k ( t + 1 ) = 1 + β τ i = 1 n { q k ( t ) ( x i ) } 1 + β π k β 2 | k ( t + 1 ) | β 2 × { 1 + τ β ( x i μ k ( t + 1 ) ) k ( t + 1 ) 1 ( x i μ k ( t + 1 ) ) }
0 ,
where
π k ( t + 1 ) = { i = 1 n { q k ( t ) ( x i ) } 1 + β w ( x i , μ k ( t + 1 ) , k ( t + 1 ) ) β } 1 1 + β = 1 K { i = 1 n { q ( t ) ( x i ) } 1 + β w ( x i , μ , ( t + 1 ) ) β } 1 1 + β .
Hence, we have
Q ( θ | θ ( t ) ) | μ k = μ k ( t + 1 ) , k = k ( t + 1 ) Q ( θ | θ ( t ) ) | μ k = μ k ( t + 1 ) , k = k ( t + 1 ) , π k = π k ( t + 1 ) 0 .
Inequations (A18), (A35) and (A40) hold for k = 1 , , K . As a result, we have
Q ( θ | θ ( t ) ) Q ( θ ( t + 1 ) | θ ( t ) ) 0 , θ
Hence we have
L τ , β ( θ ( t + 1 ) ) Q ( θ ( t + 1 ) | θ ( t ) ) Q ( θ ( t ) | θ ( t ) ) = L τ , β ( θ ( t ) ) .

Appendix C. Perspective Plots and Contour Plots for pτ,β (θ *)

Figure A1. Perspective plots (left panels) and contour plots (right panels) for p τ , β ( θ * ) .
Figure A1. Perspective plots (left panels) and contour plots (right panels) for p τ , β ( θ * ) .
Entropy 23 00518 g0a1aEntropy 23 00518 g0a1b

References

  1. Rokach, L.; Maimon, O. Clustering Methods. In Data Mining and Knowledge Discovery Handbook; Maimon, O., Rokach, L., Eds.; Springer: Boston, MA, USA, 2005. [Google Scholar]
  2. Tukey, J.W. We need both exploratory and confirmatory. Am. Stat. 1980, 314, 23–25. [Google Scholar]
  3. Dubes, R.; Jain, A.K. Clustering methodologies in exploratory data analysis. Adv. Comput. 1980, 19, 113–228. [Google Scholar]
  4. Jain, A.K. Data clustering: 50 years beyond K-means. Pattern Recognit. Lett. 2010, 31, 651–666. [Google Scholar] [CrossRef]
  5. Ghosh, S.; Dubey, S. Comparative analysis of k-means and fuzzy c-means algorithms. Int. J. Adv. Comput. Sci. Appl. 2013, 4, 35–39. [Google Scholar] [CrossRef] [Green Version]
  6. Komori, O.; Eguchi, S.; Ikeda, S.; Okamura, H.; Ichinokawa, M.; Nakayama, S. An asymmetric logistic regression model for ecological data. Methods Ecol. Evol. 2016, 7, 249–260. [Google Scholar] [CrossRef]
  7. Komori, O.; Eguchi, S.; Saigusa, Y.; Okamura, H.; Ichinokawa, M. Robust bias correction model for estimation of global trend in marine populations. Ecosphere 2017, 8, 1–9. [Google Scholar] [CrossRef]
  8. Omae, K.; Komori, O.; Eguchi, S. Quasi-linear score for capturing heterogeneous structure in biomarkers. BMC Bioinform. 2017, 18, 308. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Naudts, J. Generalised Thermostatistics; Springer: London, UK, 2011. [Google Scholar]
  10. Rose, K.; Gurewitz, E.; Fox, G.C. Statistical mechanics and phase transitions in clustering. Phys. Rev. Lett. 1990, 65, 945–948. [Google Scholar] [CrossRef]
  11. Beirlant, J.; Goegebeur, Y.; Segers, J.; Teugels, J.L.; Waal, D.D.; Ferro, C. Statistics of Extremes: Theory and Applications; Wiley: Hoboken, NJ, USA, 2004. [Google Scholar]
  12. Cox, D.R. Note on grouping. J. Am. Stat. Assoc. 1957, 52, 543–547. [Google Scholar] [CrossRef]
  13. MacQueen, J. Some methods of classification and analysis of multivariate observations. In Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability; Cam, L.M.L., Neyman, J., Eds.; University of California Press: Berkeley, CA, USA, 1967; pp. 281–297. [Google Scholar]
  14. Bezdek, J.C.; Ehrlich, R.; Full, W. FCM: The fuzzy c-means clustering algorithm. Comput. Geosci. 1984, 10, 191–2003. [Google Scholar] [CrossRef]
  15. Hathaway, R.J.; Bezdek, J.C. Optimization of clustering criteria by reformulation. IEEE Trans. Fuzzy Syst. 1995, 3, 241–245. [Google Scholar] [CrossRef]
  16. Yu, J. General C-means clustering model. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1197–1211. [Google Scholar] [PubMed]
  17. Hunter, D.R.; Lange, K. A tutorial on MM algorithms. Am. Stat. 2004, 58, 30–37. [Google Scholar] [CrossRef]
  18. Eguchi, S.; Komori, O. Path Connectedness on a Space of Probability Density Functions. In Geometric Science of Information: Second International Conference, GSI 2015; Nielsen, F., Barbaresco, F., Eds.; Springer International Publishing: Cham, Switzerland, 2015; p. 615. [Google Scholar]
  19. Komori, O.; Eguchi, S.; Saigusa, Y.; Kusumoto, B.; Kubota, Y. Sampling bias correction in species distribution models by quasi-linear Poisson point process. Ecol. Inform. 2020, 55, 1–11. [Google Scholar] [CrossRef]
  20. Nelsen, R.B. An Introduction to Copulas; Springer: New York, NY, USA, 2006. [Google Scholar]
  21. Dempster, A.P.; Laird, N.M.; Rubin, D.B. Maximum likelihood from incomplete data via the EM algorithm. J. R. Stat. Soc. Ser. 1977, 39, 1–38. [Google Scholar]
  22. Scrucca, L.; Fop, M.; Murphy, T.B.; Raftery, A.E. mclust 5: Clustering, classification and density estimation using Gaussian finite mixture models. R J. 2016, 8, 289–317. [Google Scholar] [CrossRef] [Green Version]
  23. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction, 2nd ed.; Springer: New York, NY, USA, 2009. [Google Scholar]
  24. Hartigan, J.A.; Wong, M.A. A k-means clustering algorithm. J. R. Stat. Soc. Ser. 1979, 28, 100–108. [Google Scholar]
  25. Reynolds, A.P.; Richards, G.; de la Iglesia, B.; Rayward-Smith, V.J. Clustering rules: A comparison of partitioning and hierarchical clustering algorithms. J. Math. Model. Algorithms 2006, 5, 475–504. [Google Scholar] [CrossRef]
  26. Fränti, P.; Rezaei, M.; Zhao, Q. Centroid index: Cluster level similarity measure. Pattern Recognit. 2014, 47, 3034–3045. [Google Scholar] [CrossRef]
  27. Sofaer, H.R.; Hoeting, J.A.; Jarnevich, C.S. The area under the precision-recall curve as a performance metric for rare binary events. Methods Ecol. Evol. 2019, 10, 565–577. [Google Scholar] [CrossRef]
  28. Amigó, E.; Gonzalo, J.; Artiles, J.; Verdejo, F. A comparison of extrinsic clustering evaluation metrics based on formal constraints. Inf. Retr. 2009, 12, 461–486. [Google Scholar] [CrossRef] [Green Version]
  29. Van Rijsbergen, C. Foundation of evaluation. J. Doc. 1974, 30, 365–373. [Google Scholar] [CrossRef]
  30. Hastings, W.K. Monte Carlo sampling methods using Markov chains and their applications. Biometrika 1970, 57, 97–109. [Google Scholar] [CrossRef]
  31. Chib, S.; Greenberg, E. Understanding the Metropolis-Hastings algorithm. Am. Stat. 1995, 49, 327–335. [Google Scholar]
  32. Fränti, P.; Sieranoja, S. K-means properties on six clustering benchmark datasets. Appl. Intell. 2018, 48, 4743–4759. [Google Scholar] [CrossRef]
  33. Yang, B.; Fu, X.; Sidiropoulos, N.D.; Hong, M. Towards K-means-friendly Spaces: Simultaneous Deep Learning and Clustering. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 7–9 August 2017; Precup, D., Teh, Y.W., Eds.; 2017; Volume 70, pp. 3861–3870. [Google Scholar]
  34. Mohsen, H.; El-Dahshan, E.S.A.; El-Horbaty, E.S.M.; Salem, A.B.M. Classification using deep learning neural networks for brain tumors. Future Comput. Inform. J. 2018, 3, 68–71. [Google Scholar] [CrossRef]
  35. Gorsevski, P.V.; Gessler, P.E.; Jankowski, P. Integrating a fuzzy k-means classification and a Bayesian approach for spatial prediction of landslide hazard. J. Geogr. Syst. 2003, 5, 223–251. [Google Scholar] [CrossRef]
  36. Kwok, T.; Smith, K.; Lozano, S.; Taniar, D. Parallel Fuzzy c- Means Clustering for Large Data Sets. In Euro-Par 2002 Parallel Processing; Monien, B., Feldmann, R., Eds.; Springer: Berlin/Heidelberg, Germany, 2002; pp. 365–374. [Google Scholar]
  37. Mollah, M.N.H.; Eguchi, S.; Minami, M. Robust Prewhitening for ICA by Minimizing β-Divergence and Its Application to FastICA. Neural Process. Lett. 2007, 25, 91–110. [Google Scholar] [CrossRef]
  38. Banerjee, A.; Merugu, S.; Dhillon, I.S.; Ghosh, J. Clustering with Bregman Divergences. J. Mach. Learn. Res. 2005, 6, 1705–1749. [Google Scholar]
  39. Fujisawa, H.; Eguchi, S. Robust parameter estimation with a small bias against heavy contamination. J. Multivar. Anal. 2008, 99, 2053–2081. [Google Scholar] [CrossRef] [Green Version]
  40. Notsu, A.; Eguchi, S. Robust clustering method in the presence of scattered observations. Neural Comput. 2016, 28, 1141–1162. [Google Scholar] [CrossRef] [PubMed]
  41. Pernkopf, F.; Bouchaffra, D. Genetic-based EM algorithm for learning Gaussian mixture models. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1344–1348. [Google Scholar] [CrossRef] [PubMed]
  42. Likas, A.; Vlassis, N.; Verbeek, J.J. The global k-means clustering algorithm. Pattern Recognit. 2003, 36, 451–461. [Google Scholar] [CrossRef] [Green Version]
  43. Krishna, K.; Murty, M.N. Genetic K-means algorithm. IEEE Trans. Syst. Man Cybern. Part (Cybern.) 1999, 29, 433–439. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Perspective plots (left panels) and contour plots for p τ , β ( θ * ) with boundaries marked in red for (a) τ = 0.5 , β = 1 , (b) τ = 0.5 , β = 0 and (c) τ = 10 , β = 1 .
Figure 1. Perspective plots (left panels) and contour plots for p τ , β ( θ * ) with boundaries marked in red for (a) τ = 0.5 , β = 1 , (b) τ = 0.5 , β = 0 and (c) τ = 10 , β = 1 .
Entropy 23 00518 g001aEntropy 23 00518 g001b
Figure 2. Mean squared errors (MSE) on the log scale based on 100 random samples for each method. The samples are generated based on p τ , β ( θ * ) with (a) τ = 0.5 , β = 1 , (b) τ = 0.5 , β = 0 and (c) τ = 10 , β = 1 .
Figure 2. Mean squared errors (MSE) on the log scale based on 100 random samples for each method. The samples are generated based on p τ , β ( θ * ) with (a) τ = 0.5 , β = 1 , (b) τ = 0.5 , β = 0 and (c) τ = 10 , β = 1 .
Entropy 23 00518 g002
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Komori, O.; Eguchi, S. A Unified Formulation of k-Means, Fuzzy c-Means and Gaussian Mixture Model by the Kolmogorov–Nagumo Average. Entropy 2021, 23, 518. https://doi.org/10.3390/e23050518

AMA Style

Komori O, Eguchi S. A Unified Formulation of k-Means, Fuzzy c-Means and Gaussian Mixture Model by the Kolmogorov–Nagumo Average. Entropy. 2021; 23(5):518. https://doi.org/10.3390/e23050518

Chicago/Turabian Style

Komori, Osamu, and Shinto Eguchi. 2021. "A Unified Formulation of k-Means, Fuzzy c-Means and Gaussian Mixture Model by the Kolmogorov–Nagumo Average" Entropy 23, no. 5: 518. https://doi.org/10.3390/e23050518

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop