Next Article in Journal
How Do Deep-Learning Framework Versions Affect the Reproducibility of Neural Network Models?
Previous Article in Journal
Automatic Extraction of Medication Information from Cylindrically Distorted Pill Bottle Labels
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Entropic Statistics: Concept, Estimation, and Application in Machine Learning and Knowledge Extraction

Department of Mathematics and Statistics, Mississippi State University, Mississippi State, MS 39762, USA
Mach. Learn. Knowl. Extr. 2022, 4(4), 865-887; https://doi.org/10.3390/make4040044
Submission received: 29 August 2022 / Revised: 19 September 2022 / Accepted: 26 September 2022 / Published: 30 September 2022
(This article belongs to the Section Data)

Abstract

:
The demands for machine learning and knowledge extraction methods have been booming due to the unprecedented surge in data volume and data quality. Nevertheless, challenges arise amid the emerging data complexity as significant chunks of information and knowledge lie within the non-ordinal realm of data. To address the challenges, researchers developed considerable machine learning and knowledge extraction methods regarding various domain-specific challenges. To characterize and extract information from non-ordinal data, all the developed methods pointed to the subject of Information Theory, established following Shannon’s landmark paper in 1948. This article reviews recent developments in entropic statistics, including estimation of Shannon’s entropy and its functionals (such as mutual information and Kullback–Leibler divergence), concepts of entropic basis, generalized Shannon’s entropy (and its functionals), and their estimations and potential applications in machine learning and knowledge extraction. With the knowledge of recent development in entropic statistics, researchers can customize existing machine learning and knowledge extraction methods for better performance or develop new approaches to address emerging domain-specific challenges.

1. Introduction of Entropic Statistics

Entropic statistics is a collection of statistical procedures that characterize information from non-ordinal spaces with Shannon’s entropy and its generalized functionals. Such procedures includes but not limited to statistical methods involving Shannon’s entropy (entropy) and Mutual Information (MI) [1], Kullback–Leibler divergence (KL) [2], entropic basis and diversity index [3,4], and Generalized Shannon’s Entropy (GSE) and Generalized Mutual Information (GMI) [5]. The field of entropic statistics is at the intersection of information theory and statistics. Entropic statistics quantities are also referred as information-theoretic quantities [6,7].
There are two general data types—ordinal and non-ordinal (nominal). Ordinal data are data with an inherent numerical scale. For example, {52 F, 50 F, 49 F, 53 F}—a set of daily high temperatures at Nuuk, Greenland—is ordinal. Ordinal data are generated from random variables (which map outcomes from sample space to the real numbers). For ordinal data, classical concepts, such as moments (mean, variance, covariance, etc.) and characteristics functions, are powerful tools to induce various statistical methods, including but not limited to regression analysis [8] and analysis of variance (ANOVA) [9].
Non-ordinal data are data without an inherent numerical scale. For example, {androgen receptor, clock circadian regulator, epidermal growth factor, Werner syndrome RecQ helicase-like}—a subset of human genes names—is a set of data without inherent numerical scale. Non-ordinal data are generated from random elements (which map outcomes from sample space to alphabet). Due to the absence of inherent numerical scale, the concept of random variable is undefined according to its definition. Therefore, statistical concepts involving ordinal scale (e.g., mean, variance, covariance, and characteristic functions) no longer exist. For example, consider the mentioned data of human genes names; what is the mean or variance of the data? Such questions cannot be answered because the concepts of mean and variance do not exist, while in practice, researchers need to measure the level of dependence in non-ordinal joint space between gene types and genetic phenotype to study the gene’s functionalities. One would use covariance and its generated methods in ordinal data. However, the concept of covariance no longer exists in such non-ordinal space. Furthermore, all well-established statistical methods that require ordinal scale (e.g., regression and ANOVA) cannot be directly applied anymore.
Non-ordinal data have several variant names, such as categorical data, qualitative data, and nominal data. A common situation is a dataset is mixed with ordinal and non-ordinal data. On such a dataset, a common practice is to introduce coded (dummy) variables [10]. However, introducing dummy variables is equivalent to separating the mixed dataset according to the classes in non-ordinal variables to induce multiple purely ordinal subsets and then utilizing ordinal methods (such as regression analysis) case-by-case on the induced subsets. Unfortunately, this approach sometimes could be impractical because of the curse of dimensionality, particularly when there are too many categorical variables or when some categorical variable has too many categories (classes).
With the challenges from non-ordinal data, entropic statistics methods focus on underlying probability distribution instead of associated labels. As a result, all the entropic statistical quantities are location (permutation) invariant. The main strengths of entropic statistics lie within non-ordinal alphabets, or a mixture data space that significant bulk of information lies within the non-ordinal sub-space. For ordinal spaces, although ordinal variables can be binned as categorical variables, the strength of entropic statistics are generally incapable of overcoming the loss of ordinal information during discretization. Therefore, ordinal statistical methods are preferred when they are capable of the needs. In summary, potential scenarios for entropic statistics are:
  • The data lie within non-ordinal space.
  • The data are a mixture of ordinal and non-ordinal spaces, and the non-ordinal space is expected to carry unneglectable bulk of information.
  • The data lie within ordinal space, yet the performance of ordinal statistics methods fails to meet the expectation.
The following notations are used throughout the article. They are listed here for convenience.
  • Let X = { x i ; i = 1 , 2 , } and Y = { y j ; j = 1 , 2 , } be two countable alphabets with cardinalities K 1 and K 2 , respectively.
  • Let the Cartesian product X × Y be with a joint probability distribution p X Y = { p i , j } .
  • Let the two marginal distributions be respectively denoted by p X = { p i , · } and p Y = { p · , j } where p i , · = j p i , j and p · , j = i p i , j ; hence X is a variable on X with distribution p X and Y is a variable on Y with distribution p Y .
  • For uni-variate situations, K stands for K 1 , and p stands for p X .
  • Let { X 1 , X 2 , , X n } be an independent and identically distributed ( i . i . d . ) random sample of size n from X . Let C r = i = 1 n 1 [ X i = l r ] ; hence C r is the count of occurrence of letter l r in a sample. Let p ^ = { p ^ 1 , p ^ 2 , } = { C 1 / n , C 2 / n , } . p ^ is called the plug-in estimator of p . Similarly, one can construct the plug-in estimators for p Y and p X Y and name them as p ^ Y and p ^ X Y , respectively.
  • For any two functions f and g taking values in ( 0 , ) with lim n f ( n ) = lim n g ( n ) = 0 , the notation f ( n ) = O ( g ( n ) ) means
    0 < lim inf n g ( n ) f ( n ) lim sup n g ( n ) f ( n ) < .
  • For any two functions f and g taking values in ( 0 , ) , the notation f ( n ) = O ( g ( n ) ) means
    lim n f ( n ) g ( n ) = 0 .
Many concepts discussed in the following sections have continuous counterparts under the same concept name. The results reviewed in this article focus on non-ordinal data space. Therefore, some notable results on ordinal space are not reviewed (for example, [11,12,13]). In Section 2, estimation on some classic entropic statistics quantities are discussed. Section 3 reviews estimation results and properties for some recently developed information-theoretic quantities. Entropic statistics’ application potentials in machine learning (ML) and knowledge extraction are discussed in Section 4. Finally, some remarks are given in Section 5.

2. Classic Entropic Statistics Quantities and Estimation

This section reviews three classic entropic concepts and their estimations, including Shannon’s entropy (Section 2.1.1) and mutual information (Section 2.1.2), and Kullback–Leibler divergence (Section 2.2). These three concepts are among the earliest entropic concepts and have been intensively studied over the past decades. Enormous amounts of statistical methods and computational algorithms are designed based on these three concepts [14,15,16]. Nevertheless, most of those methods and algorithms use naive plug-in estimation, which could be improved for a smaller estimation bias and better performance. For this reason, this section reviews several notable estimation methods as a reference. Some asymptotic properties are also presented as a reference. The asymptotic properties provide a theoretical guarantee for the corresponding estimators with statistical procedures such as hypothesis testing and confidence intervals.

2.1. Shannon’s Entropy and Mutual Information

2.1.1. Shannon’s Entropy

Established by Shannon in his landmark paper [1], the concept of entropy is the first and still the most important building brick in characterizing information from non-ordinal spaces. Many of the established information-theoretic quantities are linear functions of entropy. Shannon’s entropy, H, is defined as
H = i p i ln p i .
Some remarkable properties of entropy are:
Property 1
(Entropy).
1.
H is a measurement of dispersion. It is always non-negative by definition.
2.
H = 0 if and only if the probability of a letter l in X is 1; hence no dispersion.
3.
For a finite alphabet with cardinality K, H is bounded from the above by ln K , and the maximum is achieved when its distribution is uniform ( p i = 1 / K , i = 1 , 2 , , K ); hence maximum dispersion.
4.
For a countably infinite alphabet, H may not exist (See Example 4 in Section 3).

Entropy Estimation-The Plug-in Estimator

Estimation of entropy has been a core research topic for decades. Due to the curse of “High Dimensionality” and “Discrete and Non-ordinal Nature”, entropy estimation is a technically difficult problem. Advances in this area have been slow to come. The plug-in estimator of entropy (also known as empirical entropy estimator), H ^ , defined as
H ^ = i p ^ i ln p ^ i ,
is inarguably the most naive entropy estimator. H ^ has been studied thoroughly in recent decades. Ref. [17] provided the asymptotic properties for H ^ when K is finite, namely,
Theorem 1
(Asymptotic property of H ^ when K is finite).
n H ^ H / σ ^ D N ( 0 , 1 ) ,
where σ ^ = i p ^ i ln 2 p ^ i H ^ 2 .
Ref. [18] derived the bias of H ^ for finite K
E ( H ^ ) H = K 1 2 n + 1 12 n 2 1 i = 1 K 1 p i + O n 3 .
Ref. [19] derived the asymptotic properties for H ^ when K is countable infinite. Namely,
Theorem 2
(Asymptotic property of H ^ when K is countable infinite). For any nonuniform distribution p i ; i 1 satisfying i p i ln 2 p i < , if there exists an integer-valued function K ( n ) such that, as n ,
1.
K ( n ) ,
2.
K ( n ) = O ( n ) , and
3.
n i K ( n ) p i ln p i 0 ;
then
n H ^ H / σ ^ D N ( 0 , 1 ) ,
where σ ^ = k p ^ k ln 2 p ^ k H ^ 2 .
As discussed in [19], the conditions with K ( n ) hold if p i 1 / i 2 ln 2 i ; the conditions do not hold if p i 1 / i 2 ln i .

Entropy Estimation-The Miller–Madow and Jackknife Estimators

H ^ M M [20] and H ^ J K [21] are two notable entropy estimators with bias adjustments. Namely,
H ^ M M = H ^ + K ^ 1 2 n ,
where K ^ is the observed sample cardinality. For finite K, the bias of H ^ M M is
E ( H ^ M M ) H = O n 2 .
H ^ J K is calculated in three steps:
  • for each i { 1 , 2 , , n } , construct H ^ ( i ) , which is a plug-in estimator based on a sub-sample of size n 1 obtained by leaving the ith observation out;
  • obtain H ^ ( i ) = n H ^ ( n 1 ) H ^ ( i ) for i = 1 , , n ; and then
  • compute the jackknife estimator
    H ^ J K = i = 1 n H ^ ( i ) n .
Equivalently, (3) can be written as
H ^ J K = n H ^ ( n 1 ) i = 1 n H ^ ( i ) n .
When K < , it can be shown that the bias of H ^ J K is
E H ^ J K H = O n 2 .
Asymptotic properties for H ^ M M and H ^ J K were derived in [22]. H ^ M M and H ^ J K reduce the rate of bias to a higher order power-decaying. Ref. [23] proved the convergence of H ^ could be arbitrarily slow. Ref. [24] proved that for finite K, an unbiased estimator for entropy does not exist. As a result, it is only possible to reduce the bias to a smaller extent.

Entropy Estimation-The Z-Estimator

Recent studies on entropy estimation have reduced the bias to exponentially decaying. For example,
H ^ z = v = 1 n 1 1 v n 1 + v [ n ( 1 + v ) ] ! n ! i p ^ i j = 0 v 1 1 p ^ i j n
is the entropy estimator provided in [25] with an exponentially decaying bias (Interested readers may refer to [26] for discussion on an entropy estimator that is algebraically equivalent to H ^ z ). Ref. [27] derived the asymptotic properties for H ^ z . Namely,
Theorem 3
(Asymptotic property of H ^ z when K is finite).
n H ^ z H / σ ^ D N ( 0 , 1 ) ,
where σ ^ = i p ^ i ln 2 p ^ i H ^ 2 .
The following asymptotic properties for H ^ z when K is countable infinite were provided in [28].
Theorem 4
(Asymptotic property of H ^ z when K is countable infinite). For a nonuniform distribution p i ; i 1 P satisfying i p i ln 2 p i < , if there exists an integer-valued function K ( n ) such that, as n ,
1.
K ( n ) ,
2.
K ( n ) = o ( n / ln n ) , and
3.
n i K ( n ) p i ln p i 0 ;
then
n H ^ z H / σ ^ D N ( 0 , 1 ) ,
where σ ^ = H ^ 2 z H ^ z 2 and H ^ 2 z = i p ^ i v = 1 n n p ^ i s = 1 v 1 1 s ( v s ) j = 1 v 1 n p ^ i 1 n j .
The sufficient condition given in Theorem 4 for the normality of H ^ z is slightly more restrictive than that of the plug-in estimator H ^ as stated in Theorem 2, and consequently supports a smaller class of distributions. The sufficient conditions of Theorem 4 still holds for p i = C λ i λ where λ > 2 , but not for p i = C / i 2 ln 2 i , which satisfies the sufficient conditions of Theorem 2. However, it is discussed in [28] that simulation results indicate that the asymptotic normality of H ^ z in Theorem 4 may still hold for p i = C / i 2 ln 2 i for i 1 though not covered by the sufficient condition.

Remarks

Another perspective of entropy estimation is to combine H ^ z and H ^ J K . Namely, one could use H ^ z in place of each H ^ ( i ) in (3). Interested readers may refer to [25] where a single layer combination of H ^ z and H ^ J K was discussed. In addition, ref. [29] presented a non-parametric entropy estimator ( H ^ c h a o ) when there are unseen species in the sample. H ^ c h a o has a smaller sample root mean squared error than H ^ M M and a smaller bias than H ^ J K , according to the simulation study. Unfortunately, the bias decaying rate for H ^ c h a o was not theoretically offered. Based on their simulation study of H ^ c h a o , it seems that the bias decaying rate is O ( 1 / n 2 ) , which is slower than H ^ z . Asymptotic properties of H ^ c h a o are not developed in the literature.
There are several parametric entropy estimators for specific interests. For example, Dirichlet prior Bayesian estimator of entropy [30,31] and shrinkage estimator of entropy [32]. This review article focuses on results from non-parametric estimation methods. To conclude this section, a small scale comparison between H ^ and H ^ z from [33] is provided in Table 1.

2.1.2. Mutual Information

In the same paper defining Shannon’s entropy, the concept of Mutual Information (MI) was also described [1]. Shannon’s entropies for X , Y , and X × Y are defined as
H ( X ) = i p i , · ln p i , · , H ( Y ) = j p · , j ln p · , j , H ( X , Y ) = i j p i , j ln p i , j ,
and MI between X and Y is defined as
M I ( X , Y ) = H ( X ) + H ( Y ) H ( X , Y ) .
Some notable properties of MI are:
Property 2
(Mutual Information).
1.
MI is a measurement of dependence. It is always non-negative by definition.
2.
M I = 0 if and only if the two marginals are independent.
3.
M I > 0 if and only if the two marginals are dependent.
4.
A non-zero MI does not always indicate the degree (level) of dependence.
5.
MI may not exist when the cardinality of joint space is countably infinite.

MI Estimation-The Plug-in Estimator and Z-Estimator

Since MI is a function of entropy, estimation of MI is essentially entropy estimation. Let { ( X 1 , Y 1 ) , ( X 2 , Y 2 ) , , ( X n , Y n ) } be an i . i . d . random sample of size n from the joint alphabet ( X , Y ) . Based on the sample, plug-in estimators of the component entropy of MI can be obtained. Namely,
H ^ ( X ) = i p ^ i , · ln p ^ i , · , H ^ ( Y ) = j p ^ · , j ln p ^ · , j , H ^ ( X , Y ) = i j p ^ i , j ln p ^ i , j ,
where p ^ i , · is the plug-in estimator for p i , · , p ^ · , j is the plug-in estimator for p · , j , and p ^ i , j is the plug-in estimator for p i , j . Then the plug-in estimator of mutual information between X and Y is defined as
M I ^ ( X , Y ) = H ^ ( X ) + H ^ ( Y ) H ^ ( X , Y ) .
With various entropy estimation methods, one could estimate MI by replacing H ^ with a different entropy estimator. For example, using the entropy estimator with the fastest bias decaying rate, H ^ z , the resulting estimator ( M I ^ z ) also has a bias with an exponentially decaying rate [34], namely,
M I ^ z = H ^ z ( X ) + H ^ z ( Y ) H ^ z ( X , Y ) .
The asymptotic properties for M I ^ -s ( M I ^ and M I ^ z ) shall be discussed under two situations: (1) M I = 0 , and (2) M I > 0 .
The first situation of M I = 0 is used for testing independence. For example, in feature selection, irrelevant (to the outcome) non-ordinal features shall be dropped, and a feature is irrelevant if it is independent of the outcome. Let A be the potential irrelevant feature and B be the outcome; hence one must test H 0 : M I ( A , B ) = 0 against H a : M I ( A , B ) > 0 . To test such a hypothesis, one needs the asymptotic properties of M I ^ -s under the null hypothesis: M I = 0 , derived in [35]. Namely,
Theorem 5
(Asymptotic properties of M I ^ and M I ^ z when M I = 0 ). Provided that M I = 0 ,
2 n M I ^ D χ ( K 1 1 ) ( K 2 1 ) 2
and
2 n M I ^ z + ( K 1 1 ) ( K 2 1 ) D χ ( K 1 1 ) ( K 2 1 ) 2
where n is the sample size and χ ( K 1 1 ) ( K 2 1 ) 2 stands for chi-squared distribution with degrees of freedom ( K 1 1 ) ( K 2 1 ) .
For the second situation of M I > 0 (recall that M I > 0 if and only if the two marginals are dependent), the following asymptotic properties were due to [34].
Let
v ^ = p ^ 1 , , p ^ K 1 K 2 1 τ = p ^ 1 , 1 , p ^ 1 , 2 , , p ^ 1 , K 2 , p ^ 2 , 1 , p ^ 2 , 2 , , p ^ 2 , K 2 , , p ^ K 1 , 1 , p ^ K 1 , 2 , , p ^ K 1 , K 2 1 τ
be the enumeration of joint probabilities plug-in estimators. Let
Σ ( v ^ ) = p ^ 1 1 p ^ 1 p ^ 1 p ^ 2 p ^ 1 p ^ K 1 K 2 1 p ^ 2 p ^ 1 p ^ 2 1 p ^ 2 p ^ 2 p ^ K 1 K 2 1 p ^ K 1 K 2 1 p ^ 1 p ^ K 1 K 2 1 p ^ 2 p ^ K 1 K 2 1 1 p ^ K 1 K 2 1 ,
and
g ( v ^ ) = ln p ^ K 1 , · p ^ · , K 2 p ^ k ln p ^ i , · p ^ · , j p ^ K 1 , K 2 , if k s corresponding { i , j } satisfying i K 1 and j K 2 ln p ^ · , K 2 p ^ k ln p ^ · , j p ^ K 1 , K 2 , if k s corresponding { i , j } satisfying i = K 1 and j K 2 ln p ^ K 1 , · p ^ k ln p ^ i , · p ^ K 1 , K 2 , if k s corresponding { i , j } satisfying i K 1 and j = K 2 ,
where k { 1 , 2 , , K 1 K 2 1 } . Then,
Theorem 6
(Asymptotic properties of M I ^ and M I ^ z when M I > 0 ). Provided that M I > 0 ,
n ( M I ^ M I ) g τ ( v ^ ) Σ ( v ^ ) g ( v ^ ) 1 2 L N ( 0 , 1 )
and
n M I ^ z M I g τ ( v ^ ) Σ ( v ^ ) g ( v ^ ) 1 2 N ( 0 , 1 ) .
The following examples describe a proper use of MI and properties in Theorems 5 and 6.
Example 1
(Genes TMEM30A and MTCH2—data and descriptions are in Example 1 of [34]). In the example, data were from two different genes in 191 patients. It has been calculated in [34] that M I ^ z = 0.0552 . The hypothesis test in Example 1 of [35] gave a p-value of 0.0567, which suggests M I = 0 at α = 0.05 . However, one shall use the property in Theorem 6 to obtain a confidence interval of M I . One must not use the property in Theorem 5 for the purpose of the confidence interval in this situation (because the asymptotic distribution in Theorem 5 assumes a specific location for M I under the null hypothesis).
Example 2
(Genes ENAH and ENAH—data and descriptions are in Example 2 of [34]). In the example, data were from different probes of the same genes on 191 patients. It has been calculated in [34] that M I ^ z = 0.1157 . The hypothesis test in Example 2 of [35] gave a p-value of 0.0012, which suggests M I > 0 at α = 0.05 . Furthermore, one shall use the property in Theorem 6 to obtain a confidence interval of M I . One must not use the property in Theorem 5 for the purpose of the confidence interval in this situation.
Example 3
(Compare the MI between Examples 1 and 2). From Examples 1 and 2, M I ^ z ( T M E M 30 A , M T C H 2 ) = 0.0552 and M I ^ z ( E N A H 1 , E N A H 2 ) = 0.1157 . Although the second estimation value is higher, one cannot conclude that the level of dependence between E N A H 1 and E N A H 2 is higher than that between TMEM30A and MTCH2 due to the limitation described in the 4-th property in Property 2. To compare the level of dependence, one shall refer to the standardized mutual information in Section 3.1.
Recall that MI is always non-negative. For the same reason, M I ^ is always non-negative (note that M I ^ can be viewed as the M I for the distribution p ^ ). Nevertheless, M I ^ z can be negative under some scenarios. A negative M I ^ z suggests the level of dependence between the two random elements is extremely weak. If one uses the results from Theorem 5 to test if H 0 : M I = 0 , a negative M I ^ z would lead to a fail-to-reject for most settings of α (level of significance).

Remarks

There is another line of research on multivariate information-theoretic methods, the Partial Information Decomposition (PID) framework [36,37,38]. The PID may be viewed as a direct extension of MI to a measures of information provided by two or more variables about a third. Interesting applications of the PID are, for example, in explaining representation learning in neural networks [39] or in feature selection from dependent features [40]. PID aims to characterize redundancy with information decomposition. Another approach to characterize redundancy is to utilize MI on a joint feature space [33]. Additional research to compare the two approaches is needed.

2.2. Kullback–Leibler Divergence

Kullback–Leibler divergence (KL) [2], also known as relative entropy, is the distance between two probability distributions, introduced by [2], and is an important measure of information in information theory. The notations to define KL and describe its properties differ slightly from other sections. Let P = p k : k = 1 , , K and Q = q k : k = 1 , , K be two discrete probability distributions on the same finite alphabet, X = k : k = 1 , , K , where K 2 is a finite integer. KL is defined to be
K L = K L ( P Q ) = k = 1 K p k ln p k / q k = k = 1 K p k ln p k k = 1 K p k ln q k .
Note that many also use D as the notation of KL, namely, D ( P Q ) . KL is not a metric since it does not satisfy the triangle inequality and is not symmetric. Some notable properties of KL are:
Property 3
(Kullback–Leibler divergence).
1.
KL is a measurement of non-metric distance between two distributions on the same alphabet (with the same discrete support). It is always non-negative because of Gibbs’ inequality.
2.
K L = 0 if and only if the two underlying distributions are the same. Namely, P = Q for each k = 1 , , K .
3.
K L > 0 if and only if the two underlying distributions are different. Namely, p k q k for some k.
The use of KL has several variants, including but not limited to, (1) P and Q are unknown; (2) Q is known; (3) P and Q are continuous distributions. The second variant is an alternative method of the Pearson goodness-of-fit test. Interested readers may refer to [41] for more discussion on the second variant. Although utilizing entropic statistics on continuous spaces is generally not recommended, interested readers may refer to [42,43] for discussions on the third variant.

2.2.1. KL Point Estimation-The Plug-in Estimator, Augmented Estimator, and Z-Estimator

Although KL is not exactly a function of entropy, it still carries many similarities with entropy. For that reason, KL estimation is very similar to entropy estimation. For example, KL can be estimated from a plug-in perspective. Let p ^ k be the plug-in estimator of p k and q ^ k be the plug-in estimator of q k , then the KL plug-in estimator is
K L ^ = K L ^ ( P Q ) = k = 1 K p ^ k ln p ^ k k = 1 K p ^ k ln q ^ k .
Because K L ^ could have an infinite bias [44], an augmented plug-in estimator of KL was presented in [44]:
K L ^ * = i = 1 K p ^ k ln p ^ k k = 1 K p ^ k ln q ^ k * ,
where
q ^ k * = q ^ k + 1 q ^ k = 0 m ,
and m is the sample size of the sample from Q. The bias of K L ^ is no faster than O ( 1 / n ) , where n is the sample size of the sample from P [44].
Since the K L ^ could have an infinite bias, its estimation in perspectives of H ^ M M or H ^ J K will not help in reducing the bias to a finite extent. In the perspective of H ^ z , a KL estimator with exponentially decaying bias was offered in [44]:
K L ^ z = K L ^ z ( P Q ) = k = 1 K p ^ k v = 1 m m q ^ k 1 v j = 1 v 1 m q ^ k m j + 1 v = 1 n n p ^ k 1 v j = 1 v 1 n p ^ k 1 n j .

2.2.2. Symmetrized KL and Its Point Estimation

As mentioned in the first property of Property 3, KL is generally an asymmetric measurement. For certain interests that require a symmetric measurement, a symmetrized KL is defined to be
S = S ( P , Q ) = 1 2 [ K L ( P Q ) + K L ( Q P ) ] = 1 2 k = 1 K p k ln p k k = 1 K p k ln q k + 1 2 k = 1 K q k ln q k k = 1 K q k ln p k .
The symmetrized KL S, as a function of K L , can be similarly estimated in the perspective of K L ^ , K L ^ * , and K L ^ z . The respective estimators are
S ^ = 1 2 k = 1 K p ^ k ln p ^ k k = 1 K p ^ k ln q ^ k + 1 2 k = 1 K q ^ k ln q ^ k k = 1 K q ^ k ln p ^ k ,
S ^ * = 1 2 k = 1 K p ^ k ln p ^ k k = 1 K p ^ k ln q ^ k * + 1 2 k = 1 K q ^ k ln q ^ k k = 1 K q ^ k ln p ^ k * ,
where p ^ k * = p ^ k + 1 p ^ k = 0 / n (n is the sample size of the sample from P), and
S ^ z = 1 2 k = 1 K p ^ k v = 1 m m q ^ k 1 v j = 1 v 1 m q ^ k m j + 1 v = 1 n n p ^ k 1 v j = 1 v 1 n p ^ k 1 n j + 1 2 k = 1 K q ^ k v = 1 n n p ^ k 1 v j = 1 v 1 n p ^ k n j + 1 v = 1 m m q ^ k 1 v j = 1 v 1 m q ^ k 1 m j .

2.2.3. Asymptotic Properties for KL and Symmetrized KL Estimators

The asymptotic properties for K L ^ , K L ^ * , K L ^ z , S ^ , S ^ * , and S ^ z are all presented in [44]. All the asymptotic properties therein require P Q (namely, K L > 0 ). When P = Q , the asymptotic property of KL (or S) estimators are currently missing from the literature. The derivation of such asymptotic properties is not complicated yet unnecessary. The only purpose of such asymptotic property under P = Q is to test if H 0 : P = Q against H a : P Q . For such a purpose, the two-sample goodness-of-fit chi-squared test can be used (see p. 616 in [45]).

3. Recently Developed Entropic Statistics Quantities and Estimation

In this section, various recently developed entropic statistics quantities are introduced and discussed. Some quantities are quite new with limited estimation properties developed other than the plug-in estimation. Therefore, some of the following discussions focus on conceptual spirits and application potentials.

3.1. Standardized Mutual Information

Mutual information between two random elements (on non-ordinal alphabets) is similar to the covariance between two random variables (on ordinal spaces) regarding properties and drawbacks. For example, the covariance does not provide general information on the degree of correlation, and the concept correlation of coefficient was defined to fill the gap. Similarly, recall the fourth property of MI that MI generally does not provide information about the degree of dependence, standardized mutual information (SMI), κ , has been studied and defined in various ways. To name a few, provided H ( X , Y ) < ,
κ 1 = M I ( X , Y ) H ( X , Y ) , κ 2 = M I ( X , Y ) min { H ( X ) , H ( Y ) } , κ 3 = M I ( X , Y ) H ( X ) H ( Y ) , κ 4 = M I ( X , Y ) ( H ( X ) + H ( Y ) ) / 2 , κ 5 = M I ( X , Y ) max { H ( X ) , H ( Y ) } , κ 6 = M I ( X , Y ) H ( X ) .
The quantity κ 6 is also called information gain ratio [46]. The benefits of SMI are supported by Theorem 7.
Theorem 7
(Theorem 5.4 in [28]). Suppose H ( X , Y ) < . Then
0 κ 1
Moreover, (1) κ = 0 if and only if X and Y are independent, and (2) κ = 1 if and only if X and Y have a one-to-one correspondence.
Interested readers may refer to [47,48,49,50] for discussions on SMI. A detailed discussion of the estimation of various SMI may be found in [51].

3.2. Entropic Basis: A Generalization from Shannon’s Entropy

Shannon’s entropy and MI are powerful tools to quantify dispersion and dependence on non-ordinal space. More concepts and statistical tools are needed to characterize non-ordinal space information from different perspectives.
Generalized Simpson’s diversity indices were established in [52] and coined in [3].
Definition 1
(Generalized Simpson’s Diversity Indices). For a given p = p k ; k 1 and an integer pair ( u 1 , v 0 ) , let ζ u , v = k 1 p k u ( 1 p k v . Let
ζ = ζ u , v ; u 1 , v 0
be defined as the family of generalized Simpson’s diversity indices.
Generalized Simpson’s Diversity Indices are the foundation of entropic basis and entropic moment. Interested readers may refer to [53] for discussions on entropic moments and a goodness-of-fit test under permutation based on entropic moments. In estimating ζ u , v ,
z u , v = [ n ( u + v ) ] ! n u + v n ! × k 1 1 p ^ k u n i = 0 u 1 p ^ k i n j = 0 v 1 1 p ^ k j n
was derived in [52], where n is the sample size; u and v are given constants; p ^ k is the sample proportion of the k-th letter(category); 1 [ · ] stands for indicator function. z u , v is an uniformly minimum-variance unbiased estimator (UMVUE) of ζ u , v for any combination of ( u , v ) non-negative integers pair as long as u + v n , where n is the corresponding sample size.
Based on ζ u , v , ref. [3] defined the entropic basis.
Definition 2
(Entropic Basis). Given Definition 1, the entropic basis is the sub-family
ζ 1 = { ζ 1 , v ; v 0 }
of ζ.
All diversity indices can be represented as a function of ζ 1 [3] (most representations are due to Taylor’s expansion). For example,
  • Simpson’s index [54]:
    λ = k 1 p k 2 = ζ 1 , 0 ζ 1 , 1
  • Gini–Simpson index [54,55]:
    1 λ = k 1 p k 1 p k = ζ 1 , 1
  • Shannon’s entropy:
    H = k 1 p k ln p k = v = 1 1 v ζ 1 , v
  • Rényi equiv. entropy [56]:
    h r = k 1 p k r = ζ 1 , 0 + v = 1 i = 1 v i r i ζ 1 , v
  • Emlen’s index [57]:
    D = k 1 p k e p k = v = 0 e 1 v ! ζ 1 , v
  • Richness index (population size):
    K = k 1 1 p k > 0 = v = 0 ζ 1 , v
  • Generalized Simpson’s index:
    ζ u , m = k 1 p k u 1 p k m = v = 0 u 1 ( 1 ) v u 1 v ζ 1 , m + v .
In practice, plug-in estimation is used in estimating diversity indices. The representations of the diversity index on an entropic basis allow a new estimation method with a smaller bias. Namely, z 1 , v , the UMVUE for ζ 1 , v , exist for all v up to n 1 . If one replaces { ζ 1 , 0 , ζ 1 , 1 , , ζ 1 , n 1 } with { z 1 , 0 , z 1 , 1 , , z 1 , n 1 } , and let all the other ζ s (namely, { ζ 1 , n , ζ 1 , n + 1 , } ) to be zero, then the resulting estimator is exactly the same as plug-in estimator. However, the estimation can be further improved if one estimate { ζ 1 , n , ζ 1 , n + 1 , } based on { z 1 , 0 , z 1 , 1 , , z 1 , n 1 } .
For example, let K ^ (the observed number of categories) be the plug-in estimator of K. Meanwhile, v = 0 n 1 z 1 , v (the estimator in perspective of entropic basis representation) is algebraically equivalent to K ^ [58]. Namely,
K = k 1 1 p k > 0 = v = 0 ζ 1 , v = v = 0 n 1 ζ 1 , v + v = n ζ 1 , v
is a decomposition of K, then
K ^ = k 1 1 p ^ k > 0 = v = 0 n 1 ζ 1 , v ;
whereas
K ^ e n t r o p i c = v = 0 n 1 ζ 1 , v + v = n ζ 1 , v = K ^ + estimator of v = n ζ 1 , v .
K ^ e n t r o p i c has a smaller bias than K ^ . Interested readers may refer to [58] for details on the estimation of v = n ζ 1 , v . Similar estimation could benefit the estimation of Rényi equiv. entropy, Emlen’s index, and any other diversity indices or theoretical quantities which contain the terms v = n ζ 1 , v after Taylor’s expansion.

3.3. Generalized Shannon’s Entropy and Generalized Mutual Information

Because of the advantages in characterizing information in non-ordinal space, Shannon’s entropy and MI have become the building blocks of information theory and essential aspects of ML methods. Yet, they are only finitely defined for distributions with fast decaying tails on a countable alphabet.
Example 4
(Unbounded entropy). Let X = { 2 , 3 , 4 , } be a random variable following the distribution P ( X = k ) = p k = c / ( k ln 2 k ) , where k = 2 , 3 , 4 , , and c is the constant to make the distribution valid (total probability add up to 1). Such a constant uniquely exists because the summation i = 2 1 / ( i ln 2 i ) converges. Then
H ( X ) = k = 2 p k ln 1 p k = k = 2 c k ln 2 k ln k ln 2 k c = k = 2 c k ln 2 k ( ln k + 2 ln ln k ln c ) = k = 2 c k ln k + k = 2 c ( 2 ln ln k ln c ) k ln 2 k = ,
because
k = 2 c k ln k =
and
k = 2 c ( 2 ln ln k ln c ) k ln 2 k < .
Therefore the entropy H ( X ) is unbounded.
The unboundedness of Shannon’s entropy and MI over the general class of all distributions on an alphabet prevents their potential utility from being fully realized. Ref. [5] proposed GSE and GMI, which are finitely defined everywhere. To state the definition of GSE and GMI, Definition 3 is stated first.
Definition 3
(Conditional Distribution of Total Collision (CDOTC)). Given X = x i ; i 1 and p = p i , consider the experiment of drawing an identically and independently distributed (iid) sample of size m ( m 2 ) . Let C m denote the event that all observations of the sample take on the same letter in X , and let C m be referred to as the event of a total collision. The conditional probability, given C m , that the total collision occurs at the letter x i is
p m , i = p i m s 1 p s m ,
where m 2 . p m = p m , i is defined as the m-th order CDOTC.
The idea of CDOTC is to adopt a special member of the family of the escort distributions introduced in [59]. The utility of CDOTC is endorsed by Lemmas 1 and 2, which are proved in [5].
Lemma 1.
For any order m, p and p m uniquely determine each other.
Lemma 2.
For any order m, p X Y = { p i , j } = { p i , . × p . , j } if and only if p X Y , m = { p m , i , j } = { p m , i , . × p m , . , j } .
It is clear that p m is a probability distribution induced from p = p k . An example is provided to help understand Definition 3.
Example 5
(The second-order CDOTC). Let p = p i = { 6 i 2 / π 2 ; i = 1 , 2 , 3 , } , the second-order CDOTC is then defined as
p 2 = { p 2 , i } ,
where
p 2 , i = p i 2 s 1 p s 2 = 36 i 4 / π 4 s 1 36 s 4 / π 4 = i 4 s 1 s 4
for i = 1 , 2 , 3 , .
Based on Definition 3, GSE and GMI are defined as follows.
Definition 4
(Generalized Shannon’s Entropy (GSE)). Given X = x i ; i 1 , p X = p i , and p X , m = { p m , i } , generalized Shannon’s entropy (GSE) is defined as
H m ( X ) = i 1 p m , i ln p m , i ,
where p m , i is defined in Definition 3, and m = 2 , 3 , is the order of GSE. GSE with order m is called the m-th order GSE.
Definition 5
5(Generalized Mutual Information (GMI)). Let X = x i ; i 1 and Y = y j ; j 1 . Let p X Y = p i , j and p X Y , m = { p m , i , j } . Let X * and Y * be the marginal distributions of p X Y , m . The m-th order generalized mutual information (GMI) between X and Y is defined as
M I m ( X , Y ) = H ( X * ) + H ( Y * ) H m ( X , Y ) ;
or equivalently
M I m ( X , Y ) = H ( X * ) + H ( Y * ) H ( X * , Y * ) ;
To help understand Definitions 4 and 5, Examples 6 and 7 are provided as follows.
Example 6
(The second-order GSE). Let X = x i ; i 1 and p = p i = { 6 i 2 / π 2 ; i = 1 , 2 , 3 , } . The second-order GSE, H 2 ( X ) , is then defined as
H 2 ( X ) = i 1 p 2 , i ln p 2 , i ,
where
p 2 , i = i 4 s 1 s 4 ; i = 1 , 2 ,
is given in Example 5.
Example 7
(The second-order GMI). Let X = x i , . ; i 1 , Y = x . , j ; j 1 , and X × Y = x i , j ; i 1 , j 1 . Let
p X Y = p i , j = 24 ( i + j 1 ) 2 + i j + 1 2 π 2 ; i 1 , j 1 ,
and
p X Y , 2 = { p 2 , i , j } = p i , j 2 s 1 , t 1 p s , t 2 ; i 1 , j 1 .
Further, let
p X * = j p 2 , i , j ; i 1
and
p Y * = i p 2 , i , j ; j 1 .
The second-order GMI, M I 2 ( X , Y ) , is then defined as
M I 2 ( X , Y ) = H ( X * ) + H ( Y * ) H ( X * , Y * ) ,
where H ( X * ) , H ( Y * ) , and H ( X * , Y * ) are Shannon’s entropy based on p X * , p Y * , and p X Y , 2 , respectively.
GSE’s and GMI’s plug-in estimators are stated in Definitions 6 and 7.
Definition 6
(GSE’s plug-in estimators). Let X 1 , X 2 , , X n be i.i.d. random variables taking values in X = x i ; i 1 with distribution p X . The plug-in estimator for p X is p ^ X = { p ^ i } . The plug-in estimator for the m-th order GSE, H ^ m ( X ) , is
H ^ m ( X ) = i 1 p ^ m , i ln p ^ m , i = i 1 p ^ i m s 1 p ^ s m ln p ^ i m s 1 p ^ s m .
Definition 7
(GMI’s plug-in estimators). Let ( X 1 , Y 1 ) , ( X 2 , Y 2 ) , , ( X n , Y n ) be i.i.d. random variables taking values in X × Y = ( x i , y j ) ; i 1 , j 1 with distribution p X Y = { p i , j } . Let p ^ X Y = { p ^ i , j } be the plug-in estimator of p X Y . The plug-in estimator for the m-th order GMI, M I ^ m ( X , Y ) , is
M I ^ m ( X , Y ) = H ^ ( X * ) + H ^ ( Y * ) H ^ ( X * , Y * ) ,
where
H ^ ( X * ) = i 1 j 1 p ^ i , j m s 1 , t 1 p ^ s , t m ln j 1 p ^ i , j m s 1 , t 1 p ^ s , t m , H ^ ( Y * ) = j 1 i 1 p ^ i , j m s 1 , t 1 p ^ s , t m ln i 1 p ^ i , j m s 1 , t 1 p ^ s , t m , H ^ ( X * , Y * ) = i 1 , j 1 p ^ i , j m s 1 , t 1 p ^ s , t m ln p ^ i , j m s 1 , t 1 p ^ s , t m .
The following asymptotic properties for GSE’s plug-in estimators are given in [60].
Theorem 8.
Let p X = { p k ; k 1 } be a probability distribution on a countably infinite alphabet X , without any further conditions,
n H ^ m ( X ) H m ( X ) σ ^ m d N ( 0 , 1 ) ,
where
σ ^ m 2 = k = 1 m 2 p ^ k p ^ m , k ln p ^ m , k + p ^ m , k H ^ m ( Z ) 2 .
Theorem 9.
Let p X = { p k ; k = 1 , 2 , , K } be a non-uniform probability distribution on a countably finite alphabet X , without any further conditions,
n H ^ m ( X ) H m ( X ) σ ^ m d N ( 0 , 1 ) ,
where
σ ^ m 2 = k = 1 K m 2 p ^ k p ^ m , k ln p ^ m , k + p ^ m , k H ^ m ( Z ) 2 .
The properties in Theorems 8 and 9 allow interval estimation and hypothesis testing with H ^ m . The advantage of shifting the original distribution to an escort distribution is reflected in Theorem 8-the asymptotic normality requires no assumption on a countable infinite alphabet. Theorem 9 can be viewed as a special case of Theorem 8 under a finite situation, where the uniform distribution shall be omitted because a uniform distribution has no variation between different category probabilities and hence results in a zero GSE and degenerate asymptotic distribution.
Nevertheless, suppose one is certain that the cardinality of distribution is finite. In that case, one shall use Shannon’s entropy instead of GSE because Shannon’s entropy always exists under a finite distribution. There are various well-studied estimation methods with Shannon’s entropy (whereas only plug-in estimation on GSE has been studied by far).
Asymptotic properties for GMI plug-in estimator have not been studied yet. Nonetheless, a test of independence with modified GMI [61] has been studied. The test does not require the knowledge of the number of columns or rows of a contingency table; hence it yielded an alternative other than Pearson’s chi-squared test of independence, particularly when a contingency table is large or sparse.

4. Application of Entropic Statistics in Machine Learning and Knowledge Extraction

Applications of entropic statistics in ML and knowledge extraction can be clustered in two directions. The first direction is to solve an existing question from a new perspective by creating a new information-theoretic quantity [61] or revisiting an existing information-theoretic quantity for additional insights [62]. The second direction is to use different estimation methods in existing methods to improve the performance by reducing bias and/or variation [32]. Application potentials in the second direction are very promising because theoretical results from recent-developed estimation methods suggest the performance of many existing ML methods could be improved, yet not much research has been conducted in the direction. In this section, many established ML and knowledge extraction methods are discussed with their potential to improve in the second direction.

4.1. An Entropy-Based Random Forest Model

Ref. [63] proposed an entropy-importance-based random forest model for power quality feature selection and disturbance classification. The method used a greedy search based on entropy and information gain for node segmentation. Nevertheless, only the plug-in estimation of entropy and information gain was considered. The method could be improved by replacing the plug-in estimation with smaller bias estimation methods, such as H ^ z in [25]. Further, one can also combine H ^ z with the jackknife procedure in (3) to obtain
H ^ z J K = i = 1 n H ^ z ( i ) n ,
where H ^ z ( i ) = n H ^ z ( n 1 ) H ^ z ( i ) , and use H ^ z J K in place of the adopted plug-in estimation. The benefit of using H ^ z J K is the potential smaller bias, and variance [25]. However, asymptotic properties for H ^ z J K are yet developed. When asymptotic properties are desired (e.g., for confidence interval or hypothesis testing purposes), one shall consider estimators with established asymptotic properties (also called theoretical guarantee), such as H ^ , H ^ M M , H ^ J K , and H ^ z .

4.2. Feature Selection Methods

In [16,64], various information-theoretic feature selection methods were reviewed and discussed. The two review articles did not mention that all the discussed methods adopted plug-in estimators for the corresponding information-theoretic quantities. Improving the performance with different estimation methods is possible, and investigation is needed. For example, some of the discussed methods are summarized in Table 2 with suggestions to utilize smaller bias and/or variance estimation methods.

4.3. A Keyword Extraction Method

Ref. [79] proposed a keyword extraction method with Rényi’s entropy
S R ( w , q ) = 1 1 q log 2 i = 1 F w p i q ,
and used the plug-in estimator therein. Namely, S ^ R ( w , q ) = 1 1 q log 2 i = 1 F w p ^ i q . Nevertheless, S R ( w , q ) can be represented as
S R ( w , q ) = 1 1 q log 2 ζ 1 , 0 + v = 1 j = 1 v j q j ζ 1 , v = 1 1 q log 2 ζ 1 , 0 + v = 1 n 1 j = 1 v j q j ζ 1 , v + v = n j = 1 v j q j ζ 1 , v ,
where ζ 1 , v has the UMVUE z 1 , v for v = 0 , 1 , 2 , , n 1 . For v n , ζ 1 , v could be estimated based on regression analysis [58]. Hence, S R ( w , q ) can be estimated as
S ^ R z ( w , q ) = 1 1 q log 2 z 1 , 0 + v = 1 n 1 j = 1 v j q j z 1 , v + v = n j = 1 v j q j z 1 , v * ,
where the construction of z 1 , v * needs investigation using regression analysis [58]. The resulting S ^ R z ( w , q ) would have a smaller bias than that of S ^ R ( w , q ) to help improve the established keyword extraction method. Note that if one wishes to use z 1 , v up to n 1 only, the resulting estimator would become
S ^ R z * ( w , q ) = 1 1 q log 2 z 1 , 0 + v = 1 n 1 j = 1 v j q j z 1 , v .
(6) is the same as H ^ r ( n ) in [3]. Asymptotic properties for H ^ r ( n ) were provided therein (Corollary 3 in [3]) for interested readers.

5. Conclusions

Entropic statistics is effective in characterizing information from non-ordinal space. Meanwhile, it is essential to realize that non-ordinal information is inherently difficult to identify due to its non-ordinal and permutation invariant nature. This survey article aims to provide a comprehensive review of recent advances in entropic statistics, including classic entropic concepts estimation, recent-developed entropic statistics quantities, and their applications potentials in ML and knowledge extraction. This article first introduces the concept of entropic statistics and emphasizes challenges from non-ordinal data. Then this article reviews the estimation for classic entropic quantities. These classic entropic concepts, including Shannon’s entropy, MI, and KL, are widely used in established machine learning and knowledge extraction methods. Most, if not all, of the established methods use plug-in estimation, which is computation efficient yet with a large bias. The surveyed different estimation methods would help researchers to potentially improve existing methods’ performance by adopting a different estimation method or adding theoretical guarantee to the existing methods. Recent-developed entropic statistics concepts are also reviewed with their estimation and applications. These new concepts not only allow researchers to estimate existing quantities in a new perspective, but also support additional aspects in characterizing non-ordinal information. In particular, the generalized Simpson’s diversity indices (with the induced entropic basis and entropic moments) have significant application and theoretical potential to either customize existing ML and knowledge extraction methods or to establish new methods considering domain-specific challenges. Further, this article provides some examples of how to apply the surveyed results to some of the existing methods, including a random forest model, fourteen feature selection methods, and a keyword extraction model. It should be mentioned that the aim of the survey is not to claim the superiority of some estimation methods over others but to provide a comprehensive list of recent advances in entropic statistics research. Specifically, although an estimator with a faster-decaying bias seems theoretically preferred, it has a longer calculation time even with the convenient R functions, particularly when multiple layers of jackknife (bootstrap) are involved. The preference of estimation varies case by case—some may prefer an estimator with a smaller bias, some may prefer one with a smaller variance, while some may need a trade-off between them. Furthermore, the article focuses on non-parametric estimation, while parametric estimation would perform better if the specified model fits the domain-specific reality. In summary, one should always investigate if a new estimation method fits the needs.
Enormous additional works are still needed in entropic statistics. For example, (1) the asymptotic properties for many established estimators (such as H ^ c h a o and H ^ z J K ) are not clear when cardinality is infinite. (2) With the transition from original distribution to escort distribution, GSE and GMI fill the void left by Shannon’s entropy and MI. However, only plug-in estimations of GSE and GMI have been studied. The biases of these plug-in estimators have not been studied, and additional estimation methods are undoubtedly needed. (3) Calculations for many entropic statistics are not yet supported in R, such as entropic basis, GSE, and GMI. Furthermore, more work is needed to implement the new entropic statistics concepts in programming software other than R (some of the reviewed estimators are implemented in R and are listed in Appendix A as a reference), particularly in Python. With additional theoretical development and application support, entropic statistics methods would be a more efficient tool to characterize more non-ordinal information and better serve the demands arose from the emerging domain-specific challenges.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ANOVAAnalysis of Variance
GSEGeneralized Shannon’s Entropy
GMIGeneralized Mutual Information
i.i.d.independent and identically distributed
KLKullback–Leibler Divergence
MIMutual Information
MLMachine Learning
PIDPartial Information Decomposition
SMIStandardized Mutual Information
UMVUEUniformly Minimum-Variance Unbiased Estimator

Appendix A. R Functions

StatisticR Package NameFunction Name
H ^ entropy [80]entropy.plugin
H ^ M M entropyentropy.MillerMadow
H ^ J K bootstrap [81]jackknife
H ^ c h a o entropyentropy.ChaoShen
H ^ z EntropyEstimation [82]Entropy.z
σ ^ in Theorems 1 and 3EntropyEstimationEntropy.sd
M I ^ entropymi.plugin
M I ^ z EntropyEstimationMI.z
g τ ( v ^ ) Σ ( v ^ ) g ( v ^ ) 1 2 in Theorem 6EntropyEstimationMI.sd
K L ^ entropyKL.plugin
K L ^ z EntropyEstimationKL.z
S ^ EntropyEstimationSymKL.plugin
S ^ z EntropyEstimationSymKL.z
H ^ r ( n ) EntropyEstimationRenyi.z

References

  1. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  2. Kullback, S.; Leibler, R.A. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  3. Zhang, Z.; Grabchak, M. Entropic representation and estimation of diversity indices. J. Nonparametr. Stat. 2016, 28, 563–575. [Google Scholar] [CrossRef]
  4. Grabchak, M.; Zhang, Z. Asymptotic normality for plug-in estimators of diversity indices on countable alphabets. J. Nonparametr. Stat. 2018, 30, 774–795. [Google Scholar] [CrossRef]
  5. Zhang, Z. Generalized Mutual Information. Stats 2020, 3, 158–165. [Google Scholar] [CrossRef]
  6. Burnham, K.P.; Anderson, D.R. Practical use of the information-theoretic approach. In Model Selection and Inference; Springer: Berlin/Heidelberg, Germany, 1998; pp. 75–117. [Google Scholar]
  7. Dembo, A.; Cover, T.M.; Thomas, J.A. Information theoretic inequalities. IEEE Trans. Inf. Theory 1991, 37, 1501–1518. [Google Scholar] [CrossRef]
  8. Chatterjee, S.; Hadi, A.S. Regression Analysis by Example; John Wiley & Sons: Hoboken, NJ, USA, 2006. [Google Scholar]
  9. Speed, T. What is an analysis of variance? Ann. Stat. 1987, 15, 885–910. [Google Scholar] [CrossRef]
  10. Hardy, M.A. Regression with Dummy Variables; Sage: Newcastle upon Tyne, UK, 1993; Volume 93. [Google Scholar]
  11. Kent, J.T. Information gain and a general measure of correlation. Biometrika 1983, 70, 163–173. [Google Scholar] [CrossRef]
  12. Lin, J. Divergence measures based on the Shannon entropy. IEEE Trans. Inf. Theory 1991, 37, 145–151. [Google Scholar] [CrossRef]
  13. Van Erven, T.; Harremos, P. Rényi divergence and Kullback-Leibler divergence. IEEE Trans. Inf. Theory 2014, 60, 3797–3820. [Google Scholar] [CrossRef] [Green Version]
  14. Sethi, I.K.; Sarvarayudu, G. Hierarchical classifier design using mutual information. IEEE Trans. Pattern Anal. Mach. Intell. 1982, 4, 441–445. [Google Scholar] [CrossRef] [PubMed]
  15. Safavian, S.R.; Landgrebe, D. A survey of decision tree classifier methodology. IEEE Trans. Syst. Man Cybern. 1991, 21, 660–674. [Google Scholar] [CrossRef]
  16. Li, J.; Cheng, K.; Wang, S.; Morstatter, F.; Trevino, R.P.; Tang, J.; Liu, H. Feature selection: A data perspective. ACM Comput. Surv. (CSUR) 2017, 50, 1–45. [Google Scholar] [CrossRef]
  17. Basharin, G.P. On a statistical estimate for the entropy of a sequence of independent random variables. Theory Probab. Appl. 1959, 4, 333–336. [Google Scholar] [CrossRef]
  18. Harris, B. The Statistical Estimation of Entropy in the Non-Parametric Case; Technical Report; Wisconsin Univ-Madison Mathematics Research Center: Madison, WI, USA, 1975. [Google Scholar]
  19. Zhang, Z.; Zhang, X. A normal law for the plug-in estimator of entropy. IEEE Trans. Inf. Theory 2012, 58, 2745–2747. [Google Scholar] [CrossRef]
  20. Miller, G.A.; Madow, W.G. On the Maximum Likelihood Estimate of the Shannon-Weiner Measure of Information; Operational Applications Laboratory, Air Force Cambridge Research Center, Air Research and Development Command, Bolling Air Force Base: Washington, DC, USA, 1954. [Google Scholar]
  21. Zahl, S. Jackknifing an index of diversity. Ecology 1977, 58, 907–913. [Google Scholar] [CrossRef]
  22. Chen, C.; Grabchak, M.; Stewart, A.; Zhang, J.; Zhang, Z. Normal Laws for Two Entropy Estimators on Infinite Alphabets. Entropy 2018, 20, 371. [Google Scholar] [CrossRef]
  23. Antos, A.; Kontoyiannis, I. Convergence properties of functional estimates for discrete distributions. Random Struct. Algorithms 2001, 19, 163–193. [Google Scholar] [CrossRef]
  24. Paninski, L. Estimation of entropy and mutual information. Neural Comput. 2003, 15, 1191–1253. [Google Scholar] [CrossRef]
  25. Zhang, Z. Entropy estimation in Turing’s perspective. Neural Comput. 2012, 24, 1368–1389. [Google Scholar] [CrossRef]
  26. Schürmann, T. A note on entropy estimation. Neural Comput. 2015, 27, 2097–2106. [Google Scholar] [CrossRef] [PubMed]
  27. Zhang, Z. Asymptotic normality of an entropy estimator with exponentially decaying bias. IEEE Trans. Inf. Theory 2013, 59, 504–508. [Google Scholar] [CrossRef]
  28. Zhang, Z. Statistical Implications of Turing’s Formula; John Wiley & Sons: Hoboken, NJ, USA, 2016. [Google Scholar]
  29. Chao, A.; Shen, T.J. Nonparametric estimation of Shannon’s index of diversity when there are unseen species in sample. Environ. Ecol. Stat. 2003, 10, 429–443. [Google Scholar] [CrossRef]
  30. Nemenman, I.; Shafee, F.; Bialek, W. Entropy and inference, revisited. arXiv 2001, arXiv:physics/0108025. [Google Scholar]
  31. Agresti, A.; Hitchcock, D.B. Bayesian inference for categorical data analysis. Stat. Methods Appl. 2005, 14, 297–330. [Google Scholar] [CrossRef]
  32. Hausser, J.; Strimmer, K. Entropy inference and the James-Stein estimator, with application to nonlinear gene association networks. J. Mach. Learn. Res. 2009, 10, 1469–1484. [Google Scholar]
  33. Shi, J.; Zhang, J.; Ge, Y. CASMI—An Entropic Feature Selection Method in Turing’s Perspective. Entropy 2019, 21, 1179. [Google Scholar] [CrossRef]
  34. Zhang, Z.; Zheng, L. A mutual information estimator with exponentially decaying bias. Stat. Appl. Genet. Mol. Biol. 2015, 14, 243–252. [Google Scholar] [CrossRef]
  35. Zhang, J.; Chen, C. On “A mutual information estimator with exponentially decaying bias” by Zhang and Zheng. Stat. Appl. Genet. Mol. Biol. 2018, 17, 20180005. [Google Scholar] [CrossRef]
  36. Williams, P.L.; Beer, R.D. Nonnegative decomposition of multivariate information. arXiv 2010, arXiv:1004.2515. [Google Scholar]
  37. Bertschinger, N.; Rauh, J.; Olbrich, E.; Jost, J.; Ay, N. Quantifying unique information. Entropy 2014, 16, 2161–2183. [Google Scholar] [CrossRef]
  38. Griffith, V.; Koch, C. Quantifying synergistic mutual information. In Guided Self-Organization: Inception; Springer: Berlin/Heidelberg, Germany, 2014; pp. 159–190. [Google Scholar]
  39. Tax, T.M.; Mediano, P.A.; Shanahan, M. The partial information decomposition of generative neural network models. Entropy 2017, 19, 474. [Google Scholar] [CrossRef]
  40. Wollstadt, P.; Schmitt, S.; Wibral, M. A rigorous information-theoretic definition of redundancy and relevancy in feature selection based on (partial) information decomposition. arXiv 2021, arXiv:2105.04187. [Google Scholar]
  41. Mori, T.; Nishikimi, K.; Smith, T.E. A divergence statistic for industrial localization. Rev. Econ. Stat. 2005, 87, 635–651. [Google Scholar] [CrossRef] [Green Version]
  42. Wang, Q.; Kulkarni, S.R.; Verdú, S. Divergence estimation for multidimensional densities via k-Nearest-Neighbor distances. IEEE Trans. Inf. Theory 2009, 55, 2392–2405. [Google Scholar] [CrossRef]
  43. Nguyen, X.; Wainwright, M.J.; Jordan, M.I. Estimating divergence functionals and the likelihood ratio by convex risk minimization. IEEE Trans. Inf. Theory 2010, 56, 5847–5861. [Google Scholar] [CrossRef]
  44. Zhang, Z.; Grabchak, M. Nonparametric estimation of Küllback-Leibler divergence. Neural Comput. 2014, 26, 2570–2593. [Google Scholar] [CrossRef]
  45. Press, W.H.; Teukolsky Saul, A. Numerical Recipes in Fortran: The Art of Scientific Computing; Cambridge University Press: Cambridge, UK, 1993. [Google Scholar]
  46. De Mántaras, R.L. A distance-based attribute selection measure for decision tree induction. Mach. Learn. 1991, 6, 81–92. [Google Scholar] [CrossRef]
  47. Kvalseth, T.O. Entropy and correlation: Some comments. IEEE Trans. Syst. Man Cybern. 1987, 17, 517–519. [Google Scholar] [CrossRef]
  48. Strehl, A.; Ghosh, J. Cluster ensembles—A knowledge reuse framework for combining multiple partitions. J. Mach. Learn. Res. 2002, 3, 583–617. [Google Scholar]
  49. Yao, Y. Information-theoretic measures for knowledge discovery and data mining. In Entropy Measures, Maximum Entropy Principle and Emerging Applications; Springer: Berlin/Heidelberg, Germany, 2003; pp. 115–136. [Google Scholar]
  50. Vinh, N.X.; Epps, J.; Bailey, J. Information theoretic measures for clusterings comparison: Variants, properties, normalization and correction for chance. J. Mach. Learn. Res. 2010, 11, 2837–2854. [Google Scholar]
  51. Zhang, Z.; Stewart, A.M. Estimation of Standardized Mutual Information; Technical Report; UNC Charlotte Technical Report: Charlotte, NC, USA, 2016. [Google Scholar]
  52. Zhang, Z.; Zhou, J. Re-parameterization of multinomial distributions and diversity indices. J. Stat. Plan. Inference 2010, 140, 1731–1738. [Google Scholar] [CrossRef]
  53. Chen, C. Goodness-of-Fit Tests under Permutations. Ph.D. Thesis, The University of North Carolina at Charlotte, Charlotte, NC, USA, 2019. [Google Scholar]
  54. Simpson, E.H. Measurement of diversity. Nature 1949, 163, 688. [Google Scholar] [CrossRef]
  55. Gini, C. Measurement of inequality of incomes. Econ. J. 1921, 31, 124–126. [Google Scholar] [CrossRef]
  56. Rényi, A. On measures of entropy and information. In Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability, Berkeley, CA, USA, 1 January 1961; Volume 1. [Google Scholar]
  57. Emlen, J.M. Ecology: An Evolutionary Approach; Addison-Wesley: Boston, MA, USA, 1977. [Google Scholar]
  58. Zhang, Z.; Chen, C.; Zhang, J. Estimation of population size in entropic perspective. Commun.-Stat.-Theory Methods 2020, 49, 307–324. [Google Scholar] [CrossRef]
  59. Beck, C.; Schögl, F. Thermodynamics of Chaotic Systems; Cambridge University Press: Cambridge, UK, 1995. [Google Scholar]
  60. Zhang, J.; Shi, J. Asymptotic Normality for Plug-In Estimators of Generalized Shannon’s Entropy. Entropy 2022, 24, 683. [Google Scholar] [CrossRef]
  61. Zhang, J.; Zhang, Z. A Normal Test for Independence via Generalized Mutual Information. arXiv 2022, arXiv:2207.09541. [Google Scholar]
  62. Kontoyiannis, I.; Skoularidou, M. Estimating the directed information and testing for causality. IEEE Trans. Inf. Theory 2016, 62, 6053–6067. [Google Scholar] [CrossRef]
  63. Huang, N.; Lu, G.; Cai, G.; Xu, D.; Xu, J.; Li, F.; Zhang, L. Feature selection of power quality disturbance signals with an entropy-importance-based random forest. Entropy 2016, 18, 44. [Google Scholar] [CrossRef]
  64. Brown, G.; Pocock, A.; Zhao, M.J.; Luján, M. Conditional likelihood maximisation: A unifying framework for information theoretic feature selection. J. Mach. Learn. Res. 2012, 13, 27–66. [Google Scholar]
  65. Lewis, D.D. Feature selection and feature extraction for text categorization. In Proceedings of the Speech and Natural Language: Proceedings of a Workshop Held at Harriman, New York, NY, USA, 23–26 February 1992. [Google Scholar]
  66. Battiti, R. Using mutual information for selecting features in supervised neural net learning. IEEE Trans. Neural Netw. 1994, 5, 537–550. [Google Scholar] [CrossRef] [PubMed]
  67. Yang, H.; Moody, J. Feature selection based on joint mutual information. In Proceedings of the International ICSC Symposium on Advances in Intelligent Data Analysis, Rochester, NY, USA, 22–25 June 1999; Volume 1999, pp. 22–25. [Google Scholar]
  68. Ullman, S.; Vidal-Naquet, M.; Sali, E. Visual features of intermediate complexity and their use in classification. Nat. Neurosci. 2002, 5, 682–687. [Google Scholar] [CrossRef] [PubMed]
  69. Yu, L.; Liu, H. Efficient feature selection via analysis of relevance and redundancy. J. Mach. Learn. Res. 2004, 5, 1205–1224. [Google Scholar]
  70. Tesmer, M.; Estévez, P.A. AMIFS: Adaptive feature selection by using mutual information. In Proceedings of the 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No. 04CH37541), Budapest, Hungary, 25–29 July 2004; Volume 1, pp. 303–308. [Google Scholar]
  71. Fleuret, F. Fast binary feature selection with conditional mutual information. J. Mach. Learn. Res. 2004, 5, 1531–1555. [Google Scholar]
  72. Peng, H.; Long, F.; Ding, C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef]
  73. Jakulin, A. Machine Learning Based on Attribute Interactions. Ph.D. Thesis, Univerza v Ljubljani, Ljubljana, Slovenia, 2005. [Google Scholar]
  74. Lin, D.; Tang, X. Conditional infomax learning: An integrated framework for feature extraction and fusion. In Proceedings of the European Conference on Computer Vision, Graz, Austria, 7–13 May 2006; pp. 68–82. [Google Scholar]
  75. Meyer, P.E.; Bontempi, G. On the use of variable complementarity for feature selection in cancer classification. In Proceedings of the Workshops on Applications of Evolutionary Computation, Budapest, Hungary, 10–12 April 2006; pp. 91–102. [Google Scholar]
  76. El Akadi, A.; El Ouardighi, A.; Aboutajdine, D. A powerful feature selection approach based on mutual information. Int. J. Comput. Sci. Netw. Secur. 2008, 8, 116. [Google Scholar]
  77. Guo, B.; Nixon, M.S. Gait feature subset selection by mutual information. IEEE Trans. Syst. Man-Cybern.-Part Syst. Hum. 2008, 39, 36–46. [Google Scholar]
  78. Cheng, G.; Qin, Z.; Feng, C.; Wang, Y.; Li, F. Conditional Mutual Information-Based Feature Selection Analyzing for Synergy and Redundancy. Etri J. 2011, 33, 210–218. [Google Scholar] [CrossRef]
  79. Singhal, A.; Sharma, D. Keyword extraction using Renyi entropy: A statistical and domain independent method. In Proceedings of the 2021 7th International Conference on Advanced Computing and Communication Systems (ICACCS), Coimbatore, India, 19–20 March 2021; Volume 1, pp. 1970–1975. [Google Scholar]
  80. R Package Entropy. Available online: https://cran.r-project.org/web/packages/entropy/index.html (accessed on 27 September 2022).
  81. R Package Bootstrap. Available online: https://cran.r-project.org/web/packages/bootstrap/index.html (accessed on 27 September 2022).
  82. R Package EntropyEstimation. Available online: https://cran.r-project.org/web/packages/EntropyEstimation/index.html (accessed on 27 September 2022).
Table 1. Estimation comparison between H ^ and H ^ z via simulation. In the simulation, the real underlying distribution is p i = i / 2001000 , where i = 1 , 2 , , 2000 (i.e., a triangle distribution). Under this setting, the true entropy H = 7.408005 . To compare the two estimators, 10,000 samples were independently generated following the triangle distribution for each of the six sample size settings in the table (i.e., we generate 60,000 random samples in total). The average values of H ^ and H ^ z under different sample sizes are summarized and reported in the table. The simulation shows that H ^ would consistently underestimate H more than H ^ z . The underestimation is more severe when the sample size is smaller.
Table 1. Estimation comparison between H ^ and H ^ z via simulation. In the simulation, the real underlying distribution is p i = i / 2001000 , where i = 1 , 2 , , 2000 (i.e., a triangle distribution). Under this setting, the true entropy H = 7.408005 . To compare the two estimators, 10,000 samples were independently generated following the triangle distribution for each of the six sample size settings in the table (i.e., we generate 60,000 random samples in total). The average values of H ^ and H ^ z under different sample sizes are summarized and reported in the table. The simulation shows that H ^ would consistently underestimate H more than H ^ z . The underestimation is more severe when the sample size is smaller.
n100300500100015002000
avg. of H ^ 4.565.576.006.516.756.89
avg. of H ^ z 5.116.096.496.927.117.21
Table 2. Selected information-theoretic feature selection methods reviewed in [16,64] with potential perspectives to improve the performance using smaller bias and/or variance estimation methods. The proposed criterion uses the same notations as their original forms for readers’ ease in tracing them back to the original articles, except some terms are equivalently re-denoted as M I ^ or κ ^ to help readers follow the same notations in the previous sections. To further clarify the notations, M I ^ ( X 1 , X 2 ) and M I ^ ( X 1 ; X 2 ) are the same. M I ^ ( X 1 , X 2 | Y ) is the plug-in estimator of conditional mutual information. By its definition, M I ^ ( X 1 , X 2 | Y ) = H ^ ( X 1 , Y ) + H ^ ( X 2 , Y ) H ^ ( X 1 , X 2 , Y ) H ^ ( Y ) . Similarly, M I ^ z ( X 1 , X 2 | Y ) = H ^ z ( X 1 , Y ) + H ^ z ( X 2 , Y ) H ^ z ( X 1 , X 2 , Y ) H ^ z ( Y ) , where each H ^ z could be further replaced using H ^ z with jackknife procedure.
Table 2. Selected information-theoretic feature selection methods reviewed in [16,64] with potential perspectives to improve the performance using smaller bias and/or variance estimation methods. The proposed criterion uses the same notations as their original forms for readers’ ease in tracing them back to the original articles, except some terms are equivalently re-denoted as M I ^ or κ ^ to help readers follow the same notations in the previous sections. To further clarify the notations, M I ^ ( X 1 , X 2 ) and M I ^ ( X 1 ; X 2 ) are the same. M I ^ ( X 1 , X 2 | Y ) is the plug-in estimator of conditional mutual information. By its definition, M I ^ ( X 1 , X 2 | Y ) = H ^ ( X 1 , Y ) + H ^ ( X 2 , Y ) H ^ ( X 1 , X 2 , Y ) H ^ ( Y ) . Similarly, M I ^ z ( X 1 , X 2 | Y ) = H ^ z ( X 1 , Y ) + H ^ z ( X 2 , Y ) H ^ z ( X 1 , X 2 , Y ) H ^ z ( Y ) , where each H ^ z could be further replaced using H ^ z with jackknife procedure.
MIM [65]
MIFS [66]
JMI [67]
Proposed Criterion (Score) M I ^
Different Estimation MethodUse M I ^ z with jackknife procedure
IF [68]Proposed Criterion (Score) M I ^ ( C , F )
Different Estimation MethodUse M I ^ z with jackknife procedure
FCBF [69]Proposed Criterion (Score) κ ^ 4
Different Estimation MethodUse κ ^ 4 z [51] with jackknife procedure
AMIFS [70]Proposed Criterion (Score) M I ^ C ; f i β s S κ ^ 6 C ; f s M I ^ f s ; f i
Different Estimation MethodUse M I ^ z and κ ^ 6 z [51] with jackknife procedure
CMIM [71]Proposed Criterion (Score) M I ^ ( Y ; X n ) and M I ^ ( Y ; X n | X m )
Different Estimation MethodUse M I ^ z with jackknife procedure
MRMR [72]Proposed Criterion (Score) max M I ^ x i ; c 1 | s | x i , x j S M I ^ x i , x j
Different Estimation MethodUse M I ^ z with jackknife procedure
ICAP [73]Proposed Criterion (Score) arg max X A ¯ M I ^ ( X ; Y ) + A A min { 0 , M I ^ ( X ; Y ; A ) }
Different Estimation MethodUse M I ^ z with jackknife procedure
CIFE [74]Proposed Criterion (Score) argmax θ t M I ^ y ( t ) ; c u = 1 t 1 M I ^ y ( u ) ; y ( t ) M I ^ y ( u ) ; y ( t ) c
Different Estimation MethodUse M I ^ z with jackknife procedure
DISR [75]Proposed Criterion (Score) arg max X i X S X j X S κ ^ 1 X i , j ; Y
Different Estimation MethodUse κ ^ 1 z [51] with jackknife procedure
IGFS [76]Proposed Criterion (Score) arg max X X s M I ^ ( X i ; Y ) + 1 d X j X S M I ^ X i ; X j ; Y
Different Estimation MethodUse M I ^ z with jackknife procedure
SOA [77]Proposed Criterion (Score) i M I ^ X i , Y i j > i M I ^ X i , X j + i j > i M I ^ X i , X j Y
Different Estimation MethodUse M I ^ z with jackknife procedure
CMIFS [78]Proposed Criterion (Score) M I ^ C ; f i f 1 + M I ^ f n ; f i C M I ^ f n ; f i f 1
Different Estimation MethodUse M I ^ z with jackknife procedure
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, J. Entropic Statistics: Concept, Estimation, and Application in Machine Learning and Knowledge Extraction. Mach. Learn. Knowl. Extr. 2022, 4, 865-887. https://doi.org/10.3390/make4040044

AMA Style

Zhang J. Entropic Statistics: Concept, Estimation, and Application in Machine Learning and Knowledge Extraction. Machine Learning and Knowledge Extraction. 2022; 4(4):865-887. https://doi.org/10.3390/make4040044

Chicago/Turabian Style

Zhang, Jialin. 2022. "Entropic Statistics: Concept, Estimation, and Application in Machine Learning and Knowledge Extraction" Machine Learning and Knowledge Extraction 4, no. 4: 865-887. https://doi.org/10.3390/make4040044

Article Metrics

Back to TopTop