Next Article in Journal
A Multi-Scale Entropy Approach to Study Collapse and Anomalous Diffusion in Shared Mobility Systems
Previous Article in Journal
Quantum Information Entropies on Hyperbolic Single Potential Wells
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Belief Entropy Tree and Random Forest: Learning from Data with Continuous Attributes and Evidential Labels

1
Department of Automation, University of Science and Technology of China, Hefei 230027, China
2
School of Electrical Engineering, University of Jinan, Jinan 250022, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(5), 605; https://doi.org/10.3390/e24050605
Submission received: 26 March 2022 / Revised: 20 April 2022 / Accepted: 23 April 2022 / Published: 26 April 2022
(This article belongs to the Topic Machine and Deep Learning)

Abstract

:
As well-known machine learning methods, decision trees are widely applied in classification and recognition areas. In this paper, with the uncertainty of labels handled by belief functions, a new decision tree method based on belief entropy is proposed and then extended to random forest. With the Gaussian mixture model, this tree method is able to deal with continuous attribute values directly, without pretreatment of discretization. Specifically, the tree method adopts belief entropy, a kind of uncertainty measurement based on the basic belief assignment, as a new attribute selection tool. To improve the classification performance, we constructed a random forest based on the basic trees and discuss different prediction combination strategies. Some numerical experiments on UCI machine learning data set were conducted, which indicate the good classification accuracy of the proposed method in different situations, especially on data with huge uncertainty.

1. Introduction

Decision trees have been widely used for their good learning capabilities and ease of understanding. In some real world issues, instances may be ill-known for some factors such as randomness, data incompleteness and even expert’s indefinite subjective opinions; however, traditional decision trees can only handle certain samples with precise data. The incompletely observed instances are usually ignored or replaced by a precise one, despite the fact that they may contain useful information [1], which may cause a loss of accuracy.
There have been many attempts to build trees from incomplete data in the past several decades. The probability trees [2,3] were suggested based on probability theory, which is usually intuitively the first tool to modeling uncertainty in practice; however, it has been proven that probability cannot always be adequate for representing data uncertainty [4,5] (often termed epistemic uncertainty). To overcome this drawback, various approaches have been proposed, including: fuzzy decision trees [6,7], the possibilistic decision trees [8] and the uncertain decision trees [9,10]. Besides the aforementioned methods, a more general framework, called the belief function theory [11,12] (also evidential theory or Dempster–Shafer theory), has been proven to have the ability to model all kinds of knowledge. The process of embedding belief functions within decision tree techniques has already been extensively investigated [13,14,15,16,17,18,19,20,21,22,23,24,25] in recent years. Particularly, among these methods, several trees [17,18,19] estimate parameters by maximizing evidential likelihood function using the E 2 M algorithm [26,27], which is also the basis of part of the trees to be proposed in this paper.
However, the existing methods on incomplete data do not take continuous attributes into full consideration. These proposals deal with uncertain data modeled by the belief function and build trees by extending the traditional decision tree method. The imitation and transformation decides to use existing methods to handle continuous attribute values by discretization, which brings about an issue of losing the detail of the training data. For example, the information gain ratio, the attribute selecting measurement in C4.5, was transformed to adapt the evidential labels of the training set in the Belief C4.5 trees [19], in which the continuous-valued attribute is divided into four intervals of equal width before learning. This issue leads to the purpose of this paper: to learn from uncertain data with continuous attribute values without pretreatment.
To realize this purpose, we firstly, for each attribute, fit the training data to a Gaussian mixture model (GMM), which consists of normal distribution models one-by-one corresponding to class labels, by adopting the E 2 M algorithm. This step, which significantly differs from other decision trees, confirms the ability to deal with ill-known labels and original attribute values (either discrete or continuous). On the basis of these GMM models, we generate the basic belief assignment (BBA) and calculate belief entropy [28]. The attribute with minimal average entropy, which distinguishes classes from each others most, will be selected as the splitting attribute. The following decision tree induction steps are designed accordingly and logically. To our knowledge, this paper is the first to introduce GMM models and belief entropy to decision trees with evidential data.
Another part of our proposal is adopting the ensemble method for our belief entropy trees. Inspired by the idea of building bagging trees based on random sampling [29], we further choose a more efficient and popular technique—random forest [30]. Under the belief function framework, the basic trees will output either precise or mass (modeled by BBA) label predictions, while traditional random forest can only combine precise labels. Thus, a new method to summarize the basic tree predictions is proposed to combine mass labels directly, instead of voting on precise labels. This combined mass keeps the uncertain information of data as much as possible, which helps to generate a more reasonable prediction. The new combination method is discussed and compared to the traditional majority voting method later.
We note that we have proposed our early work in a shorter conference paper [31]. Compared with our initial conference paper, we have fixed the attribute selection and splitting strategy of a single tree and introduced ensemble learning to our tree method in this paper.
Section 2 recalls some basic knowledge about decision trees, belief function theory, the E 2 M algorithm and belief entropy. Section 3 details the induction procedure of belief entropy methods and proposes three different instance prediction techniques. In Section 4, we introduce how to expend the single belief entropy tree to random forests and discuss the different predicting combination strategies. In Section 5, we detail experiments that were carried out on some classical UCI machine learning data sets to compare the classification accuracies of proposed trees and random forests. Finally, conclusions are summarized in Section 6.

2. Settings and Basic Definitions

The purpose of a classification method is to build a model that maps an attribute vector X = x 1 , , x D A 1 × A 2 × × A D , which contains D attributes, to an output class y C = C 1 , , C K taking its value among K classes. Each attribute discretely has finite values or continuously takes value within an interval. The learning of classification is based on a complete training set of precise data which contains N instances, denoted as
T = X 1 , y 1 X N , y N = x 1 1 , , x 1 D , y 1 x N 1 , , x N D , y N .
However, the imperfect knowledge about the inputs (feature vector) and the outputs (classification labels) exists widely in practical applications. Traditionally and regularly, the imperfect knowledge is modeled by probability theory, which is considered to be questionable in a variety of scenarios. Hence, we model uncertainty by belief function in this paper. Typically, we consider that attribute values are precise and can be either continuous or discrete, while only the output labels are uncertain.

2.1. Decision Trees

Decision trees [32] are regarded as one of the most effective and efficient machine learning methods and widely adopted for solving classification and regression problems in practice. The success, to a great extent, relays on the easily understandable structure, for both humans and computers. Generally, a decision tree is induced top-down from a training set T, which recursively repeats the steps below:
  • Select an attribute, through a designed selection method, to generate a partition of a training set;
  • Split the current training set to several subsets and put them into child node;
  • Generate a leaf node and determine the prediction label for a child node when a stop criterion is satisfied.
Differing in the attribute selection methods, several decision tree algorithms have been proposed, such as ID3 [32], C4.5 [33] and CART [34]. Among these trees, the ID3 and C4.5 choose entropy as an information measure to compute and evaluate the quality of a node split by a given attribute.
The core of ID3 is information gain. Given training data T and an attribute A with K A modalities, the information gain will be:
G a i n T , A = I n f o T I n f o A T
where
I n f o T = i = 1 K θ i log 2 θ i
and
I n f o A T = i = 1 K A T i T I n f o T i
where θ i is the proportion of instances in T that are of class C i , T and T i are the cardinalities of the instance sets belonging to a parent node and to the child node i.
The limitation of information gain is that attributes with largest values will be most promoted  [33], which leads to the G a i n R a t i o in the C4.5 algorithm. It is given as:
G a i n R a t i o T , A = G a i n T , A S p l i t I n f o T , A
where
S p l i t I n f o T , A = i = 1 K A T i T log 2 T i T .
The attribute with the largest gain ratio will be selected for splitting.
We can easily find the the Equation (2) is actually the Shannon Entropy. Yet in this paper, concerning the feature of evidential data described by the framework of belief function, the attribute selection method is newly designed based on belief entropy [28] instead of Shannon entropy.

2.2. Random Forest

To improve the classification accuracy and generalization ability of machine learning, the ensemble model method is introduced to the learning procedure. One important branch of ensemble method is called b a g g i n g , which concurrently builds multiple basic models learning from different training sets, which are generated from original data by bootstrap sampling. On the basis of bagging decision trees, random forest (RF) [30] not only chooses the training instance randomly but also introduces randomness into attributes selection. To be specific, traditional decision trees select the best splitting attribute among all D attributes; random forest generates a random attribute subset then chooses the best one within this subset to split the tree node. The size D of this subset is adjustable and generally set as D = l o g 2 D .
A detailed description of the mathematical formulation of RF model is found in [30]. The RF model consists of a union of multiple basic trees, where each tree learns from bootstrap samples and selects attribute from a small subset of all attributes. There some advantages of RF: (a) better prediction performance, (b) resistance to overfitting, (c) low correlation of individual trees, (d) low bias and low variance and (e) small computational overhead.
Some existing works have explored the ensemble method on belief decision trees, such as bagging  [29]. In this paper, we apply the random forest technique to the proposed belief entropy trees and discuss the different prediction determining strategies.

2.3. Belief Function Theory

Let the finite set Ω denote the frame of discernment containing k possible exclusive values that a variable can take. When considering the output y, the imperfect knowledge about value of y can be modeled by mass function m y : 2 Ω 0 , 1 , such that m y = 0 , and
A Ω m y A = 1 ,
which is also called a basic belief assignment (BBA). The subset A is called a focal set where m y A > 0 , and the m y A can be interpreted as the support degree of the evidence towards the case that true value is in set A.
There are some typical mass functions need to be attended:
  • V a c u o u s mass: mass function such that m y ( Ω ) = 1 , which means total ignorance;
  • B a y e s i a n mass: for all focal set A, the cardinality A = 1 . In this case, the mass degenerates to a probability distribution;
  • L o g i c a l ( c a t e g o r i c a l ) mass: m y ( A ) = 1 for some A. In this case, the mass is equivalent to the set A.
One-to-one related to the mass function m y , the b e l i e f f u n c t i o n and p l a u s i b i l i t y f u n c t i o n are defined as:
B e l y B = A B m y A ,
P l y B = A B m y A ,
which, respectively, indicate the minimum and maximum belief degree of evidence towards set B. Typically, the function p l : Ω 0 , 1 such that p l y ω = P l y ω for all ω Ω is called c o n t o u r f u n c t i o n associated to m y .
For two mass function m 1 and m 2 induced by evidences independently, they can be combined by the D e m p s t e r s r u l e  [12]⊕ defined as:
m 1 m 2 A = 1 1 κ B C = A m 1 B m 2 C ,
for all A Ω , A , and m 1 m 2 = 0 , where
κ = B C = m 1 B m 2 C ,
is called the d e g r e e o f c o n f l i c t between m 1 and m 2 . Obviously, Dempster’s rule is commutative and associative according to the definition.
In the decision making situation, we need to determine the most reasonable hypothesis from a mass. Different decision-making strategies with belief functions [35,36] have been researched. Among these methods, in the transferable belief model (TBM), p i g n i s t i c p r o b a b i l i t y  [37] was proposed to make decision from a BBA:   
B e t P ω = A Ω , ω A m A A ,
where A is the cardinality of A.
When we model uncertain labels of evidential data with mass functions, the training set becomes
T = X 1 , m 1 X N , m N = x 1 1 , , x 1 D , m 1 x N 1 , , x N D , m N .

2.4. Evidential Likelihood

Consider a discrete random vector Y taking values in Ω with a probability mass function p Y y ; θ assumed to be associated with a parameter θ Θ . After a realization y of Y has been perfectly observed, the likelihood function of c o m p l e t e d a t a is defined as L : Θ 0 , 1 such that
L θ ; y = p Y y ; θ , θ Θ .
When the observations are uncertain, it is impossible to evaluate parameter θ from a likelihood function. In this situation, a new statistical tool [38] called evidential likelihood was proposed to perform parameter estimation. Assume that y is not precisely observed, but is known surely that y A for some A Ω . Given such i m p r e c i s e d a t a , the likelihood function will be extended to
L θ ; A = p Y A ; θ = y A p Y y ; θ , θ Θ .
Furthermore, the observation of instance y could be not only i m p r e c i s e , but also u n c e r t a i n , which is modeled by mass function m y . Thus the evidential likelihood function [27] can be defined as
L θ ; m y = A Ω L θ ; A m y A = y Ω p Y y ; θ A x m y A = y Ω p Y y ; θ p l y , θ Θ ,
where the p l is the contour function related to m y and the L θ ; m y can be remarked as L θ ; p l . According to the statement of Denoeux [27], the value 1 L θ ; p l equals to the conflict between parametric model p Y y ; θ and the uncertain observation p l , which means minimizing L θ ; p l is actually a procedure of estimating the best parameter θ to fit the parametric model to observation as closely as much.
Equation (14) also indicates that L θ ; p l can be remarked as the expectation of p l such that
L θ ; p l = E θ p l Y .
Assume that Y = y 1 , , y N is a sample set containing n cognitively independent [12] and i.i.d. uncertain observations, in which the y i is model by m y i . In the situation the Equation (15) is written as a product of n terms:
L θ ; p l = n = 1 N E θ p l n y n .

2.5. E 2 M Algorithm

Though an extension of likelihood function, the maximum likelihood estimation of evidential likelihood can not directly be computed by the broadly applied EM algorithm [39]. The E 2 M algorithms [27] introduced by Denoeux allow us to maximize the evidential likelihood iteratively, which is composed of two steps (similar to EM algorithm):
1.
The E-step require firstly a probability mass function p Y · p l ; θ q = p Y · ; θ q p l , in which the former part means the probability mass function of Y under the parameter θ q estimated from last iteration and the latter part indicates contour function p l . The expression is:
p Y y p l ; θ q = p Y y ; θ q p l y L θ q ; p l .
Then calculate the expectation of log likelihood log L c θ ; y = log p Y y ; θ of complete data with respect to p Y · p l ; θ q ,
Q θ , θ q = y Ω log L c θ ; y p Y y ; θ q p l y L θ q ; p l .
2.
The M-step is to maximize Q θ , θ q with respect to θ , obtaining a new estimation that ensures Q θ q + 1 , θ q Q θ , θ q .
The two steps repeat until L θ q + 1 L θ q ϵ , where ϵ is a set threshold.

2.6. Belief Entropy

Inspired by Shannon entropy [40], which can measure uncertainty contained by a probability distribution, a type of belief entropy called Deng entropy is proposed by Deng [28] to handle situation where the traditional probability theory is limited. When the uncertain information is described by the basic belief assignment instead of the probability distribution, Shannon entropy cannot work. Deng entropy is defined on the belief function frame, which makes it able to measure uncertain information described by the BBA efficiently.
Let A be the focal set of belief function, and A be the cardinality of A. Deng entropy E is defined as:
E m = A Ω m A log m A 2 A 1 .
We can easily learn from the definition that if the mass function is B a y e s i a n , which means A = 1 for all A, Deng entropy degenerates to Shannon entropy such that
E m = A Ω m A log m A .
The greater the cardinality of the focal set is, the bigger the corresponding Deng entropy is, so that the evidence imprecisely refers to more single elements. Thus, significant Deng entropy indicates huge uncertainty. Powered by this feature, we calculate the average Deng entropy of BBAs to select the best attribute leading to the least uncertainty. The details are shown in the next section.

3. Design of Belief Entropy Trees

Up to now, various decision tree methods have been proposed to deal with evidential data, but many of them consider categorical attributes and transform the continuous attribute values into discrete categories. Some recent works fit the continuous attributes with same class labels into normal distributions [41] and generate BBA from the normal distributions to select the best splitting attribute by calculating belief entropy [42]; however, this method divides samples into each set of certain classes, which can only handle the precise class labels. Our goal is to develop a belief decision tree method learns from data set with continuous and precise attribute values but incomplete class labels directly and efficiently.
This section explains our method in detail, specifically focusing on the procedure of attribute selection. Corresponding splitting strategy, stopping criterion and the leaf structure are also well-designed to accomplish the whole belief entropy decision tree.

3.1. The Novel Method to Select Attribute

The learning procedure of decision trees is generally to decide the split attribute and to decide how to split on this attribute on each node; our method also proceeds in this manner. As a novel decision tree, the most characteristic and core part of our method is the attribution selection, which includes three steps: firstly, for each attribute, fit the values to normal distributions corresponding to each class label, in another words, fit attribute values into K×D normal distribution models, where K is the class number and D is the attributes number of instances; secondly, for every instance, generate D BBAs from each attribute according to the normal distribution-based models; finally, calculate belief entropy from BBAs for each attribute. The attribute with minimum belief entropy will be selected to split.

3.1.1. Parameter Estimation on Data with Continuous Attributes

Powered by the idea of extracting BBAs from normal distribution-modeled attribute values [41], we try to operate similarly on data with ill-known class labels. In the situation that each instance exactly belongs to one class, the d-th attribute values set x 1 d , , x N d is divided into K subsets x n d y n = C k , k = 1 , , K corresponding to each class. It is easy to fit each subset to the normal distribution by calculating means and standard deviations.
Example 1.
Consider the Iris data set [43], a classical machine learning data set, which contains 150 training instances of three classes: ‘Setosa’, ‘Versicolor’, ‘Virginica’, with four attributes: sepal length(SL), sepal width(SW), petal length(PL) and petal width(PW). For the values of attribute SL in the class of Setosa, we can directly calculate the mean value as μ = 5.0133 and standard deviation as σ = 0.3267 . Similarly, we can obtain normal distribution parameters of class of Versicolor and Virginica. Figure 1 shows the normal distribution model of Iris data set for the SL attribute in three classes.
However, when the labels of training set are ill-known, some instances can not be allocated to a certain class assertively. The evidential likelihood and E 2 M algorithm introduced in Section 2 make it possible to generate an estimation of model parameters. Because the E 2 M algorithm uses only contour functions, the label of n-th instance will be represented by plausibility p l n = p l n k , k = 1 , , K instead of mass function m n .
For the purpose of comparing attributes, we split the whole training data into D attribute–label pairs and handle D parameter estimation problems. Consider the d-th attribute value vector X d = x 1 d , , x N d T , d 1 , , D , we assume the conditional distribution of X d when given y = C k is normal with mean μ k and standard deviation σ k :
X d | ( y = C k ) N μ k , σ k 2 , k = 1 , , K .
Actually the assumption is to build a one-dimensional G a u s s i a n m i x t u r e m o d e l ( G M M )  [44]. Similar to the application of E 2 M algorithm in l i n e a r d i s c r i m i n a n t a n a l y s i s  [45], the following discuss is practically to adopt E 2 M algorithm to estimate parameters in GMM.
Let π k be the marginal probability when y = C k , and θ = ( μ 1 , , μ K , σ 1 , , σ K , π 1 , , π K ) the parameter vector to be estimated. The complete-data likelihood is
L c θ = n = 1 N p x n d Y n = y n p y n = n = 1 N k = 1 K ϕ x n d ; μ k , σ k y n k π k y n k ,
where the ϕ is normal distribution probability density,
ϕ x ; μ , σ = 1 2 π σ e x μ 2 2 σ 2 ,
and y n k is a binary indicator variable, such that y n k = 1 if y n = C k and y i k = 0 if y N C k .
when expended to evidential data, where we use contour function to describe the labels, the evidential likelihood is drew from Equation (16) that,
L θ = n = 1 N E θ p l n y n = n = 1 N k = 1 K p l n k ϕ x n d ; μ k , σ k π k .
According to the E 2 M algorithm, we compute the expectation of complete-data log likelihood
c θ = log L c θ = n = 1 N k = 1 K y n k log ϕ x i d ; μ k , σ k + log π k
with respect to the combined mass probability function
p x d p l ; θ q = n = 1 N p x n d p l n ; θ q .
To simplify the equation, we denote
ζ n k q = p x n d p l n ; θ q = p l n k π k q ϕ x n d ; μ k q , σ k q k = 1 K p l n k π k q ϕ x n d ; μ k q , σ k q .
Finally, we obtain the to-be-maximized function
Q θ , θ q = n = 1 N k = 1 K ζ n k q log ϕ x n d ; μ k , σ k + log π k
in the E-step.
The formal of Q θ , θ q is similar to the function computed in the EM algorithm on the GMM [44]. Because of the similarity, we imitate it and learn that the optimal parameter maximizing Q θ , θ q can be iteratively computed by
π k q + 1 = 1 n n = 1 N ζ n k q , μ k q + 1 = n = 1 N ζ n k q x i d n = 1 N ζ n k q ,
σ k q + 1 = n = 1 N ζ n k q x n d μ k q + 1 2 n = 1 N ζ n k q
Finally when L θ q + 1 L θ q ϵ is satisfied for some ϵ , stop the iteration and remark θ q + 1 as θ d = μ 1 d , , μ K d , σ 1 d , , σ K d , π 1 d , , π K d , which is the estimation of parameters in the GMM extracted from d-th attribute. Repeat this procedure for every attributes of the training set, D × K normal distribution
N k d μ k d , σ k d 2 , d = 1 , , D , k = 1 , , K
will be generated.
The Algorithm 1 shows the procedure of parameter estimation and there is Example 2 to help understand it.
Algorithm 1 Parameter estimation of GMMs.
Input: evidential training set T p l = ( x , p l y ) , iteration stop threshold ϵ
Output: estimated normal distribution parameter matrix μ k d , σ k d , d = 1 , , D , k = 1 , , K
1:
for each attribute A d  do
2:
 initialize parameters as θ d ( 0 ) = μ 1 d ( 0 ) , , μ K d ( 0 ) , σ 1 d ( 0 ) , , σ K d ( 0 ) , π 1 d ( 0 ) , , π K d ( 0 ) ;
3:
q = 0 ; Initialize loop variable.
4:
for q do
5:
   update the estimation of parameters θ d ( q + 1 ) = μ 1 d ( q + 1 ) , , μ K d ( q + 1 ) , σ 1 d ( q + 1 ) , , σ K d ( q + 1 ) , π 1 d ( q + 1 ) , , π K d ( q + 1 ) by Equations (28) and (29).
6:
   if  L θ d q + 1 L θ d q ϵ  then
7:
    break; {End the loop if evidential likelihood increment is less than threshold.}
8:
   end if
9:
    q = q + 1 ;
10:
  end for
11:
  adopt μ k d ( q + 1 ) , σ k d ( q + 1 ) , k = 1 , , K as estimated normal distribution parameters under attribute A d ;
12:
end for
Example 2.
Consider the Iris data set mentioned in Example 1. To simulate the situation that labels of training set are not completely observed, we manually introduce uncertainty to the Iris data. In this example, we set that each instance has an equivalent chance (25%) to be v a c u o u s , i m p r e c i s e , u n c e r t a i n or completely observed (the detail of transformation is discussed in Section 5). Table 1 shows the attribute values and labels described by plausibility p l of some instances in evidential Iris data. Table 2 shows the mean and standard deviation pairs μ , σ calculated by E 2 M algorithm. Figure 2 shows curves of these models.

3.1.2. BBA Determination

This step is to generate D BBAs corresponding to each attribute for every instance in the training set.
Choose an instance I n with attribute vector x n = x n 1 , , x n D from the data set, calculate the intersection of x n d d = 1 , , D and the K normal distribution functions ϕ k d = ϕ x d ; μ k , σ k , k = 1 , , K , i.e., we obtain K normally distributed probability density function (PDF) values for the attribute A d and instance I n , denoted as ϕ n k d , k = 1 , , K .
Due to the property that the probability of a value x sampling from a normal distribution is proportional to the PDF ϕ x , we can infer, for the attribute d, the probability that instance x n belongs to each class is proportional to ϕ n k d = ϕ x n d ; μ k , σ k , k = 1 , , K . From this opinion of statistical analysis, the rule to assign normal PDFs to some sets was proposed to build BBAs.
Firstly, normalize the ϕ n k d with different class k such that
f k = ϕ n k d / k = 1 K ϕ n k d .
Then rank f k in decreasing order f r r = 1 , , K , whose corresponding class is denoted as C r r = 1 , , K . Assign f r to the class set by the following rule:
m C 1 = f 1 m C 1 , C 2 = f 2 m C 1 , , C K = m θ = f K .
If f i = f i + 1 = = f j , then m C 1 , , C j = p = i j f p . By this rule, we obtain a nested BBA of x n under the select attribute A d , which we denote as m n d .
Example 3.
Consider the first instance of the evidential Iris data set showed in Table 1, whose attributes are:
x S L = 5.1 , x S W = 3.5 , x P L = 1.4 , X P W = 0.2 .
For attribute SL, the intersections of x S L = 5.1 and three normal distributions are shown in Figure 3 such that
ϕ S e t o s a S L ( x S L ) = 1.0712 , ϕ V e r s i c o l o r S L ( x S L ) = 0.2354 , ϕ V i r g i n i c a S L ( x S L ) = 0.0331 .
The reader can see in the figure that this instance is closest to class ‘Setosa’, then to the ‘Versicolor’ and ‘Virginica’. Thus, we generate BBA from intersection values, which is intuitive. The BBA is assigned as:
m S e t o s a = 1.0712 1.0712 + 0.2354 + 0.0331 = 0.7996 m S e t o s a , V e r s i c o l o r = 0.2354 1.0712 + 0.2354 + 0.0331 = 0.1757 m S e t o s a , V e r s i c o l o r , V i r g i n i c a = 0.0331 1.0712 + 0.2354 + 0.0331 = 0.0247
Similarly, we build BBAs for the rest of the attributes—shown in Table 3.

3.1.3. Calculation of Belief Entropy

The last step to determine splitting attribute is to calculate the average Deng entropy
E d = E A d = 1 N n = 1 N E m n d , d = 1 , D
of all instances for each attribute. As mentioned in Section 2.6, Deng entropy measures the uncertain degree contained by BBA, which means the less E d , the more certainty the BBAs contain, and the more separate the division of classes is. Consequently, we choose the attribute A * that minimizes the average Deng entropy such that
A * = arg min A d E A d , d = 1 , , D
to be the best splitting attribute to proceed the tree building.
Example 4.
Continue the Examples 2 and 3. Calculate Deng entropy of BBAs of selected instance shown in Table 3:
E S L m = 0.7996 × log 0.7996 2 1 1 0.1757 × log 0.1775 2 2 1 0.0247 × 0.0247 2 3 1 = 0.2933 E S W m = 0.6904 × log 0.6904 2 1 1 0.2329 × log 0.2329 2 2 1 0.0767 × 0.0767 2 3 1 = 0.3687 E P L m = 1 × log 1 2 1 1 = 0 E P W m = 1 × log 1 2 1 1 = 0
Similarly proceed same calculation to all instances so that average Deng entropy for attributes are calculated that
E A S L = 1.3853 , E A S W = 2.0837 , E A P L = 0.4275 , E A P W = 0.2116 .
According to this result, attribute PW will be chosen to generate child nodes.
Comparing the Deng entropy with the curves in Figure 2, we can intuitively learn that PW has the most distinctive curves for each class, yet curves in SW overlap each other a lot, which conforms to the size of the average Deng entropy above, where PW is the lowest and SW is the highest.
As a matter of fact, Examples 1–4 in this chapter can be orderly combined as a whole calculating example, which shows the procedure of the proposed attribute selecting method.

3.2. Splitting Strategy

The splitting strategy is redesigned according to the selected attribute A * to fit the proposed attribute selection method. Branches will be associated to each class, that is to say, each node to be edged will have K branches. For an instance I n , consider the generated BBAs, the class corresponding to the maximum mass value will be the branch that this instance shall be put into. To put it simply, when splitting the tree under attribute A * , the instance I n will be assigned into the k n - t h child node, where the k n satisfies
k n = arg max k ϕ n k * x n d .
The Algorithm 2 summarizes the procedure of selecting attribute and splitting. It should be mentioned that, though the child nodes are associated to each class, this splitting strategy does not mean to determine the affiliation of instances directly and arbitrarily in this step.
Algorithm 2 Attribute selection and splitting.
Input: evidential training set T p l = ( x , p l y ) , possible splitting attribute A = A 1 , , A D
Output: selected attribute A * , instance sets in child nodes T i , i = 1 , , K
1:
compute the normal distribution parameters μ k d , σ k d for each A d and C k by E 2 M algorithm;
2:
for each attribute A d  do
3:
for each instance I n  do
4:
  generate BBA m n d from normal distributions N k d μ k d , σ k d 2 , k = 1 , , K ;
5:
   E n d = E m n d ; {Calculate Deng entropy for all generated BBAs}
6:
end for
7:
E A d = A v e r a g e E n d ;  {Calculate average Deng entropy for each attribute}
8:
end for
9:
split on attribute A * = arg min A d E A d ; {The attribute with minimum average entropy is selected}

3.3. Stopping Criterion and Prediction Decision

After designing the attribute selection and partitioning strategy, we split each decision node to several child nodes. This procedure repeats iteratively until one of the stop criterion is met:
  • No more attributes for selection;
  • The number of instances in the nodes falls below a set threshold;
  • The labels of instances are all precise and fall into the same class;
When the tree building stops at a leaf node L, a class label should be determined to predict the instances that fall into this node. We design two different prediction methods such that:
  • The first one is to generate the prediction label from the original training labels of instances contained by this node, which is a similar treatment to traditional decision trees such as C4.5 tree method. Denoting the instances in the leaf node by I 1 , I P and the corresponding evidential training labels by p l p , p = 1 , P , the leaf node will be labeled by C ^ , where
    C ^ = arg max C k p = 1 P p l p C k , k = 1 , , K ,
    which means the class label with maximal plausibility summation will represent this node. This tree predicts from the original labels of training set, which is called O r i n g i n - p r e d i c t i o n b e l i e f e n t r o p y t r e e ( O B E t r e e ) for short in this paper.
  • The first method described above in fact abandons the generated BBAs during the tree build procedure, which will be adopted to generating predicted instance label in the second method. Firstly, the splitting attributes list, which lead instance I to the leaf node from top to down, are denoted by A 1 * , , A Q * , and the BBAs generated accordingly are denoted by m 1 * , m Q * . Then combine these BBAs by Dempster rule, such that m ^ = m 1 * · · · m Q * , to predict the training instance. On this basis, we continue to combine generated BBAs of all instances in a leaf node such that m ^ l e a f = m ^ 1 · · · m ^ P , where the once again combined BBA m ^ l e a f will be the mass prediction label for the whole leaf node. To obtain a precise label for another choice, the last step is making decision on BBA by choosing the class label with maximal pignistic probability computed by Equation (11). We call this tree a L e a f - p r e d i c t i o n b e l i e f e n t r o p y t r e e ( L B E t r e e ) in this paper.
The Algorithm 3 summarizes the induction of belief entropy trees introduced in this section.
Algorithm 3 Induction of belief entropy trees (BE-tree).
Input: evidential training set T p l , classifier type T Y P E
Output: belief entropy tree T r e e
1:
construct a root node containing all instances T p l ;
2:
if stopping criterion is met then
3:
if  T Y P E = O B E  then
4:
  output precise prediction generated from original plausibility label for the whole node;
5:
else if  T Y P E = L B E   then
6:
  combine BBAs generates during each splitting m ^ = m 1 * · · · m Q * for each instance;
7:
  combine BBAs of all instances in previous node generated in step 6 that m ^ l e a f = m ^ 1 · · · m ^ P ;
8:
  output m ^ l e a f as a mass prediction for the whole leaf node;
9:
  output C ^ = P i g n i s t i c m ^ l e a f as a precise prediction for the whole leaf node;
10:
end if
11:
 return T r e e =root node;
12:
else
13:
 apply Algorithm 2 to select splitting attribute A * ;
14:
 induce each subset T p l c h i l d based on A * ;
15:
for all T p l c h i l d  do
16:
    T r e e c h i l d = B E - t r e e T p l c h i l d ; {Recursively build the tree on the new child node}
17:
   attach T r e e c h i l d to the corresponding T r e e ;
18:
  end for
19:
end if

3.4. An Alternative Method for Predicting New Instance

Two types of belief entropy trees, the OBE tree and the LBE tree, have been described in detail in the last section. Similar to traditional decision trees, a new instance will be classified in a top-down way: starting at the root node and following branches by considering its generated BBA under splitting attribute until reaching a leaf node. The prediction of leaf node will be given to this new instance.
However, differing from the idea of collecting the numerous ‘opinions’ of instances, another method to predict a new instance is considered after a tree has been built. In Section 3.1.2, we introduced how to generate each training instance’s BBA corresponding to attributes. In the same way, we can generate m 1 * , m Q * corresponding to an attributes list A 1 * , , A Q * , which orderly splits and leads the new instance to a leaf node. Then, we combine these BBAs such that m ^ = m 1 * · · · m Q * to predict the new testing instance. It is easy to find that this method performs the same way as the front part of label prediction in LBE trees, yet stops when obtaining a mass prediction from the testing instance’s own attribution values instead of the leaf node it belongs to, which also means testing instances in a same leaf node normally have different mass prediction under this design. For the sake of narrative, a tree predicting in this way is called I n s t a n c e - p r e d i c t i o n b e l i e f e n t r o p y t r e e ( I B E t r e e ) in this paper.
Figure 4 and Figure 5 show the procedure of making prediction on leaf node, where Figure 4 is the generation of mass prediction m ^ for each instance, in whether training set or testing set; Figure 5 details the different prediction making in the proposed three belief entropy trees.

4. Belief Entropy Random Forest

We have introduced the induction of belief entropy trees in the previous section, which is regarded as the basic classifier of random forest ensemble method in the following discussion.
The generalization ability of random forest draws from not only the perturbation of sampling, but also the perturbation of attributes selecting. Specific to the proposed belief entropy random forest, for each basic tree, we firstly performs bootstrap sampling on the original training set, which means randomly sampling with replacement for N times on the set T where T = N . Secondly, when training on this resampling set, for each to-be-split node, the best splitting attribute will be chosen from a subset A i , i = 1 , , D of the set of all available attributes A j , j = 1 , , D , where 1 < D < D . If D = D , the basic tree splits totally, the same as the belief entropy tree; while D = 1 means randomly selecting an attribute to split all the time.
Repeat the first and second steps above S times then a ‘forest’ containing variable basic trees will be constructed, where the repeat time S is called forest size. When making a prediction of a new instance on this forest, S primary predictions will be independently generated by S basic trees and finally summarized to one result. It should be mentioned that the OBE tree output precise label directly for testing instances while the LBE trees and IBE trees can provide mass labels described by BBAs or precise labels. This feature inspires two different strategies for making predictions in the last step: the majority voting for precise labels and belief combination of mass labels.
Algorithm 4 shows the procedure of building the complete evidential random forests based on belief entropy trees. Selecting different ensemble prediction strategies and base tree types, we build five random forest lists below:
  • label-voting O B E Random Forest ( L - O B E R F ) , which performs majority voting on precise outputs of OBE trees;
  • label-voting L B E Random Forest ( L - L B E R F ) , which performs majority voting on precise outputs of LBE trees;
  • mass-combination L B E Random Forest ( M - L B E R F ) , which combines BBAs generated by LBE trees and makes decision;
  • label-voting I B E Random Forest ( L - I B E R F ) , which performs majority voting on precise outputs of IBE trees;
  • mass-combination I B E Random Forest ( M - I B E R F ) , which combines BBAs generated by IBE trees and makes decision;
Algorithm 4 Building procedure of evidential random forests.
Input: evidential training set T p l , new instance x , base classifier number h, base classifier type T Y P E , base classifier output mode O
Output: predicted label y ^
1: 
for  i = 1 : h  do
2:
T i = R a n d o m A t t r i b u t e S a m p l i n g R a n d o m I n s t a n c e S a m p l i n g T p l ; {The resampling procedure of each base tree.}
3:
if  T Y P E = O B E  then
4:
   T r e e i = O B E T i ;
5:
else if  T Y P E = I B E  then
6:
   T r e e i = I B E T i ;
7.
else if  T Y P E = L B E  then
8:
   T r e e i = L B E T i ;
9:
end if
10:
end for
11:
for  i = 1 : h  do
12:
L i = L a b e l P r e d i c t i o n T r e e i , x ; {Generate predict labels of each base tree.}
13:
end for
14:
if  O = p r e c i s e l a b e l then
15:
y ^ = M a j o r i t y L 1 , , L h ; {Generate prediction from precise labels.}
16:
else if  O = m a s s l a b e l  then
17:
y ^ = P i g n i s t i c M a s s C o m b i n a t i o n L 1 , , L h ; {Generate prediction from mass labels.}
18:
end if
Figure 6 shows the procedure of constructing the forests, in which the Figure 6a shows generation of basic trees in a random forest, and Figure 6b shows different procedure of combining the final prediction in five forests, which will lead to a different classification performance. We will evaluate them in the next section.

5. Experiments

In this section, we detail experiments to evaluate the performance of the proposed decision tree method. The experiment settings and results are detailed below.

5.1. Experiment Settings

As there are no widely accepted evidential data sets to measure the proposed method, it is necessary to generate a data set with ill-known labels from machine learning databases taken from the UCI repository [46]. We selected several data sets, including: Iris, Wine, Balance scale, Breast cancer, Sonar and Ionosphere.
Denote the true label of a instance by C i , and give its uncertain observation m y i . Due to the characters of belief function, we can simulate several situations from precise data:
a p r e c i s e observation is such that p l y i C i = 1 , and p l y i C j = 0 , C j C i * ;
a v a c u o u s observation is such that p l y i C j = 1 , C j C ;
an i m p r e c i s e observation is such that p l y i C j = 1 if C j = C i * or C j C r m , and p l y i C j = 0 otherwise, where C r m is a set of randomly selected labels;
an u n c e r t a i n observation is such that p l y i C i * = 1 , and p l y i C j = r j , C j C i * , where r j are sampled independently from uniform distribution U 0 , 1 .
To observe the performance on evidential training data sets with different ill-known types and incomplete degrees, we set three variables, vacuousness level V 0 , 1 , imprecision level I 0 , 1 and uncertainty level U 0 , 1 , to adjust the generation procedure, where V + I + U 1 .
Example 2 shows the transformed Iris data set and listed part of instances in Table 1. In this example, labels of no.53 and no.54 instance are vacuous; labels of no.1 and no.2 instance are imprecise; labels of no.4 and no.52 instance are uncertain.
To improve the reliability and reduce the stochasticity, we performed 5-fold cross-validation on each data set and repeat ten times to compute an average classification accuracy for all experiments. Different tree induction techniques will be compared:
  • t r a d i t i o n a l C 4.5 t r e e , which only uses precise data in the training set during tree induction;
  • b e l i e f e n t r o p y t r e e s described in Section 3.3: OBE tree, LBE tree, IBE tree;
  • b e l i e f e n t r o p y r a n d o m f o r e s t described in Section 4:
    -
    l a b e l - v o t i n g O B E r a n d o m f o r e s t ( L - O B E R F ) ;
    -
    l a b e l - v o t i n g L B E r a n d o m f o r e s t ( L - L B E R F ) ;
    -
    m a s s - c o m b i n a t i o n L B E r a n d o m f o r e s t ( M - L B E R F ) ;
    -
    l a b e l - v o t i n g I B E r a n d o m f o r e s t ( L - I B E R F ) ;
    -
    m a s s - c o m b i n a t i o n I B E r a n d o m f o r e s t ( M - I B E R F ) ;
We set the maximal size of the leaf node as T / 20 to avoid overfitting in the belief entropy trees. In the random forests, the forest size was set as 50, and the size of the attributes subset was set as D = l o g 2 D .

5.2. Experiments on Vacuous Data

Assuming part of the instances in the training set are totally unobserved while others are completely observed, we performed experiments with different vacuousness levels V 0 , 1 while I = U = 0 . Generating the training sets and learning on them, the results are shown in Figure 7.
Firstly, we observe the figure as a whole. Obviously, whatever the tree induction method is, it is impossible to learn from data sets whose instances are all vacuous. Thus, the accuracy of all trees decreases gradually as V increases, yet drops sharply when the V approaches nearly to 1. On the contrary, almost all curves keep steady or decrease slightly before the vacuousness level reaches 80%, except for the OBE trees. Table 4 shows the accuracy results when V equals 90%.
Considering the basic belief entropy trees firstly, the LBE trees and IBE trees perform, most of the time, at least as well as the traditional C4.5 decision trees, and better than the traditional decision trees for some time, especially when encountering high vacuousness level V; however, the OBE preforms elusively on different data sets: it has the lowest classification accuracy in Iris, Wine and Ionosphere data set; however, it achieves better results in the Balance scale. It is possible that if all samples in a leaf node are vacuous, the direct combination of all the training labels stays vacuous, which led to the shortage of OBE tree.
It can be observed that the belief entropy random forests perform well overall for their improvement in classification accuracy compared to the corresponding basic tree and the slower accuracy decent rate as V increases. Among these forests, the ones based on IBE and making prediction by mass combination performs better than others in nearly all data sets except the Balance scale.

5.3. Experiments on Imprecise Data

The second situation is that some data are imprecisely observed, i.e., the observation is a set value, while the true value lies in this set (called superset labels [47] in some works). As mentioned before, imprecision level I controls the percentage of imprecise observations.
For the instance to be imprecise, we randomly generate a number z k 0 , 1 for each class C k except the true one. Plausibility of labels with z k < I will be set to 1. When the I = 1 , a training set becomes totally imprecise, which is, in practice, the same situation as total vacuousness; while I < 1 , instances are in a middle state of transition from precise to vacuous, which indicates a piece of similarity between the vacuous training set and the imprecise training set, i.e., we can tell that the imprecise sample contains more information than the totally vacuous ones. As a result, we can see in Figure 8, that curves of accuracy with changing I are similar to those in experiments with vacuousness in Figure 7, yet more smooth and full.
According to the Table 5, the proposed methods keep pretty good classification results under high-level imprecise observations. OBE still keeps the shortage in almost all data sets while LBE and IBE achieve similar performance. M- I B E R F keeps its advantage in most situations, especially in the Iris and Breast Cancer data; the classification accuracy is almost equal to the results on the total precise training set. The balance scale is a particular case to be discussed later.

5.4. Experiments on Uncertain Data

Another type of ill-known label is the uncertain one, which is measured by a plausibility distribution, with the true label having the highest chance among all class labels. To evaluate the performance of the proposed trees and forests in a more general situation with uncertainty, we set U 0 , 1 and V = I = 0 . For instance, to be transformed into an uncertain one, we assign a value 1 to the plausibility of the true label and random values averagely sampled from 0 , U to other labels.
Despite the inability to handle total vacuousness and imprecision, the belief entropy trees have the ability to learn from totally uncertain training data sets. The horizontal curves in Figure 9 indicate all methods proposed in this paper keep stable performance with changing uncertainty level U. On the whole, we can learn from the figure that LBE and IBE perform equally well and better than OBE as a single tree in most data sets, except in the Balance scale.
Considering the forests, for the good attribute normality of Iris, Wine and Breast cancer data, classification accuracies of the five forests on these data sets have similar performance according to Table 6, leading to a heavy overlap of curves in figure. Among these trees, the OBE trees achieve the most significant improvement by building random forest; this improvement helps OBE-RF to surpass other forests in the Ionosphere, Sonar and Balance scale data sets. Particularly, in the Balance scale, the accuracy of OBE-RF even increases slightly as the U decreases, which can be partially explained by the fact that uncertain instances are more informative then absolutely precise instances.

5.5. Summary

By carrying out experiments on training sets with different types and degrees of incomplete observation, we can conclude that the LBE trees and IBE trees, along with four types of random forests based on them, generally possess excellent learning ability on data with ill-known labels. Among the RFs, the ensemble of the IBE tree, L-IBE-RF and M-IBE RF achieve the highest classification accuracy in most situations except on samples with high uncertainty levels, especially on the Balance scale data set. We think there are two reasons: (a) compared to vacuous and imprecise samples, the learning labels of uncertain samples are more information rich, while the OBE use the learning labels to predict directly; (b) the attribute values of Ionosphere, Balance, and Sonar data sets contain less normality than others—the balance scale are totally not normal. We can conclude that the ensemble OBE RF requests less normality of the data set.
The results of experiments indicate that the application of the belief function tool to the prediction of trees and combination of forests is efficient and reasonable; yet there are also some drawbacks. Firstly, the introduction of the belief function and mass combination obviously increases the time cost of learning. The sensitivity to the normality of data makes the proposed trees and RFs unable to handle, to the greatest extent, all situations with one particular structure.

6. Conclusions

In this paper, a new classification tree method based on belief entropy is proposed to cope with uncertain data. This method directly models continuous attribute values of training data by E 2 M algorithm, and selects a splitting attribute via a new tool–belief entropy. Differing from the traditional decision trees, we redesign the splitting and prediction, making them fit the feature of uncertain labels described by the belief function. Finally, random forests with different combination strategies were constructed on the basis of the proposed tree method to seek higher accuracy and stronger generalization ability.
As the experimental results show, the proposed belief entropy trees are robust to different sorts of uncertainty. They perform closely to traditional decision trees on precise data and keep good results on data with ill-known labels. Meanwhile, the belief entropy random forests, which improve significantly when compared to the basic belief function trees, achieve excellent and stable performance even in the situation with high-level uncertainty. It is proved that the proposed trees and random forests have a potentially broad field of application. In future research, some further improvements will be investigated, such as more reasonable BBA combination methods for the incapacity of Dempster’s rule to handle huge mass conflict, and a boosting ensemble method based on the belief entropy trees.

Author Contributions

Conceptualization, K.G. and Y.W.; methodology, K.G. and L.M.; software, K.G.; validation, K.G. and L.M.; formal analysis, K.G.; investigation, K.G. and L.M.; resources, Y.W.; data curation, K.G.; writing—original draft preparation, K.G.; writing—review and editing, L.M. and Y.W.; supervision, Y.W.; funding acquisition, Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

The work described in this paper was supported by the National Natural Science Foundation of China (61973291).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors greatly appreciate the reviewers’ suggestions and the editor’s encouragement.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Josse, J.; Chavent, M.; Liquet, B.; Husson, F. Handling missing values with regularized iterative multiple correspondence analysis. J. Classif. 2012, 29, 91–116. [Google Scholar] [CrossRef] [Green Version]
  2. Quinlan, J.R. Decision trees as probabilistic classifiers. In Proceedings of the Fourth International Workshop on Machine Learning, Irvine, CA, USA, 22–25 June 1987; Elsevier: Amsterdam, The Netherlands, 1987; pp. 31–37. [Google Scholar]
  3. Tsang, S.; Kao, B.; Yip, K.Y.; Ho, W.S.; Lee, S.D. Decision trees for uncertain data. IEEE Trans. Knowl. Data Eng. 2009, 23, 64–78. [Google Scholar] [CrossRef] [Green Version]
  4. Couso Blanco, I.; Sánchez Ramos, L. Harnessing the information contained in low-quality data sources. Int. J. Approx. Reason. 2014, 1485–1486. [Google Scholar] [CrossRef]
  5. Masson, M.H.; Denoeux, T. Ranking from pairwise comparisons in the belief functions framework. In Belief Functions: Theory and Applications; Springer: Berlin/Heidelberg, Germany, 2012; pp. 311–318. [Google Scholar]
  6. Yuan, Y.; Shaw, M.J. Induction of fuzzy decision trees. Fuzzy Sets Syst. 1995, 69, 125–139. [Google Scholar] [CrossRef]
  7. Wang, X.; Liu, X.; Pedrycz, W.; Zhang, L. Fuzzy rule based decision trees. Pattern Recognit. 2015, 48, 50–59. [Google Scholar] [CrossRef]
  8. Hüllermeier, E. Possibilistic induction in decision-tree learning. In Lecture Notes in Computer Science: Proceedings of the European Conference on Machine Learning; Springer: Berlin/Heidelberg, Germany, 2002; pp. 173–184. [Google Scholar]
  9. Qin, B.; Xia, Y.; Li, F. DTU: A decision tree for uncertain data. In Lecture Notes in Computer Science: Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining; Springer: Berlin/Heidelberg, Germany, 2009; pp. 4–15. [Google Scholar]
  10. Liang, C.; Zhang, Y.; Song, Q. Decision tree for dynamic and uncertain data streams. In Proceedings of the 2nd Asian Conference on Machine Learning. JMLR Workshop and Conference Proceedings, Tokyo, Japan, 8–10 November 2010; pp. 209–224. [Google Scholar]
  11. Dempster, A.P. Upper and lower probabilities induced by a multivalued mapping. In Classic Works of the Dempster-Shafer Theory of Belief Functions; Springer: Berlin/Heidelberg, Germany, 2008; pp. 57–72. [Google Scholar]
  12. Shafer, G. A Mathematical Theory of Eevidence; Princeton University Press: Princeton, NJ, USA, 1976. [Google Scholar]
  13. Elouedi, Z.; Mellouli, K.; Smets, P. Belief decision trees: Theoretical foundations. Int. J. Approx. Reason. 2001, 28, 91–124. [Google Scholar] [CrossRef]
  14. Trabelsi, S.; Elouedi, Z.; Mellouli, K. Pruning belief decision tree methods in averaging and conjunctive approaches. Int. J. Approx. Reason. 2007, 46, 568–595. [Google Scholar] [CrossRef] [Green Version]
  15. Vannoorenberghe, P.; Denoeux, T. Handling uncertain labels in multiclass problems using belief decision trees. In Proceedings of the IPMU, Annecy, France, 1–5 July 2002; Volume 3, pp. 1919–1926. [Google Scholar]
  16. Sutton-Charani, N.; Destercke, S.; Denœux, T. Classification trees based on belief functions. In Belief Functions: Theory and Applications; Springer: Berlin/Heidelberg, Germany, 2012; pp. 77–84. [Google Scholar]
  17. Sutton-Charani, N.; Destercke, S.; Denœux, T. Learning decision trees from uncertain data with an evidential EM approach. In Proceedings of the 12th International Conference on Machine Learning and Applications, Miami, FL, USA, 4–7 December 2013; Volume 1, pp. 111–116. [Google Scholar]
  18. Sutton-Charani, N.; Destercke, S.; Denœux, T. Training and evaluating classifiers from evidential data: Application to E2M decision tree pruning. In Lecture Notes in Computer Science: Proceedings of the International Conference on Belief Functions; Springer: Berlin/Heidelberg, Germany, 2014; pp. 87–94. [Google Scholar]
  19. Ma, L.; Destercke, S.; Wang, Y. Online active learning of decision trees with evidential data. Pattern Recognit. 2016, 52, 33–45. [Google Scholar] [CrossRef]
  20. Trabelsi, A.; Elouedi, Z.; Lefevre, E. Handling uncertain attribute values in decision tree classifier using the belief function theory. In Lecture Notes in Computer Science: Proceedings of the International Conference on Artificial Intelligence: Methodology, Systems, and Aapplications; Springer: Berlin/Heidelberg, Germany, 2016; pp. 26–35. [Google Scholar]
  21. Trabelsi, A.; Elouedi, Z.; Lefevre, E. New decision tree classifier for dealing with partially uncertain data. In Proceedings of the 25th Rencontres Francophones sur la Logique Floue et ses Applications (LFA 2016), Nimes, France, 23 November 2016; pp. 57–64. [Google Scholar]
  22. Trabelsi, A.; Elouedi, Z.; Lefevre, E. Decision tree classifiers for evidential attribute values and class labels. Fuzzy Sets Syst. 2019, 366, 46–62. [Google Scholar] [CrossRef]
  23. Kim, B.; Jeong, Y.S.; Tong, S.H.; Jeong, M.K. A generalised uncertain decision tree for defect classification of multiple wafer maps. Int. J. Prod. Res. 2020, 58, 2805–2821. [Google Scholar] [CrossRef]
  24. Zou, J.; Yan, X.; Zhou, Y. Discounted Belief Decision Tree for Uncertainty data from unreliable source. In Proceedings of the 2019 International Conference on Computational Science and Computational Intelligence (CSCI), Las Vegas, NV, USA, 5–7 December 2019; pp. 382–387. [Google Scholar]
  25. Li, Y.; Chen, W. Landslide susceptibility evaluation using hybrid integration of evidential belief function and machine learning techniques. Water 2020, 12, 113. [Google Scholar] [CrossRef] [Green Version]
  26. Denœux, T. Maximum likelihood from evidential data: An extension of the EM algorithm. In Combining Soft Computing and Statistical Methods in Data Analysis; Springer: Berlin/Heidelberg, Germany, 2010; pp. 181–188. [Google Scholar]
  27. Denoeux, T. Maximum likelihood estimation from uncertain data in the belief function framework. IEEE Trans. Knowl. Data Eng. 2011, 25, 119–130. [Google Scholar] [CrossRef] [Green Version]
  28. Deng, Y. Deng entropy. Chaos Solitons Fractals 2016, 91, 549–553. [Google Scholar] [CrossRef]
  29. Ma, L.; Sun, B.; Han, C. Training instance random sampling based evidential classification forest algorithms. In Proceedings of the 2018 21st International Conference on Information Fusion (FUSION), Cambridge, UK, 10–13 July 2018; pp. 883–888. [Google Scholar]
  30. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  31. Gao, K.; Ma, L.; Wang, Y. A Classification Tree Method Based on Belief Entropy for Evidential Data. In Lecture Notes in Computer Science: Proceedings of the International Conference on Belief Functions; Springer: Berlin/Heidelberg, Germany, 2021; pp. 105–114. [Google Scholar]
  32. Quinlan, J.R. Induction of decision trees. Mach. Learn. 1986, 1, 81–106. [Google Scholar] [CrossRef] [Green Version]
  33. Quinlan, J.R. C4.5: Programs for Machine Learning; Elsevier: Amsterdam, The Netherlands, 2014. [Google Scholar]
  34. Breiman, L.; Friedman, J.H.; Olshen, R.A.; Stone, C.J. Classification and Regression Trees; Routledge: London, UK, 2017. [Google Scholar]
  35. Ma, L.; Denoeux, T. Partial classification in the belief function framework. Knowl.-Based Syst. 2021, 214, 106742. [Google Scholar] [CrossRef]
  36. Denoeux, T. Decision-making with belief functions: A review. Int. J. Approx. Reason. 2019, 109, 87–110. [Google Scholar] [CrossRef] [Green Version]
  37. Smets, P. Decision making in the TBM: The necessity of the pignistic transformation. Int. J. Approx. Rreasoning 2005, 38, 133–147. [Google Scholar] [CrossRef] [Green Version]
  38. Denoeux, T. Likelihood-based belief function: Justification and some extensions to low-quality data. Int. J. Approx. Reason. 2014, 55, 1535–1547. [Google Scholar] [CrossRef]
  39. Dempster, A.P.; Laird, N.M.; Rubin, D.B. Maximum likelihood from incomplete data via the EM algorithm. J. R. Stat. Soc. Ser. B (Methodol.) 1977, 39, 1–22. [Google Scholar]
  40. Shannon, C.E. A mathematical theory of communication. ACM Sigmobile Mob. Comput. Commun. Rev. 2001, 5, 3–55. [Google Scholar] [CrossRef]
  41. Xu, P.; Deng, Y.; Su, X.; Mahadevan, S. A new method to determine basic probability assignment from training data. Knowl.-Based Syst. 2013, 46, 69–80. [Google Scholar] [CrossRef]
  42. Li, M.; Xu, H.; Deng, Y. Evidential decision tree based on belief entropy. Entropy 2019, 21, 897. [Google Scholar] [CrossRef] [Green Version]
  43. Fisher, R.A. The use of multiple measurements in taxonomic problems. Ann. Eeugenics 1936, 7, 179–188. [Google Scholar] [CrossRef]
  44. McLachlan, G.J.; Peel, D. Finite Mixture Models; John Wiley & Sons: Hoboken, NJ, USA, 2004. [Google Scholar]
  45. Quost, B.; Denoeux, T.; Li, S. Parametric classification with soft labels using the evidential EM algorithm: Linear discriminant analysis versus logistic regression. Adv. Data Anal. Classif. 2017, 11, 659–690. [Google Scholar] [CrossRef] [Green Version]
  46. Dua, D.; Graff, C. UCI Machine Learning Repository. 2017. Available online: https://ergodicity.net/2013/07/ (accessed on 25 March 2022).
  47. Liu, L.; Dietterich, T.G. A conditional multinomial mixture model for superset label learning. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 548–556. [Google Scholar]
Figure 1. The normal distribution df three classes for the SL attribute of Iris data set.
Figure 1. The normal distribution df three classes for the SL attribute of Iris data set.
Entropy 24 00605 g001
Figure 2. The normal distribution models for each attribute in evidential Iris data set.
Figure 2. The normal distribution models for each attribute in evidential Iris data set.
Entropy 24 00605 g002
Figure 3. The normal distribution for the SL attribute in three classes.
Figure 3. The normal distribution for the SL attribute in three classes.
Entropy 24 00605 g003
Figure 4. Generation of mass prediction for each instance.
Figure 4. Generation of mass prediction for each instance.
Entropy 24 00605 g004
Figure 5. Three different ways to make prediction in belief entropy trees.
Figure 5. Three different ways to make prediction in belief entropy trees.
Entropy 24 00605 g005
Figure 6. Belief entropy in random forests. (a) Generation of basic trees and their outputs in random forest. (b) Combination strategies of different random forests.
Figure 6. Belief entropy in random forests. (a) Generation of basic trees and their outputs in random forest. (b) Combination strategies of different random forests.
Entropy 24 00605 g006
Figure 7. Classification accuracy on UCI data sets with different vacuousness levels.
Figure 7. Classification accuracy on UCI data sets with different vacuousness levels.
Entropy 24 00605 g007
Figure 8. Classification accuracy on UCI data sets with different imprecision levels.
Figure 8. Classification accuracy on UCI data sets with different imprecision levels.
Entropy 24 00605 g008
Figure 9. Classification accuracy on UCI data sets with different uncertainty levels.
Figure 9. Classification accuracy on UCI data sets with different uncertainty levels.
Entropy 24 00605 g009
Table 1. Uncertain instances in evidential Iris data set.
Table 1. Uncertain instances in evidential Iris data set.
NumberAttributesContour FunctionsTrue Label
SLSWPLPWpl (Setosa)pl (Versicolor)pl (Virginica)
15.13.51.40.2101Setosa
24.93.01.40.2101Setosa
34.73.21.30.2100Setosa
44.63.11.50.210.74980.4073Setosa
517.03.24.71.4010Versicolor
526.43.24.51.50.951910.7087Versicolor
536.93.14.91.5111Versicolor
545.52.34.01.3111Versicolor
1016.33.36.02.5001Virginica
1025.82.75.11.90.44580.50881Virginica
1037.13.05.92.1001Virginica
1046.32.95.61.8111Virginica
1505.93.05.11.8001Virginica
Table 2. Estimated normal distribution parameters for evidential Iris data set.
Table 2. Estimated normal distribution parameters for evidential Iris data set.
AttributesSetosaVersicolorVirginica
μ σ μ σ μ σ
SL5.00310.35915.91830.54196.63070.6300
SW3.38220.42642.78120.32013.01190.2765
PL1.46400.17184.30660.49615.56840.5591
PW0.24080.10111.32730.20742.04950.2522
Table 3. Generated BBAs of selected instance.
Table 3. Generated BBAs of selected instance.
AttributesBBAs
m S e t o s a = 0.7996
SL m S e t o s a , V e r s i c o l o r = 0.1757
m S e t o s a , V e r s i c o l o r , V i r g i n i c a = 0.0247
m S e t o s a = 0.6904
SW m S e t o s a , V i r g i n i c a = 0.2329
m S e t o s a , V e r s i c o l o r , V i r g i n i c a = 0.0767
m S e t o s a = 1
PL m S e t o s a , V e r s i c o l o r = 0
m S e t o s a , V e r s i c o l o r , V i r g i n i c a = 0
m S e t o s a = 1
PW m S e t o s a , V e r s i c o l o r = 0
m S e t o s a , V e r s i c o l o r , V i r g i n i c a = 0
Table 4. Classification accuracy on UCI data sets with 90% vacuousness level.
Table 4. Classification accuracy on UCI data sets with 90% vacuousness level.
IrisWineIonosphereBalance ScaleBreast CancerSonar
C4.50.8000.68140.72210.61340.80310.5247
OBE tree0.58460.40380.57750.63050.77510.5553
LBE tree0.83260.60260.77630.59540.87320.5793
IBE tree0.82880.63860.77970.62820.87930.5841
label voting OBE Random Forest0.78640.48970.74040.74530.87450.5611
label voting LBE Random Forest0.90200.85940.85550.71270.93190.6149
mass combination LBE Random Forest0.89890.82950.81820.71380.92230.6115
label voting IBE Random Forest0.90530.89400.86080.7454 0.93380.6120
mass combination IBE Random Forest0.91740.90820.86470.68910.93300.6236
Table 5. Classification accuracy on UCI data sets with 90% imprecision level.
Table 5. Classification accuracy on UCI data sets with 90% imprecision level.
IrisWineIonosphereBalance ScaleBreast CancerSonar
C4.50.80000.68140.72210.61340.80310.5247
OBE tree0.62330.53820.67860.67460.86050.5841
LBE tree0.90930.77190.78580.61460.89790.6303
IBE tree0.90400.78990.78920.63810.90510.6413
label voting OBE random forest0.86470.62080.85520.75360.92570.6351
label voting LBE random forest0.94730.91240.86210.74830.93590.6630
mass combination LBE random forest0.93270.91180.82590.74370.92620.6635
label voting IBE random forest0.9467 0.92190.86840.77090.93640.6572
mass combination IBE random forest0.94470.93260.87550.72370.93990.6702
Table 6. Classification accuracy on UCI data sets with 90% uncertainty level.
Table 6. Classification accuracy on UCI data sets with 90% uncertainty level.
IrisWineIonosphereBalance ScaleBreast CancerSonar
C4.50.80000.68140.72210.61340.80310.5247
OBE tree0.88470.77920.82190.73870.91210.6851
LBE tree0.94000.81800.79370.62350.91200.6803
IBE tree0.95130.83990.79550.63420.92780.6784
label voting OBE random forest0.95130.95900.9293 0.82330.93810.7577
label voting LBE random forest0.95470.94890.88030.76510.93440.7370
mass combination LBE random forest0.95600.94780.83900.78000.92250.7404
label voting IBE random forest0.95670.95280.88260.78290.93760.7346
mass combination IBE random forest0.95670.95960.88750.74960.93940..7433
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gao, K.; Wang, Y.; Ma, L. Belief Entropy Tree and Random Forest: Learning from Data with Continuous Attributes and Evidential Labels. Entropy 2022, 24, 605. https://doi.org/10.3390/e24050605

AMA Style

Gao K, Wang Y, Ma L. Belief Entropy Tree and Random Forest: Learning from Data with Continuous Attributes and Evidential Labels. Entropy. 2022; 24(5):605. https://doi.org/10.3390/e24050605

Chicago/Turabian Style

Gao, Kangkai, Yong Wang, and Liyao Ma. 2022. "Belief Entropy Tree and Random Forest: Learning from Data with Continuous Attributes and Evidential Labels" Entropy 24, no. 5: 605. https://doi.org/10.3390/e24050605

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop