Next Article in Journal
Fractional-Order Active Disturbance Rejection Control with Fuzzy Self-Tuning for Precision Stabilized Platform
Previous Article in Journal
Better Heisenberg Limits, Coherence Bounds, and Energy-Time Tradeoffs via Quantum Rényi Information
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Genetic Algorithm Based on a New Similarity for Probabilistic Transformation of Belief Functions †

1
Department of Artificial Intelligence, Shanghai Maritime University, Shanghai 201306, China
2
School of Cyber Science and Engineering, Southeast University, Nanjing 210096, China
*
Author to whom correspondence should be addressed.
This paper is an extended version of Dong, Y.; Li, X.; Dezert, J. A new probabilistic transformation based on evolutionary algorithm for decision making. In Proceedings of the 20th International Conference on Information Fusion (FUSION 2017), Xi’an, China, 10–13 July 2017; doi: 10.23919/ICIF.2017.8009821.
Entropy 2022, 24(11), 1680; https://doi.org/10.3390/e24111680
Submission received: 29 September 2022 / Revised: 12 November 2022 / Accepted: 15 November 2022 / Published: 17 November 2022

Abstract

:
Recent studies of alternative probabilistic transformation (PT) in Dempster–Shafer (DS) theory have mainly focused on investigating various schemes for assigning the mass of compound focal elements to each singleton in order to obtain a Bayesian belief function for decision-making problems. In the process of such a transformation, how to precisely evaluate the closeness between the original basic belief assignments (BBAs) and transformed BBAs is important. In this paper, a new aggregation measure is proposed by comprehensively considering the interval distance between BBAs and also the sequence inside the BBAs. Relying on this new measure, we propose a novel multi-objective evolutionary-based probabilistic transformation (MOEPT) thanks to global optimizing capabilities inspired by a genetic algorithm (GA). From the perspective of mathematical theory, convergence analysis of EPT is employed to prove the rationality of the GA used here. Finally, various scenarios in evidence reasoning are presented to evaluate the robustness of EPT.

1. Introduction

1.1. Background and Research Motivation

Since the pioneering work of Dempster and Shafer [1,2], known as Dempster–Shafer evidence theory (DST), belief functions are widely used in information fusion for decision making [3,4]. However, the computational complexity of reasoning with DST is one of the major points of criticism this formalism has to face. To overcome this difficulty, various approximating methods have been suggested that aim at reducing the number of focal elements in the frame of discernment (FoD) in order to maintain the tractability of computation. One common strategy is to simplify the FoD by removing or aggregating focal elements for approximating the original belief function [5]. Among these methods, probabilistic transformations (PTs) seem particularly desirable for reducing such computational complexity by means of assigning the mass of non-singleton elements to some singletons of the FoD [6,7]. The research on this probabilistic measure has received a lot of attention [8], and many efficient PTs have been proposed by scholars in recent years. Among them, a classical transformation, denoted as BetP [6], was usually adopted because it offered a compromise between the maximum of credibility (Bel) and the maximum of plausibility (Pl) for decision making. Unfortunately, BetP does not provide the highest probabilistic information content (PIC) [9], and Shenoy argued against BetP in his publication [10]. Sudano [11] also proposed series of alternatives and principles for them similar to BetP, which were called PrPl, PrBel and PrHyb. CuzzP [12], which was proposed by Cuzzolin in the framework of DST in 2009, showed its probabilistic transformation ability. Another novel transformation was proposed by Dezert and Smarandache in the framework of Dezert–Smarandache theory (DSmT) which was called Dezert–Smarandache probability (DSmP) [9], and comprehensive comparisons were made in [9] to prove the advantages of DSmP with respect to other PTs.

1.2. Challenges

Although various techniques have been proposed to evaluate PTs, these methods have limitations. On the one hand, the PIC or Shannon’s entropy criterion was applied to evaluate PTs, with less uncertainty (clarity) for the BBAs obtained from PTs being perferred in order to make decisions easily. However, Han et al. [13] and Bucci in [14] illustrated the irrationality of the overemphasizing of the Shannon entropy or PIC. On the other hand, from the perspective of fidelity, to characterize the difference between the transformed BBAs and original BBAs, various distances were applied. The most famous distance was D J [15], but it was not good enough to capture the difference between BBAs in some cases, and these can be seen in [16,17]. Thus, either the PIC or distance alone is not efficient and comprehensive enough to quantify all sorts of aspects of dissimilarity which inevitably need to be involved in PTs. To address incomprehensive evaluations, several two-dimensional measures [18] have been proposed in order to make sure that the results obtained by PTs are consistent in some manner with the original BBAs. Han in [19] proposed a 2D criteria which jointly uses distance and PIC measures to create a balance between fidelity and clarity. Liu [16] used a two-dimensional measure to effectively detect conflicts among the evidence. In [20], Liu proposed both a distance and a conflict coefficient based on probabilistic transformations B e t P to characterize the dissimilarity, which are complementary in a certain sense. That aside, Deng [21] developed a new probability transformation method called the decision-based PT in belief functions theory. Following Deng’s idea [21], Zhao [22] proposed an importance-based PT to achieve the transformation of original basic belief assignments. Recently, Ma [23] integrated a fuzzy closeness and correlation coefficient to generate a new dissimilarity measure to characterize not only the difference between BBAs but also the divergence degree of the hypothesis that two BBAs support. By analyzing the mentioned existing methods, the relationship between the techniques of PTs and their corresponding evaluations are almost independent, except in [19]; that is to say, the methods of evaluation only assess the existing PTs instead of facilitating the development of novel PTs themselves.

1.3. Contributions

In this paper, we present a novel PT method based on a multi-objective algorithm (MOEPT) using reasonable and comprehensive two-dimensional criteria in order to capture the similarity in the process of a PT. This new method has some connections with the recent algorithm proposed in [19]. However, the main differences lie in the following aspects:
  • The 2D criteria, PIC and Jousselme’s distance have pointed out its drawbacks in many references [17,24]. Thus, an efficient and different aggregation measure is proposed. Its novelty lies in considering the drawback of the past description of the distance between the evidence. In other words, up to now, most distances were defined according to the corresponding focal elements between two sources of evidence, and the sequence inside the assignments of the focal elements themselves was not considered. The sequence might also lead to dissimilarity, which is referred to as “self-conflict or self-contradiction” [25];
  • More specific steps of evolutionary-based algorithms are given in detail. Aside from that, the convergence analysis of the MOEPT is illustrated to prove the rationality of using GAs. Moreover, some bugs are detected and fixed when using the MOEPT with traditional constraints;
  • The specific application problem, target type tracking (TTT), is efficiently solved and discussed based on the proposed method with a novel simple constraint.
Compared with traditional PTs, global search replaces designing various assigning operators in classical PTs, and the evaluation criteria are embedded into an MOEPT to provide important guidance for the searching procedure. Specifically, masses of singletons are randomly generated in an evolutionary-based framework, which needs to satisfy with the basic constraints for probability distributions in evidence reasoning. Additionally, an assessment factor is presented to assess the best individual in all populations by a special objective function (desirable evaluation criteria). The simulation results in 4D FoD test cases show that in these problems, the proposed MOEPT was able to outperform other PTs from the perspective of the 2D criteria. Moreover, we propose a simple constraint-handling strategy within the MOEPT that is well-suited for two-target type tracking (2-TTT) problems, which to some extent encourages the application of MOEPTs to more complex and real-world decision-making problems.
The remainder of this paper is structured as follows. In Section 2, we briefly summarize the basis of DST. The new aggregation measure is proposed in Section 3. In Section 4, multi-objective evolutionary algorithms (EAs) based on a two-dimensional objective function are proposed. In Section 5, several examples and comprehensive comparisons are carried out. A simple pattern recognition problem and also a target type tracking problem are presented and solved in detail at the end of this section. The conclusions are drawn in Section 6.

2. Basis of Belief Functions

In this section, we introduce the belief function terminology of DST and the notations used in the sequel to this paper.

2.1. DST Basis

In DST [2], the elements θ i ( i = 1 , , N ) of the frame of discernment (FoD) Θ { θ 1 , , θ N } must be mutually exhaustive and exclusive. The power set of the FoD is denoted as 2 Θ , and a basic belief assignment (BBA), also called a mass function, is defined by the mapping: 2 Θ [ 0 , 1 ] , which satisfies m ( ) = 0 and
A 2 Θ m ( A ) = 1
where m ( A ) is defined as the BBA of A. The element A is called a focal element of m ( . ) if m ( A ) > 0 . The belief and plausibility functions, which are in one-to-one mapping with the BBA m ( . ) , are defined for all A Θ by
B e l ( A ) = B 2 Θ | B A m ( B )
P l ( A ) = 1 B e l ( A ¯ ) = A , B 2 Θ | A B m ( B )
where A ¯ Θ \ A is the complement to A in Θ . The belief interval [ B e l ( A ) , P l ( A ) ] represents the uncertainty committed to A, and the bounds of this interval are usually interpreted as lower and upper bounds of the unknown (possibly subjective) probability of A.
In order to fuse n bodies Of evidence (BOEs), Dempster’s rule of combination is usually used in the DST framework. The combination of n distinct BOEs is achieved as follows:
m ( A ) = 0 , i f A = A i = A 1 i n m i ( A i ) A i 1 i n m i ( A i ) , i f A

2.2. Classical Probabilistic Transformations

The efficiency of a probabilistic transformation (PT) in the field of decision making was analyzed in depth by Smets [6]. Various PTs have been proposed in the open literature such as BetP [6,26], CuzzP [12], DSmP [9], PrBP1 and PrBP2 [27], as well as Cobb and Shenoy’s normalization of plausibility [10]. The simple and classical transformation (BetP) is briefly recalled in this subsection.

Pignistic Transformation (BetP)

Smets in [6,26] first proposed pignistic (also called betting) probability to make decisions, which aims to transfer the mass of belief of each non-specific element onto the singletons. The classical pignistic probability is defined as B e t P ( ) = 0 , and A 2 Θ \ { } :
B e t P ( θ i ) A 2 Θ , A θ i A A m ( A ) 1 m ( )
Because in Shafer’s framework m ( ) = 0 , Equation (5) can simply be rewritten for any singleton θ i Θ as
B e t P ( θ i ) = B 2 Θ , θ i B 1 B m ( B ) = m ( θ i ) + B 2 Θ , θ i B 1 B m ( B )

2.3. Distance Proposed by Han and Dezert d B I E

The Jousselme’s distance, which was widely denoted as D J in [15], was applied in many recent works [19,28], but the particular choice of the D J distance is not a very good choice because one knows that the D J distance has bad behavior. This was clearly explained recently in [17,24]. Assuming that two independent BBAs m 1 ( · ) and m 2 ( · ) are defined on Θ = { θ 1 , θ 2 , , θ n } , for each focal element θ i Θ ( i = 1 , 2 , , 2 n 1 ) , belief intervals of θ i for m 1 ( · ) and m 2 ( · ) can be calculated, which are denoted by [ B e l 1 ( θ i ) , P l 1 ( θ i ) ] and [ B e l 2 ( θ i ) , P l 2 ( θ i ) ] , respectively. The strict distance between the interval numbers [ a , b ] and [ c , d ] is defined by [29]
d I ( [ a , b ] , [ c , d ] ) = [ a + b 2 c + d 2 ] 2 + 1 3 [ b a 2 d c 2 ] 2
Therefore, we can calculate the distance between B I 1 ( θ i ) : [ B e l 1 ( θ i ) , P l 1 ( θ i ) ] and B I 2 ( θ i ) : [ B e l 2 ( θ i ) , P l 2 ( θ i ) ] according to Equation (7). Thus, we can obtain a total of 2 N 1 belief interval distance values for all θ i Θ . Aside from that, the Euclidean family belief interval-based distance d B I E can be rewritten as
d B I E ( m 1 , m 2 ) = N c · i = 1 2 n 1 [ d I ( B I 1 ( θ i ) , B I 2 ( θ i ) ) ] 2
Here, N c = 1 / 2 n 1 is the normalization factor. In this paper, we regard d B I E as one criterion for evaluating the degree of similarity (similarity representing the degree of difference between the original BBAs and the transformed ones in [30]) between the original BBAs and the transformed ones.

3. New Evidence for Similarity Characterization

As mentioned in previous section, those distances (i.e., Jousselme’s distance [28]) and other metrics such as the PIC [31] or entropy [27] were widely applied to measure the degree of “similarity or dissimilarity” between BBAs. However, only the corresponding focal elements (or the relevant focal element set) between two sources of evidence are described or characterized. This one-sided view does not consider the order of size of the assignment of each focal element in the evidence, which might lead to “self-conflict or self-contradiction”. To consider such “information” produced by the evidence itself, here, a new evidence similarity measure is defined between two evidential sources according to the order of size of the assignment. Prior to this, to give this new similarity measure, first we define the order correlation coefficient between two sets of data:
Definition 1.
[32] Given two sets of data { x 1 , x 2 , , x n } and { y 1 , y 2 , , y n } , here, x 1 , x 2 , , x n and y 1 , y 2 , , y n are in an ascending order. After sorting, two sets of data x p 1 , x p 2 , , x p n and y q 1 , y q 2 , , y q n meet x p 1 x p 2 x p n and y q 1 y q 2 y q n , respectively, for each p i and index their positions from q 1 , q 2 , , q n , assuming it is q j ; that is, q j = p i . Note that j = f ( i ) , and the correlation coefficient is
μ = i = 1 n ( i j ) 2 i = 1 n [ n ( i 1 ) i ] 2
This satisfies 0 μ 1 . When μ = 0 , the convergence of two sets of data is the largest. When μ = 1 , this is reversed.

3.1. The Consistency of Focal Elements between Two BOEs

Definition 2.
[32] For any two sources of evidence (i.e., S 1 and S 2 ) , m 1 ( · ) and m 2 ( · ) are the basic belief assignments over the discernment framework Θ of a size n. The number of focal elements and the focal elements of m 1 ( . ) and m 2 ( . ) can be different. We denote X i and Y i as the indexes of the focal elements whose masses are sorted by increasing order. The similarity function of the evidence to characterize the order of the size of the assignments over the subsets is as follows:
S i m s e q ( m 1 , m 2 ) = 1 i = 1 n ( X i Y i ) 2 i = 1 n [ n + 1 2 i ] 2
As we all know, if there is a similarity function S i m ( m i , m j ) , which is the characterization of distance between any two evidence sources, then the following four basic conditions must be satisfied:
  • Symmetry: m i ( · ) , m j ( · ) , S i m ( m i , m j ) = S i m ( m j , m i ) ;
  • Consistency: m ( · ) , S i m ( m , m ) = 1 ;
  • Nonnegative: m i ( · ) , m j ( · ) , 0 S i m ( m i , m j ) 1 ;
  • Triangle inequality: S i m ( X , Y ) + S i m ( Y , Z ) S i m ( X , Z ) .
According to previous work [32], it is easy to find that S i m s e q satisfies symmetry, consistency and nonnegativity, but the last important condition is lost. Therefore, we prove the triangle inequality property of S i m s e q here:
Proof. 
Based on Equation (10), the triangle inequality can be rewritten as follows: 1 i = 1 n ( X i Z i ) 2 i = 1 n [ n + 1 2 i ] 2 1 i = 1 n ( X i Y i ) 2 i = 1 n [ n + 1 2 i ] 2 + 1 i = 1 n ( Y i Z i ) 2 i = 1 n [ n + 1 2 i ] 2 i = 1 n [ n + 1 2 i ] 2 i = 1 n ( X i Y i ) 2 + i = 1 n ( Y i Z i ) 2 i = 1 n ( X i Z i ) 2 .
Here, X, Y and Z denote vectors ⇒ i = 1 n [ n + 1 2 i ] 2 ( X Y ) T ( X Y ) + ( Y Z ) T ( Y Z ) ( X Z ) T ( X Z ) .
According to the squared sum formula (SSF) (squared sum formula of a natural number: 1 2 + 2 2 + 3 2 + + n 2 = n ( n + 1 ) ( 2 n + 1 ) 6 ; squared sum formula of an odd number: 1 2 + 3 2 + 5 2 + + ( 2 n 1 ) 2 = 1 3 n ( 4 n 2 1 ) ; squared sum formula of an even number: 2 2 + 4 2 + + ( 2 n ) 2 = 2 3 n ( n + 1 ) ( 2 n + 1 ) ), we have
i = 1 n [ n + 1 2 i ] 2 = i = 1 n [ ( n + 1 ) 2 + 4 i 2 4 ( n + 1 ) i ] = = n ( n + 1 ) 2 + 4 1 6 n ( n + 1 ) ( 2 n + 1 ) 4 ( n + 1 ) n ( n + 1 ) 2 = 2 3 n ( n + 1 ) ( 2 n + 1 ) n ( n + 1 ) 2 = n ( n + 1 ) · [ 4 3 n + 2 3 n 1 ] = 1 3 · n · ( n 2 1 )
Because 1 X i n , 1 Y i n , 1 Z i n , we obtain the following:
( X Y ) T ( X Y ) + ( Y Z ) T ( Y Z ) ( X Z ) T ( X Z ) 1 + ( k 1 ) 2 + ( k 3 ) 2 +
  • When n 1 = 2 k k = n 1 2 , 1 + ( k 1 ) 2 + ( k 3 ) 2 + = 1 6 · k ( k + 1 ) ( k + 2 ) , and thus
    i = 1 n [ n + 1 2 i ] 2 [ ( X Y ) T ( X Y ) + ( Y Z ) T ( Y Z ) ( X Z ) T ( X Z ) ] 1 3 · n · ( n 2 1 ) 1 6 · k ( k + 1 ) ( k + 2 ) = 1 3 · n · ( n 2 1 ) 1 6 · n 1 2 ( n 1 2 + 1 ) ( n 1 2 + 2 ) = 1 3 · n · ( n 2 1 ) 1 6 · n 1 2 · n + 1 2 · n + 3 2 = ( n 2 1 ) · ( 1 4 n + 1 6 )
    Because n 2 , thus
    i = 1 n [ n + 1 2 i ] 2 [ ( X Y ) T ( X Y ) + ( Y Z ) T ( Y Z ) ( X Z ) T ( X Z ) ] 0 Sim seq satisfy triangle inequality;
  • When n 1 = 2 k 1 k = n 2 , 1 + ( k 1 ) 2 + ( k 3 ) 2 + = 1 3 · k · ( 4 k 2 1 ) , thus
    i = 1 n [ n + 1 2 i ] 2 [ ( X Y ) T ( X Y ) + ( Y Z ) T ( Y Z ) ( X Z ) T ( X Z ) ] 1 3 · n · ( n 2 1 ) 1 3 · k · ( 4 k 2 1 ) = 1 3 · n · ( n 2 1 ) 1 3 · n 2 · ( 4 ( n 2 ) 2 1 ) = 1 3 · n · ( n 2 1 ) n 6 · ( n 2 1 ) = n 6 · ( n 2 1 )
    Because n 2 , thus i = 1 n [ n + 1 2 i ] 2 [ ( X Y ) T ( X Y ) + ( Y Z ) T ( Y Z ) ( X Z ) T ( X Z ) ] 0 Sim seq satisfies the triangle inequality.
   □
Definition 3.
[32] For any two sources of evidence (i.e., S 1 and S 2 ), m 1 ( · ) and m 2 ( · ) are the basic belief assignments over n focal elements in the discernment framework Θ (note that the BBA of each subproposition might be same). Assume that the s 1 subpropositions’ BBAs are the same in m 1 ( · ) , and the s 2 subpropositions’ BBAs are the same in m 2 ( · ) Herein, X i and Y i are the serial numbers according to the order of the size of the subpropositions’ BBAs, where the subscript i indicates the i t h subproposition, due to the BBAs of some subpropositions being the same. For the evidence S 1 , there might be s 1 kinds of sorts. For S 2 , there might be s 2 kinds of sorts. Therefore, there are s 1 × s 2 kinds of sorts for S 1 and S 2 . The similarity measure functions are redefined in this case as follows:
S i m s e q ( m 1 , m 2 ) = 1 t = 1 s 1 s 2 i = 1 n ( X i t Y i t ) 2 s 1 s 2 ( i = 1 n ( n + 1 2 i ) 2 )
Similarly, it is easy to prove that S i m s e q ( m X , m Y ) is a similarity measure function.
Example 1.
BayesianBBAs:Assuming two kinds of evidence m 1 = { θ 1 , θ 2 , θ 3 } = { 0 , 0.1 , 0.9 } and m 2 = { θ 1 , θ 2 , θ 3 } = { 0.9 , 0.1 , 0 } , then m 1 and m 2 are sorted from small to large so that X ( m 1 ) = { θ 1 , θ 2 , θ 3 } = [ 1 , 2 , 3 ] and Y ( m 2 ) = { θ 3 , θ 2 , θ 1 } = [ 3 , 2 , 1 ] , respectively. Thus, we can calculate the similarity measure based on Equation (10):
S i m s e q ( m 1 , m 2 ) = 1 ( 1 3 ) 2 + ( 2 2 ) 2 + ( 3 1 ) 2 ( 3 + 1 2 1 ) 2 + ( 3 + 1 2 2 ) 2 + ( 3 + 1 2 3 ) 2 = 0
According to S i m s e q ( m 1 , m 2 ) , we can find that m 1 and m 2 are completely different and lacking similarity.

3.2. The Inconsistency of the Focal Elements between Two BOEs

How do we calculate S i m s e q when the focal elements in BBAs are different? Let us consider the following example and put forward a different way compared with that in [32]:
Example 2.
General BBAs:Assuming two types of evidence m 1 = { θ 1 , θ 2 θ 3 , θ 2 θ 4 , θ 1 θ 3 θ 4 } = { 0.3 , 0.2 , 0.2 , 0.3 } and m 2 = { θ 1 , θ 1 θ 3 , θ 3 , θ 2 θ 3 θ 4 , θ 1 θ 2 θ 3 θ 4 } = { 0.4 , 0.1 , 0.1 , 0.2 , 0.2 } .
Borrowing ideas from Dezert ( d B I E ) in [24], for each focal element θ i Θ ( i = 1 , 2 , , 2 n 1 ) , the belief intervals of θ i for m 1 ( · ) and m 2 ( · ) can be calculated, which are denoted by [ B e l 1 ( θ i ) , P l 1 ( θ i ) ] and [ B e l 2 ( θ i ) , P l 2 ( θ i ) ] , respectively. According to the theory of evidence, the width of such an interval as [ B e l 1 ( θ i ) , P l 1 ( θ i ) ] represents the degree of uncertainty for the corresponding focal element θ i . Therefore, X i and Y i in Equation (10) are obtained, which refer to the index of the width of the interval for each focal element, whose values are sorted in increasing order. The steps of this mechanism are illustrated as follows:
  • Step 1:
    [ B e l 1 ( θ 1 ) , P l 1 ( θ 1 ) ] = [ 0.3 , 0.6 ] [ B e l 1 ( θ 2 ) , P l 1 ( θ 2 ) ] = [ 0 , 0.4 ] ,
    [ B e l 1 ( θ 3 ) , P l 1 ( θ 3 ) ] = [ 0 , 0.5 ] [ B e l 1 ( θ 4 ) , P l 1 ( θ 4 ) ] = [ 0 , 0.5 ] ,
    [ B e l 2 ( θ 1 ) , P l 2 ( θ 1 ) ] = [ 0.4 , 0.7 ] [ B e l 2 ( θ 2 ) , P l 2 ( θ 2 ) ] = [ 0 , 0.4 ] ,
    [ B e l 2 ( θ 3 ) , P l 2 ( θ 3 ) ] = [ 0.1 , 0.6 ] [ B e l 2 ( θ 4 ) , P l 2 ( θ 4 ) ] = [ 0 , 0.4 ] ;
  • Step 2: The parameter ς denotes the width of the belief interval, where ς 1 ( θ 1 ) = P l 1 ( θ 1 ) B e l 1 ( θ 1 ) = 0.3 , ς 1 ( θ 2 ) = 0.4 , ς 1 ( θ 3 ) = 0.5 , ς 1 ( θ 4 ) = 0.5 , ς 2 ( θ 1 ) = 0.3 , ς 2 ( θ 2 ) = 0.4 , ς 2 ( θ 3 ) = 0.5 and ς 2 ( θ 4 ) = 0.4 ;
  • Step 3: X 1 and Y 1 are the indexes of focal elements whose ς values are sorted in increasing order, where X 1 = { 1 , 2 , 3 , 3 } and Y 1 = { 1 , 2 , 3 , 2 } ;
  • Step 4: S i m s e q is calculated based on Equation (10).
To consider the influence of the distance of the evidence, here, based on d B I E in Equation (8), we propose a new similarity measure which is presented as follows:
C i m ( m i , m j ) = w 1 · d B I E ( m i , m j ) + w 2 · ϝ ( S i m s e q ( m i , m j ) )
where w 1 = w 2 = 0.5 . and ϝ ( · ) is the decreasing function within the interval [ 0 , 1 ] , for which in this paper ϝ ( · ) = 1 x 2 . That aside, it is easy to prove that the improved measure function C i m ( m i , m j ) is still a similarity measure function. This is because the final metric between two similarity measure functions still meets the definition of a similarity measure function.
Additionally, to consider the normalization of Equation (12), it can be rewritten as follows:
C n o r m ( m i , m j ) = w 1 ( d B I E ( m i , m j ) ) + w 2 ( ϝ ( S i m s e q ( m i , m j ) ) )
where d B I E ( m ^ ) = d B I E ( m i , m j ) m i n ( vector 1 ) m a x ( vector 1 ) m i n ( vector 1 ) , S i m s e q ( m i , m j ) = S i m s e q ( m i , m j ) m i n ( vector 2 ) m a x ( vector 2 ) m i n ( vector 2 ) ,
vector 1 = ( D 1 ( m i , m 1 ) , D 2 ( m i , m 2 ) , , D j ( m i , m j ) ) and
vector 2 = ( S i m 1 ( m i , m 1 ) , S i m 2 ( m i , m 2 ) , , S i m j ( m i , m j ) ) .

4. Multi-Objective Evolutionary Algorithm Based on Two-Dimensional Criteria

In this section, we regard a PT as a general multi-objective problem consisting of two objectives also involved in a number of inequality and equality constraints. Then, a corresponding optimization model is proposed for selecting the best Bayesian BBA in the set of candidates.

4.1. Multiple-Objective Evolutionary-Based Probabilistic Transformation

The idea to approximate any BBA into a Bayesian BBA (i.e., a subjective probability measure) using the minimization of the Shannon entropy under compatibility constraints was proposed recently by Han et al. in [13,19] using on-the-shelf optimization techniques. In this paper, we present in detail a new optimization method to achieve this PT based on a random evolutionary algorithm to acquire minimization of the new aggregation criteria, and this new comprehensive criteria represents different aspects of information in BBAs. For example, the conflict coefficient represents the degree of similarity in conflict between transformed BBAs and original BBAs (in other words, the more conflicts that exist between two BBAs, the less similarity they have). In addition, d B I E represents the interval distance between the original BBAs and the transformed ones.
Let us assume that the FoD of the original BBA m ( . ) approximated by a Bayesian BBA is Θ { θ 1 , θ 2 , , θ N } . The MOEPT method consists of the following steps, which are derived from GAs:
  • Step 0 (setting parameters): Assume t max is the maximum number of iterations, n max is the population size in each iteration, P s is the selection probability, P c is the crossover probability, and P m is the mutation probability.
  • Step 1 (population generation and encoding mechanism): A set P t of j = 1 , 2 , , n max random probability values P t j = { P j ( θ 1 ) , , P j ( θ N ) } is generated such that the constraints in Equations (14)–(16) for j = 1 , 2 , , n max are satisfied in order to make each random set of probabilities P t j compatible with the original or target BBA m ( . ) to approximate. (The lower (Bel) and upper (Pl) limits of each focal element are calculated using Equations (2) and (3) based on the value of m ( · ) .) In other words, we have
    P j ( θ i ) [ 0 , 1 ] , i = 1 , 2 , , N                                            
    i = 1 N P j ( θ i ) = 1                                                                                                          
    B e l ( θ i ) P j ( θ i ) P l ( θ i ) , i = 1 , 2 , , N
  • Step 2 (fitness assignment): For each probability set P t j , ( j = 1 , 2 , , n max ), we compute its fitness value F based on Equation (13). More precisely, one takes F ( P t j ) = C n o r m ( m ( · ) , P t j ) .
  • Step 3 (best approximation of m ( . ) ): The best probability set P t j best with a minimum fitness value is sought, and its associated index j best is stored in Best Individual and Index of BestIndividual .
  • Step 4 (selection, crossover and mutation): The tournament selection, crossover and mutation operators drawn from the evolutionary theory framework [33] are implemented to create the associated offspring population P t based on the parent population P t . If F ( P t j best ) F ( P t j best ) , then the Best Individual remains unchanged; otherwise, Best Individual = P t j best .
    -
    Crossover operator: The crossover operator is one of the most important operators in the genetic algorithm. The crossover operation is conducted for the selected pairs of individuals. The feasibility condition of each individual is described as follows. The value of each subsegment must be between 0 and 1, and the sum of the individuals should be 1. Although the initial population is formed in a way that all individuals are feasible and correct, using the standard crossover operators leads to defective sub-segments, and a normalization procedure is needed for such a situation. Consider the following two individuals to be parents: X = ( 0.1 , 0.2 , 0.3 , | 0.4 ) and Y = ( 0.2 , 0.2 , 0.1 , | 0.5 ) . (Here, the vertical bar represents the intersection point with the crossover operator.) With the single-point classic crossover operator, the following offspring will be produced: X = ( 0.1 , 0.2 , 0.3 , 0.5 ) and Y = ( 0.2 , 0.2 , 0.1 , 0.4 ) , where j = 1 4 X j is equal to 1.1, which is greater than 1, and j = 1 4 Y j is equal to 0.9, which is less than 1. Therefore, X and Y have defective values for which a normalization factor is needed, which leads to the following:
    X = X j = 1 4 X j = ( 0.1 / 1.1 , 0.2 / 1.1 , 0.3 / 1.1 , 0.5 / 1.1 ) ,
    Y = Y j = 1 4 Y j = ( 0.2 / 0.9 , 0.2 / 0.9 , 0.1 / 0.9 , 0.4 / 0.9 ) .
    -
    Mutation operator: The mutation operator randomly alters the value of a sub-segment. After applying the mutation operator, normalization of the changed individuals is required. The normalization will be performed in a similar way to the crossover operator.
  • Step 5 (stopping MOEPT): Steps 1–4 illustrate the tth iteration of the MOEPT method. If t t max , then the MOEPT method is complete; otherwise, another iteration must be performed by taking t + 1 = t and going back to step 1.
The scheme of the MOEPT method is shown in Figure 1, and its pseudo-code is given in Algorithm 1.
Algorithm 1: Multi-Objective Evolutionary-Based PT (MOEPT).
1:
Define Stopping Criteria, ( t t max ); population Size n max for each iteration; crossover probability P c , mutation probability P m and selection probability P s .  
2:
Generate an initial random population P t of consistent probabilities P t j with m ( . ) .  
3:
For each individual P t j in P t do  
4:
  Calculate Fitness F ( P t j ) = C n o r m ( m ( · ) , P t j ) of P t j  
5:
  Store the best individual P t j best  
6:
End  
7:
Repeat:  
8:
  Crossover: exchange parts of individuals with probability P c  
9:
  Mutation: mutate the child individuals with probability P m  
10:
  Selection: Select individuals based on fitness according to P s  
11:
  After these three sub-steps, the updated population P t is obtained  
12:
  Calculate the fitness of individuals of P t , and store the best individual P t j b e s t  
13:
  If F ( P t j best ) F ( P t j b e s t )  
14:
    Best-Individual remains unchanged  
15:
  else  
16:
    Best-Individual = P t j best  
17:
If t t max then stops, otherwise t + 1 t and go back to line 7  

4.2. Convergence Analysis

In order to mathematically prove the feasibility of an MOEPT, convergence analysis of our algorithm is given. First, we give a simplified description of the algorithm and also its symbolic representation for simplicity:
  • Encoding mechanism: The size of the population is n m a x , the length of individual (chromosome) is N, and the initial population is P 1 ;
  • Retain the best individual directly for the next generation;
  • Randomly select the other non-optimal individuals in P t to cross over so as to form the intermediate population Y t ;
  • The population Y t is mutated to form a population V t ;
  • The better individuals in the population V t are selected as the new generation population P t .
Specifically, three operators (crossover operator, mutation operator and selection operator) can be described by the transition probability as follows:
  • Crossover operator: For a single-point crossover, a new individual k is produced based on their parents: individuals i and j:
    P C t ( i × j , k ) = | k | p c / N , k i , j ( 1 p c ) + | k | p c / N , k = i
    where | k | is the number of individuals k, 0 p c 1 is the crossover probability and a is the minimum probability for individuals | k | :
    a = 1 p c + p c / N .
  • Mutation operator:
    P M t ( i , j ) = p m d ( i , j ) ( 1 p m ) N d ( i , j )
    where 0 p m 1 is the mutation probability, d ( i , j ) is the Hamming distance between i and j and b is the minimum probability:
    b = ( 1 p m ) N .
  • Selection operator: An MOEPT uses the strategy of retaining the elite options, and the best individual is retained for the next generation which does not participate in the competition. Assume that m individuals are selected based on the following equations:
    P S t ( P t , P t j ) = σ n ( F ( P t j ) ) k = 1 n m a x ( F ( P t j ) ) , j P t , n = 1 , 2 , .
    where σ n represents an increasing scale function. That aside, the probability of selecting the first individual in the next generation’s population is
    P S t ( P t , P t j ) = | P t | | B ( P t ) | , P t P t .
    where | P t | is the number of individuals P t in P t and B ( P t ) is the cardinality of the optimal set of P t .
In order to facilitate the convergence analysis, the changing process of the fitness value F ( P t j ) is regarded as a Markov chain. If the MOEPT obtains the best individual P t j b e s t in generation t, we can denote this as { F ^ ( P t ) } = P t j b e s t . Then, all the other populations in t + 1 generations will also reach the best fitness value due to the elite strategy [34]. Therefore, the Markov chain { F ^ ( P t ) } constitutes the lower martingale. According to the properties of the lower martingale and the convergence theorem of the lower martingale [35], convergence analysis of the MOEPT is converted into the convergence of { F ^ ( P t ) } . The following three theorems are given, in which Theorem 4 is for proving that { F ^ ( P t ) } satisfies the conditions of the martingale theorem, Theorem 5 proves the global convergence of the MOEPT and Theorem 6 constructs three conditions for the convergence of the lower martingale so that the optimal solution can almost be obtained everywhere.
Theorem 1.
The process of describing the values of the fitness functions in the MOEPT is a non-bounded martingale:
E { F ^ ( P t + 1 ) / P t } F ^ ( P t )
Proof. 
Because the algorithm retains the maximum fitness value of the previous generation for the next generation and does not participate in the genetic operation, the best individual mode is not destroyed, so the maximum fitness value of the next generation’s population will not be less than the maximum fitness value of the previous generation:
E { F ^ ( P t + 1 ) / P t } F ^ ( P t ) > 0
   □
Theorem 2.
The MOEPT converges to the global optimal solution based on the probability, which is mathematically expressed by the condition.
Proof. 
When the population P t is updated to generation t, the minimum or best fitness is recorded as P t j b e s t , and the global optimal solution is noted as F , we assume that the MOEPT can converge to a global optimal solution at generation t such that
{ F ^ ( P t ) } = F
   □
Based on Theorem 4, we have
E { F ^ ( P t + 1 ) / P t } = F
This is defined by the following conditional expectation:
E { F ^ ( P t + 1 ) / P t } = i , j P t P C t ( i × j , y ) v P M t ( y , v ) k P S t ( v , k ) F ^ ( k ) i , j P t P C t ( i × j , i ) v P M t ( y , v ) k P S t ( v , k ) F ^ ( k ) a v P M t ( y , v ) k P S t ( v , k ) F ^ ( k ) a v P M t ( y , y ) k P S t ( v , k ) F ^ ( k ) a b m { k B ( P t ) [ P S t ( v , k ) P S t ( v , k ) ] F ^ ( k ) + k B ( P t ) P S t ( v , k ) F ^ ( k ) }
When k B ( P t ) , P S t ( v , k ) = 0 , and when k B ( P t ) , F ^ ( k ) = F , E { F ^ ( P t + 1 ) / P t } can be rewritten as
E { F ^ ( P t + 1 ) / P t } a b m k B ( P t ) [ P S t ( v , k ) F ^ ( k ) + F ] a b m F .
Therefore, we obtain
a b m F F .
Because F > 0 , one obtains
a b m 1 .
Based on the above formula derivation, the MOEPT converges to the global optimal solution.
Theorem 3.
When n 1 , the following conditions are satisfied:
  • E [ F ^ ( P 1 ) ] < , F < ;
  • E [ F ^ ( P t ) / P t 1 ] = F ^ ( P t 1 ) + c t 1 F ;
  • c t [ 0 , 1 ] , lim t k = 0 t 1 c k = 1 F ^ ( P 1 ) F .
Then, we have the random sequence F ^ ( P t ) a , s F
Proof: By taking the mathematical expectation on both sides of condition (2), one has
E [ F ^ ( P t ) ] = E [ F ^ ( P t 1 ) ] + c t 1 F = E [ F ^ ( P t 2 ) ] + c t 1 F + c t 2 f = = E [ F ^ ( P 1 ) ] + F k = 0 t 1 c k .
According to conditions (1) and (3), we have
E [ F ^ ( P t ) ] < E [ F ^ ( P 1 ) ] + F < S u p t E [ F ^ ( P t ) ] < s u p t E [ F ^ ( P 1 ) ] + s u p t F <
Because F ^ ( P t ) is a non-bounded martingale, we have
F ^ ( P t ) a , s F ^ ( P ) = lim t F ^ ( P t )
lim t E [ F ^ ( P t ) ] = lim t E [ F ^ ( P 1 ) ] + F lim t k = 0 t 1 c k = E [ F ^ ( P 1 ) ] + F ( 1 F ^ ( P 1 ) F ) = F ,
F ^ ( P t ) a , s F .

5. Simulation Results

According to the first step of the MOEPT, we initially set the related parameters as follows: t m a x = 50 , n m a x = 1000 , P s = 0.3 , P c = 0.5 and P m = 0.1 .

5.1. Simple Examples

Example 3.
Let us consider the frame Θ = { θ 1 , θ 2 , θ 3 , θ 4 } and the corresponding BBA illustrated as follows:
m ( θ 1 ) = 0.16 , m ( θ 2 ) = 0.14 , m ( θ 3 ) = 0.01 , m ( θ 4 ) = 0.02 m ( θ 1 θ 2 ) = 0.20 , m ( θ 1 θ 3 ) = 0.09 , m ( θ 1 θ 4 ) = 0.04 m ( θ 2 θ 3 ) = 0.04 , m ( θ 2 θ 4 ) = 0.02 , m ( θ 3 θ 4 ) = 0.01 m ( θ 1 θ 2 θ 3 ) = 0.10 , m ( θ 1 θ 2 θ 4 ) = 0.03 m ( θ 1 θ 3 θ 4 ) = 0.03 , m ( θ 2 θ 3 θ 4 ) = 0.03 m ( Θ ) = 0.08
Based on the respective classical PTs, the original BBAs are transformed into their corresponding probabilities as illustrated in Table 1. Their corresponding C n o r m values can be calculated using Equation (13), which is already listed in Table 1. Clearly, several interesting characteristics presented in Table 1 are worth mentioning: (1) M O E P T d B I E + S i m s e q had the minimum value from the perspective of the C n o r m criteria, which consider both aspects of d B I E and S i m s e q rather than concentrating on a single aspect, and (2) compared with other PTs, especially M O E P T D J + S i m s e q , (here, to show the property of d B I E , we replaced d B I E with D J in the MOEPT to make comparisons) our method performed better than the mentioned methods. However, in practice, the suitability of various PTs depends on a number of factors, including the designer’s choices; that is, from the perspective of sequence similarity, S i m s e q plays an important role in C n o r m , but from the view of the whole distance, it involves transferring the principle role to d B I E or D J . How does one quantity this role ? Here, we depend on the parameters w 1 and w 2 in C n o r m to distinguish our ideas from Han in [19], which initially set w 1 and w 2 as 0.5. Here, we discuss three different situations: (1). w 2 is set to 0.8 so as to pay more attention to the similarity of the sequence, (2) w 1 is set to 0.8 so as to focus more on the distance, and (3) considering both sequence similarity and distance, w 1 = w 2 is set to 0.5, which is the same value used in [19], so that the similarity of both the sequence and distance are considered. This phenomenon, to some degree, reminds us of the importance of proper selections of weight in various applications when the MOEPT is applied. That aside, it is worth noting that C n o r m turned out to be S i m s e q when w 1 = 0 , and C n o r m turned out to be d B I E when w 2 = 0 .
Example 4.
Let us consider another situation in the frame Θ = { θ 1 , θ 2 , θ 3 , θ 4 } and the corresponding BBAs illustrated as follows:
m ( θ 1 ) = 0.16 , m ( θ 2 ) = 0.16 , m ( θ 3 ) = 0.16 , m ( θ 4 ) = 0.16 m ( θ 1 θ 2 ) = 0.04 , m ( θ 1 θ 3 ) = 0.04 , m ( θ 1 θ 4 ) = 0.04 m ( θ 2 θ 3 ) = 0.04 , m ( θ 2 θ 4 ) = 0.04 , m ( θ 3 θ 4 ) = 0.04 m ( θ 1 θ 2 θ 3 ) = 0.03 , m ( θ 1 θ 2 θ 4 ) = 0.03 m ( θ 1 θ 3 θ 4 ) = 0.03 , m ( θ 2 θ 3 θ 4 ) = 0.03
In actuality, Example 4 is the extension of the case studied by Han in [13], which assumes a special scenario where no difference exists between m ( θ 1 ) , m ( θ 2 ) , m ( θ 3 ) and m ( θ 4 ) and where the traditional PTs become invalid and give unreasonable results, which can be seen in Table 2. The property of the original BBA where no difference exists between m ( θ 1 ) , m ( θ 2 ) , m ( θ 3 ) and m ( θ 4 ) was almost lost when classical PTs were applied. When a “sequence” is not considered in an MOEPT, which is denoted as M O E P T D i s t a n c e , the feature of equal mass in the original BBAs was also missing, as with other classical PTs. Fortunately, when information of the “sequence” was added into the objective function, the MOEPT performed better in keeping the original information, as expected.
Example 5.
Θ = { θ 1 , θ 2 , θ 3 , θ 4 }
To investigate the robustness of the MOEPT from a statistical point of view, in this example, we randomly generate BBAs and compare the MOEPT with classical PTs (BetP [6,26], CuzzP [12], DSmP [9], PrBP1 and PrBP2 [27]). The original BBAs for approximation are generated according to Algorithm 2 of [36].
Algorithm 2: Random generation of BBAs.
1:
Input: Frame of Discernment Θ = { θ 1 , θ 2 , θ 3 , θ 4 }  
2:
N m a x :Maximum number of focal element  
3:
Output: BBA-m  
4:
Generate K ( Θ ) , which is the power set of Θ  
5:
Generate a random permutation of K ( Θ ) R ( Θ )  
6:
Generate an integer between 1 and N m a x l  
7:
For each First k elements of R ( Θ ) do  
8:
Generate a value within [ 0 , 1 ] m i , i = 1 , , l  
9:
End  
10:
Normalize the vector m = m 1 , m 2 , , m l m  
11:
m ( θ i ) = m i  
In our test, we set the cardinality of the FoD to 4 and fixed the number of focal elements to l = N m a x = 15 . We randomly generated L = 100 BBAs. Six PT methods were tested, and C n o r m was used to evaluate the quality of their corresponding results, shown in Figure 2. As we can see, and as was naturally expected, the MOEPT significantly outperformed the other methods based on the minimum C n o r m criterion, which is absolutely normal because the method was developed for this aim.

5.2. Example of Pattern Classification Using the MOEPT

In this example, we used the evaluation of decision making under the evidence theory framework to indirectly evaluate the MOEPT. We considered seven classes of aircraft, which are illustrated in Figure 3, and the classifier used in this example was the probabilistic neural network (PNN). For each test example, the output of the classifier was represented by a BBA. The corresponding BBA for each test sample was generated according to Li’s previous work [37].
1. First, the image was preprocessed with binarization, and then multiple features were extracted, such as Hu moments, the normalized moment of inertia, affine invariant moments, discrete outline parameters and singular values. Secondly, five BBAs could be assigned to the evidence sources for each PNN. (Specifically, the transfer functions in five PNNs were set to a Gaussian function, the weighting function was set to the Euclidean distance; the input function was netprod, and the output function was compet.) Third, all five of these BBAs were fused by PCR6 [7] to form a single BBA m ( · ) .
2. For the two classes t 1 and t 2 ( t 1 , t 2 1 , 2 , 3 , , 7 , t 1 t 2 ), with the top two values of m ( i ) , i = 1 , 2 , 3 , , 7 , the corresponding updated mass assignments were generated according to [38]:
m ( i ) = m ( i ) , i = t 1 , t 2
The remaining mass was assigned to the total set Θ :
m ( Θ ) = 1 m ( t 1 ) m ( t 2 ) .
For example, for a test sample t a r g e t 1 , we obtained the corresponding BBA from the PNNs, where m ( 1 ) = 0.7 , m ( 2 ) = 0.05 , m ( 3 ) = 0.2 , m ( 4 ) = 0.01 , m ( 5 ) = 0 , m ( 6 ) = 0.02 and m ( 7 ) = 0.02 . The dominant class was class 1, and class 3 was in second place. The updated corresponding BBA was m ( 1 ) = 0.7 , m ( 3 ) = 0.2 and m ( 2 , 4 , 5 , 6 , 7 ) = 0.1 .
There were 100 samples for each class, with a total of 700 samples. For each class, 50 samples were randomly selected for training PNNs, and the remaining samples were used for testing. For the MOEPT, the decision result would be class t f i n a l if
t f i n a l = a r g m a x ( M O E P T )
As we can see from Figure 4, the MOEPT performed well in this task of pattern classification.

5.3. Example of Target Type Tracking Using the MOEPT

To further discuss the practicality of the proposed MOEPT, a target type tracking (TTT) problem in the area of decision making was used, which is briefly described below [39].

5.3.1. Target Type Tracking Problem (TTT)

1. Consider ζ = 1 , 2 , , ζ m a x as the time index, and let there be N possible target types T a r ζ Θ = θ 1 , θ 2 , , θ N in the surveillance area. For instance, in normal air target surveillance systems, the FoD could be Θ = F i g h t e r , C a r g o ; that is, T a r 1 = θ 1 F i g h t e r , and T a r 2 = θ 2 C a r g o . Similarly, the FoD in a ground target surveillance system could be Θ g r o u n d = T a n k , T r u c k , C a r , B u s . In this paper, we just considered the air target surveillance systems to prove the practicability of EPT.
2. At every time ζ , the true type of the target T a r ζ Θ was immediately observed by an attribute-sensor (here, we assumed a possible target probability).
3. A defined classifier was applied to process the attribute measurement of the sensor, which provided the probability T a r d ζ for the type of observed target at each instance ζ .
4. The sensor was, in general, not totally reliable and was characterized by an N × N confusion matrix:
M = M i j = P T a r d = T a r j | T r u e T y p e = T a r i
where 0 i N ; 0 j N .
Here, we briefly summarize the main steps of the TTT using MOEPT:
1. Initialization: Determine the target type frame Θ = θ 1 , θ 2 , , θ N and set the initial BBA m i n i t i a l θ 1 θ 2 θ N = 1 , since there is no information about the first target type that will be observed;
2. Updating the BBA: An observed BBA m o b s ( . ) on the types of unknown observed targets is defined from the current target type declaration and confusion matrix M ;
3. Combination: We combine the current BBA m o b s ( · ) with the initial BBA m i n i t i a l ( · ) according to the PCR6 combination rule [7]: m P C R 6 ( · ) = m o b s ( · ) m i n i t i a l ( · ) ;
4. Approximation: Use M O E P T ( · ) to approximate m P C R 6 ( · ) into a Bayesian BBA;
5. Decision making: Make a final decision about the type of the target at the current observation time based on the obtained Bayesian BBA;
6. Updating the BBA: Set m i n i t i a l ( · ) = m P C R 6 ( · ) , and increase the time index ζ = ζ + 1 before going back to step 2.

5.3.2. Raw Dataset of TTT

We tested our MOEPT-based TTT on a very simple scenario for a 2D TTT problem, namely Θ = F i g h t e r , C a r g o , for two types of classifiers. The matrix M 1 corresponds to the confusion matrix of the good classifier, and M 2 corresponds to the confusion matrix of the poor classifier:
M 1 = 0.95 0.05 0.05 0.95 ; M 2 = 0.75 0.25 0.25 0.75
In our scenario, a true target type sequence over 120 scans was generated according to Figure 5. We can observe clearly from Figure 5 that Cargo (which is denoted as Type 2) appeared first in the sequence, and then the observation of the target type switched three times onto the Fighter type (Type 1) during different time durations (namely 20 s, 10 s and 5 s).
A pathological case for TTT: Our analysis showed that MOEPT can nevertheless be troublesome for tracking two target types, as proven in this particularly simple example (when 0 m ( θ 1 θ 2 ) 0.1 ). Let us consider the following BBA:
m t a r g e t ( . ) = θ 1 , θ 2 , θ 1 θ 2 = 0 , 1 , 0
According to the compatibility constraints in Equations (14)–(16), the population P t was obtained from P t through a selection procedure. Next, an individual P t j in P t , which is denoted as P t j = m ( θ 1 ) , m ( θ 2 ) , was subjected to the initial constraint in Equations (1) and (36):
m ( θ 1 ) ( B e l ( θ 1 ) = m ( θ 1 ) = 0 ) m ( θ 1 ) ( P l ( θ 1 ) = m ( θ 1 ) + m ( θ 1 θ 2 ) = 0 + 0 = 0 ) ; m ( θ 2 ) ( B e l ( θ 2 ) = m ( θ 2 ) = 1 ) m ( θ 2 ) ( P l ( θ 2 ) = m ( θ 2 ) + m ( θ 2 θ 1 ) = 1 + 0 = 1 ) ;
From the above inequalities, one can see that only one probability measure, P t S = m ( θ 1 ) , m ( θ 2 ) = 0 , 1 (where the superscript index S means single), satisfied this constraint (the constraint was m ( θ 1 ) [ B e l ( θ 1 ) , P l ( θ 1 ) ] = [ 0 , 0 ] , m ( θ 2 ) [ B e l ( θ 2 ) , P l ( θ 2 ) ] = [ 1 , 1 ] ). However because of the mechanism of MOEPT Equations (14)–(16), the P t j in population P t , which was randomly generated in the interval B e l ( θ i ) , P l ( θ i ) , i = 1 , 2 , · , N , would be unable to generate enough candidates for evolutionary computation. (A sufficient number of candidate sets is a prerequisite for ensuring the global optimization performance of evolutionary algorithms.) That is why MOEPT becomes inefficient in this case, which occurs with a probability of 1 / n max , where n max denotes the size of the population P t . (In our simulation, we had n max = 1000 .) Unfortunately, in TTT decision-making problems, such a case cannot be avoided because it can really happen.
To circumvent this problem and make the MOEPT approach work in most circumstances, we needed modify the MOEPT method a bit to generate enough individuals for making the selection steps efficient when the bounds of the belief interval B e l , P l took their minimum and maximum values ( 0.9 , 0.05 , 0.05 and 0.05 , 0.9 , 0.05 , respectively). To achieve this, we proposed enlarging this particular interval through a parameter λ and maintaining the property of the original interval to some degree at the same time. More precisely, the modified belief interval, denoted as B e l , P l , was heuristically computed by a simple thresholding technique as follows.
First, we assume that the original BBA we consider here for the FoD ( Θ = { θ 1 , θ 2 } ) is [ θ 1 , θ 2 , θ 1 θ 2 ] = [ a , b , c ] , with (a+b+c) = 1 and 0 c 0.1 :
  • Step 1: Let m ( θ 1 θ 2 ) = c + λ ;
  • Step 2: If a > b , then
    m ( θ 1 ) = a λ ; m ( θ 2 ) = b ; m ( θ 1 θ 2 ) = c + λ ;
  • Step 3: If a b , then
    m ( θ 1 ) = a ; m ( θ 2 ) = b λ ; m ( θ 1 θ 2 ) = c + λ ;
Therefore, the values of [ B e l ( θ 1 ) , P l ( θ 1 ) ] and [ B e l ( θ 2 ) , P l ( θ 2 ) ] can be calculated based on Equations (37) and (38), which are presented as follows. When a > b , we have
P l ( θ 1 ) = m ( θ 1 ) + m ( θ 1 θ 2 ) = a λ + c + λ = a + c ; B e l ( θ 1 ) = 1 P l ( θ ¯ 1 ) = 1 ( b + c + λ ) = a λ .
P l ( θ 2 ) = m ( θ 2 ) + m ( θ 1 θ 2 ) = b + c + λ = b + c + λ ; B e l ( θ 2 ) = 1 P l ( θ ¯ 2 ) = 1 ( a λ + c + λ ) = 1 ( a + c ) = b .
When a b , we have
P l ( θ 1 ) = m ( θ 1 ) + m ( θ 1 θ 2 ) = a + c + λ ; B e l ( θ 1 ) = 1 P l ( θ ¯ 1 ) = 1 ( b λ + c + λ ) = 1 ( b + c ) = a .
P l ( θ 2 ) = m ( θ 2 ) + m ( θ 1 θ 2 ) = b λ + c + λ = b + c ; B e l ( θ 2 ) = 1 P l ( θ ¯ 2 ) = 1 ( a + c + λ ) = b λ .
Explanation: Through step 1, one computes the total singleton mass one has in the entire BBA, and the threshold value of 0.9 allows one to evaluate if the percentage of the singleton mass is big enough or not. Here, we not only consider the unique extreme case m t a r g e t ( · ) = θ 1 , θ 2 , θ 1 θ 2 = 0 , 1 , 0 but also other possible cases, such as m t a r g e t ( · ) = θ 1 , θ 2 , θ 1 θ 2 = 0.0001 , 0.9998 , 0.0001 . Why do we consider the concept of this percentage? Actually, the higher the percentage of singleton mass, the smaller the interval for P t j ; in other words, the higher value of m θ 1 θ 2 , the bigger interval for P t j , which can be seen in Equation (36). Then, step 2 and step 3 give the method for calculating the updated upper bound of the belief interval B e l , P l , and Equations (39)–(42) prove that the parameter λ determines the range of the interval. Next, we give two examples to show how the above method works.
Pathological case one for TTT (revisited with a modified MOEPT):
m t a r g e t ( . ) = θ 1 , θ 2 , θ 1 θ 2 = 0.0001 , 0.9998 , 0.0001 .
Here, the parameter λ is arbitrarily set to 0.4. (The value of the parameter λ can be chosen to be any value in 0 , 1 by the designer for his or her own reasons to ensure the alternative interval effectively in the modified MOEPT.) Then, one computes in step 2 the modified plausibility bounds B e l ( θ 1 ) = 0.0001 , P l ( θ 1 ) = 0.0001 + 0.0001 + λ = 0.4002 and B e l ( θ 2 ) = 0.9998 0.4 = 0.5998 , P l ( θ 2 ) = 0.9999 . Therefore, we obtain B e l ( θ 1 ) , P l ( θ 1 ) = 0.0001 , 0.4002 and B e l ( θ 2 ) , P l ( θ 2 ) = 0.5998 , 0.9999 . The relationship between the original interval [ B e l , P l ] and the updated interval [ B e l , P l ] is illustrated in Figure 6.
Consequently, any Bayesian BBA P t j = m ( θ 1 ) , m ( θ 2 ) must be generated according the (modified) compatibility constraints:
m ( θ 1 ) B e l ( θ 1 ) , P l ( θ 1 ) = [ 0.0001 , 0.4002 ] m ( θ 2 ) B e l ( θ 2 ) , P l ( θ 2 ) = [ 0.5998 , 0.9999 ]
Pathological case two for TTT (revisited with a modified MOEPT):
m t a r g e t ( . ) = θ 1 , θ 2 , θ 1 θ 2 = 0.45 , 0.48 , 0.07 .
Here, the parameter λ is set to 0.2. Then, any Bayesian BBA P t j = m ( θ 1 ) , m ( θ 2 ) must be generated according the (modified) compatibility constraints:
m ( θ 1 ) B e l ( θ 1 ) , P l ( θ 1 ) = [ 0.45 , 0.72 ] m ( θ 2 ) B e l ( θ 2 ) , P l ( θ 2 ) = [ 0.28 , 0.55 ]
In order to evaluate the influence of parameter λ , we reexamined all the pathological cases based on the following procedure:
  • The value of parameter λ was set to five possible values: 0, 0.1, 0.2, 0.3, 0.4 and 0.5;
  • We randomly generated an initial population P t based on λ , which was also subject to the constraints in Equations (14)–(16).
With this simulation, we can observe in Figure 7 and Figure 8 the impact of the λ value on the number of P t j in P t . When we set λ = 0 (in which the original MOEPT was applied), there existed no suitable P t j for case one, which demonstrates the necessity to circumvent the pathological case problem. Obviously, the number of P t j increased with the increase in the λ value, which efficiently proves the advantage of using the modified MOEPT approach to make the selection step of the evolutionary algorithm more efficient. One point we need to clarify is that the intervals (i.e., B e l ( θ 1 ) , P l ( θ 1 ) , B e l ( θ 2 ) , P l ( θ 2 ) ) induced from the parameter λ above aim at guaranteeing enough of a number of P t j in P t in the implementation of the MOEPT. Another point we also need to mention is that the number of P t j in P t was not influenced by the weight. (Here, the weight equals w 2 in Equation (13). And thus, w 1 = 1 W e i g h t , which to some degree guarantees the implementation of the MOEPT.)

5.3.3. Simulation Results of TTT Based on the Modified MOEPT

Our simulation consisted of 100 Monte Carlo runs, and we show in the sequel the averaged performances of the MOEPT. Figure 9 and Figure 10 illustrate the Bayesian BBAs obtained by our new MOEPT method for solving the TTT problem using the PCR6 fusion rule. One can see that regardless of the good classifier M 1 (recognition rate: 90.83 %) or poor classifier M 2 (recognition rate: 80.83%) being used, the MOEPT was able to track properly the quick changes in target type.

6. Conclusions

A multi-objective evolutionary-based algorithm for probabilistic transformation (MOEPT) was proposed in this paper. It uses a genetic algorithm to obtain a Bayesian belief function and offer a comprehensive consideration concerning the closeness of distance between the orignal BBA and the Bayesian approximate one. In addition, a new aggregation measure was proposed in this paper to be combined into a more accurate “distance closeness” measure for MOEPT. More importantly, the convergence analysis of the MOEPT was given to prove the rationality of our proposed method. The effectiveness of the MOEPT was compared with respect to several probabilistic transformations proposed in the literature. Furthermore, the shortcomings of the original MOEPT version were clearly identified in two target type tracking problems, and they were solved thanks to modification of the belief interval constraints. As for future works, we would like to establish an adaptive scheme for the selection of weights in an MOEPT and make more comparisons between the performance of this MOEPT approach and other recently proposed evolutionary algorithms. We would also conduct more investigations to extend the MOEPT to DSmT using the DSm cardinal of elements. That aside, the current work in this paper is mainly for verifying the effectiveness of the algorithm through simulation examples from a theoretical perspective, and the feasibility of the proposed evolutionary-based PT will be verified through the practical and real problems in our future discussions.

Author Contributions

Conceptualization, Y.D. and L.C.; methodology, Y.D. and L.C.; software, Y.D. and L.C.; validation, Y.D. and L.C.; formal analysis, Y.D.; investigation, Y.D.; writing—original draft preparation, Y.D.; writing—review and editing, Y.D. and L.C.; visualization, Y.D. and K.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Nature Science Fund, grant number 62103258.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dempster, A.P. Upper and lower probabilities induced by a multivalued mapping. Ann. Math. Stat. 1967, 38, 325–339. [Google Scholar] [CrossRef]
  2. Shafer, G. A Mathematical Theory of Evidence; Princeton University Press: Princeton, VA, USA, 1976. [Google Scholar]
  3. Ghaffari, M.; Ghadiri, N. Ambiguity-driven fuzzy C-means clustering: How to detect uncertainty clustered records. Appl. Intell. 2016, 2, 293–304. [Google Scholar] [CrossRef] [Green Version]
  4. Su, X.Y.; Mahadevan, S.; Han, W.H.; Deng, Y. Combining dependent bodies of evidence. Appl. Intell. 2015, 3, 634–644. [Google Scholar] [CrossRef]
  5. Denœux, T. Inner and outer approximation of belief structures using a hierarchical clustering approach. Int. J. Uncertain. Fuzziness Knowl. Based Syst. 2001, 9, 437–460. [Google Scholar] [CrossRef]
  6. Smets, P. The Combination of Evidence in the Transferable Belief Model. IEEE Trans. PAMI 1990, 5, 29–39. [Google Scholar] [CrossRef]
  7. Smarandache, F.; Dezert, J. (Eds.) Advances and Applications of DSmT for Information Fusion; American Research Press: Rehoboth, NM, USA, 2015; pp. 1–4. [Google Scholar]
  8. Daniel, M. Probabilistic Transformations of Belief Functions. In Proceedings of the 8th European Conference on Symbolic and Quantitative Approaches to Reasoning and Uncertainty (ECSQARU), Barcelona, Spain, 6–8 July 2005; pp. 539–551. [Google Scholar]
  9. Dezert, J.; Smarandache, F. A new probabilistic transformation of belief mass assignment. In Proceedings of the11th International Conference on Information Fusion, Cologne, Germany, 30 June–3 July 2008; pp. 1–8. [Google Scholar]
  10. Cobb, B.R.; Shenoy, P.P. On the plausibility transformation method for translating belief function models to probability models. Int. J. Approx. Reason. 2006, 41, 314–330. [Google Scholar] [CrossRef] [Green Version]
  11. Sudano, J. Equivalence between belief theories and naive bayesian fusion for systems with independent evidential data: Part I, the theory. In Proceedings of the Fusion 2003, Cairns, Australia, 8–11 July 2003. [Google Scholar]
  12. Cuzzolin, F. The Intersection probability and its properties. In Proceedings of the ECSQARU Conference, Verona, Italy, 1–3 July 2009. [Google Scholar]
  13. Han, D.Q.; Dezert, J.; Han, C.; Yang, Y. Is entropy enough to evaluate the probability transformation approach of belief function? In Proceedings of the 13th Conference on Information Fusion, Edinburgh, UK, 26–29 July 2010; pp. 286–293. [Google Scholar]
  14. Bucci, D.J.; Acharya, S.; Pleskac, T.J.; Kam, T.J. Performance of probability transformations using simulated human opinions. In Proceedings of the 17th International Conference on Information Fusion, Salamanca, Spain, 7–10 July 2014; pp. 1–8. [Google Scholar]
  15. Yang, Y.; Han, D.Q. A new distance-based total uncertainty measure in the theory of belief function. Knowl.-Based Syst. 2016, 94, 114–123. [Google Scholar] [CrossRef]
  16. Liu, W.R. Analyzing the degree of conflict among belief functions. Artif. Intell. 2006, 170, 909–924. [Google Scholar] [CrossRef] [Green Version]
  17. Han, D.Q.; Dezert, J.; Yang, Y. New Distance Measures of Evidence based on Belief Intervals. In Proceedings of the International Conference on Belief Functions, Oxford, UK, 26–28 September 2014. [Google Scholar]
  18. Frikha, A. On the use of a multi-criteria approach for reliability estimation in belief function theory. Inf. Fusion 2014, 18, 20–32. [Google Scholar] [CrossRef]
  19. Han, D.Q.; Dezert, J.; Duan, Z.S. Evaluation of Probability Transformations of Belief Functions for Decision Making. IEEE Trans. Syst. Man Cybern. Syst. 2015, 99, 93–108. [Google Scholar] [CrossRef]
  20. Liu, Z.G.; Dezert, J.; Pan, Q.; Mercier, G. Combination of sources of evidence with different discounting factors based on a new dissimilarity measure. Decis. Support Syst. 2011, 52, 133–141. [Google Scholar] [CrossRef]
  21. Deng, Z.; Wang, J. A novel decision probability transformation method based on belief interval. Knowl.-Based Syst. 2020, 208, 427–450. [Google Scholar] [CrossRef]
  22. Zhao, K.Y.; Chen, Z.Q.; Li, L.; Li, J.Y.; Sun, R.Z.; Yuan, G. DPT: An importance-based decision probability transformation method for uncertain belief in evidence theory. Expert Syst. Appl. 2022, 11, 120–133. [Google Scholar] [CrossRef]
  23. Ma, M.M.; An, J.Y. Combination of evidence with different weighting factors: A novel probabilistic-based dissimilarity measure approach. J. Sens. 2015, 2015, 509385. [Google Scholar] [CrossRef] [Green Version]
  24. Dezert, J.; Han, D.Q.; Tacnet, J.M.; Carladous, S.; Yang, Y. Decision-Making with Belief Interval Distance. In Proceedings of the International Conference on Belief Functions, Prague, Czech Republic, 21–23 September 2016. [Google Scholar]
  25. Han, D.Q.; Deng, Y.; Han, C.Z. Sequential weighted combination for unreliable evidence based on evidence variance. Decis. Support Syst. 2013, 56, 387–393. [Google Scholar] [CrossRef]
  26. Smets, P. Decision making in the TBM: The necessity of the pignistic transformation. Int. Approx. Reason. 2005, 38, 133–147. [Google Scholar] [CrossRef] [Green Version]
  27. Pan, W.; Hong, J.Y. New methods of transforming belief functions to pignistic probability functions in evidence theory. In Proceedings of the 2009 International Workshop on Intelligent Systems and Applications, Wuhan, China, 23–24 May 2009. [Google Scholar]
  28. Jousselme, A.L.; Liu, C.S.; Grenier, D.; Bossé, E. Measuring Ambiguity in the Evidence Theory. IEEE Trans. Syst. Man Cybern. Part Syst. Hum. 2009, 36, 890–903. [Google Scholar] [CrossRef]
  29. Irpino, A.; Verde, R. Dynamic clustering of interval data using a wasserstein-based distance. Pattern Recognit. Lett. 2008, 29, 1648–1658. [Google Scholar] [CrossRef]
  30. Burger, T. Geometric views on conflicting mass functions: From distances to angles. Int. J. Approx. Reason. 2016, 70, 36–50. [Google Scholar] [CrossRef]
  31. Sudano, J. The system probability information content (PIC) relationship to contributing components, combining independent multisource beliefs, hybrid and pedigree pignistic probabilities. In Proceedings of the 5th International Conference on Information Fusion. FUSION 2002.(IEEE Cat. No. 02EX5997), Annapolis, MA, USA, 8–11 July 2002; pp. 1277–1283. [Google Scholar]
  32. Li, X.; Wang, F. A Clustering-Based Evidence Reasoning Method. Int. J. Intell. Syst. 2016, 31, 698–721. [Google Scholar] [CrossRef]
  33. Srinivas, N.; Deb, K. Multiobjective function optimization using nondominated sorting genetic algorithms. Evol. Comput. 1995, 2, 221–248. [Google Scholar] [CrossRef]
  34. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A Fast and Elitist Multiobjective Genetic Algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef] [Green Version]
  35. Zhu, Q.B. Analysis of Convergence of Ant Colony Optimization Algorithms. Control Decis. 2006, 21, 764–770. [Google Scholar] [CrossRef]
  36. Jousselme, A.L.; Maupin, P. On some properties of distances in evidence theory. In Proceedings of the Workshop on Theory of Belief Functions, Brest, France, 31 March–2 April 2010; pp. 1–6. [Google Scholar]
  37. Li, X.D.; Yang, W.D.; Dezert, J. An Airplane Image Target’s Multi-feature Fusion Recognition Method. Acta Autom. Sin. 2012, 38, 1299–1307. [Google Scholar] [CrossRef]
  38. Bi, Y.; Bell, D.; Guan, J. Combining evidence from classifiers in text categorization. In Proceedings of the 8th International Conference on Knowledge-Based and Intelligent Information and Engineering Systems, Wellington, New Zealand, 20–25 September 2004; pp. 521–528. [Google Scholar]
  39. Dezert, J.; Tchamova, A.; Smarandache, F.; Konstantinova, P. Target Type Tracking with PCR5 and Dempster’s rules: A Comparative Analysis. In Proceedings of the 2006 9th International Conference on Information Fusion, Florence, Italy, 10–13 July 2006. [Google Scholar]
Figure 1. Scheme of MOEPT algorithm.
Figure 1. Scheme of MOEPT algorithm.
Entropy 24 01680 g001
Figure 2. Comparisons between MOEPT and state-of-the-art PTs.
Figure 2. Comparisons between MOEPT and state-of-the-art PTs.
Entropy 24 01680 g002
Figure 3. The binary images of seven kinds of airplanes.
Figure 3. The binary images of seven kinds of airplanes.
Entropy 24 01680 g003
Figure 4. Recognition rate of MOEPT.
Figure 4. Recognition rate of MOEPT.
Entropy 24 01680 g004
Figure 5. Raw sequence of true target type.
Figure 5. Raw sequence of true target type.
Entropy 24 01680 g005
Figure 6. The principle of modified-constraint MOEPT ( λ = 0.4 ).
Figure 6. The principle of modified-constraint MOEPT ( λ = 0.4 ).
Entropy 24 01680 g006
Figure 7. Impact of λ (x-axis) on individuals in P t (y-axis) for case one.
Figure 7. Impact of λ (x-axis) on individuals in P t (y-axis) for case one.
Entropy 24 01680 g007
Figure 8. Impact of λ (x-axis) on individuals in P t (y-axis) for case two.
Figure 8. Impact of λ (x-axis) on individuals in P t (y-axis) for case two.
Entropy 24 01680 g008
Figure 9. Results of MOEPT for Cargo and Fighter types using M 1 .
Figure 9. Results of MOEPT for Cargo and Fighter types using M 1 .
Entropy 24 01680 g009
Figure 10. Results of MOEPT for Cargo and Fighter types using M 2 .
Figure 10. Results of MOEPT for Cargo and Fighter types using M 2 .
Entropy 24 01680 g010
Table 1. Results of different PTs in Example 3 ( w 1 = w 2 = 0.5).
Table 1. Results of different PTs in Example 3 ( w 1 = w 2 = 0.5).
θ 1 θ 2 θ 3 θ 4 C norm
CuzzP 0.38600.33820.16070.11510.2800
BetP 0.39830.34330.15330.10500.2799
DSmP 0 0.51760.40510.03030.04700.1897
DSmP 0 . 001 0.51620.40430.03190.04770.1896
PrBP 1 0.54190.39980.02430.03400.1918
PrBP 2 0.55780.38420.02260.03530.1933
MOEPT D J + Sim seq 0.39800.33220.11560.15410.1849
MOEPT d BI E + Sim seq 0.39850.39830.06230.14090.0733
Table 2. Results of different PTs in Example 4.
Table 2. Results of different PTs in Example 4.
θ 1 θ 2 θ 3 θ 4 C norm
B e t P 0.39830.34330.15330.10500.3974
D S m P 0 0.25000.25000.25000.25000.5458
P r B P 1 0.54190.39980.02430.03400.6368
P r B P 2 0.55780.38420.02260.03530.6412
M O E P T D i s t a n c e 0.25000.15970.35780.23250.3415
MOEPT D J + Sim seq 0.24830.24850.24960.25360.1708
MOEPT d BI E + Sim seq 0.24840.24880.24890.25390.0450
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Dong, Y.; Cao, L.; Zuo, K. Genetic Algorithm Based on a New Similarity for Probabilistic Transformation of Belief Functions. Entropy 2022, 24, 1680. https://doi.org/10.3390/e24111680

AMA Style

Dong Y, Cao L, Zuo K. Genetic Algorithm Based on a New Similarity for Probabilistic Transformation of Belief Functions. Entropy. 2022; 24(11):1680. https://doi.org/10.3390/e24111680

Chicago/Turabian Style

Dong, Yilin, Lei Cao, and Kezhu Zuo. 2022. "Genetic Algorithm Based on a New Similarity for Probabilistic Transformation of Belief Functions" Entropy 24, no. 11: 1680. https://doi.org/10.3390/e24111680

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop