Next Article in Journal
On Finite-Time Blow-Up Problem for Nonlinear Fractional Reaction Diffusion Equation: Analytical Results and Numerical Simulations
Next Article in Special Issue
Fractional Transformation-Based Decentralized Robust Control of a Coupled-Tank System for Industrial Applications
Previous Article in Journal
Local Stability, Global Stability, and Simulations in a Fractional Discrete Glycolysis Reaction–Diffusion Model
Previous Article in Special Issue
On Fuzzy and Crisp Solutions of a Novel Fractional Pandemic Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

New Classifier Ensemble and Fuzzy Community Detection Methods Using POP Choquet-like Integrals

1
School of Mathematics and Data Science, Shaanxi University of Science and Technology, Xi’an 710016, China
2
Shaanxi Joint Laboratory of Artificial Intelligence, Shaanxi University of Science and Technology, Xi’an 710016, China
*
Author to whom correspondence should be addressed.
Fractal Fract. 2023, 7(8), 588; https://doi.org/10.3390/fractalfract7080588
Submission received: 17 June 2023 / Revised: 27 July 2023 / Accepted: 28 July 2023 / Published: 30 July 2023

Abstract

:
Among various data analysis methods, classifier ensemble (data classification) and community network detection (data clustering) have aroused the interest of many scholars. The maximum operator, as the fusion function, was always used to fuse the results of the base algorithms in the classifier ensemble and the membership degree of nodes to classes in the fuzzy community. It is vital to use generalized fusion functions in ensemble and community applications. Since the Pseudo overlap function and the Choquet-like integrals are two new fusion functions, they can be combined as a more generalized fusion function. Along this line, this paper presents new classifier ensemble and fuzzy community detection methods using a pseudo overlap pair (POP) Choquet-like integral (expressed as a fraction). First, the pseudo overlap function pair is proposed to replace the product operator of the Choquet integral. Then, the POP Choquet-like integrals are defined to perform the combinatorial step of ensembles of classifiers and to generalize the GN modularity for the fuzzy community network. Finally, two new algorithms are designed for experiments, and some computational experiments with other algorithms show the importance of POP Choquet-like integrals. All of the experimental results show that our algorithms are practical.

1. Introduction

A decade ago, data seemed primarily the preserve of researchers and a few professional managers. It has entered our everyday lexicon with the constant refinement and iteration of hardware and software, and our digital world generates and consumes vast amounts of data daily. Data analysis topics include classification, clustering, mining, data association analysis, etc. This paper mainly studies data classification and clustering algorithms in data analysis.

1.1. Classifier Ensembles and Choquet Integrals

Data classification is a model that describes and distinguishes data classes based on existing data sets. Usually, such a model is also called a classifier. The commonly used classifier algorithms include the Naive Bayes method [1], support vector machine [2], decision tree [3], artificial neural network [4], k-nearest neighbor [5], etc. At the same time, in real life, a single classification algorithm can only handle some specific types of data, which makes the classification work tedious. Therefore, people are eager to implement a classification algorithm that can deal with various types of data, and classifier ensembles can do this work well.
An ensemble of classifiers is a mechanism that fuses multiple classifiers, which consists of two steps: classification and fusion. It combines the benefits of multiple classifiers so that any type of data set can be classified well. It also reduces the error generated by a single classifier, namely overfitting. Recently, in the literature, classifier ensemble has been used in areas such as disease detection [6], social networks [7] and mood recognition [8]. In these articles, the authors adopted a classifier ensemble for the current context, which yielded good results. However, they simply replaced or improved the base classifier algorithm of classifier ensembles and chose a simple approach to the fusion process, such as voting and weighting.
Obviously, using the Choquet integral as a fusion operator is more complex, and many scholars have used this ensemble approach to study in different contexts. In the literature, the author of [9] used the Choquet integral to fuse multiple classifiers to design credit score models. The results show that this model improves the accuracy of the classification algorithm. In  [8], the author used the Choquet integral to integrate the segmentation results of multiple classifiers to obtain a better image segmentation scheme. In [10], the author used a classifier ensemble based on the Choquet integral to classify malware in smartphones and experimentally verified that the method is superior to a single classifier, with an accuracy of 95.08%. Especially, in [11], Batista et al. thought that it was more appropriate to replace the product operator with the same property in the Choquet integral with a quasi-overlap function that did not require associativity and continuity, so the Choquet-like integral based on quasi-overlap functions (CQO integral) was constructed. The author used the CQO integral to solve the classifier ensemble problem and compared it with other ensemble algorithms, and proved that it has advantages in solving the classifier ensemble problem.
Inspired by the CQO integral in [11] which is shown as
C m O ( l ) = i = 1 n O ( l ( i ) , m ( A ( i ) ) ) O ( l ( i 1 ) , m ( A ( i ) ) ) ,
we find the following two urgent tasks:
(1)
the CQO integral did not map to the [0,1] interval, similarly to many fusion functions, but rather to the [0,n] interval;
(2)
the CQO integral used the same O before and after replacing the product in the Choquet integral, which may lead to inconspicuous results for different inputs.
Therefore, this paper aims to design a new Choquet-like integral that not only naturally maps to the [0,1] interval but also replaces the product operator of the Choquet integral with two different pseudo overlap functions, which we named the pseudo overlap pair (POP) Choquet-like integral. We face two important questions for that: does the POP Choquet-like integral meet the boundary conditions and contacting (pre)aggregate functions? Does the POP Choquet-like integral still play a good role in classifier ensemble?
To solve these problems, we have certain requirements for selecting a pseudo-function pair, and some theories prove that the POP Choquet-like integral is a pre-aggregate function under some conditions. Moreover, we will use the POP Choquet-like integral as the fusion function for the classifiers ensemble and design an algorithm for experimental verification.

1.2. Community Network Detection and Choquet Integrals

Scholar Jain pointed out in [12] that clustering is classifying patterns (observations, data items, or feature vectors) without supervision. There are many clustering methods, and the common ones are nearest neighbor clustering [13]; fuzzy clustering [14]; artificial neural network clustering [15], etc. All of the above clustering methods can be used for complex network detection.
Complex network detection has been a popular topic in recent years because it involves a wide and deep range, including but not limited to telecommunications networks [16], community networks [17], traffic networks [18], and biological networks [19]. Among them, the nodes of the community networks are usually composed of people, and the connection between the nodes is dominated by familiarity, emotion, information, and other factors. Generally speaking, a community in a network is a subgraph characterized by close connections between nodes within the subgraph but sparse connections between subgraphs. The study of community structure is significant in detecting community networks.
Considering the overlapping character of the complex community networks, the fuzzy community detection (FCD) methods were designed to calculate the membership degree of nodes in each community network. Many FCD methods previously proposed are only partially suitable for fuzzy frameworks, such as [20,21]. Specifically, in [22], the author used the overlap and the grouping functions to aggregate the membership of nodes to classes, generalized the classical GN modularity, and designed a community network detection algorithm that is completely suitable for fuzzy frameworks. However, when designing this algorithm, the author still selected the maximum as a group function; using the maximum function as an average could limit the results. Using fusion functions with non-average properties in most applications is more appropriate.
Generalized Choquet integrals are a class of non-average sum vector aggregation functions with good performance. Therefore, in this paper, the POP Choquet-like integral is used to replace the maximum operation in the definition of modularity in [22], which improves the membership processing of nodes after fuzzy clustering and can effectively detect overlapping communities. Moreover, we replace the overlap function in [22] with a pseudo overlap function that does not require commutativity, which is more in line with the actual situation. Then we will design a community network detection algorithm to prove our FCD method is effective.

1.3. Organizational Structure of This Paper

The rest of this article is framed as follows: Section 2 reviews pseudo overlap functions, Choquet integral and its generalizations, and preaggregate function. In Section 3, we introduce the pseudo overlap function pair and define the POP Choquet-like integral, discussing its properties in some cases. In Section 4 and Section 5, we design a new ensemble classification algorithm and a new community network detection algorithm by using the defined POP Choquet-like integrals and design some experiments to prove their performance. Finally, Section 6 summarizes the full paper.

2. Preliminaries

This section aims to introduce the basic theories necessary for this article.
Definition 1
([23]). Let l = ( l 1 , , l n ) be a non-zero real vector of n-dimension. If for any u = ( u 1 , , u n ) [ 0 , 1 ] n and c > 0 , u + c l = ( u 1 + c l 1 , , u n + c l n ) [ 0 , 1 ] n , the function Z : [ 0 , 1 ] n [ 0 , 1 ] satisfies Z ( u 1 + c l 1 , , u n + c l n ) > Z ( u 1 , , u n ) , then Z is l -increasing.
Definition 2
([23]). P A F : [ 0 , 1 ] n [ 0 , 1 ] is said to be an n-dimensional pre-aggregate function if it satisfies the all following conditions.
( P A F 1 ) directionally l-increasing: there exists a non-zero vector l = ( l 1 , , l n ) [ 0 , 1 ] n such that the function P A F is l increasing;
( P A F 2 )   P A F ( 0 , , 0 ) = 0 ;
( P A F 3 )   P A F ( 1 , , 1 ) = 1 .
Definition 3
([24]). A pseudo overlap function P O : [ 0 , 1 ] 2 [ 0 , 1 ] is a binary function that satisfies all of the following properties:
( P O 1 )   P O ( x , y ) = 0 if and only if x y = 0 ;
( P O 2 )   P O ( x , y ) = 1 if and only if x y = 1 ;
( P O 3 )   P O is incremental;
( P O 4 )   P O is continuous.
In Table 1, we give some examples of pseudo overlap functions.
Definition 4
([25]). Let N = { 1 , 2 , , n } . X Y N . If the following conditions are met, the function m : P ( N ) [ 0 , 1 ] will be called a fuzzy measure.
( m 1 )   m ( Ø ) = 0 ;
( m 2 )   m ( N ) = 1 ;
( m 3 ) Incremental: m ( X ) m ( Y ) .
Definition 5
([25]). Let N = { 1 , 2 , , n } , H N . The most classical fuzzy measure is the uniform fuzzy measure, which is defined as follows:
m U ( H ) = H n .
Some of the functions in Table 1 are of the same family, such as pseudo-overlapping functions with serial numbers 2 and 8, where the arguments α are 1 and 0.5 for pseudo-overlapping function family
P O ( x , y ) = 2 x y α x + ( 2 α ) y , α x + ( 2 α ) y 0 , 0 , α x + ( 2 α ) y = 0 . α 0 , 2 ,
respectively. Due to the subsequent need to distinguish the size relationships between functions, it is necessary to determine the parameters of each family of functions. After many experiments, the pseudo-overlapping functions of these deterministic parameters in Table 1 are best fused in the two algorithms in this paper.
Another classical fuzzy measure is the most commonly used one, the  g λ fuzzy measure.
Definition 6
([26]). Let λ 1 , N = { 1 , 2 , , n } . Function m : P ( N ) [ 0 , 1 ] satisfies the g λ law that for any disjoint sequence { E 1 , E 2 , , E n , } in P ( N ) , and their union is also in P ( N ) , there is
m ( i = 1 E i ) = 1 λ i = 1 1 + λ m ( E i ) 1 , λ 0 , i = 1 m ( E i ) , λ = 0 .
Here, the λ coefficient is obtained by solving the following equation:
λ + 1 = x A ( 1 + λ m ( { x } ) ) .
The formula (1) is called g λ fuzzy measure. When g λ fuzzy measure is used in our experiment, each x E i initial measure value m ( { x } ) is determined by the membership value in the initial classification result.
Definition 7
([25]). Let m : P ( N ) [ 0 , 1 ] be a fuzzy measure. For any l = ( l 1 , l 2 , , l n ) [ 0 , 1 ] n , the discrete Choquet integral C m : [ 0 , 1 ] n [ 0 , 1 ] is constructed as follows:
C m ( l ) = i = 1 n l ( i ) · m ( A ( i ) ) l ( i 1 ) · m ( A ( i ) ) ,
where ( l ( 1 ) , , l ( n ) ) is an increasing permutation of l , i.e.,  0 l ( 1 ) l ( n ) , and the initial value l ( 0 ) = 0 . A ( i ) = ( i ) , , ( n ) is the subset of indices corresponding to the n i + 1 largest components of l .
Lucca et al. [23] constructed a series of Choquet-like integrals by replacing the product operators of Choquet integrals with t-norm, copula, and fusion function pair, which are proven to have good performance in fuzzy-rule classification systems (FRBCS). In this paper, Lucca’s recently-constructed Choquet-like integral based on pair of fusion functions is referred to as C m ( F 1 , F 2 ) . The definition of C m ( F 1 , F 2 ) is as follows:
Definition 8
([23]). Let m : P ( N ) [ 0 , 1 ] be a fuzzy measure, ( F 1 , F 2 ) be a pair of fusion functions satisfying x , y 0 , 1 , F 1 x , y F 2 x , y . For any l = ( l 1 , l 2 , , l n ) [ 0 , 1 ] n , the integral C m F 1 , F 2 : [ 0 , 1 ] n [ 0 , 1 ] is constructed as follows:
C m F 1 , F 2 ( l ) = min 1 , l ( 1 ) + i = 2 n F 1 ( l ( i ) , m ( A ( i ) ) ) F 2 ( l ( i 1 ) , m ( A ( i ) ) ) ,
where ( l ( 1 ) , , l ( n ) ) is an increasing permutation of l , i.e.,  0 l ( 1 ) l ( n ) , and the initial value l ( 0 ) = 0 . A ( i ) = ( i ) , , ( n ) is the subset of indices corresponding to the n i + 1 largest components of l .
To improve the quality of the classifier ensemble, in [11], the CQO integral is used as the fusion function in an ensemble. The definition of CQO is as follows
Definition 9
([11]). Let m : P ( N ) [ 0 , 1 ] be a fuzzy measure, O : [ 0 , 1 ] 2 [ 0 , 1 ] be a quasi overlap function. for any l = ( l 1 , l 2 , , l n ) 0 , 1 n , the CQO integral C m O : [ 0 , 1 ] n [ 0 , n ] is constructed as follows:
C m O ( l ) = i = 1 n O ( l ( i ) , m ( A ( i ) ) ) O ( l ( i 1 ) , m ( A ( i ) ) ) ,
where ( l ( 1 ) , , l ( n ) ) is an increasing permutation of l , i.e.,  0 l ( 1 ) l ( n ) , and the initial value l ( 0 ) = 0 . A ( i ) = ( i ) , , ( n ) is the set of index that corresponds to the previous n i + 1 largest element.

3. POP Choquet-like Integral

Although several Choquet-like integrals have been proposed for various scenarios, their structures have always been somewhat unnatural. For example, C m ( F 1 , F 2 ) directly maps integral values greater than 1 to 1, making it impossible to compare the results. In addition, the integral C m O maps the value to [0, n]. However, the codomain of the common fusion function is [0, 1], so it needs some restrictions to reduce its range to [0, 1], which makes the application more difficult.
In addition, since the product operator in the original Choquet integral does not force commutativity and associativity, the pseudo overlap function is a good substitution function. Additionally, the non-average function is more competitive than the average function in many application scenarios. Therefore, in this section, we use pseudo overlap function pair P O 1 , P O 2 to generalize the Choquet integral to obtain the POP Choquet-like integral. The following is its construction procedure.
Definition 10.
For given two pseudo overlap functions P O 1 , P O 2 : 0 , 1 2 0 , 1 . P O 1 , P O 2 is called a pseudo overlap function pair as long as it satisfies P O 1 x , y P O 2 x , y for any x , y 0 , 1 .
We compare the size of the 16 functions in Table 1 and list 15 pseudo-overlap function pairs in Table 2.
For the n-dimensional incremental vector l = l 1 , l 2 , , l n 0 , 1 n , since its components may contain duplicates, the order of components of the incremental vector l can be controversial in constructing the POP Choquet-like integral afterward. Hence, vector l needs to be reduced in dimension to ensure that it has no duplicate components.
Definition 11.
For a given l = l 1 , l 2 , , l n 0 , 1 n ( n N , n 2 ) . The dimensionality reduction function is defined as D R : 0 , 1 n 0 , 1 0 , 1 2 0 , 1 n , in the form:
D R l 1 , l 2 , , l n = y = y 1 , y 2 , , y k , ( 1 k n )
which satisfies the following conditions:
( D R 1 )   y i < y i + 1 , i = 1 , 2 , , k 1 ;
( D R 2 )   y i y , l j , l j + 1 , , l j + p , 1 j n , 0 p n j , s . t . y i = l j = l j + 1 = = l j + p .
After defining the pseudo overlap function pair and dimensionality reduction function, we can design a POP Choquet-like integral. After that, we will also explore the boundary conditions for POP Choquet-like integral and the conditions under which the POP Choquet-like integral is a pre-aggregate function.
Definition 12.
Let N = { 1 , 2 , n } be a finite set, m : P ( N ) [ 0 , 1 ] be a fuzzy measure, P O 1 , P O 2 be a pseudo overlap function pair that satisfying x , y 0 , 1 , P O 1 x , y P O 2 x , y , D R : 0 , 1 n 0 , 1 0 , 1 2 0 , 1 n be a dimensionality reduction function. For any x = x 1 , x 2 , , x n 0 , 1 n , POP Choquet-like integral C m P O 1 , P O 2 : 0 , 1 n 0 , 1 is defined as follows:
C m P O 1 , P O 2 ( x ) = j = 1 k P O 1 ( y ( j ) , m ( A ( j ) ) ) P O 2 ( y ( j 1 ) , m ( A ( j ) ) ) k ,
of which
y = y 1 , y 2 , , y k = D R x 1 , x 2 , , x n ,
and ( y ( 1 ) , , y ( k ) ) is an increasing permutation on y , that is, 0 y ( 1 ) y ( k ) , where y ( 0 ) = 0 and A ( j ) = { ( j ) , , ( k ) } is the subset of indices corresponding to the k j + 1 largest components of y .
Table 3 lists fifteen POP Choquet-like integrals based on pseudo overlap function pairs.
Remark 1.
C m P O 1 , P O 2 is well defined for arbitrary pairs of pseudo overlap functions and fuzzy measures m.
Obtained by Definition 12, for any given x = x 1 , x 2 , , x n 0 , 1 n , dimensionality reduction function
D R x 1 , x 2 , , x n = y 1 , y 2 , , y k .
and ( y ( 1 ) , , y ( k ) ) is an increasing permutation on y , so
C m P O 1 , P O 2 x 1 , x 2 , , x n = j = 1 k P O 1 ( y ( j ) , m ( A ( j ) ) ) P O 2 ( y ( j 1 ) , m ( A ( j ) ) ) k j = 1 k P O 1 ( y ( j ) , m ( A ( j ) ) ) P O 1 ( y ( j 1 ) , m ( A ( j ) ) ) k 0 .
C m P O 1 , P O 2 x 1 , x 2 , , x n = j = 1 k P O 1 ( y ( j ) , m ( A ( j ) ) ) P O 2 ( y ( j 1 ) , m ( A ( j ) ) ) k j = 1 k P O 1 ( y ( j ) , m ( A ( j ) ) ) k 1 .
Proposition 1.
Let P O 1 , P O 2 be a pair of pseudo overlap functions, m be a fuzzy measure, D R be a dimensionality reduction function. We have C m P O 1 , P O 2 0 , 0 , , 0 = 0 , and C m P O 1 , P O 2 ( 1 , 1 , ,   1 ) = 1 .
Proof. 
C m P O 1 , P O 2 0 , 0 , , 0 = C m P O 1 , P O 2 0 = P O 1 ( 0 , m ( A ( 1 ) ) ) 0 = 0 .
C m P O 1 , P O 2 1 , 1 , , 1 = C m P O 1 , P O 2 1 = P O 1 ( 1 , m ( A ( 1 ) ) ) 0 = P O 1 ( 1 , 1 ) = 1 .
   □
Proposition 2.
Let P O 1 , P O 2 be a pair of pseudo overlap functions, m be a fuzzy measure, and  D R be a dimensionality reduction function. C m P O 1 , P O 2 is idempotent if and only if P O 1 has a neutral element.
Proof. 
Let l [ 0 , 1 ] , then
C m P O 1 , P O 2 ( l , l , , l ) = C m P O 1 , P O 2 ( l ) = P O 1 ( l , m ( A ( 1 ) ) ) P O 2 ( 0 , m ( A ( 1 ) ) ) = P O 1 ( l , m ( A ( 1 ) ) ) = P O 1 ( l , 1 ) .
So C m P O 1 , P O 2 is idempotent if and only if P O 1 ( l , 1 ) = l .    □
Example 1.
POP Choquet-like integral C m P O α , P O φ is idempotent because P O α ( x , y ) = x y satisfies P O α ( x , 1 ) = x .
Proposition 3.
Let P O 1 , P O 2 be a pair of pseudo overlap functions and m be a fuzzy measure. C m P O 1 , P O 2 is a pre-aggregate function if the following conditions are true:
(1) 
P O 1 ( v , w ) v P O 2 ( u , w ) u , u , v , w [ 0 , 1 ] , u < v ;  
(2) 
P O 1 ( k u , v ) = k P O 1 ( u , v ) , P O 2 ( k u , v ) = k P O 2 ( u , v ) , x , y [ 0 , 1 ] and k ( 0 , 1 ] .
Proof. 
We only need to prove C m P O 1 , P O 2 is 1 = ( 1 , , 1 ) -increasing, that is, for each u , v , w , c , u + c , v + c [ 0 , 1 ] and u < v , P O 1 ( v , w ) P O 2 ( u , w ) P O 1 ( v + c , w ) P O 2 ( u + c , w ) is established.
From condition 1, for each sufficiently small c, we can obtain the following:
P O 1 ( v + c , w ) P O 1 ( v , w ) c P O 2 ( u + c , w ) P O 2 ( u , w ) c ,
P O 1 ( v + c , w ) P O 1 ( v , w ) P O 2 ( u + c , w ) P O 2 ( u , w ) , P O 1 ( v + c , w ) P O 2 ( u + c , w ) P O 1 ( v , w ) P O 2 ( u , w ) .
From condition 2, for each not sufficiently small c, always exist k ( 0 , 1 ] makes c 1 = c / k sufficiently small and greater than 0, then we have
k P O 1 ( v / k + c 1 , w ) k P O 1 ( v / k , w ) c 1 k P O 2 ( u / k + c 1 , w ) k P O 2 ( u / k , w ) c 1 ,
P O 1 ( v / k + c 1 , w ) P O 1 ( v / k , w ) P O 2 ( u / k + c 1 , w ) P O 2 ( u / k , w ) ,
think of v / k + c 1 as v + c , u / k + c 1 as u + c , then we have
P O 1 ( v + c , w ) P O 1 ( v , w ) P O 2 ( u + c , w ) P O 2 ( u , w ) , P O 1 ( v + c , w ) P O 2 ( u + c , w ) P O 1 ( v , w ) P O 2 ( u , w ) .
Therefore, for each u , v , w , c , u + c , v + c [ 0 , 1 ] , and  u < v , P O 1 ( v , w ) P O 2 ( u , w ) P O 1 ( v + c , w ) P O 2 ( u + c , w ) is established, C m P O 1 , P O 2 is a pre-aggregate function.    □
Remark 2.
Condition 2 of Proposition 3 is a special kind of homogeneity, similar to the homogeneity proposed in [27]. Specifically, according to the definition given in [27], when P O 1 and P O 2 are overlap functions, we have
P O 1 ( k x , k y ) = k P O 1 ( x , k y ) = k P O 1 ( k y , x ) = k 2 P O 1 ( y , x ) = k 2 P O 1 ( x , y ) .
The same is true for P O 2 . Therefore, condition 2 of Proposition 3 is a special homogeneity.
Example 2.
C m ( P O α , P O π ) is a pre-aggregate function because
(1) 
P O α ( x , y ) x = y , P O π ( x , y ) x = y 2 and x , y [ 0 , 1 ] , so P O α ( x , y ) x P O π ( x , y ) x .
(2) 
P O α ( k x , y ) = k x y , P O π ( k x , y ) = k x y 2 , satisfying the (2) of Proposition 3.
Similarly, all POP integrals C m P O 1 , P O 2 that satisfy P O 1 = P O 2 are pre-aggregate functions.
To illustrate the advantages of C m P O 1 , P O 2 proposed by us, an example is given below to preliminarily compare it with integral C m ( F 1 , F 2 ) and integral C m O . To facilitate calculation, the uniform fuzzy measure is selected.
Example 3.
Given two randomly increasing arrays a = ( 0.6 , 0.7 , 0.8 , 0.9 ) and b = ( 0.8 , 0.8 , 0.9 , 0.9 ) . We take them as input and calculate them in C m ( F 1 , F 2 ) , C m O , and  C m P O 1 , P O 2 , and analyze the results.
First, to make the comparison easier, select integral C m ( G M , T L ) in [23], C m O C in [11], and  C m ( P O χ , P O λ ) defined in Section 3. It is easy to obtain the fuzzy measure value of a as (1.0, 0.75, 0.5, 0.25). For input, the three integrals calculation steps are as follows:
C m ( G M , T L ) ( a ) = min 1 , G M ( 0.6 , 1.0 ) + G M ( 0.7 , 0.75 ) T L ( 0.6 , 0.75 ) + G M ( 0.8 , 0.5 ) T L ( 0.7 , 0.5 ) + G M ( 0.9 , 0.25 ) T L ( 0.8 , 0.25 ) = min { 1 , 2.01 } = 1 .
C m O C ( a ) = O C ( 0.6 , 1.0 ) + O C ( 0.7 , 0.75 ) O C ( 0.6 , 0.75 ) + O C ( 0.8 , 0.5 ) O C ( 0.7 , 0.5 ) + O C ( 0.9 , 0.25 ) O C ( 0.8 , 0.25 ) = 0.70 .
C m ( P O χ , P O λ ) ( a ) = P O χ ( 0.6 , 1.0 ) + P O χ ( 0.7 , 0.75 ) P O λ ( 0.6 , 0.75 ) + P O χ ( 0.8 , 0.5 ) P O λ ( 0.7 , 0.5 ) + P O χ ( 0.9 , 0.25 ) P O λ ( 0.8 , 0.25 ) 4 0.59 .
For input b , in the integral C m O C , its fuzzy measure value still is (1.0, 0.75, 0.5, 0.25). In the integral C m ( G M , T L ) and the integral C m ( P O χ , P O λ ) , its fuzzy measure value becomes [1.0, 0.5] because of the function of dimensionality reduction function. For input b , the three integrals calculation steps are as follows:
C m ( G M , T L ) ( b ) = min 1 , G M ( 0.8 , 1.0 ) + G M ( 0.9 , 0.5 ) T L ( 0.8 , 0.5 ) = min { 1 , 1.265 } = 1 .
C m O C ( b ) = O C ( 0.8 , 1.0 ) + O C ( 0.8 , 0.75 ) O C ( 0.8 , 0.75 ) + O C ( 0.9 , 0.5 ) O C ( 0.8 , 0.5 ) + O C ( 0.9 , 0.25 ) O C ( 0.9 , 0.25 ) = O C ( 0.8 , 1.0 ) + O C ( 0.9 , 0.5 ) O C ( 0.8 , 0.5 ) = 0.78 .
C m ( P O χ , P O λ ) ( b ) = P O χ ( 0.8 , 1.0 ) + P O χ ( 0.9 , 0.5 ) P O λ ( 0.8 , 0.5 ) 2 0.75 .
As you can see, whether we input a or b , the integral C m ( G M , T L ) will compute 1, which is inappropriate. Additionally, it is easier to see the change in integral C m ( P O χ , P O λ ) large than in integral C m O C when the input changes.

4. Ensemble Algorithm Based on POP Choquet-like Integrals

A classifier ensemble is a means of fusing multiple classifiers’ classification results. It can compensate with other classifiers when one classifier classification is not effective.
Classifier ensemble can be viewed as a two-tier pattern recognition structure, as shown in Figure 1. The first layer selects multiple classifiers as base classifiers, in which all base classifiers accept input conditions and output separately. This paper will give each base classifier the same training set as input. The second layer is the fusion method, which receives the output of multiple base classifiers, fuses them with a given fusion operator, and finally obtains a clear decision.
This paper uses the POP Choquet-like integral proposed in Section 3 as the fusion operator. This strategy is described in more detail below.

4.1. Algorithmic Framework

In the experiment of this algorithm, we test the pairs of pseudo-overlap functions by transforming them to obtain the C m P O 1 , P O 2 integral(Using the g λ fuzzy measure) with the best fusion function. This process is described in detail below, and the overall algorithmic framework is given.
Let H be a set of a given group of classifiers I 1 , I 2 , I n , H = I 1 , I 2 , I n . The first step in the ensemble is to enter the same training set into the classifier I 1 , I 2 , I n to train the respective models, then test these models with the same test set and obtain a classification membership matrix for each base classifier. Suppose the input x = ( x 1 , x 2 , , x r ) is classified into k classes, let P = P 1 , , P j , , P n , where P j is the classification membership matrix I j corresponding to the classifier. The row of P j represents each element of the test set, and the column’s index represents the class in which it is located. Each element of P j represents the extent to which the test item corresponding to the row belongs to the class corresponding to the current column.
Next is the most critical step in an ensemble. Take the fused operator C m ( P O 1 , P O 2 ) proposed in this article to de-fuse the elements of the coherence of the ranks of each component matrix of vector P = P 1 , , P j , , P n . Then we obtain a matrix M of size r × k , where element M r , k represents the degree to which the r-th data item belongs to class k. Let F be a matrix of size r × 1 , and its element consists of an index of columns corresponding to the maximum value of each row of M, representing the final classification result of the r-th data item. The model framework of Algorithm 1 is constructed below to describe this process more clearly.
Algorithm 1: Pseudo-code for our proposed ensemble model
Input: A pseudo overlap function pair ( P O 1 , P O 2 ) , classifier set
    H = I 1 , I 2 , , I n .
Output: F.
Fractalfract 07 00588 i001

4.2. Experimental Framework

  • Step 1. Selection of data sets
This paper will select the data sets in the UCI database (https://archive.ics.uci.e.,du/ml\/datasets.php (accessed on 11 December 2022)). This algorithm is aimed at multiple categories of data, so 12 data sets with a classification greater than or equal to 3 will be selected, with the number of feature items ranging from a few to a few hundred, as shown in Table 4.
  • Step 2. Experimental preprocessing
(1) Deleting Missing Items
Because some data sets with incomplete data elements make the algorithm fail, delete the entire row of those data elements.
(2) Ten-fold cross validation
Before the experiment, the data items for each data set category are divided into ten pieces on average, resulting in ten subsets so that each subset has a different category of data items. Each time a subset is used as a test set, and the remaining nine are trained as training sets until each subset is selected once as a test set. There are 15 pseudo-overlap function pairs selected in this article. Each data set runs at least 150 times to obtain the final result.
  • Step 3. Base algorithms and experiment details
This paper selects four basic classifiers: Naive Bayes, KNN (Euclidean distance, K = 10), neural network (three layers), and CART. The four base classifiers act as the four elements of the classifier set in a 1:1:1:1 scale, the step that is conducted through Python’s scikit-learn library. To better demonstrate the experiment’s results, multiply the F1-Measure value by 100.
Definition 13.
Set the accuracy rate as P, and recall rate as R. The F-Measure is the weighted harmonic average of P and R, and is defined as:
F α = ( 1 + α 2 ) · P · R α 2 · ( P + R ) .
When the parameter α = 1 , it is the most common F1-Measure, defined as:
F 1 = 2 · P · R P + R .
The values of P and R are in the interval [0,1].

4.3. Experimental Results and Analysis

This subsection demonstrates the experimental results of the classifier ensemble algorithm of POP Choquet-like integral, and then we compare and analyze this algorithm with other classification algorithms.
In the first experiment of this algorithm, we test the performance of POP Choquet-like integrals C m P O 1 , P O 2 based on different pseudo-overlap function pairs in the classifier ensemble. For the POP Choquet-like integrals in Table 5, we use pseudo overlap function pairs constructed by two identical pseudo overlap functions. In contrast, for the POP Choquet-like integrals in Table 6, we use pseudo overlap function pairs constructed by two different pseudo overlap functions. Table 5 and Table 6 measure the classifier ensemble’s performance in this paper. For each data set, the best-performing calculation schemes are shown in bold.
It can be seen that the results of ensemble classifiers based on pseudo-overlap function pairs are similar, which indicates the stability of this algorithm. Even if the pseudo overlap function pair with optimal performance is not selected, the final classification result is still acceptable. In particular, POP Choquet-like integrals that choose pairs with two different pseudo-overlap functions perform slightly better in the ensemble than POP Choquet-like integrals constructed with two identical pseudo-overlap functions. In Table 6, when the pseudo-overlap function pair P O γ , P O δ is selected, the average F1-Measure value of the final classification of the 12 data sets is the largest (83.27), and the classification effect is the best. In the following experiment, the experimental results of POP Choquet-like integral C m P O γ , P O δ are compared with those of other classification algorithms.
To illustrate the advantages of this algorithm in classification applications, Table 7 compares this algorithm with other advanced classifier ensemble algorithms in the literature: classifier ensemble algorithms based on C m O (CQO) [11], the generalized mixed function classifier (GM) [28], random forest trees (RT) [29], XGBoost (XGB) [30], META-DES (META) [31], lightGBM (LGBM) [32], randomized reference classifier (RRC) [33], and CatBoost (CA) [34]. Note that some of the comparative experiments for eight out of the twelve datasets we selected can be found in [11], and the details of the other comparative experiments will be introduced in the next paragraph.
For CQO, we chose the configuration in [11]: Naive Bayes, classical k-NN with Euclidean distances, multilayer perceptron, and CART with a proportion of 1:3:3:3, and we compare the results of the best performing overlap function and fuzzy measure in [11] with the results of our algorithm. For GM, we chose the configuration in [28]: k-NN, decision trees, MLP, Naive Bayes, SVM with a proportion of 3:3:2:1:1. This article is implemented with sklearn, the dataset situation determines the number of trees, and other parameters use the default values of the library for RT. For XGB, using the softmax loss function, the maximum depth of the tree is set between 3 and 10, and the number of trees is set between 100 and 500, depending on the size of the data sets. For META and RRC, we selected the same configuration as our algorithm. For LGBM, the maximum depth of the tree is set to 10, and the number of leaves is less than 2 10 . For CA, the maximum number of iterations is set to 3000.
In Table 7, we use the F1 measure, score, and the p value of the t-test to compare our ensemble algorithm with other algorithms. The two values of the score (win-loss) indicate the number of data sets whose F1 measure averages are greater than and less than the F1 measure averages of other algorithms, respectively.
It can be seen from Table 7 that our proposed ensemble algorithm has a better classification effect on most data sets(BS, WR, CE, MH, WW) than others. In particular, some data sets can greatly improve classification accuracy by using our ensemble algorithm in the case of other ensemble algorithms with general classification effects. Because the classification qualities of our algorithms on these data sets are much greater than that of other ensemble algorithms, the results of the sample T-test with them are significantly different.

5. Community Network Detection Algorithm Based on POP Choquet-like Integrals

5.1. Modularity

As the focus of many scholars in recent years, community network has been widely studied. In previous explorations, scholars have noted that those communities and structures inherent in a social network are the main goals of understanding the network. A social network structure is often unclear in a real complex network. There are always overlapping parts between communities, which adds to the charm of the community network detection problem and is also why scholars are interested in this problem.
Based on the above questions, the authors of [22] pointed out that social networks fall into three categories: classical community networks (where there is no overlap at all), crisp community networks (where there is overlap and each node of the overlap can belong to more than one community), and fuzzy community networks (where each node belongs to each community to some extent).
When the three situations of network community are defined, the optimal division method should be found. Because the number of communities to be divided is unknown, the problem of community network division is a clustering problem. As for how to evaluate the quality of the community clustering algorithm, modularity is a commonly used measurement method. To find the optimal solution, we can judge the quality of the network community division according to the modularity value.
Modularity, proposed by Girvan and Newman in [35], is used to measure the classical network community division scheme. In this paper, it is represented by Q G N and defined as
Q G N = 1 2 m i , j V A i j k i k j 2 m δ ( c i c j ) ,
where G = V , E is a given network, C = c 1 , , c r is a partition, m is the number of edges of the network, k i represents the degree of node i, A i j is the adjacency matrix of the network before partition, if node i and node j belong to the same community after partition, δ ( c i c j ) = 1 ; otherwise, it is 0. Under the premise of an unknown number of community partitions, the Q G N modularity is the most classic and most commonly used measure of community network partition schemes.
The value range of modularity is [−0.5, 1], and within this interval, the larger the modularity value, the better the clustering effect will be. In particular, when the value of modularity is greater than 0.3, the superiority of the current algorithm can be explained.
Considering that Q G N modularity is more suitable for classical network communities, the network community required to be measured has no overlap. However, in the real network, a node often belongs to several division areas simultaneously, so the original Q G N modularity cannot accurately determine the division scheme of these fuzzy networks. Many scholars are committed to improving the Q G N modularity to make it more suitable for the fuzzy network community.
In [21], Nepusz et al. defined crisp modularity in fuzzy scenarios, represented by Q T in this paper. It is defined as follows:
Q T = 1 2 m i , j V A i j k i k j 2 m s i j .
The Q T modularity improves the classic Q G N modularity. The author replaced δ ( c i c j ) with s i j , which represents the sum of the product of membership degrees of node i and j belonging to the same community, s i j = c = 1 , , r μ C c ( i ) μ C c ( j ) , and satisfies i V , c = 1 , , r μ C c ( i ) = 1 . Finally, other symbols are consistent with the Q G N modularity.
Although in [21], the author had extended the classical Q G N modularity to the fuzzy scene, it also has some shortcomings. If there is a node membership degree c = 1 r μ C c ( i ) > 1 in the fuzzy network community after partition, then Q T modularity is not applicable.
Considering the above shortcomings, Gomez had improved the Q G N modularity in [22] to fully fit the fuzzy framework. In this paper, the modularity proposed by Gomez is represented by Q D , which is defined as follows
Q D = 1 2 m i , j V A i j k i k j 2 m G G G O μ C c ( i ) , μ C c ( j ) c C ,
where G O is a two-dimensional overlapping function, G G is an n-dimensional grouping function, and other symbols have the same meaning as Q G N modularity.
Gomez, though, is mindful of the overlap of the web community and used overlap and group function with good performance to improve Q G N modularity. However, in processing node membership, the group function continues to use the maximum value to realize, which is not a non-average mean, which is the shortcoming of the research.
Considering the above deficiency, this paper uses the non-average POP Choquet-like integral to improve the modularity to obtain new modularity, which is defined as follows:
Q ˜ = 1 2 m i , j V A i j k i k j 2 m C m ( P O 1 , P O 2 ) P O μ C c ( i ) , μ C c ( j ) c C ,
where P O is a two-dimensional pseudo overlap function, and  C m ( P O 1 , P O 2 ) are the Choquet-like POP integrals constructed in this paper’s third section. The meaning of the remaining symbols is consistent with the modularity Q G N .
Note that our proposed modularity is not just a solution to the fuzzy network community problem. When the network community is classic or clear, the modularity and Q G N modularity has the same effect.
Example 4.
In [20], the author designed a simple fuzzy graph network, as shown in Figure 2. In this network, nodes are naturally divided into three communities, with overlaps between the three communities. Nodes 4 and 8 belong to multiple communities at the same time.
In [20], the author gives the degree of membership to each community after soft clustering of each node, as shown in Table 8. In this paper, the author stated that when the class number C = 3 , the network modularity in the figure above was the highest (0.326).
Bring the soft clustering results given in Table 8 into the modularity we defined for calculation, select the pseudo overlap function min { x 1 / 2 , y 1 / 2 } , and randomly select the POP Choquet-like integral C m ( P O β , P O α ) (Using the g λ fuzzy measure). Finally, when the partition number C = 3, the network modularity in the example figure is the highest (0.420), which can preliminarily verify that our modularity is reliable and advanced.

5.2. Experimental Framework

For several different network communities, many scholars have proposed different detection algorithms. However, few people have proposed non-average network community detection algorithms, so on the basis of [20,22], we propose a new community network detection algorithm based on the new modularity defined in Section 5.1. This process is explained in detail, and the overall algorithm framework is given below.
Algorithm 2: Pseudo-code for our proposed network community detection model
Input: An upper bound K and an adjacent matrix A = ( a i j ) n × n for the number of
   clusters in a given network.
Output: The largest Q ˜ ( U k ) and its corresponding k.
Fractalfract 07 00588 i002
Fuzzy C-means (FCM) mentioned in the Algorithm 2 is a clustering method commonly used by scholars. It was proposed by Dunn as early as 1973. Later scholars tried to improve and put forward different FCM algorithms many times, especially the version proposed by Bezdek in [36], which has been used until now. FCM can allow each datapoint in the current cluster to belong to multiple classes simultaneously; a clustering algorithm fully adapted to the fuzzy framework. The main idea is to minimize the function
J m = i = 1 n j = 1 k u i j m | | x i c j | | 2 ,
where x i is an n-dimensional data point to be clustered, and c j is the n-dimensional clustering center of class j. u i j is the degree to which x i belongs to class j and satisfies j u i j = 1 , m [ 1 , ) . represents any norm of similarity between any data point to be clustered and any clustering center.
Parameter m needs to be set for the FCM algorithm. An m value that is too large will lead to a poor clustering effect; an m value that is too small will make the algorithm similar to HCM and cannot highlight the fuzziness. In [37], the author proved that the best value range of m is [1.5,2.5]. When the only parameter to be set is confirmed, the FCM algorithm does not need human intervention in the implementation process.
In matrix theory [38], the generalized and ordinary eigenvalues of the same matrix are the same, and their eigenvectors are the same after normalization by Euclidean norm. However, it is a more stable numerical method to calculate the eigenvectors of generalized eigensystems. In the Algorithm 2 of this paper, the eigenvector of the k-1 dimensional generalized characteristic system of the diagonal matrix we calculated represents the k-1 dimensional values of the network graph, and these values serve as the numerical form of the points to be clustered.

5.3. Experimental Results and Analysis

To further test the benefits of our proposed social network detection algorithm, in this section, we compare it to other classic social network detection algorithms using two well-known reality networks: the Karate Club network and the Les Miserables network.
We use Python to implement the Algorithm 2 in this paper and select some classic network community detection algorithms and clear network community detection algorithms, such as GN [35] and D&L [39]. In addition, some advanced fuzzy network community detection algorithms are also used for comparisons, such as OCD [22] and NeSiFC [40].
In the process of algorithm implementation, we chose the pseudo overlap function min { x 1 / 2 , y 1 / 2 } . Since there are too many POP Choquet-like integrals proposed in this paper, it would be too tedious to carry out experiments on all of them, so we chose four non- average POP Choquet-like integrals(Using the g λ fuzzy measure) to carry out experiments, namely: C m ( P O α , P O ν ) , C m ( P O α , P O φ ) , C m ( P O φ , P O ν ) , C m ( P O α , P O π ) .
(1)
Network of Karate Club.
One of the most commonly used test networks for social network detection algorithms is the Karate Club network [41] (Figure 2), which Zachary observed over two years and has been mentioned in many articles. Karate Club is a real network with 34 members, each interacting with another member. The corresponding graph in the network is an undirected graph with 34 nodes and 78 edges. During these two years, because the relationship between the administrator and the coach broke down (nodes 0 and 34), the network was split into two smaller groups in reality, as shown in Figure 3.
It should be noted that social relationship networks will lead to fission because people’s emotions and transmitted information are not equal in society, which is constrained by interpersonal relationships and poor information, and other aspects. This inequality can lead to a network being divided into subgroups over time, with smaller subgroups being more stable. We want to identify potential information nodes where fission is likely to occur and reasonably predict what will happen next.
Table 9 shows the Karate Club network processing with the Algorithm 2 proposed in this paper, and the maximum modularity of each algorithm is shown in bold. The results show that the feedback obtained by the Algorithm 2 under the action of the four POP Choquet-like integrals is very close, and the differences are negligible. The measurement results of the Karate Club network show that when the partition number C = 3, the modularity value is the highest.
To illustrate the practicability of the proposed algorithm, Table 10 compares the best results obtained in Table 9 with other advanced algorithms. For the Karate Club network, the GN algorithm has the best partition effect when the partition number C = 5 (modularity value is 0.385). D&L and OCD algorithms perform best when the partition number C = 4 (modularity value: 0.416,0.437). NeSiFC algorithm is a recently proposed network community detection algorithm based on neighbor similarity. It does not need to determine the partition number. In [40], the author used it to calculate the Karate Club network had a maximum modularity value of 0.372. As you can see, our network algorithm obtains a higher modularity value than other algorithms.
Figure 4 shows the node division of the algorithm in this paper. We use yellow, blue, and green to represent three different partitions to see which community each node belongs to. The criteria are that a node belongs to the community if its membership to the current community is greater than 0.25. The nodes in red represent overlapping nodes with membership greater than 0.25 to several communities simultaneously. Under the current partition, all three overlapping nodes belong to both yellow and green communities.
(2)
The Les Miserables network
The Les Miserables network comes from the famous novelist Hugo’s novel Les Miserables. The original data can be found at http://www.personal.umich.edu/~mejn/netdata/ (accessed on 12 January 2023). The network has 77 nodes and 254 edges corresponding to the characters and their relationships in the novel (Figure 5).
The degree of connection between wired nodes is the same as in the Karate Club network. That is, the values of its adjacency matrix are only 0 and 1. Unlike the Karate Club network, the Les Miserables network is much more complex and larger. That is, the value of the adjacency matrix of the Les Miserables network is not only 0 or 1 but also larger than 1.
Table 11 shows the measurement results of the Les Miserables network based on four different POP Choquet-like integrals in this paper. Similarly, it can be observed that the results of the four algorithms are similar, which shows the stability of our algorithm on the side. Table 11 shows that when the partition number C = 5, the modularity value is the highest, which means dividing five communities is the best fit for Les Miserables Network.
In Table 12, we compare the best results obtained in Table 11 with other advanced algorithms. Notice that for the Les Miserables network, the GN algorithm works best when C = 11 (modularity value is 0.538); D&L and OCD algorithms performed best when the partition number C = 7 (the modularity values were 0.556 and 0.564, respectively). In [40], the highest modularity value of the Les Miserables network calculated by the author using NeSiFC as 0.573. It can be seen that our proposed network algorithm obtains a higher modularity value (0.585) than other algorithms.
Figure 6 shows the partition of the Les Miserables Network by our proposed network algorithm. The five communities are shown in pink, purple, blue, green, and yellow. Similarly, the threshold of 0.25 is used to determine whether the node belongs to the current community. The red nodes represent overlapping nodes. Node Marguerite belongs to both purple and yellow communities. Node Perpetue belongs to both the pink and yellow communities.

6. Conclusions

In this paper, we proposed the POP Choquet-like integrals and applied them to classifier ensemble and fuzzy community detection, which makes good results. The contributions of this paper are listed as follows:
  • The pseudo overlap function pair is introduced to replace products in discrete Choquet integral. So, the POP Choquet-like integral C m P O 1 , P O 2 is obtained.
  • Two new algorithms using the POP Choquet-like integral are designed. One is the ensemble algorithm, a branch of the classification algorithm. We use the defined C m P O 1 , P O 2 as the fusion operator and the classification results of the base algorithms as inputs into the fusion operator to obtain a clear classification result. Another is the network community detection algorithm, a typical clustering algorithm. We use the defined C m P O 1 , P O 2 to act on the results after each node’s soft clustering, improving the classical modularity. Theoretically, our algorithm considers the non-average node membership degree in fuzzy community networks, which is more practical.
  • Many experiments were conducted on multiple datasets, proving the advantages of the two algorithms.
In future research, the following topics deserve attention: (1) The POP Choquet-like integrals will be combined with other decision making and aggregation methods, such as neural networks, decision trees, etc. (2) The integration of the research approach of this paper with the latest research findings (such as [42,43,44,45]).

Author Contributions

Conceptualization, methodology, designing computer programs, X.Z.; writing—original draft preparation, designing computer programs, H.J.; writing—reviewing and editing, validation, J.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work is funded by the National Natural Science Foundation of China (Nos. 12271319 and 12201373), and China Postdoctoral Science Foundation 2023T160402.

Data Availability Statement

All the data we used has been presented in the paper Table 4, Figure 2, Figure 3 and Figure 5.

Conflicts of Interest

No conflict of interest exists in the submission of this manuscript, and all authors approve the manuscript for publication. The work was original research that has not been published previously and is not under consideration for publication elsewhere, in whole or in part. All the authors listed have approved the manuscript that is enclosed.

References

  1. Adankon, M.; Cheriet, M.; Biem, A. Semisupervised learning using Bayesian interpretation: Application to LS-SVM. IEEE Trans. Neural Netw. 2011, 22, 513–524. [Google Scholar] [CrossRef]
  2. Vapnik, V. The Nature of Statistical Learning Theory; Springer Science & Business Media: New York, NY, USA, 1999. [Google Scholar]
  3. Swaney, D.; McAlister, G.; Coon, J. Decision tree–driven tandem mass spectrometry for shotgun proteomics. Nat. Methods 2008, 5, 959–964. [Google Scholar] [CrossRef]
  4. Valova, I.; Harris, C.; Mai, T. Optimization of convolutional neural networks for imbalanced set classification. Procedia Comput. Sci. 2020, 176, 660–669. [Google Scholar] [CrossRef]
  5. Denoeux, T. A k-nearest neighbor classification rule based on Dempster-Shafer theory. IEEE Trans. Syst. Man Cybern. 1995, 25, 804–813. [Google Scholar] [CrossRef] [Green Version]
  6. Tarus, J.; Ogada, K.; Mwalili, T. An Ensemble Based Model for Detecting Genetically Inherited Disorder. In Proceedings of the 2022 12th International Conference on Advanced Computer Information Technologies (ACIT), Ruzomberok, Slovakia, 26–28 September 2022; IEEE: New York, NY, USA, 2022; pp. 121–125. [Google Scholar]
  7. AminiMotlagh, M.; Shahhoseini, H.; Fatehi, N. A reliable sentiment analysis for classification of tweets in social networks. Soc. Netw. Anal. Min. 2023, 13, 1–11. [Google Scholar] [CrossRef]
  8. Qiu, H.; Su, P.; Jiang, S. Learning from Human Uncertainty by Choquet Integral for Optic Disc Segmentation. In Proceedings of the 2021 4th International Conference on Control and Computer Vision, Macau, China, 13–15 August 2021; pp. 7–12. [Google Scholar]
  9. Tripathi, D.; Edla, D.; Cheruku, R. A novel hybrid credit scoring model based on ensemble feature selection and multilayer ensemble classification. Comput. Intell. 2019, 35, 371–394. [Google Scholar] [CrossRef]
  10. Taha, A.; Barukab, O.; Malebary, S. Fuzzy Integral-Based Multi-Classifiers Ensemble for Android Malware Classification. Mathematics 2021, 9, 2880. [Google Scholar] [CrossRef]
  11. Batista, T.; Bedregal, B.; Moraes, R. Constructing multi-layer classifier ensembles using the Choquet integral based on overlap and quasi-overlap functions. Neurocomputing 2022, 500, 413–421. [Google Scholar] [CrossRef]
  12. Jain, A.; Murty, M.; Flynn, P. Data clustering: A review. ACM Comput. Surv. 1999, 31, 264–323. [Google Scholar] [CrossRef]
  13. Sharma, K.; Seal, A. Spectral embedded generalized mean based k-nearest neighbors clustering with s-distance. Expert Syst. Appl. 2021, 169, 114326. [Google Scholar] [CrossRef]
  14. Zhang, H.; Li, H.; Chen, N. Novel fuzzy clustering algorithm with variable multi-pixel fitting spatial information for image segmentation. Pattern Recognit. 2022, 121, 108–201. [Google Scholar] [CrossRef]
  15. Esfandian, N.; Hosseinpour, K. A clustering-based approach for features extraction in spectro-temporal domain using artificial neural network. Int. J. Eng. 2021, 34, 452–457. [Google Scholar]
  16. Liu, M.; Qi, X.; Pan, H. Construction of network topology and geographical vulnerability for telecommunication network. Comput. Netw. 2022, 205, 108764. [Google Scholar] [CrossRef]
  17. Ouyang, X.; Huo, J.; Xia, L. Dual-sampling attention network for diagnosis of COVID-19 from community acquired pneumonia. IEEE Trans. Med. Imaging 2020, 39, 2595–2605. [Google Scholar] [CrossRef] [PubMed]
  18. Liu, N.; Hu, X.; Ma, L. Vulnerability assessment for coupled network consisting of power grid and EV traffic network. IEEE Trans. Smart Grid 2021, 13, 589–598. [Google Scholar] [CrossRef]
  19. Liu, X.; Maiorino, E.; Halu, A. Robustness and lethality in multilayer biological molecular networks. Nat. Commun. 2020, 11, 6043. [Google Scholar] [CrossRef]
  20. Zhang, S.; Wang, R.; Zhang, X. Identification of overlapping community structure in complex networks using fuzzy c-means clustering. Phys. A Stat. Mech. Appl. 2007, 374, 483–490. [Google Scholar] [CrossRef]
  21. Nepusz, T.; Petróczi, A.; Négyessy, L. Fuzzy communities and the concept of bridgeness in complex networks. Phys. Rev. E 2008, 77, 016107. [Google Scholar] [CrossRef] [Green Version]
  22. Gomez, D.; Rodríguez, J.; Yanez, J. A new modularity measure for fuzzy community detection problems based on overlap and grouping functions. Int. J. Approx. Reason. 2016, 74, 88–107. [Google Scholar] [CrossRef] [Green Version]
  23. Lucca, G.; Dimuro, G.; Fernández, J. Improving the Performance of Fuzzy Rule-Based Classification Systems Based on a Nonaveraging Generalization of CC-Integrals Named CF1F2-Integrals. IEEE Trans. Fuzzy Syst. 2018, 27, 124–134. [Google Scholar] [CrossRef] [Green Version]
  24. Zhang, X.; Liang, R.; Bustince, H. Pseudo overlap functions, fuzzy implications and pseudo grouping functions with applications. Axioms 2022, 11, 593. [Google Scholar] [CrossRef]
  25. Choquet, G. Theory of capacities. Ann. Inst. Fourier 1954, 5, 131–295. [Google Scholar] [CrossRef] [Green Version]
  26. Li, X.; Wang, F.; Chen, X. Support vector machine ensemble based on Choquet integral for financial distress prediction. Int. J. Pattern Recognit. Artif. Intell. 2015, 29, 1550016. [Google Scholar] [CrossRef]
  27. Qiao, J.; Hu, B. On homogeneous, quasi-homogeneous and pseudo-homogeneous overlap and grouping functions. Fuzzy Sets Syst. 2019, 357, 58–90. [Google Scholar] [CrossRef]
  28. Costa, V.; Farias, A.; Bedregal, B. Combining multiple algorithms in classifier ensembles using generalized mixture functions. Neurocomputing 2018, 313, 402–414. [Google Scholar] [CrossRef] [Green Version]
  29. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  30. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd ACM Sigkdd International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  31. Cruz, R.; Sabourin, R.; Cavalcanti, G. META-DES: A dynamic ensemble selection framework using meta-learning. Pattern Recognit. 2015, 48, 1925–1935. [Google Scholar] [CrossRef] [Green Version]
  32. Ke, G.; Meng, Q.; Finley, T. Lightgbm: A highly efficient gradient boosting decision tree. Adv. Neural Inf. Process. Syst. 2017, 30, 3146–3154. [Google Scholar]
  33. Woloszynski, T.; Kurzynski, M. A probabilistic model of classifier competence for dynamic ensemble selection. Pattern Recognit. 2011, 44, 2656–2668. [Google Scholar] [CrossRef]
  34. Prokhorenkova, L.; Gusev, G.; Vorobev, A. CatBoost: Unbiased boosting with categorical features. Adv. Neural Inf. Process. Syst. 2018, 31, 6639–6649. [Google Scholar]
  35. Girvan, M.; Newman, M. Community structure in social and biological networks. Proc. Natl. Acad. Sci. USA 2002, 99, 7821–7826. [Google Scholar] [CrossRef]
  36. Bezdek, J.; Ehrlich, R.; Full, W. FCM: The fuzzy c-means clustering algorithm. Comput. Geosci. 1984, 10, 191–203. [Google Scholar] [CrossRef]
  37. Pal, N.; Bezdek, J. On cluster validity for the fuzzy c-means model. IEEE Trans. Fuzzy Syst. 1995, 3, 370–379. [Google Scholar] [CrossRef]
  38. Franklin, J. Matrix Theory; Courier Corporation: North Chelmsford, MA, USA, 2012. [Google Scholar]
  39. Gomez, D.; Zarrazola, E.; Yanez, J. A divide-and-link algorithm for hierarchical clustering in networks. Inf. Sci. 2015, 316, 308–328. [Google Scholar] [CrossRef]
  40. Roy, U.; Muhuri, P.; Biswas, S. NeSiFC: Neighbors’ similarity-based fuzzy community detection using modified local random walk. IEEE Trans. Cybern. 2021, 52, 10014–10026. [Google Scholar] [CrossRef]
  41. Zachary, W. An information flow model for conflict and fission in small groups. J. Anthropol. Res. 1977, 33, 452–473. [Google Scholar] [CrossRef] [Green Version]
  42. Wang, J.; Zhang, X. A novel multi-criteria decision-making method based on rough sets and fuzzy measures. Axioms 2022, 11, 275. [Google Scholar] [CrossRef]
  43. Zhang, X.; Sheng, N.; Borzooei, R.A. Partial residuated implications induced by partial triangular norms and partial residuated lattices. Axioms 2023, 12, 63. [Google Scholar] [CrossRef]
  44. Zhang, X.; Li, M.; Liu, H. Overlap Functions-Based Fuzzy Mathematical Morphological Operators and Their Applications in Image Edge Extraction. Fractal Fract. 2023, 7, 465. [Google Scholar] [CrossRef]
  45. Jing, M.; Zhang, X. Pseudo-Quasi Overlap Functions and Related Fuzzy Inference Methods. Axioms 2023, 12, 217. [Google Scholar] [CrossRef]
Figure 1. General architecture of classifier ensemble.
Figure 1. General architecture of classifier ensemble.
Fractalfract 07 00588 g001
Figure 2. A simple fuzzy graph network for testing.
Figure 2. A simple fuzzy graph network for testing.
Fractalfract 07 00588 g002
Figure 3. The Karate Club network.
Figure 3. The Karate Club network.
Fractalfract 07 00588 g003
Figure 4. The clustering result of the Karate Club network.
Figure 4. The clustering result of the Karate Club network.
Fractalfract 07 00588 g004
Figure 5. The Les Miserables network.
Figure 5. The Les Miserables network.
Fractalfract 07 00588 g005
Figure 6. Clustering result of Les Miserables network.
Figure 6. Clustering result of Les Miserables network.
Fractalfract 07 00588 g006
Table 1. Pseudo overlap functions.
Table 1. Pseudo overlap functions.
Sequence NumberNomenclatureDefinition
1 P O α P O α x , y = x y .
2 P O β P O β x , y = 2 x y x + y , x + y 0 , 0 , x + y = 0 .
3 P O χ P O χ x , y = 2 x y 1 + x y .
4 P O δ P O δ x , y = max min x , y 2 , x + y 1 .
5 P O ε P O ε x , y = 0 · 1 x y 2 + 0.9 max 0 , x + y 1 .
6 P O ϕ P O ϕ x , y = x 2 y 0.5 x 2 y 1 x 1 y .
7 P O φ P O φ x , y = min x 2 , y 2
8 P O γ P O γ x , y = 2 x y 0.5 x + 1.5 y , 0.5 x + 1.5 y 0 , 0 , o t h e r s .
9 P O η P O η x , y = 2 x 2 y 2 1 + x 2 y 2 .
10 P O ι P O ι x , y = x y 0.5 x + 0.5 y , 0.5 x + 0.5 y 0 , 0 , o t h e r s .
11 P O κ P O κ x , y = 3 x y x + 2 y , x + 2 y 0 , 0 , o t h e r s .
12 P O λ P O λ x , y = 2 x 10 y 10 1 + x 10 y 10
13 P O μ P O μ x , y = x 2 y + x 2 y ( 1 x ) ( 1 y )
14 P O ν P O ν x , y = min x 3 , y 3
15 P O π P O π x , y = x y 2
Table 2. The size relation of pseudo overlap functions in Table 1.
Table 2. The size relation of pseudo overlap functions in Table 1.
Sequence NumberPseudo Overlap Function PairExpression
1 ( P O β , P O α ) P O β ( x , y ) P O α ( x , y ) , x , y [ 0 , 1 ]
2 ( P O β , P O χ ) P O β ( x , y ) P O χ ( x , y ) , x , y [ 0 , 1 ]
3 ( P O β , P O δ ) P O β ( x , y ) P O δ ( x , y ) , x , y [ 0 , 1 ]
4 ( P O β , P O ε ) P O β ( x , y ) P O ε ( x , y ) , x , y [ 0 , 1 ]
5 ( P O β , P O ϕ ) P O β ( x , y ) P O ϕ ( x , y ) , x , y [ 0 , 1 ]
6 ( P O χ , P O λ ) P O χ ( x , y ) P O λ ( x , y ) , x , y [ 0 , 1 ]
7 ( P O φ , P O ν ) P O φ ( x , y ) P O ν ( x , y ) , x , y [ 0 , 1 ]
8 ( P O α , P O φ ) P O α ( x , y ) P O φ ( x , y ) , x , y [ 0 , 1 ]
9 ( P O α , P O ν ) P O α ( x , y ) P O ν ( x , y ) , x , y [ 0 , 1 ]
10 ( P O ι , P O α ) P O ι ( x , y ) P O α ( x , y ) , x , y [ 0 , 1 ]
11 ( P O κ , P O α ) P O κ ( x , y ) P O α ( x , y ) , x , y [ 0 , 1 ]
12 ( P O γ , P O α ) P O γ ( x , y ) P O α ( x , y ) , x , y [ 0 , 1 ]
13 ( P O γ , P O δ ) P O γ ( x , y ) P O δ ( x , y ) , x , y [ 0 , 1 ]
14 ( P O γ , P O η ) P O γ ( x , y ) P O η ( x , y ) , x , y [ 0 , 1 ]
15 ( P O α , P O π ) P O α ( x , y ) P O π ( x , y ) , x , y [ 0 , 1 ]
Table 3. Example of POP Choquet-like integrals.
Table 3. Example of POP Choquet-like integrals.
Sequence NumberPOP IntegralExpression
1 C m ( P O β , P O α ) C m ( P O β , P O α ) ( x ) = j = 1 k 2 y ( j ) · m ( A ( j ) ) y ( j ) + m ( A ( j ) ) y ( j 1 ) · m ( A ( j ) ) k , y ( j ) + m ( A ( j ) ) 0 0 , y ( j ) + m ( A ( j ) ) = 0
2 C m ( P O β , P O χ ) C m ( P O β , P O χ ) ( x ) = j = 1 k 2 y ( j ) · m ( A ( j ) ) y ( j ) + m ( A ( j ) ) 2 y ( j 1 ) · m ( A ( j ) ) 1 + y ( j 1 ) · m ( A ( j ) ) k , y ( j ) + m ( A ( j ) ) 0 0 , y ( j ) + m ( A ( j ) ) = 0
3 C m ( P O β , P O δ ) C m ( P O β , P O δ ) ( x ) = j = 1 k 2 y ( j ) · m ( A ( j ) ) y ( j ) + m ( A ( j ) ) max min y ( j 1 ) , m ( A ( j ) ) 2 , y ( j 1 ) + m ( A ( j ) ) 1 k , y ( j ) + m ( A ( j ) ) 0 0 , y ( j ) + m ( A ( j ) ) = 0
4 C m ( P O β , P O ε ) C m ( P O β , P O ε ) ( x ) = j = 1 k 2 y ( j ) · m ( A ( j ) ) y ( j ) + m ( A ( j ) ) 0.1 · y ( j 1 ) · m ( A ( j ) ) 2 + 0.9 · max 0 , y ( j 1 ) + m ( A ( j ) ) 1 k , y ( j ) + m ( A ( j ) ) 0 0 , y ( j ) + m ( A ( j ) ) = 0
5 C m ( P O β , P O ϕ ) C m ( P O β , P O ϕ ) ( x ) = j = 1 k 2 y ( j ) · m ( A ( j ) ) y ( j ) + m ( A ( j ) ) y ( j 1 ) 2 · m ( A ( j ) ) + 0.5 · y ( j 1 ) 2 · m ( A ( j ) ) · ( 1 y ( j 1 ) ) · ( 1 m ( A ( j ) ) k , y ( j ) + m ( A ( j ) ) 0 0 , y ( j ) + m ( A ( j ) ) = 0
6 C m ( P O χ , P O λ ) C m ( P O χ , P O λ ) ( x ) = j = 1 k 2 y ( j ) · m ( A ( j ) ) 1 + y ( j ) · m ( A ( j ) ) 2 y ( j 1 ) 10 · m ( A ( j ) ) 10 1 + y ( j 1 ) 10 · m ( A ( j ) ) 10 k
7 C m ( P O α , P O ν ) C m ( P O α , P O ν ) ( x ) = j = 1 k y ( j ) · m ( A ( j ) ) min y ( j 1 ) 3 , m 3 ( A ( j ) ) k
8 C m ( P O φ , P O ν ) C m ( P O φ , P O ν ) ( x ) = j = 1 k min y ( j ) 2 , m 2 ( A ( j ) ) min y ( j 1 ) 3 , m 3 ( A ( j ) ) k
9 C m ( P O α , P O φ ) C m ( P O α , P O φ ) ( x ) = j = 1 k y ( j ) · m ( A ( j ) ) min y ( j 1 ) 2 , m 2 ( A ( j ) ) k
10 C m ( P O ι , P O α ) C m ( P O ι , P O α ) ( x ) = j = 1 k y ( j ) · m ( A ( j ) ) 0.5 · y ( j ) + m ( A ( j ) ) y ( j 1 ) · m ( A ( j ) ) k , y ( j ) + m ( A ( j ) ) 0 0 , y ( j ) + m ( A ( j ) ) = 0
11 C m ( P O κ , P O α ) C m ( P O κ , P O α ) ( x ) = j = 1 k 3 · y ( j ) · m ( A ( j ) ) y ( j ) + 2 · m ( A ( j ) ) y ( j 1 ) · m ( A ( j ) ) k , y ( j ) + 2 · m ( A ( j ) ) 0 0 , y ( j ) + 2 · m ( A ( j ) ) = 0
12 C m ( P O γ , P O α ) C m ( P O γ , P O α ) ( x ) = j = 1 k 2 · y ( j ) · m ( A ( j ) ) 0.5 · y ( j ) + 1.5 · m ( A ( j ) ) y ( j 1 ) · m ( A ( j ) ) k , 0.5 · y ( j ) + 1.5 · m ( A ( j ) ) 0 0 , 0.5 · y ( j ) + 1.5 · m ( A ( j ) ) = 0
13 C m ( P O γ , P O δ ) C m ( P O γ , P O δ ) ( x ) = j = 1 k 2 · y ( j ) · m ( A ( j ) ) 0.5 · y ( j ) + 1.5 · m ( A ( j ) ) max min y ( j 1 ) , m ( A ( j ) ) 2 , y ( j 1 ) + m ( A ( j ) ) 1 k , 0.5 · y ( j ) + 1.5 · m ( A ( j ) ) 0 0 , 0.5 · y ( j ) + 1.5 · m ( A ( j ) ) = 0
14 C m ( P O γ , P O η ) C m ( P O γ , P O η ) ( x ) = j = 1 k 2 · y ( j ) · m ( A ( j ) ) 0.5 · y ( j ) + 1.5 · m ( A ( j ) ) 2 · y ( j 1 ) 2 · m 2 ( A ( j ) ) 1 + y ( j 1 ) 2 · m 2 ( A ( j ) ) k , 0.5 · y ( j ) + 1.5 · m ( A ( j ) ) 0 0 , 0.5 · y ( j ) + 1.5 · m ( A ( j ) ) = 0
15 C m ( P O α , P O π ) C m ( P O α , P O π ) ( x ) = j = 1 k y ( j ) · m ( A ( j ) ) y ( j 1 ) 5 · m ( A ( j ) ) 5 k
Table 4. Description of the datasets used in our ensemble.
Table 4. Description of the datasets used in our ensemble.
Name of DatasetInstances NumberAttributes NumberClasses Number
Iris (IR)15043
Balance Scale (BS)62543
Winequality-red (WR)1599116
Waveform (WF)5000213
Optical recognition (OR)56206410
Cnae-9 (C9)10808579
Wireless indoor locatization (WI)200074
Splice junction Gene sequences (SG)2552613
Car evaluation (CE)172864
Maternal health risk (MH)101466
Winequality-white (WW)4898117
Page-blocks (PB)5472105
Table 5. Ensemble algorithm performance of POP Choquet-like integrals that choose pairs with two identical pseudo overlap functions.
Table 5. Ensemble algorithm performance of POP Choquet-like integrals that choose pairs with two identical pseudo overlap functions.
Dataset
Integral
IRBSWRWFORC9WISGCEMHWWPBMean
C m ( P O α , P O α ) 95.9785.7345.5284.1395.5193.9297.5399.1984.3567.4042.4694.6282.19
C m ( P O β , P O β ) 95.9785.7348.3684.3095.1993.6497.4399.0083.7570.3643.9694.5282.68
C m ( P O χ , P O χ ) 95.9785.7350.4284.1095.1593.8797.1299.3884.0772.7144.1194.4583.09
C m ( P O δ , P O δ ) 96.6485.7343.5284.0695.3193.3497.4398.9785.0367.8643.0894.2182.10
C m ( P O ε , P O ε ) 97.3186.6544.5383.2396.2194.0397.5199.7288.4772.0042.8894.6283.10
C m ( P O ϕ , P O ϕ ) 95.9786.6543.1684.2896.5394.0497.7499.4186.2570.2242.7294.7482.64
C m ( P O φ , P O φ ) 95.9786.6541.2182.7095.2294.1497.5299.0382.8868.9041.1494.4080.81
C m ( P O γ , P O γ ) 95.9785.6247.5584.5795.5994.0197.5899.6285.6973.6343.6994.4383.16
C m ( P O η , P O η ) 95.9786.6540.6683.9896.2493.6797.7499.3185.7265.6142.0594.6981.86
C m ( P O ι , P O ι ) 95.9785.6248.3684.3095.1993.5497.4399.0083.7570.3643.9694.5282.67
C m ( P O κ , P O κ ) 95.9785.6248.2184.3395.4194.1097.4399.3185.4671.2543.5394.4082.92
C m ( P O λ , P O λ ) 95.9785.6242.3183.9295.3493.5897.1299.0582.9870.1242.3694.4281.90
C m ( P O μ , P O μ ) 95.9785.6240.5883.9895.1893.6697.5289.7283.5771.2541.5494.2181.07
C m ( P O ν , P O ν ) 96.6485.4343.6282.7095.2093.9797.4399.0383.7569.3642.3194.2081.97
C m ( P O π , P O π ) 95.9785.4342.3182.6595.2393.4897.5399.2184.8570.2441.8594.3581.93
Table 6. Ensemble algorithm performance of POP Choquet-like integrals that choose pairs with two different pseudo overlap functions.
Table 6. Ensemble algorithm performance of POP Choquet-like integrals that choose pairs with two different pseudo overlap functions.
Dataset
Integral
IRBSWRWFORC9WISGCEMHWWPBMean
C m ( P O β , P O α ) 97.3182.8249.5084.8895.6795.8997.9498.6387.9950.3845.5995.0881.81
C m ( P O β , P O χ ) 97.3182.7750.0085.0094.4995.9897.8490.4487.7044.1945.2394.0880.42
C m ( P O β , P O δ ) 95.9283.0449.9784.7495.4995.9098.0497.5087.9451.2945.5695.2781.72
C m ( P O β , P O ε ) 95.2683.0349.8184.5695.9096.1897.9495.8788.1264.2845.4195.3982.65
C m ( P O β , P O ϕ ) 95.2782.7849.8884.9895.5496.0998.9498.6387.7251.0845.5395.0081.79
C m ( P O χ , P O λ ) 95.9786.3446.9483.7096.7194.9897.5998.3486.8565.4841.8995.2582.50
C m ( P O μ , P O ϕ ) 95.9786.3544.6784.3896.7493.8597.6499.3787.0472.5142.2794.8482.97
C m ( P O φ , P O ν ) 95.9786.3541.4883.7596.4793.5697.6498.9086.3362.8140.7194.7581.56
C m ( P O α , P O φ ) 96.6481.1348.6085.1596.0095.4797.0499.5988.0366.9942.6894.4682.65
C m ( P O ι , P O α ) 93.2360.7850.5081.5492.4895.5598.9790.4685.0435.8643.6795.0876.93
C m ( P O κ , P O α ) 91.8160.5451.6081.0692.6795.8398.0491.5485.3435.6746.1695.1477.12
C m ( P O γ , P O α ) 97.3182.9749.1785.3695.6795.6184.0498.6387.8466.0245.3495.3181.94
C m ( P O γ , P O δ ) 98.6682.9749.4085.2395.9595.8998.0498.7887.8465.8245.1895.4483.27
C m ( P O γ , P O η ) 96.6482.7049.0285.3496.0495.8098.0499.2587.9466.4245.4794.8683.13
C m ( P O α , P O π ) 95.9786.3447.8184.0696.6993.7897.7498.5986.8064.6742.0895.0682.47
Table 7. Comparison of the results of the ensemble algorithm.
Table 7. Comparison of the results of the ensemble algorithm.
Algorithm
Dataset
C m ( P O γ , P O δ ) CQOGMRTXGBMETALGBMRRCCAT
IR98.6695.3195.9896.6595.9895.9894.6495.3195.31
BS82.9761.0855.8745.5651.6155.2148.9256.0056.14
WR49.4029.0729.7025.6529.7427.4729.8128.3926.37
WF85.2384.9384.0081.2984.8677.5384.6276.9385.28
OR95.9598.5398.3795.5397.4898.4097.6898.3795.19
C995.8995.0893.8890.4290.8293.8184.3494.1987.62
WI98.0497.5498.3795.6396.8997.3294.3295.4796.54
SG98.7894.8494.8791.3195.3794.7195.4694.5395.15
CE87.8458.4355.9952.2155.0551.4159.1851.7546.38
MH65.8235.9932.4530.6833.5731.4737.8731.2929.84
WW45.1825.7124.5523.1825.4522.5425.0122.4322.15
PB95.4495.2195.3394.8994.5495.4295.3694.0793.52
Average83.2772.6471.6168.5870.5970.1170.5569.8969.12
Win-loss9–31–111–110–120–120–120–120–121–11
p-value-0.00130.0140.0060.0100.0090.0060.0070.008
Table 8. Soft clustering results of example graphs.
Table 8. Soft clustering results of example graphs.
NodesSoft Clustering ResultsCrisp C
0[0.9951, 0.0026, 0.0023][1, 0, 0]
1[0.9804, 0.0108, 0.0088][1, 0, 0]
2[0.9984, 0.0008, 0.0008][1, 0, 0]
3[0.9984, 0.0008, 0.0008][1, 0, 0]
4[0.4327, 0.1133, 0.4540][1, 0, 1]
5[0.0039, 0.0037, 0.9924][0, 0, 1]
6[0.0039, 0.0037, 0.9924][0, 0, 1]
7[0.0012, 0.0013, 0.9975][0, 0, 1]
8[0.0715, 0.1519, 0.7766][0, 1, 1]
9[0.0020, 0.9959, 0.0022][0, 1, 0]
10[0.0012, 0.9976, 0.0011][0, 1, 0]
11[0.0020, 0.9959, 0.0022][0, 1, 0]
12[0.0054, 0.9899, 0.0047][0, 1, 0]
Table 9. Our algorithm measures the Karate Club network.
Table 9. Our algorithm measures the Karate Club network.
Choquet-like POP Integers
Clasess
C m ( P O α , P O ν ) C m ( P O α , P O φ ) C m ( P O φ , P O ν ) C m ( P O α , P O π )
20.3870.3670.3680.376
30.4620.4600.4570.448
40.4610.4570.4570.446
50.4170.3890.3850.402
60.1470.0310.0240.074
70.1000.0280.0370.070
80.1040.0210.0100.061
90.1220.0190.0350.065
100.1000.0200.0310.060
110.1070.0170.0350.066
120.0910.0140.0340.056
130.0730.0120.0290.051
140.0510.0100.0140.048
150.0440.0070.0120.026
160.0420.0040.0060.008
170.0070.0010.0020.004
Table 10. Karate Club network experimental comparison results.
Table 10. Karate Club network experimental comparison results.
ClasessGND&LOCDNeSiFC C m ( P O α , P O ν )
20.3600.3150.340 0.387
30.3490.3850.400 0.462
40.3630.4160.437 0.461
50.3850.4130.434 0.417
60.3520.4060.405 0.147
70.3760.3980.310Best: 0.3720.100
80.3580.3890.215 0.104
90.3420.3770.213 0.122
100.3250.3620.318 0.100
110.3160.3510.230 0.107
120.2990.3340.120 0.091
130.2800.3170.221 0.073
140.2630.3000.251 0.051
150.2480.2820.346 0.044
160.2270.2520.208 0.042
170.2090.2310.172 0.007
Table 11. Our algorithm measures the Les Miserables network.
Table 11. Our algorithm measures the Les Miserables network.
Choquet-like POP Integers
Clasess
C m ( P O α , P O ν ) C m ( P O α , P O φ ) C m ( P O φ , P O ν ) C m ( P O α , P O π )
20.3740.2440.3880.324
30.4930.4120.4660.417
40.5340.5290.5320.517
50.5850.5680.5740.561
60.1780.1100.0210.104
70.1750.0840.0150.087
80.1320.0710.0120.074
90.1200.0620.0100.066
100.1090.0560.0100.060
110.0940.0490.0100.051
120.0870.0450.0100.047
130.0740.0400.0040.041
140.0670.0470.0260.036
150.0330.0470.0260.130
160.0300.0220.0140.004
170.0110.0060.0030.001
Table 12. Les Miserables Club network experimental comparison results.
Table 12. Les Miserables Club network experimental comparison results.
ClasessGND&LOCDNeSiFC C m ( P O α , P O ν )
20.0750.3720.233 0.374
30.2600.4640.264 0.493
40.2670.5110.494 0.534
50.4160.5520.553 0.585
60.4590.5540.556 0.178
70.4560.5560.564 0.175
80.4540.5560.276Best: 0.5730.132
90.4520.5530.260 0.120
100.4520.5510.113 0.109
110.5380.5480.233 0.094
120.5350.5460.174 0.087
130.5310.5430.115 0.074
140.5280.5400.061 0.067
150.5250.5370.041 0.033
160.5230.5250.026 0.030
170.5200.5200.041 0.011
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, X.; Jiang, H.; Wang, J. New Classifier Ensemble and Fuzzy Community Detection Methods Using POP Choquet-like Integrals. Fractal Fract. 2023, 7, 588. https://doi.org/10.3390/fractalfract7080588

AMA Style

Zhang X, Jiang H, Wang J. New Classifier Ensemble and Fuzzy Community Detection Methods Using POP Choquet-like Integrals. Fractal and Fractional. 2023; 7(8):588. https://doi.org/10.3390/fractalfract7080588

Chicago/Turabian Style

Zhang, Xiaohong, Haojie Jiang, and Jingqian Wang. 2023. "New Classifier Ensemble and Fuzzy Community Detection Methods Using POP Choquet-like Integrals" Fractal and Fractional 7, no. 8: 588. https://doi.org/10.3390/fractalfract7080588

Article Metrics

Back to TopTop