Next Article in Journal
Some Classical and Quantum Aspects of Gravitoelectromagnetism
Previous Article in Journal
High Efficiency Continuous-Variable Quantum Key Distribution Based on ATSC 3.0 LDPC Codes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Geometrical Divide of Data Particle in Gravitational Classification of Moons and Circles Data Sets

Institute of Information Technology and Technical Sciences, Stefan Batory State University, 96-100 Skierniewice, Poland
*
Author to whom correspondence should be addressed.
Entropy 2020, 22(10), 1088; https://doi.org/10.3390/e22101088
Submission received: 18 August 2020 / Revised: 24 September 2020 / Accepted: 25 September 2020 / Published: 27 September 2020

Abstract

:
Thus far, the Universal Law of Gravitation has found application in many issues related to pattern classification. Its popularity results from its clear theoretical foundations and the competitive effectiveness of the classifiers based on it. Both Moons and Circles data sets constitute distinctive types of data sets that can be found in machine learning. Despite the fact that they have not been formally defined yet, on the basis of their visualization, they can be defined as sets in which the distribution of objects of individual classes creates shapes similar to circles or semicircles. This article makes an attempt to improve the gravitational classifier that creates a data particle based on the class. The aim was to compare the effectiveness of the developed Geometrical Divide method with the popular method of creating a class-based data particle, which is described by a compound of 1 ÷ 1 cardinality in the Moons and Circles data sets classification process. The research made use of eight artificially generated data sets, which contained classes that were explicitly separated from each other as well as data sets with objects of different classes that did overlap each other. Within the limits of the conducted experiments, the Geometrical Divide method was combined with several algorithms for determining the mass of a data particle. The research did also use the k-Fold Cross-Validation. The results clearly showed that the proposed method is an efficient approach in the Moons and Circles data sets classification process. The conclusion section of the article elaborates on the identified advantages and disadvantages of the method as well as the possibilities of further research and development.

1. Introduction

Thus far, there is no formal definition in the literature to describe Moons and Circles data sets. However, as stated in the original work about the library implementing methods that generate the aforementioned data sets, their hallmark is the fact that in the feature space in the 2 dimension, the distribution of objects described by different labels, belonging to individual classes, creates shapes similar to circles or semicircles, which is illustrated by the visualization of such a data set. Other characteristics of these data sets are the dearth of a linear decision boundary, the proximity of centroids of different classes to each other, and the proximity of objects of different classes in the vicinity of the centroids. By reason of the above-mentioned features, these data sets generate problems in the process of their classification with the application of centroid-based algorithms and linear classifiers [1].
One of many popular groups of centroid-based classifiers is the one applying the Universal Law of Gravitation approach [2]. Their growing popularity is due to their simplicity and high efficiency [3]. As a result, the gravity model finds application in both supervised [4,5], and unsupervised [6,7] machine learning techniques. As Gorecki and Luczak stated in the article devoted to the variant of gravitational classification [8], the gravitational data model was first proposed by Wright in 1977 [9]. The analysis of the currently published fieldworks allows supplementing the cited thesis that the complete explanation of the theory of Data Gravitation-based Classification (DGC)—the explanation of key definitions and lemmas—was made only in 2005 in the pages of [5]. The following key concepts were clarified at that time: data particle, data mass, data centroid, atomic data particle, data gravity, gravitational data field, and a number of lemmas for Data Gravitation-based Classification. Based on that, Peng et al. developed Data Gravitation-based Classification. In the same year, Wang and Chen [10], attempted to solve the problem of a significant decrease in the performance of the Nearest Neighbor (NN) method in data sets, in which the distribution of attribute values of objects of individual classes determines the overlapping of objects with different labels in the feature space, which in the case of particular data sets, may also determine the aforementioned problem of too near distribution of class centroids [1]. Then, a new method of gravitational classification, namely Simulated Gravitational Collapse (SGC), was developed.
Throughout the years, the DGC received a number of amendments and hands-on applications. In 2006, the practical application of the DGC in the Intrusion Detection System (IDS) prototype was demonstrated [11]. By adapting the classic DGC [5] to solve the problem of classification of unbalanced data sets, the following methods were developed: the Imbalanced DGC (IDGC) [12], Under-Sampling Imbalanced Data Gravitation Classification (UI-DGC) [13], and Synthetic Minority Over-sampling Technique methods integrated with DGC (SMOTE-DGC) [14]. Carrying on with work on the IDGC method, Peng et al. (2017) [15] implemented an algorithm with a view to identifying problems and threats in unbalanced network traffic data. In the literature, one can also observe several gravity classifiers implemented to the problem of small data sets classification, one of them being the Cognitive Gravitation Model (CGM) [16].
Thus far, three methods have been proposed in terms of creating a data particle. The first is a method that creates a data particle on the basis of a single element of the set, which is described by a compound of 1 ÷ 1 cardinality (1OT1P) [5].
The Maximum Distance Principium (MDP) algorithm creates a data particle based on the distance within the elements with the same label [5]. The approach creates a data particle on the basis of the object class with a 1 ÷ 1 compound (1CT1P), finding application, among others, in [17,18]. A number of approaches can be distinguished in the literature that focus on the determination of the centroid, e.g., [19,20], as mentioned in [17], and several approaches used to determine the mass of a data particle [17,18].
The popular Density-Based Spatial Clustering of Applications with Noise (DBSCAN) algorithm is the method worth recalling in considerations concerning the Moons and Circles data sets, although it does not belong to the gravitational approaches group. It is a clustering algorithm that is suitable for the spatial databases, in which the groups of the objects form the arbitrary shapes [21]. The mentioned approach has two main disadvantages. The first is the rapidly growing computational complexity—in the two-dimensional feature space, it is logarithmic, but in higher dimension and/or when the number of data set objects is large, the algorithm rapidly leads to quadratic complexity [22]. This article confirms that DBSCAN is a good approach to data sets whose objects create the arbitrary shapes, but on the condition that they are clearly separated from each other in the feature space, which is the second disadvantage of this algorithm.
Taking into account the fact that there is no linear decision boundary in Moons and Circles data sets and the fact that the gravitational classifiers connecting the data particle to the class do not allow for a sufficient curvature of the decision boundary, it can be hypothesized that the gravity model using the aforementioned method of creating a data particle may not be effective in the Moons and Circles data sets classification process. The above-mentioned fact became a direct inspiration for the authors of this article to take up the research problem, in which the authors were looking for an answer to the following question: Is it possible to effectively adapt gravity classifiers to the Moons and Circles data sets classification process in the feature space in the 2 dimension, using the Geometrical Divide method?
The purpose of the article was to compare the effectiveness of the developed Geometrical Divide (GD) method with the popular method of creating a class-based data particle, which was described by a compound of 1 ÷ 1 cardinality (1CT1P) in the Moons and Circles data sets classification process in the feature space in the 2 dimension.
As part of the research study, the authors conducted an experiment in which they compared the results obtained on the basis of the Geometrical Divide (GD) algorithm with the method that creates a data particle based on the class described by a compound of 1 ÷ 1 cardinality (1CT1P). They combined both methods with algorithms for determining the mass of a data particle, which were developed after the year 2017. The authors used the k-Fold Cross-Validation method and a number of classifier quality assessment measures.
The obtained results showed that centroid-based classifiers that apply the gravity model have the potential in the Moons and Circles data sets classification process described in the 2 dimension, where there is no linear decision boundary, the centers of gravity are located in close proximity to each other, and there are objects in their surroundings that belong to different classes.

2. Materials and Methods

The origins of the proposed Geometrical Divide method is based on the mathematical foundations of the Universal Law of Gravitation, which was developed by Isaac Newton in 1687 [2]; it says that between any pair of points with mass, on the line intersecting their centers, there is a force whose value is proportional to the product of their masses and inversely proportional to the square of the distance between their centers. In [5], adapting this theory to the area of machine learning, the principles of gravitational data classification, on which the proposed GD method is based, were defined.

2.1. Gravitation-Based Classification Principles

The data particle definition is the starting point for the Geometrical Divide method. In compliance with the methods developed thus far, a data particle may consist of a single element o or a set of elements O = { o 1 ,   ,   o N } belonging to a certain, defined feature space S . Each element o is described by a feature vector F = { f 1 ,   ,   f M } and label l . According to the theory idea of [5], the data particle P is the fundamental entity subject to analysis and ought to be considered as a defined type of data unit with three properties: data mass m , such that m \ { 0 } , data centroid μ , and label l . Therefore, the data particle is expressed as a vector, as represented by the Formula (1):
P = { m ,   μ ,   l } .
Prediction of the label l X of the analyzed atomic data particle X , i.e., according to the definition of [5]—indivisible data element with mass m X = 1 , is based on the determination of the value of the gravitational force of the data F between the data particle X and all data particles belonging to the examined feature space S . Thus, the following relationship (Equation (2)) is valid:
P S F \ { 0 } F ( X , P i ) > 0 .
It needs to be noted that in the process of determining the force of gravity of data between data particles, unlike the classical theory of gravity, the gravity constant G is ignored [6]. As a result, the value of the gravitational force of the data between the atomic particle X and the i -th data particle P i is expressed by the following relationship, as shown in Equation (3):
F ( X , P i ) = m i r 2 ,
where r is the distance between the atomic data particle X and the data particle P i .
It is also possible to apply a different formula proposed in [17], which replaces the reciprocal of the square of the distance 1 r 2 with the similarity function S i m ( X , P i ) , according to this Equation (4):
F ( X , P i ) = m i · S i m ( X , P i ) .
Based on the data gravity force values determined in such way between the atomic data particle X and all N data particles belonging to the feature space S , the data particle P i , for which the data gravity force was the highest, is selected [17], which is expressed by the following relationship, as shown in Equation (5):
P y = a r g m a x i     i < N F ( X , P i ) .
In the last step, a decision is made to assign the data particle P y label to the classified atomic data particle X .

2.2. Data Particle Creation Based on Class

The approach to creating a data particle that has found application in the fieldworks is the one mentioned in the introduction, namely 1CT1P, which creates a data particle P on the basis of all objects o with the same label, belonging to the set O . The method assumes that in the input data set D , consisting of an N amount of objects o , there exist an M amount of sets O described with the label l O , and the object o belongs to the set O only if the label of the object l o is identical with the label of the set l O , which is expressed by the following relationship, Equation (6):
o i O j D l o i = l O j ,
where i = 1 ,   2 ,   ,   N and j = 1 ,   2 ,   ,   M .
Then, for each set of objects O , a data particle P is created, which is expressed by the following relationship, Equation (7):
O P ! P = [ m ,   o O F o , l o ] ,
where F o is a feature vector describing the object o .
The presented method shares similarities with the MDP method [5]: in relation to the 1OT1P method, it reduces the number of data particles subject to further analysis. Therefore, it also curtails information about the relative position of objects described by the same label, which may have a negative bearing on the relevance of the data gravity computed in subsequent steps and ultimately on the accuracy of the entire classification process.

2.3. Determining Mass of Data Particle

Two algorithms for determining the mass of a data particle—Stochastic Learning Algorithm (GM SLA) and Batch-update Learning Algorithm (GM BLA) were proposed in 2017 [17]. It was then emphasized that the effectiveness of centroid-based algorithms depends on the characteristics of the data set.
Stochastic Learning Algorithm in the classifier learning phase, with each sample assigned incorrectly, modifies the mass values of two data particles m i and m j by the defined constant, ξ . The value of mass m i of a given data particle P i , into which the sample X ought to be classified, is increased, which is expressed by the following relationship, as shown in Equation (8):
m i = m i + ξ ,
while the value of the mass m j of the data particle P j , to which the sample X has been assigned incorrectly, is decreased, which is described by the following Equation (9):
m j = m j ξ .
Batch-update Learning Algorithm modifies the mass value of individual classes once the classification run is complete, in the course of which it counts the number β i 1 of objects of the class C i , which was classified to other classes, and also the number β i 2 of objects of the other classes erroneously classified to the class C i . Then, based on the accumulated values and the constant ξ , it modifies the i -th mass of the data particle m i created on the basis of the class C i , which is expressed by the following Formula (10):
m i = m i + ( β i 1 β i 2 ) · ξ .
The authors of the presented algorithms indicated that the direction of further research may be to proofread the classifier, in which the mass of the data particle will depend on the numerical amount of the class. Therefore, in 2019, the effectiveness of the approach was put to the test, in which the mass value of a data particle results from the size of a particular class [18]. At that time, the aim of the research was to find out how the mass function influences the effectiveness of the Centroid-Based Classifier, which applies Newton’s Law of Universal Gravity [2]. The results of the research that was carried out on the Iris set showed that there is a potential for linear binding the mass of a data particle m i derived from the class of objects of its size N i , which is expressed by the following relationship, as shown in Equation (11):
m i = N i .

2.4. Geometrical Divide of Data Particle

The Geometrical Divide (GD) method presented by the authors of this article has its mathematical foundations in analytic geometry and assumes that each data particle P consists of at least 2 atomic data particles A , which is described by the feature vector F = [ f 1 ,   f 2 ] , belonging to the feature space S in the 2 dimension. The process of splitting a single piece of data P is described below.
In the first step, the value of the depth level d , such that d + , which satisfies the following equation, is defined in Equation (12):
l o g 2 N m i n < d ,
where N m i n is the size of the smallest class C creating the data particle P .
Subsequently, for each existing data particle P based on the values of the atomic attributes of data particles f 1 and f 2 that assemble the data particle P , the data particle centroid μ is determined, which expresses the following relationship, as shown in Equation (13):
μ = [ i = 1 | P | A i f 1 · | P | 1 , i = 1 | P | A i f 2 · | P | 1 ] .
In the third step, based on the dispersion of the values of the atomic attributes of the data particles that assemble the data particle P , the geometric center C is determined for each of them, which is expressed by the following Formula (14):
C = [ f 1 m a x f 1 m i n ,   f 2 m a x f 2 m i n ] ,
where f 1 m a x is the maximum value of the feature f 1 of the atomic data particle A creating the particle P ; f 1 m i n is the minimum value of the feature f 1 of the atomic data particle A creating the particle P ; f 2 m a x is the maximum value of the feature f 2 of the atomic data particle A creating the particle P ; and f 2 m i n is the minimum value of the feature f 2 of the atomic data particle A creating the particle P .
In the fourth step, the method determines the equation of the line constructed through the given points μ = [ μ 1 ,   μ 2 ] and C = [ c 1 ,   c 2 ] , which is expressed by the following relationship, as shown in Equation (15):
( y μ 2 ) ( c 1 μ 1 ) ( c 2 μ 2 ) ( x μ 1 ) = 0 .
In the last, fifth step, the data particle is divided in regard to the defined straight line. Then, if the analyzed data particle A is above the constructed line, it will create a new data particle P i , whereas if it is on the line or below it, it will create a data particle P j . For this purpose, each atomic data particle A = [ a 1 , a 2 ] , assembling a data particle P , is placed at the beginning of Formula (16):
A { P i   i f   ( a 2 μ 2 ) ( c 1 μ 1 ) ( c 2 μ 2 ) ( a 1 μ 1 ) > 0 P j   i f   ( a 2 μ 2 ) ( c 1 μ 1 ) ( c 2 μ 2 ) ( a 1 μ 1 ) 0 .
The pseudocode of the Geometrical Divide (GD) algorithm is illustrated in Algorithm 1:
Algorithm 1 Geometrical Divide (GD)
dataParticles: list of data particles
featureVectors: list of data particle’s feature vectors F = [ f 1 ,   f 2 ]
newDataParticles: list of data particles created by divide
N m i n : size of the smallest class
d :   d + AND   l o g 2 N m i n < d
FOR i = 2 ; i d
   FOREACH   P d a t a P a r t i c l e s DO
     μ P     centroid ( P )          / /   μ   = [ μ 1 ,   μ 2 ]
     C P     geometricCentroid ( P )     / /   C = [ c 1 ,   c 2 ]
dataParticle1 = List()           // create two empty lists dataParticle1
    dataParticle2 = List()          // and dataParticle2 for new data particles
     FOREACH   F f e a t u r e V e c t o r s DO
       IF   ( f 2 μ 2 ) ( c 1 μ 1 ) ( c 2 μ 2 ) ( f 1 μ 1 ) > 0 THEN
         dataParticle 1     add ( returnParticle ( F ) )
      ELSE dataParticle 2     add ( returnParticle ( F ) )
      END IF
    END FOREACH
     newDataParticles     add(dataParticle1)
newDataParticles     add(dataParticle2)
  END FOREACH
   dataParticles     newDataParticles
   newDataParticles     clear()
END FOR

2.5. Computational Complexity of Geometrical Divide

The computational complexity of the proposed algorithm depends on the value of the depth level d and the size of the data set n . It is known that each element of the data set corresponds to exactly one atomic data particle, and each of them belongs to only one data particle. In reference to that, the deepest nested FOREACH loops, of which the outer will be executed i fold ( i equals the number of existing data particles) and the inner will be executed j fold ( j is the number of atomic data particles belonging to the data particle) will be executed n times in total—each data set object will be analyzed.
The analysis of the outer FOR loop shows that it will be executed d 1 fold, because at the level of the partition depth d = 1 , data particle division is not performed. Thus, the computational complexity of this algorithm is presented by Formula (17):
C o m p u t a t i o n a l   C o m p l e x i t y G D = O ( ( d 1 ) n ) .
It should be emphasized that d n .

2.6. Data Sets

In the conducted research, 8 artificially generated data sets were applied—4 in which the classes were explicitly separated from each other (data sets without ns (not separated) in their name) and 4 in which the objects of different classes did overlap each other (data sets with ns in their name). The distribution of feature values in the feature space in the 2 dimension was defined in a typical manner for Moons and Circles data sets. The sizes of the generated data sets were determined based on the characteristics of 42 real data sets used in previous publications in this field. The types of the generated data sets are presented in Figure 1a–h.
The characterization of the generated data sets is presented in Table 1.
In addition, the popular real data sets were processed as well. Two data sets from the repository of the University of California in Irvine (UCI) were chosen: The Banknote Authentication and The Breast Cancer Wisconsin [23]. In order for these data sets to meet the defined assumptions, their pre-processing was carried out. The preparation consisted in selecting two pairs of attributes from each data set that described them in the 2 feature space. In this way, 4 data sets were created. The first is banknote_authentication_12, which is based on the values of the variance and the skewness obtained as a result of the image wavelet transformation. The second data set is banknote_authentication_13, in which objects are described through variance and the kurtosis of the wavelet transformed image. The third is the breast_cancer_wisconsin_37, which described objects through the value of the cell size’s uniformity and the number of the bare nuclei. The last one is the breast_cancer_wisconsin_49, which is based on the value of the cell shape’s uniformity and the number of the normal nucleoli.

2.7. Evaluation of the Method

In the evaluation process of the proposed method, k -Fold Cross-Validation was used. It was also used to evaluate the models in many new publications [17,24]. The application of k -Fold Cross-Validation enables avoiding the model overfitting and improves the model’s generalization property [24]. The value of the parameter k was set at k = 10 . Such a parameterized method divided, in each iteration, the entire data set into two subsets: the test set, which constituted 10% of the entire base set, and the training set, consisting of the remaining 90% of the original set samples. In course of the first experiment, two data particle determination algorithms were compared: 1CT1P and GD.
The most popular and simplest metrics for pattern recognition system evaluation are accuracy (ACC) and error rate (ERR). It results from their key advantages—low complexity, the possibility of applying them in multi-class and multi-label classification processes, and easy-to-process scoring [25]. However, in the context of the Geometrical Divide evaluation in the Moons and Circles data sets classification process, in which the information about the prediction of individual classes is relevant, the dearth of informativeness of said measures in the subject issue, which excludes their use, needs be pointed out [25,26]. Therefore, said measures are not capable of providing relevant information that could enable an objective evaluation of the developed solution.
On account of the aforementioned downside, scientists apply complementary metrics in the field of machine learning, which was developed in 1955 and used to evaluate the information retrieval systems [27], including P R E C I S I O N , as shown in Equation (18):
P R E C I S I O N = T P · ( T P + F P ) 1 ,
where T P is true positive and F P is false positive, and R E C A L L , as shown in Equation (19):
R E C A L L = T P · ( T P + F N ) 1 ,
where F N is false negative.
In the process of model optimization, two objective functions may be taken into account, namely the maximization of the precision value and the maximization of the reference value. In practice, this task is difficult to implement, as most of the time, these functions are negatively correlated, which means that increasing the precision value often reduces the reference value and vice versa [28].
P R E C I S I O N and R E C A L L can be reduced to a single value F β - s c o r e , which is the harmonic mean of the P R E C I S I O N and R E C A L L [29], as shown in Equation (20):
F β - s c o r e = ( β 2 + 1 ) · P R E C I S I O N · R E C A L L · ( β 2 · ( P R E C I S I O N + R E C A L L ) ) 1 .

3. Results

In the first stage, the parameter defining the depth of division in the proposed GD method was set at d = 8 . Each of the algorithms was combined with the Nearest Centroid (NC) method and three weight determination algorithms. The first one was the Stochastic Learning Algorithm (GM SLA), in which the parameter defining the maximum number of iterations in the learning process was defined as M a x I t e r s = 50 , the mass value update coefficient was assigned the value ξ = 0.0001 , while the expected error threshold was set at ε = 0.00 . The second algorithm was the Batch-update Learning Algorithm (GM BLA), in which the data particle mass value update parameter was set at ξ = 0.0001 , while the last method of mass determination was the n-Mass Model (N-MM). The DBSCAN clustering algorithm with the minimum number of points required to create a dense region M i n P t s = 10 and the parameter defining the maximum admissible distance between points within the same region E p s = 0.150 was added to the comparison. The values of the F 1 - s c o r e obtained as a result of the first experiment for each individual method are presented in Table 2. The best result within each data set is highlighted in bold.
Having analyzed the acquired results, it can be concluded that the gravity classifier using the proposed GD approach obtained the best results on all artificially generated data sets. On data sets in which classes are separated—their objects do not overlap each other—the GD algorithm in all combinations and the DBSCAN algorithm obtained the measure result F 1 - s c o r e = 1.000 , whereas for data sets in which the objects create the unseparated arbitrary shapes—sets with postfix ns—the best measure results of F 1 - s c o r e were returned by the GD approach in combination with the n-Mass Model (n-MM GD), which significantly outperforms the class-based data particle creation approach and the DBSCAN algorithm.
In the next step, the similar experiment was conducted on the real data sets whose objects do not create the shapes in the feature space. The parameters of the algorithms were set up identical to the first stage. Only the value of the d parameter was different, and in this case, d = 3 . The results of that part of the research are presented in Table 3. The best result within each data set is highlighted in bold.
Through the analysis of the results presented in Table 3, it can be observed that in the case of the tested real data sets, the F 1 - s c o r e results achieved by the algorithms using the proposed Geometrical Divide method outperform the results achieved by the algorithms that use the method of creating a class-based data particle by a compound of 1 ÷ 1 cardinality. The next conclusion based on an analysis of Table 3 is that the data sets, whose objects do not create the shapes in the feature space, are problematic for the DBSCAN algorithm.
The results presented in Table 2 and Table 3 were subjected to the Friedman Aligned Ranks statistical test [30] with the Holm post-hoc test [31] and with a statistical significance level of p = 0.05 . The test showed a significant improvement in the results obtained by the approaches using the proposed data particle divide algorithm in relation to the results obtained by the methods of creating a class-based data particle by a compound of 1 ÷ 1 cardinality and the popular DBSCAN algorithm. The same test showed that there is no statistically significant difference in the results obtained by the approaches belonging to the group of the methods that applied the Geometrical Divide algorithm (approaches with GD in the name).
In the second stage of the research, the influence of the d parameter value on the F 1 - s c o r e obtained by the n-MM (GD) method was examined. On account of the characteristics of the examined data sets, the parameter d assumed integer values in the interval of [1 ÷ 8]. The results of this experiment are presented in Figure 2.
Based on the analysis of Figure 2, it can be noticed that the greatest improvement in the classifier’s results occurs already in the second iteration of the algorithm, i.e., in the first division of data particles. In each subsequent iteration, the improvement in results is relatively smaller. In some cases, noticeable decreases can be seen in the F 1 - s c o r e . At the depth level of d = 3 , the said decrease can be observed for the two_moons_mirrored_ns and two_moons_ns data sets. At the depth level of d = 4 , similar behavior occurred for the circle_and_moon data set. Along with the subsequent levels of the depth d , the above-mentioned decreases are being gradually improved. In terms of all the examined data sets, at the depth level d = 4 , the F 1 - s c o r e value was higher than the value obtained before the splitting of the data particle.
In the third stage of the experiment, the n-MM GD method was examined while paying special attention to the impact of the value of the depth parameter d on the value of the P R E C I S I O N and R E C A L L measures. The results of this stage of the research are visualized in Figure 3, Figure 4 and Figure 5.
For d = 2 (Figure 3), the biggest difference in the results of both measures occurs in the case of the circle_in_ring data set and amounts to | P R E C I S I O N R E C A L L | = 0.226 , whereas the mean difference between the above-mentioned measures within the entire experiment was 0.061 .
At the depth level d = 4 (Figure 4), in case of the circle_in_ring and two_moons_mirrored data sets, the significant difference between the P R E C I S I O N and R E C A L L values is not noticeable. For the circle_and_moon and two_moons data sets, the difference amounts to 0.090 and 0.082 respectively, giving the two highest difference scores between these measures within the experiment. In data sets where the classes are not separated from each other, the difference value between P R E C I S I O N and R E C A L L fluctuates in the interval of [0.009 ÷ 0.065]. On the other hand, the average difference value of said measures for the parameter d = 4 , in terms of all the analyzed data sets, amounted to 0.037 . Thus, a balanced improvement of the classification can be noticed in the light of P R E C I S I O N and R E C A L L measures in contrast to the previous experiment, where the depth threshold d was set at d = 2 .
At the maximum depth level d = 8 (Figure 5) that was examined, in case of data sets with clearly separated classes—circle_in_ring, circle_and_moon, two_moons and two_moons_mirrored—there is no significant difference between the obtained P R E C I S I O N and R E C A L L values. In terms of data sets in which the classes are not separated from each other (data sets with ns-postfix in its name), the differences in the values of both measures fluctuate in the interval of [0.015 ÷ 0.050], while the mean result of the difference between P R E C I S I O N and R E C A L L within the entire experiment with the parameter d = 8 amounted to 0.015 .
Based on the analysis of Figure 3, Figure 4 and Figure 5, it can be pointed out that the proposed method of splitting the GD data particle aims at a sustainable improvement of the results of the P R E C I S I O N and R E C A L L , which proves that the quality improvement is not based on the discrimination of any classes, and the classifier with similar effectiveness can predict all classes within a given data set. While analyzing Figure 2, Figure 3, Figure 4 and Figure 5, it can also be observed that with the increase of the depth of data particle division d , the difference between P R E C I S I O N and R E C A L L decreases, and the F 1 - s c o r e value increases concurrently.

4. Discussion

The article solves the problem of the low efficiency of classifiers based on the gravity model, which creates a class-based data particle described by a compound of 1 ÷ 1 cardinality. The disadvantage of such a method of binding is noticeable in the data sets classification process, in which the centers of gravity of classes with different labels are located in the proximity to each other, and there are objects belonging to other classes in their surrounding. The data sets possessing the above-mentioned features are Moons and Circles data sets.
The evaluation of the developed method was conducted with the use of approved classifier evaluation measures P R E C I S I O N , R E C A L L , and F 1 - s c o r e . The obtained values of the above-mentioned measures showed an improvement in the results in contrast to the results obtained by applying the method of creating a class-based data particle with a compound of 1 ÷ 1 cardinality.
The first stage of the experiment demonstrated the main advantages of the Geometrical Divide algorithm. The first one is its resistance to a short distance between class centroids. In accordance to that, the GD method outperforms the data particle creation method based on class with a compound 1 ÷ 1 cardinality. The second advantage of the proposed algorithm is its resistance to the presence of objects of different classes in the vicinity of centroids. In this criterion, the GD algorithm significantly outperforms DBSCAN algorithm. It is also worth noting that in comparison to the Maximum Distance Principium method, the Geometrical Divide does not calculate the distance between atomic data particles and centroids in the phase of determining the data particles. In order to significantly improve the results of the classifier, the division of a data particle in the Geometrical Divide method does not always have to apply the maximum available depth level d .
The second stage of the experiment showed that for the value of the parameter d = 4 , which in terms of the analyzed data sets made up approximately the middle value of the interval of the available values, the results did not differ significantly from those obtained at the depth level d 5 . A noticeable improvement in the results can be observed already in the second iteration of the algorithm, during the first splitting of the data particle.
The third stage of the experiment displayed that the Geometrical Divide improves the gravity classifier in a balanced way; therefore, the improvement of its quality is not based on the discrimination of any of the classes, and the classifier is capable of predicting all classes within a given data set with similar effectiveness.
As far as further research is concerned, it is planned to apply the Geometrical Divide method in the process of classifying unbalanced data sets and/or higher dimension data sets. Another direction of analysis may be an attempt to develop a method of mass inheritance in the course of the splitting process. Furthermore, it is worth endeavoring to develop a method for selecting the parameter d based on the distribution of class attributes of the data set, which may contribute toward the further development of the presented method.

Author Contributions

Ł.R. provided the concept of the work, prepared the research methodology and resources, implemented software for the evaluation process and wrote the paper under the guidance of J.D. J.D. was the supervisor of project, participated in formal analysis, lead the project administration, and supported the review and analyzed the results. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pedregosa, F.; Varoquax, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Duborg, V.; et al. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  2. Newton, I. Matematyczne Zasady Filozofii Naturalnej; Brzezowski, S., Translator; Cracow Copernicus Center Press: Krakow, Poland, 2015. [Google Scholar]
  3. Peng, L.; Liu, Y. Gravitation theory based model for multi-label classification. Int. J. Comput. Commun. Control 2017, 12, 689–703. [Google Scholar] [CrossRef]
  4. Cano, A.; Zafra, A.; Ventura, S. Weighted data gravitation classification for standard and imbalanced data. IEEE Trans. Cybern. 2013, 43, 1672–1687. [Google Scholar] [CrossRef] [PubMed]
  5. Peng, L.; Chen, Y.; Yang, B.; Chen, Z. A Novel Classification Method Based on Data Gravitation. In Proceedings of the 2005 International Conference on Neural Networks and Brain, Beijing, China, 13 October 2005; pp. 667–672. [Google Scholar] [CrossRef]
  6. Aghajanyan, A. Introduction to gravitational clustering. Pattern Anal. Mach. Intell. 2015, 10, 1–3. [Google Scholar]
  7. Gomez, J.; Dasgupta, D.; Nasraoui, O. A New Gravitational Clustering Algorithm. In Proceedings of the 2003 SIAM International Conference on Data Mining, San Francisco, CA, USA, 1–3 May 2003. [Google Scholar]
  8. Górecki, T.; Łuczak, M. A variant of gravitational classification. Biometr. Lett. 2014, 51. [Google Scholar] [CrossRef] [Green Version]
  9. Wright, W.E. Gravitational clustering. Pattern Recogn. 1977, 9, 151–166. [Google Scholar] [CrossRef]
  10. Wang, C.; Chen, Y.Q. Improving Nearest Neighbor Classification with Simulated Gravitational Collapse. In Proceedings of the First international Conference on Advances in Natural Computation, Changsha, China, 27–29 August 2005; pp. 845–854. [Google Scholar] [CrossRef]
  11. Yang, B.; Peng, L.; Chen, Y.; Liu, H.; Yuan, R. A DGC-based data classification method used for abnormal network intrusion detection. In Proceedings of the 13th International Conference on Neural Information Processing, Hong Kong, China, 3–6 October 2006; pp. 209–216. [Google Scholar] [CrossRef]
  12. Peng, L.; Zhang, H.; Yang, B.; Chen, Y. A new approach for imbalanced data classification based on data gravitation. Inform. Sci. Inform. Comput. Sci. Intell. Syst. Appl. Int. J. 2014, 288, 347–373. [Google Scholar] [CrossRef]
  13. Peng, L.; Yang, B.; Chen, Y.; Zhou, X. An Under-Sampling Imbalanced Learning of Data Gravitation Based Classification. In Proceedings of the 12th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery, Changsha, China, 13–15 August 2016; pp. 419–425. [Google Scholar] [CrossRef]
  14. Peng, L.; Yang, B.; Chen, Y.; Zhou, X. SMOTE-DGC: An Imbalanced Learning Approach of Data Gravitation Based Classification. In Proceedings of the 12th International Conference on Intelligent Computing: Intelligent Computing Theories and Application, Lanzhou, China, 2–5 August 2016; Volume 9772, pp. 133–144. [Google Scholar] [CrossRef]
  15. Peng, L.; Zhang, H.; Chen, Y.; Yang, B. Imbalanced traffic identification using an imbalanced data gravitation-based classification model. Comput. Commun. 2017, 102, 177–189. [Google Scholar] [CrossRef]
  16. Wena, G.; Wei, J.; Wang, J.; Zhou, T.; Chen, L. Cognitive gravitation model for classification on small noisy data. Neurocomputing 2013, 118, 245–252. [Google Scholar] [CrossRef]
  17. Liu, C.; Wang, W.; Tu, G.; Xiang, Y.; Wang, S.; Lv, F. A new Centroid-Based Classification model for text categorization. Knowl. Based Syst. 2017, 136, 15–26. [Google Scholar] [CrossRef]
  18. Rybak, L.; Dudczyk, J. Various approaches to modelling of the mass using the size of the class in the Centroid Based Classification. Elektron. Konstr. Technol. Zastos. 2019, 60, 62–65. [Google Scholar] [CrossRef]
  19. Guan, H.; Zhou, J.; Guo, M. A Class-Feature-Centroid Classifier for Text Categorization. In Proceedings of the 18th International Conference on World Wide Web, Madrid, Spain, 20–24 April 2009; pp. 201–210. [Google Scholar] [CrossRef]
  20. Tan, S. An improved centroid classifier for text categorization. Exp. Syst. Appl. Int. J. 2008, 35, 279–285. [Google Scholar] [CrossRef]
  21. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. A Density-Based Algorithm for Discovering Clusters in Large Spatial Databases with Noise. In Proceedings of the Second International Conference on Knowledge Discovery and Data Mining (KDD-96), Portland, OR, USA, 2–4 August 1996; pp. 226–231. [Google Scholar]
  22. Gan, J.; Tao, Y. DBSCAN Revisited: Mis-claim, un-fixability, and approximation. In SIGMOD’15, Proceedings of the 2015 ACM SIGMOD International Conference on Management of Data, Melbourne, Australia, 31 May–4 June 2015; Association for Computing Machinery: New York, NY, USA, 2015; pp. 519–530. [Google Scholar] [CrossRef]
  23. UCI Machine Learning Repository Datasets. Available online: https://archive.ics.uci.edu/ml/datasets.php (accessed on 24 August 2020).
  24. Chae, J.; Kang, Y.J.; Noh, Y. A deep-learning approach for foot-type classification using heterogeneous pressure data. Sensors 2020, 20, 4481. [Google Scholar] [CrossRef] [PubMed]
  25. Hossin, M.; Sulaiman, M.N. A Review on evaluation metrics for data classification evaluations. Int. J. Data Min. Knowl. Manag. Process 2015, 5, 1–11. [Google Scholar] [CrossRef]
  26. Hossin, M.; Sulaiman, M.; Mustapha, A.; Mustapha, N.; Rahmat, R. A Hybrid Evaluation Metric for Optimizing Classifier. In Proceedings of the 3rd Conference on Data Mining and Optimization, Putrajaya, Malaysia, 28–29 June 2011; pp. 165–170. [Google Scholar] [CrossRef]
  27. Kent, A.; Berry, M.; Luehrs, F.U.; Perry, J.W. Machine literature searching VIII. Operational criteria for designing information retrieval systems. J. Assoc. Inf. Sci. Technol. 1955, 6, 93–101. [Google Scholar] [CrossRef]
  28. Raghavan, V.; Bollmann, P.; Jung, G.S. A critical investigation of recall and precision as measures of retrieval system performance. ACM Trans. Inf. Syst. 1989, 7, 205–229. [Google Scholar] [CrossRef]
  29. Flach, P.; Kull, M. Precision-Recall-Gain Curves: PR Analysis Done Right. In Proceedings of the 28th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 838–846. [Google Scholar] [CrossRef]
  30. Hodges, J.L.; Lehmann, E.L. Ranks methods for combination of independent experiments in analysis of variance. Ann. Math. Stat. 1962, 33, 482–497. [Google Scholar] [CrossRef]
  31. Holm, S. A simple sequentially rejective multiple test procedure. Scand. J. Stat. 1979, 6, 65–70. [Google Scholar]
Figure 1. This figure presents the data sets that were used in the research: (a) circle_in_ring; (b) circle_in_ring_ns; (c) circle_and_moon; (d) circle_and_moon_ns; (e) two_moons; (f) two_moons_ns; (g) two_moons_mirrored; (h) two_moons_mirrored_ns.
Figure 1. This figure presents the data sets that were used in the research: (a) circle_in_ring; (b) circle_in_ring_ns; (c) circle_and_moon; (d) circle_and_moon_ns; (e) two_moons; (f) two_moons_ns; (g) two_moons_mirrored; (h) two_moons_mirrored_ns.
Entropy 22 01088 g001aEntropy 22 01088 g001b
Figure 2. Change of the F 1 - s c o r e values for the subsequent values of the depth level d within individual data sets.
Figure 2. Change of the F 1 - s c o r e values for the subsequent values of the depth level d within individual data sets.
Entropy 22 01088 g002
Figure 3. Values of P R E C I S I O N and R E C A L L measures for individual data sets, method: the n-Mass Model (n-MM) (Geometrical Divide, GD), parameter d = 2 .
Figure 3. Values of P R E C I S I O N and R E C A L L measures for individual data sets, method: the n-Mass Model (n-MM) (Geometrical Divide, GD), parameter d = 2 .
Entropy 22 01088 g003
Figure 4. Values of P R E C I S I O N and R E C A L L measures for individual data sets, method: n-MM (GD), parameter d = 4 .
Figure 4. Values of P R E C I S I O N and R E C A L L measures for individual data sets, method: n-MM (GD), parameter d = 4 .
Entropy 22 01088 g004
Figure 5. Values of P R E C I S I O N and R E C A L L measures for individual data sets, method: n-MM (GD), parameter d = 8 .
Figure 5. Values of P R E C I S I O N and R E C A L L measures for individual data sets, method: n-MM (GD), parameter d = 8 .
Entropy 22 01088 g005
Table 1. The results of the F1-score obtained by the algorithms applied on individual data sets.
Table 1. The results of the F1-score obtained by the algorithms applied on individual data sets.
Data SetPOSITIVESNEGATIVES
circle_in_ring18621864
circle_in_ring_ns23882457
circle_and_moon25842697
circle_and_moon_ns24252500
two_moons28861938
two_moons_ns23842402
two_moons_mirrored25122610
two_moons_mirrored_ns22132252
Table 2. The results of the F 1 - s c o r e obtained by the algorithms applied on artificially generated data sets.
Table 2. The results of the F 1 - s c o r e obtained by the algorithms applied on artificially generated data sets.
Data SetDBSCANNC (1CT1P)GM SLA (1CT1P)GM BLA (1CT1P)n-MM (1CT1P)NC (GD)GM SLA (GD)GM BLA (GD)n-MM (GD)
circle_in_ring1.0000.5070.5730.5050.4841.0001.0001.0001.000
circle_in_ring_ns0.5090.4970.4850.4980.3140.7240.7250.7240.735
circle_and_moon1.0000.8510.6740.8760.8371.0001.0001.0001.000
circle_and_moon_ns0.4920.7440.7240.7390.7460.8040.8010.8040.806
two_moons1.0000.6860.5230.7050.6731.0001.0001.0001.000
two_moons_ns0.4960.6010.6040.5880.5940.7340.7360.7340.738
two_moons_mirrored1.0000.8240.8220.8260.8241.0001.0001.0001.000
two_moons_mirrored_ns0.4980.7760.7760.7750.7760.8990.9000.9000.907
Table 3. The results of the F 1 - s c o r e obtained by the algorithms applied on real data sets.
Table 3. The results of the F 1 - s c o r e obtained by the algorithms applied on real data sets.
Data SetDBSCANNC (1CT1P)GM SLA (1CT1P)GM BLA (1CT1P)n-MM (1CT1P)NC (GD)GM SLA (GD)GM BLA (GD)n-MM (GD)
banknote_authentication_120.3720.8640.8620.8640.8660.9040.9050.9040.897
banknote_authentication_130.3030.8100.8100.8100.8000.8300.8310.8290.838
breast_cancer_wisconsin_370.5180.9190.9190.9190.8810.9370.9340.9370.890
breast_cancer_wisconsin_490.4690.8710.8710.8710.8160.8870.8860.8860.872

Share and Cite

MDPI and ACS Style

Rybak, Ł.; Dudczyk, J. A Geometrical Divide of Data Particle in Gravitational Classification of Moons and Circles Data Sets. Entropy 2020, 22, 1088. https://doi.org/10.3390/e22101088

AMA Style

Rybak Ł, Dudczyk J. A Geometrical Divide of Data Particle in Gravitational Classification of Moons and Circles Data Sets. Entropy. 2020; 22(10):1088. https://doi.org/10.3390/e22101088

Chicago/Turabian Style

Rybak, Łukasz, and Janusz Dudczyk. 2020. "A Geometrical Divide of Data Particle in Gravitational Classification of Moons and Circles Data Sets" Entropy 22, no. 10: 1088. https://doi.org/10.3390/e22101088

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop