Next Article in Journal
Entropy and Cities: A Bibliographic Analysis towards More Circular and Sustainable Urban Environments
Previous Article in Journal
An Innovative Possibilistic Fingerprint Quality Assessment (PFQA) Filter to Improve the Recognition Rate of a Level-2 AFIS
Previous Article in Special Issue
Diversity-Aware Marine Predators Algorithm for Task Scheduling in Cloud Computing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Particle Swarm Optimization Algorithm with Dynamic Adjustment of Inertia Weight Based on a New Feature Selection Method to Optimize SVM Parameters

1
School of Media Science & School of Journalism, Northeast Normal University, Jingyue, Changchun 130117, China
2
Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, Jilin University, Qianjin, Changchun 130012, China
3
College of Computer Science and Technology, Jilin University, Qianjin, Changchun 130012, China
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(3), 531; https://doi.org/10.3390/e25030531
Submission received: 8 December 2022 / Revised: 17 March 2023 / Accepted: 17 March 2023 / Published: 19 March 2023
(This article belongs to the Special Issue Information Theory and Swarm Optimization in Decision and Control)

Abstract

:
Support vector machine (SVM) is a widely used and effective classifier. Its efficiency and accuracy mainly depend on the exceptional feature subset and optimal parameters. In this paper, a new feature selection method and an improved particle swarm optimization algorithm are proposed to improve the efficiency and the classification accuracy of the SVM. The new feature selection method, named Feature Selection-score (FS-score), performs well on data sets. If a feature makes the class external sparse and the class internal compact, its FS-score value will be larger and the probability of being selected will be greater. An improved particle swarm optimization model with dynamic adjustment of inertia weight (DWPSO-SVM) is also proposed to optimize the parameters of the SVM. By improving the calculation method of the inertia weight of the particle swarm optimization (PSO), inertia weight can decrease nonlinearly with the number of iterations increasing. In particular, the introduction of random function brings the inertia weight diversity in the later stage of the algorithm and the global searching ability of the algorithm to avoid falling into local extremum. The experiment is performed on the standard UCI data sets whose features are selected by the FS-score method. Experiments demonstrate that our algorithm achieves better classification performance compared with other state-of-the-art algorithms.

1. Introduction

Support vector machine (SVM) is one of the most popular machine learning methods, which is based on the statistical learning theory and the structural risk minimization principle [1]. Moreover, it has unique advantages in solving practical problems, such as small sample data sets as well as nonlinear and high-dimensional pattern recognition [2,3,4]. In SVM, the kernel function selection is a key part for nonlinear problems. The Radial Basis Function (RBF), as the most common kernel function, has the characteristics of fewer optimized parameters and better classification performance in terms of accuracy and stability, especially for high-dimensional data. However, the classification effect of the SVM is related to two factors. The first one is the quality of the input data set. The greater the degree of data set differentiation, the better the algorithm effects. The second one is the value of the error penalty parameters and the kernel parameters. In order to achieve the best classification effect and the maximum generalization of the SVM, many researchers have proposed different solutions to solve the two issues above.
One of the most widespread methods for selecting these two key parameters is using the grid algorithm to traverse and compare iteratively. This method can achieve relatively good results, but it is time-consuming [5]. Therefore, some researchers proposed using the meta-heuristic algorithm to find the best parameter value within the specified range [6,7,8,9]. The most commonly used meta-heuristic algorithms are the genetic algorithms. They perform well in solving large-scale nonlinear problems [10,11]. However, they are easy to fall into the local optimal value and fail to find the global optimal solution [12,13,14], since the algorithm follows the natural law of information exchange and evolution among populations. Another meta-heuristic algorithm used to find the optimal parameters is particle swarm optimization (PSO), which mimics the information exchange between birds in searching for food to find the optimal result in the population. Each iteration of the algorithm will be adjusted according to the results of the previous iteration, which greatly improves the efficiency. Compared with the cross-mutation of the genetic algorithm, it is more targeted, which weakens its randomness. However, the PSO algorithm is as easy as the genetic algorithm to fall into local optimum, resulting in large errors in the results [15,16]. Mafarja et al. proposed the Binary Dragonfly Algorithm for minimizing all features using the wrapper feature selection algorithms to enhance the classification accuracy [17]. By taking a mutational approach and scanning the area surrounding the explorers, Binary Al-Biruni Earth Radius Optimization is able to deliver cutting-edge exploration capabilities. It ensures diversity and thorough exploration by shuffling the order of responses between repetitions [18]. Khafaga et al. proposed the Binary Dipper Throated Optimization (bDTO) method to optimize the weighted ensemble model by the DTO algorithm. Once the significant features are selected, the optimized weighted ensemble model is employed to predict the gain values of metamaterial antenna [19]. The Binary Cat Swarm Optimization (BCSO) algorithm includes two major modes: tracing and seeking modes. However, the traditional BCSO has the chance of becoming trapped in the local optima [20].
Many scholars have improved the PSO algorithm in recent years. The most common method is the change of inertia weight, which can improve the global search ability of the algorithm [21,22,23]. Inertial weight w is a very important parameter in the PSO algorithm, which plays an important role in the global searching ability of the algorithm. It increases the diversity of the population and speeds up the convergence speed. Clerc proposed that the value of w was 0.729 [24,25]. Trelea suggested that w should be 0.6 [14,26]. The smaller w is, the stronger the local search ability of the algorithm is and, consequently, the faster the algorithm converges. However, it is easier to fall into local convergence, and its search ability in new regions is relatively deficient. The larger w is, the stronger the global search ability of the algorithm is, and the more diversity the population reflects. However, it is easy to miss the optimal point when w is large, thus leading to non-convergence. Therefore, it is of importance to adjust the value of w dynamically to obtain the right value at different times. At the earliest, Shi and Eberhart proposed adjusting the inertia weight w by the way of linear decline [27]. In their view, the minimum value of w is taken as the bottom line, and the value of w decreases linearly from the maximum to the minimum with the number of iterations. As the actual optimization process is nonlinear and complex, the linear decline way is not in line with the actual rules of optimization. It did not consider the situation of all the particles in the population and had no strong adjustment, so it is easy to fall into local optimum and cannot obtain better results.
Later, Shi and Eberhart introduced the random inertia weighting strategy, which changed the value of w by introducing a random function [27]. Some scholars put forward the nonlinear decreasing method to adjust the inertia weight, such as the decreasing inertia weight adjustment strategy in the form of a parabola by Chatterjee and Siarry [28]. Jiang et al. used the change of cosine function to realize the nonlinear decreasing strategy [24]. This nonlinear decreasing method improves the searching efficiency of the algorithm to a certain extent. It can overcome the premature convergence of the algorithm and reduce the probability of falling into local extreme value to a certain extent [29,30,31]. Lu et al. introduced the concept of personalized inertia weight. This method judges the current position of each particle by comparing its fitness with the average fitness of the population. Based on the state, the inertia weight of a particle with a better position is reduced, and the particle is focused on the development of the current region. For the particles in poor positions, the inertia weight of the particle is increased, causing the particle to focus on potential area exploration [32].
In the above algorithms, the values of w in the early stage are large, so they have a strong global search ability. However, with the decrease of w in the later stage, the algorithm easily falls into the local optimum, resulting in the loss of population diversity, and the convergence speed of the algorithm will also slow down. Therefore, in the later part of the algorithm, the w value will still become very small. In this way, the group does not have diversity, it is easy to fall into a local extreme value, and it does not have a strong global search ability. At the same time, the cosine function cannot be adjusted in time for specific situations. The flexibility of the equation is poor [33,34,35].
The application of SVM relies on training in corresponding data sets to build corresponding models. The rapid development of modern society leads to the explosive growth of all kinds of information and data [36]. The growth is not only in quantity but also in more and more diversified forms of data, such as text documents, pictures and gene sequences [37]. Faced with such massive data and complex forms of expression, it is impossible to carry out the practical application without filter and process. At the same time, the relationship and characteristics of massive data in modern society also have wider application space [38,39]. Therefore, data screening and dimension reduction of complex data are becoming more and more important, while the feature selection method is the key to achieving the above applications. Feature selection is a widely used data processing method, which selects the conforming features and excludes the non-conforming features according to the set evaluation criteria [40]. New data sets are obtained by the feature selection of the original data sets, which can improve the accuracy and efficiency of the algorithm in classification, regression, prediction and other tasks. It makes the models we generate more accurate and easier to understand.
The main methods of feature selection include wrapping, filtering and embedded methods. Most feature selection methods are derived and improved from these three methods. Some scholars use SVM classifier for feature selection. When redundant features or interference features exist, the performance of the classifier will be significantly reduced. In addition, forward feature selection and backward feature selection are commonly used in feature selection by classifier. Different feature selection methods should be applied to data sets with different characteristics.
In order to further improve the accuracy of SVM classification, we intend to solve it from two different perspectives. First, an efficient new multi-class feature selection method is proposed, which is simple and intuitive. When using the hybrid model of particle swarm optimization support vector machine (PSO-VM) to evaluate the proposed feature selection method, not only the accuracy is greatly improved but also the training time is shortened, and a good effect is achieved. Then, to address the drawback of the elementary PSO algorithm where it is easy to fall into local optimum, a method based on the dynamic change of the inertia weight is proposed to optimize the PSO, which can enhance the global searching ability of PSO and expand the diversity of the population. The improved PSO algorithm is used to optimize the parameters of SVM, which improves the performance of the model and enhances the generalization ability of the model. In summary, the main contributions of this paper are presented as follows:
  • A new feature selection method is proposed. In the formula, the numerator is the sum of the mean values of the variances between classes, and the denominator is the dispersion coefficient to measure the dispersion degree of each eigenvalue within the class. When the numerator is larger and the denominator is smaller, the gap between classes will be larger and the gap within each class will be smaller. The discretization coefficient can describe the differences between classes more accurately and have a better performance than other ways. The proposed feature selection method can improve the classification accuracy and shorten the training time of the classifier.
  • An improved method for the inertia factor in PSO is proposed, which dynamically changes the inertia factor. We use the logarithmic function and random number to improve the changing process. It not only ensures that the inertia weight can decrease nonlinearly but also meets the necessary conditions for its convergence. In addition, a random function is introduced to ensure that the algorithm has a strong search ability in the early stage, and it will not be premature in the later stage. This algorithm can improve the accuracy of the search.
  • The proposed feature selection method is combined with the proposed Dynamic Weighted Particle Swarm Optimization (DWPSO) algorithm to improve the classification accuracy of the SVM and shorten the whole experiment time.
This paper is organized as follows. Section 2 introduces the method of the feature selection. Section 3 makes a brief introduction to SVM, PSO and introduces the DWPSO-SVM model. Section 4 describes the experiments performed and the obtained results. Section 5 discusses the main conclusions and future work.

2. Feature Selection

In this section, the new feature selection method called FS-score is detailed. By using this method, the key features can be selected, and the redundant or irrelevant features can be eliminated so as to build a better representation of data and improve both the accuracy and computation efficiency of classification algorithms.
The feature selection method selects the most effective feature for classification. Multi-classification feature selection refers to the feature selection on the data sets with multiple classes. To solve the problem of low classification accuracy of multi-classes and multi-feature data sets, the method to improve the efficiency of the algorithm and shorten the running time is proposed. The feature selection method is described as follows:
Given data set Q R m , x k is the k-th sample in the data set, where k = 1 , 2 , 3 n ( n 2 ) , and n j is the number of samples of class j. Then, the FS-score value of the i-th feature in the data set is calculated below.
F i = j = 1 p ( x ¯ i j x i ¯ ) 2 q i
q i = j = 1 p k = 1 n j ( ( x k ) i j x ¯ i j ) 2 x j ¯
In Equations (1) and (2), x ¯ i j is the average characteristic value of the i-th feature in the j-th class. x i ¯ is the average characteristic value of the i-th feature in the whole data set. p is the total number of classes. ( x k ) i j is the characteristic value, where j, k and i refer to the j-th class, the k-th sample and the i-th feature, respectively.
The proposed feature selection method is based on the value of F i . In Equation (1), the numerator is the sum of the mean values of the variances between classes, and the denominator is the dispersion coefficient to measure the dispersion degree of each eigenvalue within the class. The discretization coefficient can be used to describe the differences between classes more accurately. When the numerator is larger and the denominator is smaller, the gap between classes will be bigger and the gap within classes will be smaller. This feature will play an important role in the classification, and the discrimination ability of the feature will be stronger. Therefore, when the F i value is larger, the probability of this feature being selected is higher.
The principle upheld by the feature selection method is that the larger the gap between categories and the smaller the gap within categories, the better the classification effect will be. In many other methods, numerators represent the approximate sum of distances between classes. If there is an extreme value, it is very easy to cause a too large or too small result, which will cause interference in the process of feature selection. In the FS-score method, the numerator is the sum of the mean values of various inter-class variances; it can avoid the interference caused by extreme values on feature selection. The denominator is the discrete coefficient to measure the dispersion degree of each feature value within the class. Using the discrete coefficient will make the description of the difference between classes more accurate and more intuitive.
The algorithm process will be elaborated in order, and the specific process is described in detail in Algorithm 1.
Algorithm 1 An effective feature extraction method
Input: The data set Q R m , the penalty parameters C, the kernel parameter γ , the number of iterations m.
Procedure:
  1:
Value preprocessing: using g = g min a max a min a to scale the eigenvalue. g is the scaled value, g is the original value of the feature, m a x a and m i n a are the upper and lower bounds of the original eigenvalue, respectively.
  2:
For the i t h sample, i = 1 , 2 , L . Calculating F i using Equations (1) and (2). According to the F i value of features, the features are divided into u groups in descending order ( u = 5 10 ).
  3:
Calculating the a c c u r a c y for each group.
Output:  a c c u r a c y R m

3. DWPSO-SVM Model

3.1. Support Vector Machines

The initial theory of SVM is based on the principle of maximizing the interval and minimizing the structural risk. For practical problems, there are a variety of SVM-derived algorithms, such as the linear SVM, the nonlinear SVM and the least squares SVM. For the traditional dichotomous linear SVM, the hyperplane is used to classify the linearly separable dichotomous data sets [12,41]. Given a dichotomous data set T { x i , y i } ( i = 1 , 2 n ) , where x i R n , y i { 1 , 1 } , for the dichotomous linearly separable data set in the R n space, there will be multiple linear classification hyperplanes with the expression ω T x + b = 0 . The convex quadratic programming model of the linear separable SVM is described as follows:
min ω , b ω 1 2 s . t . y i ω T x i + b 1 , i = 1 , 2 n
According to this constraint condition, the training sample points of the two classes are divided into both sides of the two hyperplanes parallel to the classification hyperplane [11,42]. Sometimes, some sample points of the two classes are too close. In this case, the hyperplane obtained according to Equation (3) cannot properly separate these sample points. To enhance the generalization ability and robustness of the model, the convex quadratic programming problem of the linear SVM is obtained by adding the relaxation variable ε i and the penalty parameter C.
min ω , b , ε 1 2 ω 2 + C i = 1 n ε i s . t . y i ω T x i + b + ε i 1 , i = 1 , 2 n ,
where ε i 0 , i = 1 , 2 n . Equation (4) is a minimization problem that is difficult to solve. It is difficult to find a hyperplane with a large margin while controlling the total amount of relaxation variables [43,44,45]. The penalty parameter C is used to balance the two terms. This model can be solved by maximizing the dual Lagrangian formula L D ( α ) as follows.
max L D α = i n α i i , j = 1 m α i α j y i y j x i x j 1 2 ,
where 0 α < C , i = 1 , 2 n , i n α i y i = 0 . ω and b can be solved directly. In practical problems, the samples of many data sets cannot be separated linearly. Hence, a nonlinear classifier is needed to separate the data sets through the nonlinear mapping. So, the SVM can solve the problem in a higher-dimensional space. We implement this process by using the following kernel function instead of the dot product.
max L D α = i n α i 1 2 i , j = 1 m α i α j y i y j K x i , x j ,
where 0 α < C , i = 1 , 2 n , i n α i y i = 0 . α is the Lagrange multiplier. In this paper, the kernel function is chosen to be the radial basis, while the PSO algorithm and the improved PSO algorithm are used to find the best penalty parameter C and the kernel parameter γ . The radial basis kernel function is described as follows.
K ( x i , x j ) = exp ( γ | | x i x j | | 2 )
Small samples, high latitude, nonlinearity, and local minimum points are all difficulties that the SVM model excels at handling when using optimization theory. It solves the problems of pattern categorization, dimensionality disaster, and over-learning at the same time, while it has great generalization, global optimal performance and learning ability. It has been extensively utilized and has effectively solved numerous pattern recognition challenges.
Many scholars developed strategies for improving and modifying the SVM model. Osuan et al. devised and implemented a decomposition technique for face detection in 1997. A sequential optimization approach was suggested by Platt in 1998. The purpose of the technique is to supplement the original method with some variables or functions [46]. Suykens et al., for example, presented the LS-SVM approach, which increased support vector sparsity and proposed a sparse approximation strategy for regression issues [47]. In addition, there are more algorithms, such as the dynamic weighted LS-SVM, Huber approximation algorithm and Do Lagrangian multiplier collaborative optimization algorithm. To improve the efficiency of optimization, S S. Keerth proposed the nearest point algorithm to improve the efficiency [48]. Mavroforakis et al. developed an intuitive application for understanding geometric optimization algorithms for the geometric framework of SVM classification issues in terms of geometric methods [49]. In terms of incremental algorithms, Ahmed presented the SVM incremental training method, which selected a small batch of traditional quadratic programming algorithms as increments, allowing the support vector and new samples to be blended for training [50].
Scholars have discovered that the SVM method offers numerous advantages as a result of ongoing in-depth research. It can ensure that the extreme value solution has been found to be the global best solution. This also proves that the SVM method is capable of generalizing unknown samples. Some SVM models and approaches have shown strong promotion ability in numerous application domains and have been widely employed as a result of these advantages. Face detection and recognition, handwritten numbers recognition, text classification, speech recognition, picture recognition, image classification, and the like are some of the most common applications in the field of pattern recognition. Several pattern recognition challenges have been solved satisfactorily.

3.2. DWPSO-SVM

The PSO algorithm was put forward by Dr. Eberhart and Dr. Kennedy in 1995. It is a population intelligent algorithm designed by imitating the study on the predation behavior of birds while observing the predation situation [51,52]. In the proposed method, DWPSO is used to find the two best parameters values of the SVM. Scholars have put forward improvement measures to the PSO algorithm from different perspectives; most of them improve the change way of inertia weight, which uses different types of calculation methods to slow down the speed of inertia weight reduction in the later period of the algorithm. However, these methods all have the limitations and shortcomings. In particular, although these algorithms delay the decreasing trend of the inertia factor of the PSO algorithm in the later period, the trend does not change. It is still likely to lead to the local extreme value. The DWPSO-SVM improves the problem that the inertia factor of the PSO algorithm easily falls into local convergence in the later stage of the algorithm, and it achieves a better effect. However, compared with the grid algorithm, the DWPSO-SVM still gives non-optimal parameter values with some probability. However, the efficiency of the time-consuming grid algorithm is poor, and its practical application effect is not good.
Different from the genetic algorithm, the DWPSO-SVM does not need to go through steps such as crossover, mutation and evolution, and it avoids the complex evolutionary operation. It uses mobile search in the whole area, and it can dynamically adjust its search strategy according to the current situation. As all the particles are searching and moving at the same time, it is an efficient parallel search strategy. Moreover, the algorithm is not sensitive to the size of the population, and the number of the population has no obvious effect on the effectiveness of the algorithm [24,27]. Here, we introduce the fundamental PSO algorithm, which has the following formula.
v i = v i + c 1 r a n d ( ) ( p b e s t i x i ) + c 2 r a n d ( ) ( g b e s t i x i )
x i = x i + v i
In Equations (8) and (9), i = 1 , 2 N , N denotes the total number of particles in this swarm. v i denotes the velocity of the particle, r a n d ( ) denotes a random number whose value is between (0, 1), x i denotes the current position of the particle in space, c 1 and c 2 are learning factors, and usually, a fixed value will be selected. The maximum value of v i is V m a x . If v i is greater than V m a x during iteration, then v i = V m a x . p b e s t i and g b e s t i , respectively, represent the best-known position of the particle itself and the best-known position of particle swarm globally [53,54].
Equations (8) and (9) are called elementary particle swarm equations. The standard particle swarm equation will be introduced below. The specific equation is expressed as follows.
v i = w v i + c 1 r a n d ( ) ( p b e s t i x i ) + c 2 r a n d ( ) ( g b e s t i x i )
Compared with Equation (8), Equation (10) is only modified in the first term. w is called the inertial factor, and its value is non-negative. The magnitude of its value will affect the motion of particles in space. When its value is large, the global optimization ability is strong, but the local optimization ability is weak conversely. When its value is small, the global optimization ability is weak, but the local optimization ability is strong. The introduction of w has greatly improved the performance of the PSO algorithm, and it can flexibly adjust the global and local search ability according to different actual situations. Equations (9) and (10) refer to the standard PSO algorithm together [55,56].
Next, we present the proposed PSO algorithm based on the improved inertial weight in detail. At the same time, the improved PSO algorithm is used to optimize the parameters of the SVM classifier. This section mainly includes the model elaboration and the algorithm.
A method is proposed about dynamically changing the inertia factor, which can adjust the variation trend of the equation according to the specific situation. The specific method is described as follows:
The natural logarithm is used to control the reduction of inertia weight. r a n d ( ) is a random generation function, which can increase the diversity of the population in the later stage of the algorithm to avoid premature maturity. The specific formula is presented in detail as follows.
w ( t ) = w min + ( w max w min 2 ) 1 ln ( e + ( k z ) 2 ) + ( w max w min ) 2 r a n d ( )
where the value of w max is 0.9 and w min is 0.4 [45], z = t / T , k is a random number from 0 to 1. In Equation (11), the second part is used to reduce the inertia weight gradually through the logarithmic function 1 ln ( e + ( k z ) 2 ) . k is the regulating factor, which can adjust the decreasing trend of inertia weight. The weight of inertia is large in the early period, and as z becomes larger, the first term of the equation becomes smaller. In this case, the third term of the equation adjusts the value of w ( t ) in time. The value will increase randomly to ensure the diversity of the population and to avoid premature convergence. Therefore, it can be seen from Equation (11) that this algorithm not only ensures that the inertia weight can decrease nonlinearly but also meets the necessary conditions for its convergence. In addition, r a n d ( ) (random function) is introduced to ensure that the algorithm has a strong search ability in the early stage, and it will not be premature in the later stage, so as to improve the accuracy of the search.
The DWPSO-SVM classifier is utilized to classify the data set after feature selection by using the proposed FS-score method. The kernel selected by the SVM classifier is RBF kernel. The improved particle swarm optimization algorithm is used to find the optimal penalty factor C and the parameter γ in the kernel function of SVM. We evaluate the DWPSO-SVM method according to the evaluation indexes. As the data sets are used after feature selection, there is no need to scale the eigenvalues of each sample in the data set. We divide the data sets into three independent parts: training set, monitoring set and test set. The training set is used to train the DWPSO-SVM classifier including the feature selection step. The monitoring set is used to optimize the model such as the number of features and other parameters in the model to avoid overfitting. The test set is used to test the predictive ability of the model independently. The whole algorithm is introduced detailed in Algorithm 2.
Algorithm 2 The DWPSO algorithm to optimize the parameters of the SVM
Input: the data set after feature selection Q R m , the range of the penalty parameter C, the range of kernel parameter γ , the number of iterations m.
Procedure:
  1:
Calculating w ( t ) using Equation (11) in the PSO algorithm.
  2:
Q is divided into the training set and the testing set, and it is trained on the DWPSO-SVM classifier.
  3:
For the i t h iteration, i = 1 , 2 , m . If the iteration number = m, testing begins.
  4:
Using Equations (12), (13) and (14) to calculate a c c u r a c y , T N R and T P R .
Output:  a c c u r a c y R m , T N R ,   T P R [ 0 ,   1 ]

4. Experimental Results and Discussions

4.1. Experiments Settings

All experiments in this paper are conducted on a personal computer with an Intel 8-generation i5 processor, 2.4 GHz, 8 GB of RAM, and Windows 10. The code runs on Matlab 2016 and is used to experiment with the Libsvm toolkit. All the data sets used in the proposed FS-score method are the standard UCI data sets, while the data sets used in the DWPSO-SVM method are the data sets after being feature selected by the FS-score method on the standard UCI data sets. The values of the two parameters of the RBF kernel function in SVM used in our experiment [57] are C in [0, 32,000] and γ in [0, 10]. The characteristics of the data sets, such as the numbers of features and instances, are shown in Table 1. According to the preliminary experimental results, the maximum number of iterations is set to 150, and all experiments converge within 150 generations. The initialization of agents used in the proposed method refers to [58]. Various evaluation criteria are used to evaluate the effectiveness of the proposed approach. The specific evaluation criteria will be described in detail. To verify the effectiveness of our proposed method, the method of averaging ten random trials is used. The original data set is divided into K parts. All of the parts except the K- t h part are trained. The K- t h part tests the classification effect and returns the value of the classification evaluation index being used. Here, K is set to 10.

4.2. Evaluation Criteria

In this subsection, the PSO-SVM classifier is used to demonstrate the effectiveness of our proposed feature selection method. The kernel function selected by the classifier is the RBF. The accuracy of classification is used as the criterion to evaluate the effectiveness of the proposed feature method. Its value is calculated according to the confusion matrix. Table 2 indicates the actual category of the samples as well as the predicted category, where 1 denotes positive and −1 denotes negative.
According to the confusion matrix above, the accuracy of the classifier is defined as Equation (12).
A c c u r a c y = T P + T N T P + F P + N P + T N
At the same time, the training time of the test samples is considered as a part of the evaluation criteria. When the accuracy difference is within the acceptable range, the fewer features selected, the shorter the training time, and the higher the efficiency of the algorithm.
To evaluate the effectiveness of our proposed DWPSO method, we take advantage of the hybrid matrix mentioned in Table 2 to evaluate our proposed model. Among all the evaluation indexes, accuracy is considered to be the most important metric to evaluate the classification effect. In addition to the accuracy, other corresponding indicators for different types of data sets are also adopted to accurately evaluate the effects. Sensitivity (TPR) and specificity (TNR) are two common evaluation indexes for dichotomous data sets. The specific Equations (13) and (14) are described in detail.
T P R = p r e c i s i o n = T P T P + F P
T N R = T N T N + F P
For binary data sets, the ROC curve and AUC value can also be used to evaluate the classification performance. The ROC curve is the receiver operating characteristic curve. Each point on the curve reflects the sensitivity to the same signal stimulus. The vertical axis of the curve is TPR, and the horizontal axis is FPR. ROC shows the tradeoff between sensitivity and specificity. The closer the curve is to the left and upper boundary of the coordinate axis, the better the classification effect of the classifier will be. The area contained under the ROC curve is the AUC value. Similarly, according to the definition of the ROC curve, the larger the AUC value is, the better the classification effect will be.

4.3. Results of Feature Selection with FS-Score

In this part of the experiment, 15 data sets from the standard UCI data set are used (Australia, CMC, Diabetes, Dnatest, Germen, Heart, Iris, Sonar, Vehicle, WDBC, Wine, Air, Farm Ads, Oclar, Arcene). First, FS-scores of the features are calculated, and the features are sorted from the largest to the smallest according to the FS-score value. Then, they are divided into four to ten groups based on the number of features. The first group has at least one feature and at most 15 ranked top by the FS-score. The actual number is determined based on the number of features in the data set. The second group, based on the first group, adds one or more features whose FS-score just stays behind the first group. For example, the data set Iris has a total of four features, and these features are grouped into four groups. The first group has a feature, the second group adds a feature to the first, and the third group adds a feature to the second. The data set Sonar has a total of sixty features, which are divided into six groups. The first set of features is ten, and each subsequent set of features adds ten to the previous set of features.The order of addition is from large to small by F i . The feature groups are formed in this way until the last feature group contains all the features. The classification results of some data sets are presented in Figure 1. Finally, the best feature group is determined according to the criteria of the highest classification accuracy and the least selected features. Table 3 shows the optimal number of features and the feature subsets selected for all data sets.
It can be seen from Figure 1 that the accuracy of all data sets fluctuates with the increase of the number of selected features. This means that many or few features do have a big impact on the accuracy of the data set. The few features make the model poorly learned. The excessive features will increase the training time and affect the efficiency. Moreover, the useless or redundant features will decrease the accuracy of the classifier. For the 14 selected data sets, the accuracy changes with the number of selected features, and the accuracy does not always increase with the number of features. Take the data set Sonar as an example: it has the highest accuracy at 40 features and decreases when the number of features continues to increase. As shown in Table 1 and Table 3, the number of features selected from some data sets (Iris, Vehicle, Diabetes, Australia) is close to that of the original data set when the accuracy reaches the highest. For example, the data set Diabetes has only one feature reduced compared to the original data set, while the data set Australia has only three features reduced compared to the original data set, although the number of features in these data sets is small. However, for some data sets (Farm Ads, Oclar, Arcene) with a large number of original features, the number of selected features at the time of reaching the highest accuracy is much lower than before after feature selection. For example, the data set Farm Ads decreases from 54,877 features to 1858 features, and the data set Oclar decreases from 3916 features to 435 features.
As the number of features decreases, the training time of the algorithm reduces significantly. As shown in Table 4, the training time of the data sets with a large number of features decreases significantly. For example, the training time of the Germen data set is reduced by nearly 15 min, and the training time of the CMC data set is reduced by nearly 7 min. Therefore, in the case of improved accuracy, the training time is greatly reduced, while the efficiency of the algorithm is greatly improved.
Accuracy is the most important evaluation index of the algorithm. It can be seen from Table 5 (optimal values are marked in bold) that the accuracy of most data sets has been greatly improved after feature selection. The accuracy is one to four percent higher. For example, the accuracy of data set Australia increases by 0.89%, and the accuracy of CMC increases by 4.26%. Some data sets have little or no improvement in accuracy as the data set itself has few features, and each of its features has a significant impact on the accuracy of classification. Therefore, the feature selection method reduces the number of features by only one or even no reduction, so its accuracy is improved relatively low.

4.4. Evaluation on DWPSO-SVM

In this part of the experiment, the DWPSO-SVM algorithm we proposed will be compared with the combination of the basic PSO algorithm and SVM, namely the PSO-SVM algorithm, and different evaluation indexes will be used for comparison. In addition to the PSO-SVM algorithm, the improved PSO-SVM algorithms and other types of algorithms related to SVM are also be compared. The UCI data sets listed in Table 1 are used for the experiment, and all features are selected by the FS-score method.
As shown in Table 6, experiments on 16 standard UCI data sets are conducted with feature selection in the improved PSO. The evaluation criteria we used are accuracy, TPR, and TNR. Ten experiments are performed for each data set to obtain the average value of all indicators. Among them, the values of TPR and TNR for data sets of three or more categories are expressed by N/A.
The results in Table 6 show that compared with the PSO algorithm, the DWPSO algorithm significantly improves the performance of the classifier on 16 data sets. Based on the accuracy, TPR, TNR, and other evaluation indexes, the proposed algorithm performs better than the basic PSO algorithm in most of the 16 data sets. For example, on the Vowel and Diabetes data sets, the accuracy of the indicators is 6.58% and 4.8% higher, respectively, than the algorithm before improved. For the Sonar data set, the TPR and TNR values for the DWPSO-SVM algorithm are 90% and 94.44%, respectively. Compared with the PSO-SVM algorithm, the results are respectively improved 5.48% and 1.11%. At the same time, its accuracy improves 4.77%. Although the accuracy is not high in some data sets, it also drops slightly for TPR and TNR. However, in general, compared with the PSO-SVM algorithm, the DWPSO-SVM algorithm has improved significantly in terms of classification accuracy. Meanwhile, the nonparametric statistical significance analysis is conducted to prove the strength of the proposed algorithm. As shown in Table 7, DWPSO-SVM is not significantly different from PSO-SVM at 0.05 level by the Kruskal–Wallis variance analysis.
In addition, other improved algorithms concerning the particle swarm inertia factor in recent years are also compared. It is combined with SVM in the same way as the DWPSO-SVM; other experimental conditions, parameters and configurations are the same. The DWPSO-SVM algorithm and other eight comparison algorithms are run on 12 data sets selected by the FS-score method, and its evaluation index is the accuracy of the classifier. The generation number and iteration number of all the compared algorithms have been listed in Table 8 (optimal values are marked in bold). As shown by the experimental results in Table 9 (optimal values are marked in bold), the accuracy of the proposed DWPSO-SVM algorithm is higher than that of other comparison algorithms in nine data sets. However, in the other three data sets, the classification accuracy is inferior to other comparison algorithms. For example, on the data sets Australia and Breast Cancer, the classification accuracy of the RAND-SVM method is the highest.
Table 10 (optimal values are marked in bold) shows the comparison between the DWPSO-SVM algorithm and other types of algorithms. Each algorithm adopts its default parameters and settings. The highest value of the classification accuracy on each data set has been shown in bold. It can be seen that in the nine data sets we adopted, the DWPSO-SVM algorithm has the highest accuracy.
Three data sets with different orders of magnitude have been selected for the memory experiment. For Iris, the data set with the least features, the memory occupied by the DWPSO-SVM method during training is 237 M due to the smallest number of samples. For Oclar, the data set with thousands of features, it takes up 2.32 GB of memory for training. For Farm Ads with about 50,000 features, it takes up 7.37 G of memory, which is almost the full memory. The time complexity of the algorithm is O ( n 2 ) , and the running time of the DWPSO-SVM algorithm on different data sets are is shown in Table 11. The spatial complexity of the algorithm is O ( n ) .
To evaluate the classifier’s ability to distinguish categories, the ROC curves of the data sets Australia and Breast Cancer are drawn. As shown in Figure 2 and Figure 3, the closer the curve is to the left and upper part of the coordinate axis, the better the classification ability of the classifier is. The larger the area of the curve is, the better the performance of the classifier. The area under the curve is the AUC value. We also calculate the AUC values of the Australian and Breast Cancer data sets under the DWPSO-SVM and PSO-SVM algorithms, which are illustrated in Figure 2 and Figure 3, respectively. From the perspective of the AUC value, the AUC value of the DWPSO-SVM algorithm is greater than that of the previous two data sets.
The iterative process of the comparable algorithms is similar to that of DWPSO-SVM. Figure 4 and Figure 5 show the generation process of the DWPSO-SVM, in which the generation comes to stabilization when it reaches about 100 on almost all data sets. Table 8 shows the number of generations when other comparable algorithms reach convergence. As with the DWPSO-SVM, the comparable algorithms converge in about 100 generations. Table 11 shows the running time of DWPSO-SVM on different data sets. The DWPSO-SVM has some real-time performance on the data sets with a small number of features.
In general, the performance of our DWPSO-SVM classifier based on the PSO algorithm is better than that of the original elementary PSO algorithm in terms of the accuracy, TPR, TNR, and other evaluation criteria. Compared with other types of PSO and other swarm intelligence algorithms, the DWPSO-SVM has better classification accuracy on most data sets.

5. Conclusions

In this paper, a new feature selection method is proposed, which can be used for the feature selection of the multi-class data sets. The PSO-SVM algorithm is used to test the proposed feature selection method on some UCI data sets. The experimental results demonstrate that the proposed method is effective on the data sets with a large number of features. Compared with the MVO-SVM method [2], the proposed method performs better on some data sets, such as the Hear and Ionosphere data sets, with an accuracy rate 2.73% and 2.56% higher than MVO-SVM, respectively. Similarly, compared with the GA-SVM method [3], in the terms of the two indicators TNR and TPR, the proposed method also performs better on most data sets. For example, DWPSO is 17% higher on TPR and 1.33% higher on TNR than GA-SVM on the data sets Diabetes and Ionosphere. The classification accuracy is significantly improved, and the training time is significantly reduced.
Using the proposed feature selection method, an improved PSO algorithm is proposed to optimize the SVM parameters and RBF kernel parameters. By improving the variation of the inertia factor in the PSO algorithm, the population diversity in the later stage of the PSO algorithm can be increased to avoid premature convergence. In addition, the inertia factor can be adjusted and changed according to the actual situation. The experimental results show that compared with the hybrid algorithm of elementary particle swarm and SVM, the DWPSO-SVM is improved on all data sets in terms of accuracy, TPR, and TNR. Compared with other PSO algorithms and the meta-heuristic algorithms, our proposed algorithm is also optimal on most data sets, which proved that our improved algorithm has a significant effect on improving the accuracy of SVM.
The instability is the limitation of PSO, while the grid algorithm has good stability but poor efficiency. How to combine the advantages of these two algorithms needs to be explored in the next step. In fact, there are some redundant and invalid samples when applying the proposed feature selection algorithm. How to efficiently screen the redundant samples in the data set is also a problem that we want to focus on in the future. The training time can be decreased, since the accuracy of the DWPSO-SVM has satisfied almost all of the situations. For example, some scholars have proposed using the meta-learning method to reduce the training time. However, the meta-learning methods always make the accuracy of the classifier decrease. Therefore, we will attempt to reduce the training time of the classifier on the premise of maintaining accuracy and stability.

Author Contributions

Conceptualization, J.W. and X.L.; Data curation, X.W.; Funding acquisition, J.W. and J.Y.; Methodology, J.W., X.L. and J.Y.; Software, X.W.; Supervision, X.L.; Validation, J.Y.; Visualization, J.W. and X.W.; Writing—original draft, J.W. and J.Y.; Writing—review and editing, J.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science and Technology Development Plan Project of Jilin Province under Grant 20220508033RC; and the National Key Research and Development Program of China under Grant 2020YFA0714100; and the Fundamental Research Funds for the Central Universities.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yu, S.; Li, X.; Zhang, X.; Wang, H. The OCS-SVM: An objective-cost-sensitive SVM with sample-based misclassification cost invariance. IEEE Access 2019, 7, 118931–118942. [Google Scholar] [CrossRef]
  2. Faris, H.; Hassonah, M.A.; Ala’M, A.Z.; Mirjalili, S.; Aljarah, I. A multi-verse optimizer approach for feature selection and optimizing SVM parameters based on a robust system architecture. Neural Comput. Appl. 2018, 30, 2355–2369. [Google Scholar] [CrossRef]
  3. Phan, A.V.; Le Nguyen, M.; Bui, L.T. Feature weighting and SVM parameters optimization based on genetic algorithms for classification problems. Appl. Intell. 2017, 46, 455–469. [Google Scholar] [CrossRef]
  4. Galmeanu, H.; Andonie, R. Weighted Incremental-Decremental Support Vector Machines for concept drift with shifting window. Neural Netw. Off. J. Int. Neural Netw. Soc. 2022, 152, 528–541. [Google Scholar] [CrossRef] [PubMed]
  5. Gallagher, K.; Sambridge, M. Genetic algorithms: A powerful tool for large-scale nonlinear optimization problems. Comput. Geosci. 1994, 20, 1229–1236. [Google Scholar] [CrossRef] [Green Version]
  6. Gomes, T.A.; Prudêncio, R.B.; Soares, C.; Rossi, A.L.; Carvalho, A. Combining meta-learning and search techniques to select parameters for support vector machines. Neurocomputing 2012, 75, 3–13. [Google Scholar] [CrossRef] [Green Version]
  7. Meng, Z.; Li, G.; Wang, X.; Sait, S.M.; Yildiz, A.R. A Comparative Study of Metaheuristic Algorithms for Reliability-Based Design Optimization Problems. Arch. Comput. Methods Eng. 2021, 28, 1853–1869. [Google Scholar] [CrossRef]
  8. Meng, Z.; Yildiz, A.R.; Mirjalili, S. Efficient decoupling-assisted evolutionary metaheuristic framework for expensive reliability-based design optimization problems. Expert Syst. Appl. 2022, 205, 117640. [Google Scholar] [CrossRef]
  9. Meng, Z.; Li, H.; Zeng, R.Q.; Mirjalili, S.; Yildiz, A.R. An efficient two-stage water cycle algorithm for complex reliability-based design optimization problems. Neural Comput. Appl. 2022, 34, 20993–21013. [Google Scholar] [CrossRef]
  10. Yu, X.; Liong, S.Y.; Babovic, V. EC-SVM approach for real-time hydrologic forecasting. J. Hydroinform. 2004, 6, 209–223. [Google Scholar] [CrossRef] [Green Version]
  11. Krishnakumar, K.; Goldberg, D.E. Control system optimization using genetic algorithms. J. Guid. Control Dyn. 1992, 15, 735–740. [Google Scholar] [CrossRef]
  12. Zhang, X.; Chen, X.; He, Z. An ACO-based algorithm for parameter optimization of support vector machines. Expert Syst. Appl. 2010, 37, 6618–6628. [Google Scholar] [CrossRef]
  13. Carrizosa, E.; Martin-Barragan, B.; Morales, D.R. A nested heuristic for parameter tuning in support vector machines. Comput. Oper. Res. 2014, 43, 328–334. [Google Scholar] [CrossRef] [Green Version]
  14. Le, T.; Tran, G.; Le, B. Feature learning for representing sparse networks based on random walks. Intell. Data Anal. 2020, 24, 873–886. [Google Scholar] [CrossRef]
  15. Rezaei, S.; Behnamian, J. Benders decomposition-based particle swarm optimization for competitive supply networks with a sustainable multi-agent platform and virtual alliances. Appl. Soft Comput. 2022, 114, 114. [Google Scholar] [CrossRef]
  16. Qiu, C.; Xiang, F. Feature selection using a set based discrete particle swarm optimization and a novel feature subset evaluation criterion. Intell. Data Anal. 2019, 23, 5–21. [Google Scholar] [CrossRef]
  17. Mafarja, M.M.; Eleyan, D.; Jaber, I.; Hammouri, A.; Mirjalili, S. Binary dragonfly algorithm for feature selection. In Proceedings of the 2017 International Conference on New Trends in Computing Sciences (ICTCS), Amman, Jordan, 11–13 October 2017; pp. 12–17. [Google Scholar] [CrossRef] [Green Version]
  18. Abdelhamid, A.A.; El-Kenawy, E.S.M.; Khodadadi, N.; Mirjalili, S.; Khafaga, D.S.; Alharbi, A.H.; Ibrahim, A.; Eid, M.M.; Saber, M. Classification of Monkeypox Images Based on Transfer Learning and the Al-Biruni Earth Radius Optimization Algorithm. Mathematics 2022, 10, 3614. [Google Scholar] [CrossRef]
  19. Khafaga, D.S.; El-Kenawy, E.; Karim, F.K.; Alshetewi, S.; Ibrahim, A.; Aabdelhamid, A. Optimized Weighted Ensemble Using Dipper Throated Optimization Algorithm in Metamaterial Antenna. Comput. Mater. Contin. 2022, 73, 5771–5788. [Google Scholar] [CrossRef]
  20. Rajasekhar, B.; Kamaraju, M.; Sumalatha, V. Cat swarm optimized ensemble technique for emotion recognition in speech signals. Concurr. Comput. Pract. Exp. 2022, 34, e7319. [Google Scholar] [CrossRef]
  21. Chen, Z.; Li, X.; Wang, M.; Yang, S. Domain sentiment dictionary construction and optimization based on multi-source information fusion. Intell. Data Anal. 2020, 24, 229–251. [Google Scholar] [CrossRef]
  22. El Moussawi, A.; Giacometti, A.; Labroche, N.; Soulet, A. MAPK-means: A clustering algorithm with quantitative preferences on attributes. Intell. Data Anal. 2020, 24, 459–489. [Google Scholar] [CrossRef]
  23. Hou, Y.; Hao, G.; Zhang, Y.; Gu, F.; Xu, W. A multi-objective discrete particle swarm optimization method for particle routing in distributed particle filters. Knowl.-Based Syst. 2022, 240, 108068. [Google Scholar] [CrossRef]
  24. Clerc, M. The swarm and the queen: Towards a deterministic and adaptive particle swarm optimization. In Proceedings of the 1999 Congress on Evolutionary Computation-CEC99 (Cat. No. 99TH8406), Washington, DC, USA, 6–9 July 1999; Volume 3, pp. 1951–1957. [Google Scholar] [CrossRef]
  25. Liu, L.; Wang, B.s.; Yu, B.; Zhong, Q.x. Automatic malware classification and new malware detection using machine learning. Front. Inf. Technol. Electron. Eng. 2017, 18, 1336–1347. [Google Scholar] [CrossRef]
  26. Trelea, I.C. The particle swarm optimization algorithm: Convergence analysis and parameter selection. Inf. Process. Lett. 2003, 85, 317–325. [Google Scholar] [CrossRef]
  27. Shi, Y.; Eberhart, R. A modified particle swarm optimizer. In Proceedings of the 1998 IEEE International Conference on Evolutionary Computation Proceedings. IEEE World Congress on Computational Intelligence (Cat. No. 98TH8360), Anchorage, AK, USA, 4–9 May 1998; pp. 69–73. [Google Scholar] [CrossRef]
  28. Chatterjee, A.; Siarry, P. Nonlinear inertia weight variation for dynamic adaptation in particle swarm optimization. Comput. Oper. Res. 2006, 33, 859–871. [Google Scholar] [CrossRef]
  29. Prastyo, P.H.; Hidayat, R.; Ardiyanto, I. Enhancing sentiment classification performance using hybrid Query Expansion Ranking and Binary Particle Swarm Optimization with Adaptive Inertia Weights. ICT Express 2022, 8, 189–197. [Google Scholar] [CrossRef]
  30. Kordestani, J.K.; Rezvanian, A.; Meybodi, M.R. An efficient oscillating inertia weight of particle swarm optimisation for tracking optima in dynamic environments. J. Exp. Theor. Artif. Intell. 2016, 28, 137–149. [Google Scholar] [CrossRef]
  31. Kalintha, W.; Ono, S.; Numao, M.; Fukui, K.i. Kernelized evolutionary distance metric learning for semi-supervised clustering. Intell. Data Anal. 2019, 23, 1271–1297. [Google Scholar] [CrossRef]
  32. Lu, S.X.; Liu, J.H.; Fu, X.W. A Fast Globally Convergent Particle Swarm Optimization for Defect Profile Inversion Using MFL Detector. Machines 2022, 10, 1091. [Google Scholar] [CrossRef]
  33. Ye, Z.; Ma, L.; Wang, M.; Chen, H.; Zhao, W. Texture image classification based on support vector machine and bat algorithm. In Proceedings of the 2015 IEEE 8th International Conference on Intelligent Data Acquisition and Advanced Computing Systems: Technology and Applications (IDAACS), Warsaw, Poland, 24–26 September 2015; Volume 1, pp. 309–314. [Google Scholar] [CrossRef]
  34. Sheta, A.F.; Ahmed, S.E.M.; Faris, H. A comparison between regression, artificial neural networks and support vector machines for predicting stock market index. Soft Comput. 2015, 7, 2. [Google Scholar] [CrossRef]
  35. Mirjalili, S.; Mirjalili, S.M.; Hatamlou, A. Multi-verse optimizer: A nature-inspired algorithm for global optimization. Neural Comput. Appl. 2016, 27, 495–513. [Google Scholar] [CrossRef]
  36. Chen, H.; Tai, X.; Wang, W. Multi-view subspace clustering with inter-cluster consistency and intra-cluster diversity among views. Appl. Intell. 2022, 52, 9239–9255. [Google Scholar] [CrossRef]
  37. Yu, S.; Li, X.; Feng, Y.; Zhang, X.; Chen, S. An instance-oriented performance measure for classification. Inf. Sci. 2021, 580, 598–619. [Google Scholar] [CrossRef]
  38. Yu, S.; Li, X.; Wang, H.; Zhang, X.; Chen, S. BIDI: A classification algorithm with instance difficulty invariance. Expert Syst. Appl. 2021, 165, 113920. [Google Scholar] [CrossRef]
  39. Yu, S.; Li, X.; Wang, H.; Zhang, X.; Chen, S. C_CART: An instance confidence-based decision tree algorithm for classification. Intell. Data Anal. 2021, 25, 929–948. [Google Scholar] [CrossRef]
  40. Syed, F.H.; Tahir, M.A.; Rafi, M.; Shahab, M.D. Feature selection for semi-supervised multi-target regression using genetic algorithm. Appl. Intell. 2021, 51, 8961–8984. [Google Scholar] [CrossRef]
  41. Zhang, H.w.; Xie, J.w.; Lu, W.l.; Sheng, C.; Zong, B.f. A scheduling method based on a hybrid genetic particle swarm algorithm for multifunction phased array radar. Front. Inf. Technol. Electron. Eng. 2017, 18, 1806–1816. [Google Scholar] [CrossRef]
  42. Wu, X.c.; Qin, G.h.; Sun, M.h.; Yu, H.; Xu, Q.y. Using improved particle swarm optimization to tune PID controllers in cooperative collision avoidance systems. Front. Inf. Technol. Electron. Eng. 2017, 18, 1385–1395. [Google Scholar] [CrossRef]
  43. He, C.H. Tabu search based resource allocation in radiological examination process execution. Front. Inf. Technol. Electron. Eng. 2018, 19, 446–458. [Google Scholar] [CrossRef]
  44. Chao, C.F.; Horng, M.H. The construction of support vector machine classifier using the firefly algorithm. Comput. Intell. Neurosci. 2015, 2015, 229–236. [Google Scholar] [CrossRef] [Green Version]
  45. Hu, Z.; Bao, Y.; Xiong, T. Comprehensive learning particle swarm optimization based memetic algorithm for model selection in short-term load forecasting using support vector regression. Appl. Soft Comput. 2014, 25, 15–25. [Google Scholar] [CrossRef]
  46. Osuna, E.; Freund, R.; Girosi, F. An improved training algorithm for support vector machines. In Proceedings of the Neural Networks for Signal Processing VII. Proceedings of the 1997 IEEE Signal Processing Society Workshop, Amelia Island, FL, USA, 24–26 September 1997; pp. 276–285. [Google Scholar]
  47. Suykens, J.A.; Vandewalle, J. Least squares support vector machine classifiers. Neural Process. Lett. 1999, 9, 293–300. [Google Scholar] [CrossRef]
  48. Keerthi, S.S.; Shevade, S.K.; Bhattacharyya, C.; Murthy, K.R. A fast iterative nearest point algorithm for support vector machine classifier design. IEEE Trans. Neural Netw. 2000, 11, 124–136. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. Mavroforakis, M.E.; Theodoridis, S. A geometric approach to support vector machine (SVM) classification. IEEE Trans. Neural Netw. 2006, 17, 671–682. [Google Scholar] [CrossRef]
  50. Bo, L.; Wang, L.; Jiao, L. Training hard-margin support vector machines using greedy stagewise algorithm. IEEE Trans. Neural Netw. 2008, 19, 1446–1455. [Google Scholar] [CrossRef] [Green Version]
  51. Chandrashekar, G.; Sahin, F. A survey on feature selection methods. Comput. Electr. Eng. 2014, 40, 16–28. [Google Scholar] [CrossRef]
  52. Firdaus, A.; Anuar, N.B.; Karim, A.; Ab Razak, M.F. Discovering optimal features using static analysis and a genetic search based method for Android malware detection. Front. Inf. Technol. Electron. Eng. 2018, 19, 712–736. [Google Scholar] [CrossRef]
  53. Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2003, 3, 1157–1182. [Google Scholar] [CrossRef]
  54. Liu, H.; Motoda, H. Feature Extraction, Construction and Selection: A Data Mining Perspective; Springer Science & Business Media: Berlin/Heidelberg, Germany, 1998; Volume 453, pp. 157–182. [Google Scholar] [CrossRef]
  55. Wang, G.G.; Gandomi, A.H.; Alavi, A.H.; Hao, G.S. Hybrid krill herd algorithm with differential evolution for global numerical optimization. Neural Comput. Appl. 2014, 25, 297–308. [Google Scholar] [CrossRef]
  56. Wang, G.G.; Gandomi, A.H.; Zhao, X.; Chu, H.C.E. Hybridizing harmony search algorithm with cuckoo search for global numerical optimization. Soft Comput. 2016, 20, 273–285. [Google Scholar] [CrossRef]
  57. Ding, X.J.; Liu, J.; Yang, F.; Cao, J. Random radial basis function kernel-based support vector machine. J. Frankl. Inst.-Eng. Appl. Math. 2021, 58, 10121–10140. [Google Scholar] [CrossRef]
  58. Gao, Y.L.; Duan, Y.H. A New Particle Swarm Optimization Algorithm with Random Inertia Weight and Evolution Strategy. In Proceedings of the International Conference on Computational Intelligence & Security Workshops, Harbin, China, 15–19 December 2007. [Google Scholar] [CrossRef]
Figure 1. Changes in the accuracy of each data set with the increase in the number of features.
Figure 1. Changes in the accuracy of each data set with the increase in the number of features.
Entropy 25 00531 g001aEntropy 25 00531 g001bEntropy 25 00531 g001c
Figure 2. ROC curve of the Australia data set, (a) the ROC curve before improvement, (b) the ROC curve after improvement.
Figure 2. ROC curve of the Australia data set, (a) the ROC curve before improvement, (b) the ROC curve after improvement.
Entropy 25 00531 g002
Figure 3. ROC curve of Breast cancer data set, (a) the ROC curve before improvement, (b) the ROC curve after improvement.
Figure 3. ROC curve of Breast cancer data set, (a) the ROC curve before improvement, (b) the ROC curve after improvement.
Entropy 25 00531 g003
Figure 4. Running process of the DWPSO-SVM on 11 UCI data sets.
Figure 4. Running process of the DWPSO-SVM on 11 UCI data sets.
Entropy 25 00531 g004
Figure 5. Running process of the DWPSO-SVM on 3 high-dimensional data sets.
Figure 5. Running process of the DWPSO-SVM on 3 high-dimensional data sets.
Entropy 25 00531 g005
Table 1. The data sets from the UCI repository.
Table 1. The data sets from the UCI repository.
Data SetFeaturesInstancesClassesFeature CharacteristicsData Set Characteristics
Australia146902Categorical, RealMultivariate, Time-Series
Breast cancer92772CategoricalMultivariate
CMC914733Categorical, IntegerMultivariate
Diabetes97682Categorical, IntegerMultivariate, Time-Series
Germen2410004Integer, RealMultivariate
Heart142702Categorical, Integer, RealMultivariate
Ionosphere343512Integer, RealMultivariate
Iris41502RealMultivariate
Dnatest18011863IntegerMultivariate, Time-Series
Sonar602082IntegerMultivariate, Text
Vehicle188464RealMultivariate, Integer
Vowel38716N/AImage
Wdbc305692RealMultivariate, Data-Generator
Wine131783Integer, RealMultivariate
Air643593RealMultivariate, Time-Series
Farm Ads54,87771432N/AText
Oclar391639165IntegerText
Arcene10,0009002RealMultivariate
Table 2. The confusion matrix.
Table 2. The confusion matrix.
Actual Class
1−1
Predicted class1True Positive (TP)False Negative (NP)
−1False Positive (FP)True Negative (TN)
Table 3. The optimal feature subset.
Table 3. The optimal feature subset.
Data SetNo. of Selected FeaturesSelected Features
Australia112, 3, 4, 5, 6, 7, 8, 10, 12, 13, 14
CMC51, 2, 3, 4, 8
Diabetes71, 2, 3, 4, 5, 6, 8
Dnatest285, 18, 44, 47, 50, 68, 72, 82, 83, 85, 90, 93, 97,
99, 100, 105, 111, 120, 123, 126, 129, 150, 156, 162, 171,
174, 180
Germen151, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 17
Heart121, 2, 3, 4, 5, 7, 8, 9, 10, 11, 12, 13
Iris41, 2, 3, 4
Sonar408, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20,
21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36,
37, 38, 39, 41, 42, 43, 44, 45, 46, 47, 48
Vehicle101, 2, 3, 4, 7, 8, 9, 10, 11, 12, 13, 14, 16
Wdbc251, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 19, 21,
22, 23, 24, 25, 26, 27, 28, 29, 30
Wine81, 3, 4, 5, 7, 10, 12, 13
Air571, 2, 3, 4, 5, 6, 7, 15, 16, 17, 18, 19, 20, 21, 22,
23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37,
38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52,
53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64
Farm Ads1858N/A
Oclar435N/A
Arcene175N/A
Table 4. The comparison of the training time before and after feature selection.
Table 4. The comparison of the training time before and after feature selection.
Data SetTraining Time (s)
Full FeaturesSelected FeaturesPercentage of Improvement (%)
Australia32124431.56
CMC138298740.02
Diabetes3052953.38
Dnatest4156497836.28
Germen2245132269.82
Heart105987.14
Iris440.00
Sonar624440.91
Vehicle78261028.20
Wdbc47843011.16
Wine544617.39
Farm Ads103,47887581081.52
Oclar86343287162.67
Arcene64381265408.93
Table 5. The accuracy of PSO-SVM before and after feature selection.
Table 5. The accuracy of PSO-SVM before and after feature selection.
Data Set Accuracy (%)
Full FeaturesSelected Features
Australia83.56 ± 1.7284.45 ± 1.81
CMC49.59 ± 2.1553.85 ± 4.55
Diabetes75.08 ± 3.0777.54 ± 4.8
Dnatest54.92 ± 0.7690.91 ± 3.13
Germen49.50 ± 3.1053 ± 2.43
Heart80.57 ± 2.1583.57 ± 2.87
Iris95.4 ± 3.2797 ± 3.33
Sonar85.36 ± 4.2889.29 ± 4.25
Vehicle71.83 ± 2.9473.63 ± 4.13
Wdbc96.81 ± 2.4498.55 ± 2.83
Wine98.42 ± 2.22100 ± 0
Farm Ads47.48 ± 2.3277.78 ± 5.43
Oclar62.13 ± 7.1874.74 ± 4.23
Arcene67.37 ± 7.3481.54 ± 5.86
Table 6. The comparison between the DWPSO-SVM (improved algorithm) and the PSO-SVM.
Table 6. The comparison between the DWPSO-SVM (improved algorithm) and the PSO-SVM.
Data SetDWPSO-SVMPSO-SVM
TPR (%)TNR (%)Accuracy (%)TPR (%)TNR (%)Accuracy (%)
Australia88.8991.1187.41 ± 3.7597.1487.2784.45 ± 1.81
Breast cancer21.0594.8374.89 ± 4.492592.9871.43 ± 6.47
Diabetes8087.1879.66 ± 4.5656.7687.6574.86 ± 5.44
GermenN/AN/A57.00 ± 5.29N/AN/A57.00 ± 3.43
Heart88.8976.4785.71 ± 2.8683.7875.7682.86 ± 3.78
Ionosphere90.4810093.46 ± 2.268510091.50 ± 2.26
IrisN/AN/A98.67 ± 1.15N/AN/A96.00 ± 2.00
Sonar9094.4491.67 ± 2.0684.6293.3386.90 ± 2.10
VehicleN/AN/A78.08 ± 4.31N/AN/A75.11 ± 5.14
VowelN/AN/A84.51 ± 3.45N/AN/A77.93 ± 5.33
WdbcN/AN/A99.03 ± 1.45N/AN/A98.07 ± 0.84
WineN/AN/A100 ± 0N/AN/A96.86 ± 0.12
CMCN/AN/A54.38 ± 1.18N/AN/A53.48 ± 1.94
Farm Ads83.6592.7387.33 ± 3.4277.8984.4679 ± 4.21
OclarN/AN/A81.46 ± 3.12N/AN/A77.35 ± 2.45
Arcene86.4389.6789.65 ± 4.6583.8385.9483.75
Table 7. Kruskal–Wallis variance analysis.
Table 7. Kruskal–Wallis variance analysis.
Descriptive statistics
NMinQ1MedianQ3Max
DWPSO-SVM1354.3876.48585.7196.065100
PSO-SVM1353.4873.14582.8693.7598.07
Rank
NMean rankSum rank
DWPSO-SVM1314.88462193.5
PSO-SVM1312.11538157.5
Test statistics
Chi-SquareDFP > Chi-Square
0.8523610.35588
Table 8. Number of generations when the algorithm reaches convergence.
Table 8. Number of generations when the algorithm reaches convergence.
Data SetDWPSO-SVMRAND-SVMSWISPSO-SVMALGPSO-SVMAIWPSO-SVMIIWPSO-SVMIDWPSO-SVMDSOPSO-SVMPSO-A
Australia93 ± 2.3190 ± 3.21103 ± 2.4393 ± 1.3292 ± 2.33114 ± 3.42115 ± 3.38117 ± 3.12110 ± 2.12
Breast cancer77 ± 1.5467 ± 3.4586 ± 3.21102 ± 4.33130 ± 2.63127 ± 2.39132 ± 2.93140 ± 2.45128 ± 3.31
Heart87 ± 2.5397 ± 3.32102 ± 3.7380 ± 2.67100 ± 3.5694 ± 1.5493 ± 1.5688 ± 2.15100 ± 3.21
Ionosphere98 ± 2.1292 ± 2.13108 ± 3.11100 ± 2.4580 ± 2.1594 ± 3.21102 ± 3.1894 ± 2.1598 ± 2.16
Iris96 ± 2.41100 ± 1.12118 ± 2.54110 ± 1.15103 ± 2.12111 ± 2.15115 ± 3.12110 ± 2.15102 ± 3.17
Sonar89 ± 3.12110 ± 1.2795 ± 3.1295 ± 2.17114 ± 3.22109 ± 1.2492 ± 2.12107 ± 2.87100 ± 2.65
Vehicle79 ± 2.1290 ± 2.7485 ± 3.5290 ± 2.1584 ± 1.2882 ± 3.8586 ± 2.1685 ± 1.6783 ± 1.42
Vowel84 ± 1.6797 ± 2.1290 ± 1.43100 ± 3.1690 ± 3.1295 ± 2.1498 ± 1.43104 ± 3.14112 ± 3.16
Wdbc97 ± 2.15108 ± 4.12100 ± 1.41110 ± 2.15113 ± 3.15120 ± 1.34118 ± 2.14112 ± 1.45117 ± 2.65
Wine100 ± 0.87120 ± 2.67139 ± 3.56108 ± 1.44115 ± 1.78112 ± 2.45136 ± 3.12114 ± 1.54115 ± 1.34
Farm Ads87.33 ± 3.4277.43 ± 3.2381.24 ± 2.3463.34 ± 4.5377.62 ± 4.2963.45 ± 4.2371.51 ± 3.6283.56 ± 3.6761.26 ± 4.18
Oclar81.46 ± 3.1282.34 ± 4.3479.23 ± 3.2453.34 ± 3469.34 ± 3.6152.45 ± 5.2369.23 ± 6.1675.57 ± 3.1752.77 ± 3.67
Arcene89.65 ± 4.6581.34 ± 4583.56 ± 5.2366.56 ± 2578.21 ± 1.6671.23 ± 3.3585. ± 4.3287.23 ± 5.1967.72 ± 3.82
Table 9. The accuracy comparison of the DWPSO-SVM algorithm and other types of the PSO-SVM.
Table 9. The accuracy comparison of the DWPSO-SVM algorithm and other types of the PSO-SVM.
Data SetDWPSO-SVMRAND-SVMSWISPSO-SVMALGPSO-SVMAIWPSO-SVMIIWPSO-SVMIDWPSO-SVMDSOPSO-SVMPSO-A
Australia87.41 ± 3.5789.26 ± 2.2382.59 ± 6.385.56 ± 3.889.26 ± 3.0183.33 ± 0.7483.33 ± 2.5482.59 ± 1.4284.07 ± 3.06
Breast cancer74.89 ± 4.4989.26 ± 2.8173.05 ± 2.0870.56 ± 4.8667.97 ± 1.8267.97 ± 5.6674.19 ± 3.5373.76 ± 5.1574.33 ± 4.51
Germen57.00 ± 5.2953.5 ± 4.3547.00 ± 2.4546.00 ± 5.0849.50 ± 8.1749.00 ± 4.7250.00 ± 3.8951.5 ± 4.3557.00 ± 3.87
Heart85.71 ± 2.8685.67 ± 3.3078.10 ± 3.6077.14 ± 5.1582.38 ± 4.5984.29 ± 5.1583.33 ± 3.6084.29 ± 1.4382.86 ± 1.43
Ionosphere93.46 ± 2.2695.42 ± 1.1390.20 ± 1.9696.07 ± 3.4096.73 ± 1.1395.42 ± 3.0092.16 ± 5.1995.42 ± 3.0094.12 ± 1.96
Iris98.67 ± 1.1596.67 ± 1.1593.33 ± 4.6294.67 ± 1.1596.67 ± 1.1595.33 ± 0.9796.00 ± 2.3195.33 ± 1.0096.00 ± 1.33
Sonar91.67 ± 2.0685.71 ± 7.1488.10 ± 5.4588.10 ± 7.4385.71 ± 3.5685.71 ± 7.1489.28 ± 3.5784.52 ± 5.4686.90 ± 14.4
Vehicle78.08 ± 4.3171.49 ± 3.7774.32 ± 0.6871.69 ± 3.3874.66 ± 4.1771.58 ± 7.8872.60 ± 4.1771.46 ± 1.4376.25 ± 3.09
Vowel84.51 ± 3.4576.06 ± 5.6378.87 ± 6.1476.06 ± 2.4482.16 ± 2.9380.28 ± 2.8276.69 ± 4.3077.46 ± 6.1475.12 ± 1.63
Wdbc99.03 ± 1.4597.10 ± 2.5198.06 ± 1.6797.59 ± 0.8497.10 ± 1.4596.14 ± 2.2196.62 ± 3.0297.10 ± 1.4596.45 ± 0.84
Wine100 ± 098.25 ± 1.5297.37 ± 2.6399.12 ± 1.5299.12 ± 1.5299.12 ± 1.3697.37 ± 2.5199.12 ± 1.5299.11 ± 1.52
CMC54.38 ± 1.1853.15 ± 1.2952.81 ± 4.1752.38 ± 1.1250.92 ± 2.6051.10 ± 0.6849.74 ± 1.2749.45 ± 1.7350.74 ± 4.94
Table 10. The accuracy comparison between the DWPSO-SVM algorithm and other types of algorithms.
Table 10. The accuracy comparison between the DWPSO-SVM algorithm and other types of algorithms.
Data SetDWPSO-SVMGA-SVMDAGOASCASSAWOABAT
Australia87.7468.668.666.6771.0271.0167.4567.45
Diabetes79.6671.4964.4771.9378.9560.0979.8270.18
Germen578078.677279.6773.677978.33
Heart85.7174.2166.6756.7978.3175.3167.875.31
Iris98.6795.5595.5597.7895.5693.3395.56100
Sonar86.98083.3386.3383.3378.3386.3378.33
Vehicle78.0880.9579.7643.2579.7682.5479.3784.13
Vowel84.5198.7210098.7299.3610099.3698.72
Wine10062.7472.5576.4784.3186.2762.7594.12
Table 11. The running time of DWPSO-SVM on different data sets.
Table 11. The running time of DWPSO-SVM on different data sets.
Data SetDWPSO-SVM Running Time (s)
Austtalia266 ± 12
CMC1122 ± 33
Diabetes324 ± 23
Dnatest505 ± 45
Germen1445 ± 126
Heart100 ± 5
Iris4 ± 1
Sonar36 ± 3
Vehicle593 ± 44
Wdbc465 ± 14
Wine56 ± 8
Farm Ads8426 ± 445
Oclar3354 ± 321
Arcene2478 ± 435
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, J.; Wang, X.; Li, X.; Yi, J. A Hybrid Particle Swarm Optimization Algorithm with Dynamic Adjustment of Inertia Weight Based on a New Feature Selection Method to Optimize SVM Parameters. Entropy 2023, 25, 531. https://doi.org/10.3390/e25030531

AMA Style

Wang J, Wang X, Li X, Yi J. A Hybrid Particle Swarm Optimization Algorithm with Dynamic Adjustment of Inertia Weight Based on a New Feature Selection Method to Optimize SVM Parameters. Entropy. 2023; 25(3):531. https://doi.org/10.3390/e25030531

Chicago/Turabian Style

Wang, Jing, Xingyi Wang, Xiongfei Li, and Jiacong Yi. 2023. "A Hybrid Particle Swarm Optimization Algorithm with Dynamic Adjustment of Inertia Weight Based on a New Feature Selection Method to Optimize SVM Parameters" Entropy 25, no. 3: 531. https://doi.org/10.3390/e25030531

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop