Next Article in Journal
ADHD-AID: Aiding Tool for Detecting Children’s Attention Deficit Hyperactivity Disorder via EEG-Based Multi-Resolution Analysis and Feature Selection
Next Article in Special Issue
A Random Particle Swarm Optimization Based on Cosine Similarity for Global Optimization and Classification Problems
Previous Article in Journal
A Novel Artificial Electric Field Algorithm for Solving Global Optimization and Real-World Engineering Problems
Previous Article in Special Issue
Credit and Loan Approval Classification Using a Bio-Inspired Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

BGOA-TVG: Binary Grasshopper Optimization Algorithm with Time-Varying Gaussian Transfer Functions for Feature Selection

1
College of Artificial Intelligence, Guangxi Minzu University, Nanning 530006, China
2
Guangxi Key Laboratories of Hybrid Computation and IC Design Analysis, Nanning 530006, China
3
Faculty of Information Science and Technology, Universiti Kebangsaan Malaysia, Bangi 43600, Selangor, Malaysia
*
Author to whom correspondence should be addressed.
Biomimetics 2024, 9(3), 187; https://doi.org/10.3390/biomimetics9030187
Submission received: 1 February 2024 / Revised: 9 March 2024 / Accepted: 11 March 2024 / Published: 20 March 2024
(This article belongs to the Special Issue Nature-Inspired Metaheuristic Optimization Algorithms 2024)

Abstract

:
Feature selection aims to select crucial features to improve classification accuracy in machine learning and data mining. In this paper, a new binary grasshopper optimization algorithm using time-varying Gaussian transfer functions (BGOA-TVG) is proposed for feature selection. Compared with the traditional S-shaped and V-shaped transfer functions, the proposed Gaussian time-varying transfer functions have the characteristics of a fast convergence speed and a strong global search capability to convert a continuous search space to a binary one. The BGOA-TVG is tested and compared to S-shaped and V-shaped binary grasshopper optimization algorithms and five state-of-the-art swarm intelligence algorithms for feature selection. The experimental results show that the BGOA-TVG has better performance in UCI, DEAP, and EPILEPSY datasets for feature selection.

1. Introduction

Researchers from all over the world have been paying more and more attention to how to handle massive datasets in recent years, thanks to data exploration. These databases can contain duplicate and pointless features. Thus, in order to solve this issue and increase the effectiveness of both supervised and unsupervised learning algorithms, feature selection is crucial [1,2,3]. It is obviously a challenging task to find the best subset of features in high-dimensional feature datasets because there are 2 N 1 feature subsets in a dataset with N features. Traditional mathematical techniques cannot produce the desired outcome in an acceptable amount of time; hence, meta-heuristic algorithms are frequently used for feature selection issues.
Metaheuristic algorithms have been successfully used to solve complex engineering and science computation problems, such as function optimization [4,5,6], engineering optimization [7,8,9], and feature selection problems [10]. Researchers have proposed binary metaheuristic algorithms or improved versions for feature selection, such as binary swarm optimization (BPSO) [11], the binary artificial bee colony (BABC) [12], the binary gravitational search algorithm (BGSA) [13], binary grey wolf optimization (BGWO) [14], the binary salp swarm algorithm (BSSA) [15], the binary bat algorithm (BBA) [16], the binary whale optimization algorithm (BWOA) [17], binary spotted hyena optimization (BSHO) [18], binary emperor penguin optimization (BEPO) [19], binary Harris hawks optimization (BHHO) [20], binary equilibrium optimization (BEO) [21], binary atom search optimization (BASO) [22], the binary dragonfly algorithm (BDA) [23], the binary jaya algorithm (BJA) [24], binary coronavirus herd immunity optimization (BCHIO) [25], the binary butterfly optimization algorithm (BBOA) [26], binary black widow optimization (BBWO) [27], the binary slime mould algorithm (BSMA) [28], binary golden eagle optimization (BGEO) [29], and so on.
To find the best solution to complex optimization problems, metaheuristics are employed. A system of solutions that develops over many iterations utilizing a set of rules or mathematical equations can be used by several agents to facilitate the search process. These iterations continue until the result satisfies a set of predetermined requirements. This final solution (the near-optimal solution) is called the optimal solution, and the system is considered to have reached a state of convergence [30].
In contrast to exact methods that find optimal solutions but require a long computational time, heuristic methods find near-optimal solutions quite quickly [31]. However, most of these methods are problem-specific. As the word “meta” in metaheuristic methods indicates, metaheuristics are one level higher than heuristics. Metaheuristics have been very successful because they have the potential to provide solutions at an acceptable computational cost. By mixing good heuristics with classical metaheuristics, very good solutions can be obtained for many real-world problems.
Binary particle swarm optimization (BPSO) and its variants have been widely used in the FS problem. In 2020, the authors of ref. [32] proposed self-adaptive PSO with a local search strategy to find less-correlated feature subsets. The authors of ref. [33] proposed an improved version of the SSA algorithm called the ISSA to solve the FS problem. Furthermore, a binary chaotic horse herd optimization algorithm for feature selection (BCHOAFS) was proposed by Esin Ayşe Zaimoğlu [34]. The authors of ref. [35] used four families of transfer functions in the binary AOA (BAOA) to test 10 low-dimensional and 10 high-dimensional datasets. In feature selection problems, several wrapping-based algorithms using binary meta-heuristic algorithms have been proposed. The S-shaped and V-shaped transfer functions are mostly used, such as modified binary SSA (MBSSA) [36].
In recent years, to increase the efficiency of transfer functions, the time-varying S-shaped and V-shaped transfer functions have been proposed and applied to the binary dragonfly algorithm (BDA) [37] and BMPA [38]. The results showed that the time-varying S-shaped transfer function has a better performance than the V-shaped one. Besides the single-objective algorithm, the muti-objective algorithm also plays an important role in the feature selection problem. Multi-objective whale algorithm optimization (WOA) [39] was proposed for data classification for which filter and wrapper fitness functions were simultaneously optimized. Another paper studies a multi-label feature selection algorithm using improved multi-objective particle swarm optimization (PSO) [40], with the purpose of searching for a Pareto set of non-dominated solutions (feature subsets). Liu proposes a novel feature selection method utilizing a filtered and supported sequential forward search technique called multi-objective ant colony optimization (MOACO) in the context of support vector machines (SVMs) [41] to solve the feature selection problem. A new method known as improved multi-objective salp swarm algorithm (IMOSSA) is tested for a feature selection task [42]. Also, other kinds of transfer functions are used in feature selection problems. An automata-based improved BEO [43] (AIEOU) used a U-shaped transfer function to select the best subset of features. This algorithm applied both learning-based automata and adaptive β-hill climbing (AβHC) to find the best parameters to form a better equilibrium pool. Many binary metaheuristic algorithms have been introduced to select the best subset of features. However, due to the importance of feature selection in many fields, it is necessary to design algorithms that can obtain a higher accuracy with a smaller subset of features.
An important step in the feature selection problem is mapping a continuous space to binary ones, and the transfer function plays a significant role in this process. Moreover, using transfer functions is one of the easiest ways to convert an algorithm from continuous to binary without modifying its structure. The shapes of transfer functions are classified into two families: S-shaped and V-shaped [44]. The S-shaped transfer function is a transformation function that increases monotonically within the [0, 1] interval, and its main attribute is that it is asymmetric and has a single fixed value throughout its entire area. The V-shaped transformation function is symmetric and has two numbers with equal values in the [0, 1] interval, which enhances the diversity of the population in the region within the range.
A common drawback of common transfer functions used in binary algorithms is that they do not explore and develop evolutionarily during the search for the optimal solution; that is, the process to obtain a solution involves changing the probability of the parameter values in a nonadaptive manner. Thus, they have poor exploration or exploitation and are static functions which cannot change as time goes on.
In this paper, a powerful transfer function is proposed to convert continuous search spaces to binary ones. The transfer function is dynamic during this process, which can enhance the search ability of the BGOA in the exploration phase. Then, the transfer function gradually changes while the proposed algorithm switches from exploration to exploitation and finally reaches good result in the end, the K-nearest neighbors (KNN) algorithm is applied to classify it.
The main contributions for this paper can be summarized as follows:
  • A time-varying Gaussian transfer function is introduced.
  • A new binary grasshopper optimization algorithm based on time-varying Gaussian transfer functions (BGOA-TVG) is proposed.
  • The BGOA-TVG achieves a balance between its exploration and exploitation capabilities and improves the convergence speed of the algorithm.
  • The BGOA-TVG can effectively deal with high-dimensional feature selection problems.
  • Compared with proposed binary metaheuristic optimization algorithms in recent years, the excellent performance of the BGOA-TVG is verified.
The rest of the paper is organized as follows: Section 2 presents a brief introduction to the feature selection problem. In Section 3, the basic grasshopper optimization algorithm is discussed. The enhanced transfer function is presented in Section 4. Section 5 shows the results of the test. In Section 6, the proposed method is demonstrated within the EEG analysis field. Finally, Section 7 concludes the paper and suggests some directions for future studies.

2. Feature Selection Problem

The feature selection problem is an NP-hard optimization problem [45], in which as the features of a dataset increase, the search space of the problem exponentially grows. It is a useful way to find a relevant subset of fewer features from an initial dataset to reduce dimensions and training times [46]. However, traditional mathematic methods cannot solve high-dimensional feature selection problems in a reasonable time, and according to tests, metaheuristic algorithms are better at finding subsets of features [47,48,49,50,51,52,53,54]. There are three selection strategies in feature selection: wrapper-based, filter-based, and hybrid filter–wrapper-based methods [55]. The precision of the learning algorithm in the wrapper-based strategy creates the optimum subset. The chosen subset in the filter-based method is unrelated to the learning process. These methods are combined in the hybrid approach. The wrapper-based method outperforms the others in terms of accuracy, but it requires more CPU resources and a longer testing time. The authors of [56,57,58] provided a new method to extract optimal feature subset to enhance accuracy of the calculation. Others have proposed correlation feature selection [59] and in-depth analyses on the usage of searching methods like the best-first, greedy step-wise, genetic, linear forward selection, and rank searches [60].
A feature selection module is applied prior to the classification method to optimize efficiency and precision by eliminating irrelevant features and to reduce the time complexity to find the classification to which a document belongs [61].
The most important thing in feature extraction is to extract subsets and determine whether to select an element in the set according to the accuracy rate. In the binary algorithm, individuals traverse a set and display whether to select an element using 0 s and 1 s. In Equation (1), N is the number of the dataset and S F is a subset selected by the algorithms; in the binary algorithm, it is selected by the value of an individual.
Feature selection is a multi-objective optimization problem for which we aim to minimize the subset of the selected features and maximize the classification accuracy, which is described as a fitness function as follows:
F i t n e s s = α × E r r + β × S F N
where E r r is the resulting classification error. S F is the number of selected features of the subset, and N is the total number of features of the dataset. S F N is the feature selection ratio of the subset to the total set. α and β are parameters in the interval of [0, 1] and α = 1 β .

3. Grasshopper Optimization Algorithm (GOA)

The grasshopper optimization algorithm is a population-based swarm intelligence algorithm introduced by Mirjalili et al. in 2017 [62], which models the behaviour of grasshopper swarms in nature. There are two essential phases in this algorithm: the exploration and exploitation of the search space. Through social interactions during the food search process, the swarm of grasshoppers changes between the phases. The swarm moves slowly and goes a small distance in the larval stage. In contrast, the swarm moves quickly and goes a large distance in adulthood.
There are three evolutionary operators in the position-updating process of individuals in swarms [62]: the social interaction operator, S i in Equation (2); the gravity force operator, G i in Equation (2); and the wind advection operator, A i in Equation (2). The movement of individuals in the swarm is describes as follows:
X i = S i + G i + A i
where X i defines the position of the i t h grasshopper.
S i = j = 1 j i N S | X j X i | X j X i d i j
where N is the number of grasshoppers in the swarm, d i j represents the distance between the i t h and the j t h grasshopper, S is a function that defines the strength of the social forces and is calculated as shown in Equation (4), and X j X i d i j is the unit vector from the i t h grasshopper to the j t h .
S r = f e r l e r
where f and l are two constants that indicate the intensity of attraction and the attraction length scale, respectively, and r is a real value.
G i in Equation (2) is calculated as shown in Equation (5) below:
G i = g × e g
where g is the gravitational constant, and e g shows a unity vector towards the center of the earth. The effect of an individual’s flight to overcome gravity is represented by the symbol preceding it.
A i in Equation (2) is calculated as shown in Equation (6) below:
A i = u × e w
where u is a constant drift, and e w is a unity vector in the direction of the wind.
Equation (2) can be expanded to Equation (7) as follows:
X i = j = 1 j i N S | X j X i | X j X i d i j g × e g + u × e w
However, the mathematical model using Equation (7) cannot be used directly to solve optimization problems, mainly because the grasshoppers quickly reach their comfort zone and the swarm does not converge to a specified point, according to a test in ref. [62]. The author of the GOA algorithm suggested a modified version of Equation (7) as shown in Equation (8) to solve optimization problems [62], where the gravity operator is unconsidered, the gravity factor is set to 0, and the wind direction is always defined as moving towards a target. Accordingly, Equation (2) becomes Equation (8) as follows:
X i d = c j = 1 j i N c u b d l b d 2 S ( | X j X i | ) X j X i d i j + T d
where u b d is the upper bound in the dth dimension, and l b d is the lower bound in the d t h dimension. T d is the value of the d t h dimension in the target (the best solution found so far). The coefficient c reduces the comfort zone proportional to the number of iterations and is calculated in Equation (9) as follows.
c = C m a x l C m a x C m i n L
where C m a x is the maximum value, C m i n is the minimum value, l indicates the current iteration, and L is the maximum number of iterations. In ref. [62], they use C m a x = 1 and C m i n = 0.00001 . Equation (8) shows that the next position of a grasshopper is defined based on its current position, the position of all other grasshoppers, and the position of the target. Algorithm 1 shows the pseudocode of the GOA algorithm.

4. Our Proposed BGOA-TVG Method

A binary search space is commonly considered as a hypercube [63]. The space is four-dimensional, which is formed by moving three-dimensional objects. In this search space, the search agents of the binary optimization algorithm can only move to nearer and farther corners of this hypercube by flipping various numbers of bits. Therefore, to design the binary version of the GOA, the concepts of the velocity- and position-updating process should be modified.
In the continuous version of the GOA, the swarm of grasshoppers moves around the search space by utilizing direction vectors, and the value of position is in the continuous real domain. In the binary space, due to dealing with only two numbers (“0” and “1”), the position cannot be updated using Equation (6). The way to change the position and velocity is outlined below.
In binary spaces, position updating means switching between “0” and “1” values. This switching should be based on the probability of updating the binary solution’s elements from 0 to 1 and vice versa. The main problem here is how to change the concept of velocity in the real world to a binary space.
In order to achieve this, a transfer function is important to map velocity values to probability values to update the positions. In other words, a transfer function defines the probability of changing a position element from 0 to 1 and vice versa. In general, transfer functions force predators to move in a binary space. According to ref. [64], the following concepts should be taken into consideration when selecting a transfer function in order to map velocity values to probability values:
(1) The range of a transfer function should be bounded in the interval [0, 1], as this represents the probability that a particle will change its position.
(2) A transfer function should have a high probability of changing position for large absolute values of velocity. Particles with large absolute values for their velocities are probably far from the best solution, so they should switch their positions in the next iteration.
(3) A transfer function should also have a small probability of changing position for small absolute values of velocity.
(4) The return value of a transfer function should increase as the velocity rises. Particles that are moving away from the best solution should have a higher probability of changing their position vectors in order to return to their previous positions.
(5) The return value of a transfer function should decrease as the velocity is reduced.
(6) These concepts guarantee that a transfer function is able to map the process of searching from a continuous search space to a binary search space while preserving simi-lar concepts of the search for a particular evolutionary algorithm. The GOA is simulated by PSO, the changed part in Equation (8). Defined as Δ X , Equation (10) is analogous to the velocity vector (step) in PSO [65]. The transfer function defines the probability of updating the binary solution’s elements from 0 to 1 and vice versa. In the BGOA, the probability of changing the positions of elements is based on the step vector values.
Δ X ( u b d , l b d , X i , X j ) = c 1 j = 1 j i N c 2 u b d l b d 2 S ( | X j X i | ) X j X i d i j
New time-varying transfer functions are proposed to enhance the ability of the BGOA in the search space. Algorithm 1 shows the pseudocode of the BGOA-TVG. The first transfer function (time-varying sin) is proposed to convert positions in the continuous space into the binary search space. The position is in the range of [ l b d , u b d ]. The binary position is mapped as follows:
T i m e v a r y i n g p o s i t i o n , α = sin p o s i t i o n α
B i n a r y p o s i t i o n = 1         i f   r a n d < T i m e v a r y i n g p o s i t i o n , α 0         i f   r a n d T i m e v a r y i n g p o s i t i o n , α
where α is in the range of [ α m i n , α m a x ], and the linear increase in Equation (11) switches the algorithm smoothly from the exploration to the exploitation phases.
α = α m i n + i t e r a t i o n α m a x α m i n / M a x _ I t e r a t i o n s
Figure 1 shows the time-varying transfer function. It enhances the capability of the exploration in the first phase, as shown by the blue curve in Figure 1. In this phase, the diversity is extremely high, so the swarm can search all of the space. The red curve shows the phase between exploration and exploitation, which has a lower level of diversity than the first phase and searches more around the good solutions. The last phase, shown by the purple curve, changes slowly for the last iterations.
To avoid the local optima, the GOA uses Equation (8) to update the best solution. In the BGOA-TVG, a new time-varying V-shaped transfer function combined with a Gaussian mutation is proposed, as shown in Figure 2. The binary solutions are generated based on the TVG as shown in Equation (14) and are defined as follows:
T V G p o s i t i o n , β , s i g m a = | 2 π a r c t a n ( π 2 p o s i t i o n / β ) | + G a u s s i a n 0 , s i g m a
β = β m a x i t e r a t i o n β m a x β m i n M a x I t e r a t i o n s
s i g m a = s i g m a m a x i t e r a t i o n s i g m a m a x s i g m a m i n / M a x _ I t e r a t i o n s
where β is in the range of [0.05, 10], and sigma is in the range of [0.01, 10] to switch efficiently from the exploration to the exploitation phases over time.
B i n a r y p o s i t i o n = ¬ B i n a r y p o s i t i o n         i f   r a n d < T V G p o s i t i o n , β , s i g m a B i n a r y p o s i t i o n                 i f   r a n d T V G p o s i t i o n , β , s i g m a
In Figure 2, the blue curve is the initial status of the combined function, which has both exploration and exploitation, and the purple curve is the last status, which has maximal exploitation. Because the parameters β and s i g m a in the function are constantly changing with each iteration, the intermediate conversion function is also constantly changing. The yellow curves depict diverse scenarios characterized by varying parameters.
Algorithm 1 shows the pseudocode of the BGOA-TVG algorithm. In ref. [65], it showed that normalizing the distance of grasshoppers in [1, 4], individuals can have both attraction and repulsion forces, which balance exploration and exploitation in the algorithm. Hence, we set the distance of individuals within the closed interval [1, 4].
Figure 3 shows the flowchart of the proposed algorithm.
Algorithm 1: Pseudocode of the BGOA-TVG algorithm.
Initialize C m a x , C m i n , and Max_Iterations
Initialize a population of solutions X i (i = 1, 2, …, n)
Evaluate each solution in the population
Set T as the best solution
While (t < Max_Iterations)
     Update c using Equation (9)
     For each search agent
          Normalize the distances between grasshoppers in [1, 4]
          Update the step vector ΔX of the current solution using Equation (10)
          For i = 1: dim
                Use Equation (8) to obtain the current position
                Use Equations (10)–(13) to obtain the binary position
                Use Equations (14)–(17) to obtain the final position
                Calculate α , β , s i g m a based on Equations (13), (15), and (16)
          End
          Reevaluate the fitness of each individual in the population
          If there is a better solution, replace T with it
          Update T
     End
End
Return T

Computational Complexity

The proposed transfer functions do not change the computational complexity of the algorithm during each iteration. Moreover, the core program of the grasshopper optimization algorithm is to find the current optimal value in a loop. Factors that affect the overall complexity include the population number, number of individuals, and number of iterations. Therefore, the maximum of the BGOA-TVG is O ( M a x _ I t e r a t i o n × N × D ), where D shows the number of dimensions.

5. Experimental Results and Discussion

5.1. Experimental Simulation Platform

For this experiment, we used Windows10 on a computer with the following specifications: a main frequency of 3.30 GHz; a memory of 16.0 GB; and an Inter(R) Core (TM)i3_6100 CPU. All algorithm codes were run in MATLAB2022a.

5.2. UCI Datasets

For this section, we selected 10 University of California at Irvine (UCI) datasets with different characteristics to verify the BGOA-TVG from different perspectives in terms of its performance. The name of the dataset, number of features, and number of instances the dataset has been used are shown in Table 1.

5.3. Parameter Settings

In order to verify the feasibility and effectiveness of the BGOA-TVG, we adopted some binary meta-heuristic algorithms: the BDA [23], BHHO [20], BPSO [11], BGWO [14], the BGBO [66], the BWOA [17], BGA [67], BBOA [68] and the BGOA-TVG. To make a fair comparison, the population size of the seven algorithms is set to 40, and the number of iterations is set to 100. Table 2 shows the main parameter settings of the seven algorithms.

5.4. Evaluation Criteria

The experimental results are evaluated in terms of the following criteria:
(1)
Average fitness function
The datasets are tested dependently 30 times, and the average fitness shows the stability of the proposed algorithm, as calculated in Equation (18):
M e a n = 1 N i = 1 N f i t n e s s i
where N is the number of runs of the optimization algorithm, and f i t n e s s i is the optimal solution resulting from the i t h run.
(2)
Average classification accuracy
The result is formulated in Equation (19) as follows:
A c c = 1 N i = 1 N A c c i
where N is the total number of runs of the proposed algorithm to select the subset. A c c i is the accuracy of the best solution from the i t h run.
(3)
Average feature selection size
This criterion can be calculated as in Equation (20), and the result is shown in Table 3.
A f s s = 1 N i = 1 N s i z e ( i ) D
where N is the total number of runs of the proposed algorithm to select the subset. s i z e ( i ) returns the number of features selected in the best solution from the i t h run, and D is the size of the original dataset.
The results of the average classification accuracy are shown in Table 3. Among them, the BGOA-TVG achieved the best results. It managed to reach the highest precision in nine datasets, and in addition, it had the highest precision value among all the methods in five datasets, so it came in first place in the overall ranking.
The performance of the proposed algorithm will be more clearly demonstrated by comparing it with the other 7 algorithms, as depicted in Figure 4, Figure 5, Figure 6, Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13.

5.5. Different Transfer Functions

In order to demonstrate the impact of different types of conversion functions on the final data results, we selected four classic S-shaped transfer functions and four classic V-shaped transfer functions, which are shown in Table 4. We compared these with our proposed time-varying transfer function. The grasshopper optimization algorithm was used, and the results are shown in Figure 14, Figure 15, Figure 16, Figure 17, Figure 18, Figure 19, Figure 20, Figure 21, Figure 22, Figure 23, Figure 24 and Figure 25.

6. Electroencephalogram (EEG) Dataset Analysis

In this section, we present the proposed approach for channel selection for EEG-based signal acquisition. An EEG is very valuable for the diagnosis, identification, and treatment monitoring of epilepsy. Epilepsy often manifests itself as uncontrollable convulsions and involuntary behaviors. By observing and analyzing the signals recorded via EEGs, not only does it not harm a patient, but it can also help physicians determine if a patient has epilepsy through mapping analyses. In cases such as brain injury or stroke, EEGs can provide doctors with critical information to help them plan patients’ treatment. This is why EEGs have important clinical applications.

6.1. EEG Dataset

6.1.1. DEAP Dataset

The EEG signals used in this work were obtained from the EEG Motor Movement/Imagery dataset. The data were collected from 109 healthy volunteers using the BCI2000 System, which makes use of 64 channels (sensors) and provides a separated EDF (European data format) file for each of them. The subjects performed different motor/imagery tasks. These tasks are mainly used in BCI (brain–computer interface) applications and neurological rehabilitation and consist of imagining or simulating a given action, like opening and closing one’s eyes, for example. Each subject performed four tasks according to the position of a target that appeared on the screen placed in front of them (if the target appears on the right or left side, the subject opens and closes the corresponding fist; if the target appears on the top or bottom, the subject opens and closes both fists or both feet, respectively). In short, the four experimental tasks were as follows:
  • To open and close their left or right fist;
  • To imagine opening and closing their left or right fist;
  • To open and close both their fists or both their feet;
  • To imagine opening and closing both their fists or both their feet.
Each of these tasks were performed three times, thus generating 12 recordings for each subject for a two-minute run, and the 64 channels were sampled at 160 samples per second. The features of the twelve recordings are extracted by means of an AR model with three output configurations for each EEG channel: 5, 10, and 20 features. Further, the average of each configuration is then computed in order to obtain just one feature per EEG channel (sensor). In short, for each sensor, we extracted three different numbers of AR-based features, with the output of each sensor being the average of their values. Henceforth, we have adopted the following notation for each of the dataset configurations: AR5 for the five autoregression coefficients extracted, and AR10 and AR20 for the ten and twenty autoregression coefficients, respectively. All the datasets we used were processed and can be found at https://openneuro.org/ (accessed on 31 January 2024).

6.1.2. Dataset of EEG Recordings of Pediatric Patients with Epilepsy

Self-limiting epilepsy with central temporal spikes is a common focal epilepsy in childhood, mainly characterized by paroxysmal seizures in the mouth, pharynx, and on one side of the face. It is often accompanied by tongue stiffness, as well as speech and swallowing difficulties. Today, electroencephalography and other methods are the primary diagnostic tools for self-limiting epilepsy with central temporal spikes, and the prognosis is generally favorable. An epileptic electroencephalogram (EEG) refers to a special type of brain wave phenomenon induced by sleep, which is close to sustained spike slow-wave emissions and occurs more frequently during the SELECTS seizure period. In order to further investigate the impact of epileptic electrical persistence during sleep on patients’ pediatric symptoms, we selected the publicly available EEG dataset of pediatric epilepsy syndromes in a feature selection and accuracy analysis. In total, 88 subjects recorded EEGs with their closed eyes during the resting state in this dataset. A total of 36 of the participants were diagnosed with Alzheimer’s disease (AD group), 23 were diagnosed with frontotemporal fementia (FTD group), and 29 were healthy subjects (CN group). Using the international Mini-Mental State Examination (MMSE), their cognitive and neuropsychological state was measured. Lower MMSE scores indicate more severe cognitive decline, with scores ranging from 0 to 30. Months were used to measure the duration of the disease, with an average of 25 and an interquartile range (IQR) of 24 to 28.5 months. No comorbidities related to dementia were reported in the AD groups. The AD group experienced an average MMSE of 17.75 (SD = 4.5), while the FTD group had an average of 22.17 (SD = 8.22), and the CN group had an average of 30. The AD group averaged 66.4 years, the FTD group averaged 63.6 years, and the CN group averaged 67.9 years, which was the SD for the whole group. The data we used were analyzed and are now available at https://openneuro.org/ (accessed on 31 January 2024).

6.2. Compared Methods

The BGOA-TVG is an improved version of the grasshopper optimization algorithm (GOA) for multitask problems of pattern recognition. Furthermore, it is characterized as a swarm intelligence (SI) algorithm. SI has been proven to be a technique that can solve NP-hard computational problems, such as feature selection. Although a considerable number of new swarm-inspired algorithms have emerged in recent years, particle swarm optimization (PSO) is still the most widely used SI algorithm for solving feature selection problems [52]. In addition, the individual expression in SI for feature selection is typically a bit string, whereby the dimensionality of an individual is equal to the total number of features in the dataset. Binary encoding is more commonly used for feature selection than real encoding. Therefore, for feature selection, we compared the BGOA-TVG with BPSO [11,12], the BGA [24], BHHO [20], the BWOA [17], BACO [21] BGWO [14], and a method with all the features obtained from the time, frequency, and time–frequency domains. Their characteristics are shown in Table 5. All of three methods adopt 40 individuals for 100 iterations.

6.3. Classification Indices

In the experiment, five classification indices are used for the validation of the compared methods, including the true positive rate (recall, TPR), the positive predictive value (precision, PPV), the true negative rate (specificity, TNR), the negative predictive value (NPV), and the classification accuracy (ACC). They can be respectively defined as follows:
T P R = T P T P + F N
P P V = T P T P + F P
T N R = T N T N + F P
N P V = T N T N + F N
A C C = T P + T N T P + T N + F P + F N

6.4. Analysis of Results

6.4.1. Selection of Classifier

The algorithms chosen are directly applied to every subject dataset with a fixed classification method, i.e., the SVM, KNN, Bayes, DT, RF, or Adaboost methods. Their results for the DEAP dataset are recorded in Figure 26, Figure 27, Figure 28, Figure 29, Figure 30, Figure 31, Figure 32 and Figure 33, respectively. It can be seen that the classification accuracies of the five classifiers fluctuate prominently even for the same subject dataset and the same pattern recognition scheme. In terms of the highest classification accuracy, the best classifier is chosen artificially for every subject dataset, as shown in Table 6, respectively, corresponding to the DEAP dataset. It should be noted that the emotion recognition results for the same dataset depend on the classification method to some extent.
The training dataset comprises 50% of the original dataset, while 30% is allocated for the validation sets and the remaining 20% is reserved for the test sets. When considering the selection of classifiers for optimization, the BGOA-TVG undoubtedly realizes emotion recognition in a more efficient way. Moreover, according to the data in Figure 26, Figure 27, Figure 28, Figure 29, Figure 30, Figure 31, Figure 32 and Figure 33, the RF, DT, and Bayes methods are chosen as the best classifiers for emotion recognition with a high probability, while the KNN method is the worst one. This indicates that the RF, DT, and Bayes methods are more suitable for EEG-based emotion recognition than the KNN method.

6.4.2. Selection of Features and Parameters

To further identify the optimal results of the BGOA-TVG, we simultaneously recorded five classification indices for Allfeat, BPSO, the BGA, BHHO, the BWOA, BACO, and BGWO for all subject datasets. The parameters were the same as those in the previous experiment. The results are shown in Figure 34, Figure 35, Figure 36, Figure 37 and Figure 38. Firstly, the TPR (true positive rate), describes the ratio of all identified positive cases to all positive cases. It can be seen in Figure 34 that the classification index curve of the BGOA-TVG is better than the other algorithms most of time, and BPSO and BHHO have lower TPR values. Secondly, the PPV (positive predictive value) is shown in Figure 35. BGWO achieves the best performance in subject 3; however, it exhibits inferior results compared to BGOA-TVG in other subjects. Moreover, the volatility of the curve is too large, proving the poor robustness of the algorithm. In contrast, the BGOA-TVG is very robust. Thirdly, the TNR describes the proportion of identified negative cases to all negative cases. The BGOA-TVG shows the best robustness compared to the other algorithms, as can be seen in Figure 36. Fourthly, the NPV is the negative predictive value. In Figure 37, BGWO shows a similar classification rate as that of BHHO for subjects 1–9, but in subject 10, BGWO performs worse than the others, while BPSO, BACO, and the BGOA-TVG obtain a ratio of 1. Furthermore, the average classification accuracy values of the DEAP datasets found by these algorithms are recorded in Figure 38. It can been seen that the BGOA-TVG can choose an appropriate method based on the specific characteristics of an emotion dataset. The mean classification accuracy of the BGOA-TVG for DEAP is respectively higher than those of the other algorithms. This verifies that the BGOA-TVG can efficiently recognize emotion patterns.

6.4.3. Analysis of Results

The horizontal lines in Figure 34, Figure 35, Figure 36, Figure 37 and Figure 38 represent the different datasets that we used, with subjects no. 1–10 corresponding to each of the ten datasets and the points representing the outcomes obtained using diverse algorithms. The lines indicate the compatibility and stability of the different algorithms. As a result, the proposed binary GOA is more stable than the other algorithms, and the application of the proposed algorithm for different DEAP datasets is universal.
Figure 39 depicts the mean feature numbers for all optimization techniques regarding the learning algorithm. As we did not consider the feature extraction procedure, i.e., the autoregression coefficient computation, the feature numbers chosen over all dataset configurations are quite similar for BPSO and the BGA. In Table 7, it is possible to observe that the BGOA-TVG is the fastest technique in all situations, since it only updates one agent per iteration. Although this may be a drawback in terms of convergence, it is still the fastest approach with lowest number of features.
However, in Figure 38, the proposed binary GOA is the most accurate method compared to other algorithms and also selects fewer features. In general, the proposed algorithm can get results exhibit the most minimal error rate than other algorithms in EEG signal analysis.
Finally, we still need to deal with the trade-off between the number of features and the computational efficiency. Using all of the sensors does not lead to very different results, which supports the idea of this work: one can find a subset of sensors that can obtain reasonable results.

6.4.4. Analysis of Results for Epilepsy EEG

Every subject dataset receives a direct application of the chosen algorithms through the SVM method, a fixed classification method. The epilepsy dataset results are documented in Figure 40, Figure 41, Figure 42, Figure 43, Figure 44, Figure 45, Figure 46, Figure 47, Figure 48, Figure 49, Figure 50 and Figure 51, respectively. The classification accuracy shows significant fluctuations for the various subject datasets. In terms of the highest classification accuracy, the best classifier is chosen artificially for every subject dataset. The results shown in Table 8 correspond to the epilepsy EEG datasets.
A total of 50% of the original dataset is present in the training dataset, while 30% and 20% are present in the validation and test sets, respectively. Without a doubt, the BGOA-TVG recognizes emotions in a more efficient manner. The data presented in Figure 39, Figure 40, Figure 41, Figure 42, Figure 43, Figure 44, Figure 45, Figure 46, Figure 47, Figure 48, Figure 49 and Figure 50 suggest that the BGOA-TVG is superior to the other algorithms when it comes to EEG-based emotion recognition.

7. Conclusions and Future Work

This paper proposed a new binary version of the grasshopper optimization algorithm, called the binary grasshopper optimization algorithm, using time-varying Gaussian mixed transfer functions (BGOA-TVG). Compared with the original GOA, which has a slow convergence speed and can easily fall into a local optimum, the BGOA-TVG has a good global search capability and accuracy in the local search space. Furthermore, the improved version of the GOA balances the relationship between exploration and exploitation in the search space and effectively avoids premature convergence. In order to verify the effectiveness and feasibility of the BGOA-TVG, 10 UCI datasets and a well-known DEAP dataset were tested for the algorithm. And the application of the K-nearest neighbor method further proves that the improved BGOA-TVG has a faster convergence speed in the global search space and a better accuracy in the local search space than other binary algorithms. For the BGOA-TVG, in future works, there are two possible research avenues. First, one could use different strategies to improve the original BGOA to increase the speed of searching the global space or obtain more accurate results in the local search space. Second, one could apply the BGOA-TVG to solve more complex optimization problems, such as image segmentation or network optimization configuration problems. Specific feature extraction applications are recommended for most mental illnesses, as they can quickly process a patient’s symptoms. Furthermore, augmenting the database volume will help us conduct more rapid and accurate data analyses.

Author Contributions

Conceptualization, Y.Z. and Q.L.; methodology, Y.Z.; software, M.L.; validation, M.L., Q.L. and Y.Z.; formal analysis, M.L.; investigation, M.L.; resources, Q.L.; data curation, M.L.; writing—original draft preparation, M.L.; writing—review and editing, Y.Z.; visualization, Q.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under grant nos. U21A20464 and 62066005.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data can be found at https://openneuro.org/ (accessed on 31 January 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Meenachi, L.; Ramakrishnan, S. Metaheuristic Search Based Feature Selection Methods for Classification of Cancer. Pattern Recognit. 2021, 119, 108079. [Google Scholar] [CrossRef]
  2. Şahin, C.B.; Abualigah, L. A novel deep learning-based feature selection model for improving the static analysis of vulnerability detection. Neural Comput. Appl. 2021, 33, 14049–14067. [Google Scholar] [CrossRef]
  3. Maleki, N.; Zeinali, Y.; Niaki, S.T.A. A k-NN method for lung cancer prognosis with the use of a genetic algorithm for feature selection. Expert Syst. Appl. 2021, 164, 113981. [Google Scholar] [CrossRef]
  4. Abualigah, L.; Yousri, D.; Elaziz, M.A.; Ewees, A.A.; Al-qaness, M.A.A.; Gandomi, A.H. Aquila optimizer: A novel meta-heuristic optimization algorithm. Comput. Ind. Eng. 2021, 157, 107250. [Google Scholar] [CrossRef]
  5. Sharma, P.; Raju, S. Metaheuristic optimization algorithms: A comprehensive overview and classification of benchmark test functions. Soft Comput. 2024, 28, 3123–3186. [Google Scholar] [CrossRef]
  6. Beheshti, Z.; Shamsuddin, S.M.; Sulaiman, S. Fusion Global-Local-Topology Particle Swarm Optimization for Global Optimization Problems. Math. Probl. Eng. 2014, 2014, 107–136. [Google Scholar] [CrossRef]
  7. Zhao, F.; Bao, H.; Wang, L.; Cao, J.; Tang, J. A multipopulation cooperative coevolutionary whale optimization algorithm with a two-stage orthogonal learning mechanism. Knowl. Based Syst. 2022, 246, 108664. [Google Scholar] [CrossRef]
  8. Pan, J.-S.; Zhang, L.-G.; Wang, R.-B.; Snášel, V.; Chu, S.-C. Gannet optimization algorithm: A new metaheuristic algorithm for solving engineering optimization problems. Math. Comput. Simul. 2022, 202, 343–373. [Google Scholar] [CrossRef]
  9. Lu, K.-D.; Wu, Z.-G. Constrained-Differential-Evolution-Based Stealthy Sparse Cyber-Attack and Countermeasure in an AC Smart Grid. IEEE Trans. Ind. Inform. 2022, 18, 5275–5285. [Google Scholar] [CrossRef]
  10. Dokeroglu, T.; Deniz, A.; Kiziloz, H.E. A comprehensive survey on recent metaheuristics for feature selection. Neurocomputing 2022, 494, 269–296. [Google Scholar] [CrossRef]
  11. Kılıç, F.; Kaya, Y.; Yildirim, S. A novel multi population based particle swarm optimization for feature selection. Knowl. Based Syst. 2021, 219, 106894. [Google Scholar] [CrossRef]
  12. Beheshti, Z. BMNABC: Binary Multi-Neighborhood Artificial Bee Colony for High-Dimensional Discrete Optimization Problems. Cybern. Syst. 2018, 49, 452–474. [Google Scholar] [CrossRef]
  13. Bostani, H.; Sheikhan, M. Hybrid of binary gravitational search algorithm and mutual information for feature selection in intrusion detection systems. Soft Comput. 2015, 21, 2307–2324. [Google Scholar] [CrossRef]
  14. Rajalaxmi, R.R.; Mirjalili, S.; Gothai, E.; Natesan, P. Binary grey wolf optimization with mutation and adaptive k-nearest neighbor for feature selection in Parkinson’s disease diagnosis. Knowl. Based Syst. 2022, 246, 108701. [Google Scholar]
  15. Faris, H.; Mafarja, M.M.; Heidari, A.A.; Aljarah, I.; Al-Zoubi, A.M.; Mirjalili, S.; Fujita, H. An efficient binary Salp Swarm Algorithm with crossover scheme for feature selection problems. Knowl. Based Syst. 2018, 154, 43–67. [Google Scholar] [CrossRef]
  16. Nakamura, R.Y.M.; Pereira, L.A.M.; Rodrigues, D.; Costa, K.A.P.; Papa, J.P. 9—Binary bat algorithm for feature selection. Swarm Intell. Bio-Inspired Comput. 2013, 225–237. [Google Scholar]
  17. Zhou, R.; Zhang, Y.; He, K. A novel hybrid binary whale optimization algorithm with chameleon hunting mechanism for wrapper feature selection in QSAR classification model: A drug-induced liver injury case study. Expert Syst. Appl. 2023, 234, 121015. [Google Scholar] [CrossRef]
  18. Kumar, V.; Kaur, A. Binary spotted hyena optimizer and its application to feature selection. J. Ambient. Intell. Humaniz. Comput. 2020, 11, 2625–2645. [Google Scholar] [CrossRef]
  19. Dhiman, G.; Oliva, D.; Kaur, A.; Singh, K.K.; Vimal, S.; Sharma, A.; Cengiz, K. BEPO: A novel binary emperor penguin optimizer for automatic feature selection. Knowl. Based Syst. 2021, 211, 106560. [Google Scholar] [CrossRef]
  20. Hussain, K.; Neggaz, N.; Zhu, W.; Houssein, E.H. An efficient hybrid sine-cosine Harris hawks optimization for low and high-dimensional feature selection. Expert Syst. Appl. 2021, 176, 114778. [Google Scholar] [CrossRef]
  21. Rahab, H.; Haouassi, H.; Laouid, A. Rule-Based Arabic Sentiment Analysis using Binary Equilibrium Optimization Algorithm. Arab. J. Sci. Eng. 2023, 48, 2359–2374. [Google Scholar] [CrossRef]
  22. Too, J.; Abdullah, A.R. Binary atom search optimisation approaches for feature selection. Connect. Sci. 2020, 32, 406–430. [Google Scholar] [CrossRef]
  23. Too, J.; Mirjalili, S. A Hyper Learning Binary Dragonfly Algorithm for Feature Selection: A COVID-19 Case Study. Knowl. Based Syst. 2021, 212, 106553. [Google Scholar] [CrossRef]
  24. Chaudhuri, A.; Sahu, T.P. A hybrid feature selection method based on Binary Jaya algorithm for micro-array data classification. Comput. Electr. Eng. 2021, 90, 106963. [Google Scholar] [CrossRef]
  25. Alweshah, M.; Alkhalaileh, S.; Al-Betar, M.A.; Abu Bakar, A. Coronavirus herd immunity optimizer with greedy crossover for feature selection in medical diagnosis. Knowl. Based Syst. 2022, 235, 107629. [Google Scholar] [CrossRef]
  26. Sadeghian, Z.; Akbari, E.; Nematzadeh, H. A hybrid feature selection method based on information theory and binary butterfly optimization algorithm. Eng. Appl. Artif. Intell. 2021, 97, 104079. [Google Scholar] [CrossRef]
  27. Al-Saedi, A.; Mawlood-Yunis, A.R. Binary Black Widow Optimization Algorithm for Feature Selection Problems. Learn. Intell. Optim. 2022, 13621, 93–107. [Google Scholar]
  28. Hu, J.; Gui, W.; Heidari, A.A.; Cai, Z.; Liang, G.; Chen, H.; Pan, Z. Dispersed foraging slime mould algorithm: Continuous and binary variants for global optimization and wrapper-based feature selection. Knowl. Based Syst. 2022, 237, 107761. [Google Scholar] [CrossRef]
  29. Eluri, R.K.; Devarakonda, N. Binary Golden Eagle Optimizer with Time-Varying Flight Length for feature selection. Knowl. Based Syst. 2022, 247, 108771. [Google Scholar] [CrossRef]
  30. Hussain, K.; Salleh, M.N.M.; Cheng, S.; Shi, Y. Metaheuristic research: A comprehensive survey. Artif. Intell. Rev. 2019, 52, 2191–2233. [Google Scholar] [CrossRef]
  31. Liu, Y.; Wang, Y.; Ren, X.; Zhou, H.; Diao, X. A Classification Method Based on Feature Selection for Imbalanced Data. IEEE Access 2019, 7, 81794–81807. [Google Scholar] [CrossRef]
  32. Khorashadizade, M.; Hosseini, S. An intelligent feature selection method using binary teaching-learning based optimization algorithm and ANN. Chemom. Intell. Lab. Syst. 2023, 240, 104880. [Google Scholar] [CrossRef]
  33. Hegazy, A.E.; Makhlouf, M.; El-Tawel, G.S. Improved salp swarm algorithm for feature selection. J. King Saud Univ. Comput. Inf. Sci. 2020, 32, 335–344. [Google Scholar] [CrossRef]
  34. Zaimoğlu, E.A.; Yurtay, N.; Demirci, H.; Yurtay, Y. A binary chaotic horse herd optimization algorithm for feature selection. Eng. Sci. Technol. Int. J. 2023, 44, 101453. [Google Scholar] [CrossRef]
  35. Khodadadi, N.; Khodadadii, E.; Al-Tashi, Q.; El-Kenawy, E.S.M.; Abualigah, L.; Abdulkadir, S.J.; Alqushaibi, A.; Mirjalili, S. BAOA: Binary Arithmetic Optimization Algorithm with K-Nearest Neighbor Classifier for Feature Selection. IEEE Access 2023, 11, 94094–94115. [Google Scholar] [CrossRef]
  36. Ghazali, S.M.; Alizadeh, M.; Mazloum, J.; Baleghi, Y. Modified binary salp swarm algorithm in EEG signal classification for epilepsy seizure detection. Biomed. Signal Process. Control. 2022, 78, 72–93. [Google Scholar]
  37. Chantar, H.; Tubishat, M.; Essgaer, M.; Mirjalili, S. Hybrid binary dragonfly algorithm with simulated annealing for feature selection. SN Comput. Sci. 2021, 2, 295. [Google Scholar] [CrossRef]
  38. Beheshti, Z. BMPA-TVSinV: A Binary Marine Predators Algorithm using time-varying sine and V-shaped transfer functions for wrapper-based feature selection. Knowl. Based Syst. 2022, 252, 109446. [Google Scholar] [CrossRef]
  39. Got, A.; Moussaoui, A.; Zouache, D. Hybrid filter-wrapper feature selection using whale optimization algorithm: A multi-objective approach. Expert Syst. Appl. 2021, 183, 115312. [Google Scholar] [CrossRef]
  40. Rostami, M.; Forouzandeh, S.; Berahmand, K.; Soltani, M. Integration of multi-objective PSO based feature selection and node centrality for medical datasets. Genomics 2020, 112, 4370–4384. [Google Scholar] [CrossRef]
  41. Del Sagrado, J.; Del Águila, I.M.; Orellana, F.J. Multi-objective ant colony optimization for requirements selection. Empir. Softw. Eng. 2015, 20, 577–610. [Google Scholar] [CrossRef]
  42. Aljarah, I.; Habib, M.; Faris, H.; Al-Madi, N. A dynamic locality multi-objective salp swarm algorithm for feature selection. Comput. Ind. Eng. 2020, 147, 106628. [Google Scholar] [CrossRef]
  43. Ahmed, S.; Ghosh, K.K.; Mirjalili, S.; Sarkar, R. AIEOU: Automata-based improved equilibrium optimizer with U-shaped transfer function for feature selection. Knowl. Based Syst. 2021, 228, 107283. [Google Scholar] [CrossRef]
  44. Mirjalili, S.; Lewis, A. S-shaped versus V-shaped transfer functions for binary Particle Swarm Optimization. Swarm Evol. Comput. 2013, 9, 1–14. [Google Scholar] [CrossRef]
  45. Yedukondalu, J.; Sharma, L.D. Cognitive load detection using circulant singular spectrum analysis and Binary Harris Hawks Optimization based feature selection. Biomed. Signal Process. Control. 2023, 79, 104006. [Google Scholar] [CrossRef]
  46. Gunia, A.; Moraresku, S.; Janča, R.; Ježdík, P.; Kalina, A.; Hammer, J.; Marusič, P.; Vlček, K. The brain dynamics of visuospatial perspective-taking captured by intracranial EEG. NeuroImage 2024, 285, 120487. [Google Scholar] [CrossRef] [PubMed]
  47. Emary, E.; Zawbaa, H.M.; Hassanien, A.E. Binary ant lion approaches for feature selection. Neurocomputing 2016, 213, 54–65. [Google Scholar] [CrossRef]
  48. Faris, H.; Hassonah, M.A.; Ala, M.A.-Z.; Mirjalili, S.; Aljarah, I. A multi-verse optimizer approach for feature selection and optimizing SVM parameters based on a robust system architecture. Neural Comput. Appl. 2018, 30, 2355–2369. [Google Scholar] [CrossRef]
  49. Emary, E.; Zawbaa, H.M.; Hassanien, A.E. Binary grey wolf optimization approaches for feature selection. Neurocomputing 2016, 172, 371–381. [Google Scholar] [CrossRef]
  50. Mafarja, M.; Aljarah, I.; Heidari, A.A.; Faris, H.; Fournier-Viger, P.; Li, X.; Mirjalili, S. Binary dragonfly optimization for feature selection using time-varying transfer functions. Knowl. Based Syst. 2018, 161, 185–204. [Google Scholar] [CrossRef]
  51. Rao, H.; Shi, X.; Rodrigue, A.K.; Feng, J.; Xia, Y.; Elhoseny, M.; Yuan, X.; Gu, L. Feature selection based on artificial bee colony and gradient boosting decision tree. Appl. Soft Comput. 2019, 74, 634–642. [Google Scholar] [CrossRef]
  52. Sayed, G.I.; Hassanien, A.E.; Azar, A.T. Feature selection via a novel chaotic crow search algorithm. Neural Comput. Appl. 2019, 31, 171–188. [Google Scholar] [CrossRef]
  53. Hancer, E.; Xue, B.; Zhang, M.; Karaboga, D.; Akay, B. Pareto front feature selection based on artificial bee colony optimization. Inf. Sci. 2018, 422, 462–479. [Google Scholar] [CrossRef]
  54. Liu, H.; Yu, L. Toward integrating feature selection algorithms for classification and clustering. IEEE Trans. Knowl. Data Eng. 2005, 17, 491–502. [Google Scholar]
  55. Baig, M.Z.; Aslam, N.; Shum, H.P.H. Filtering techniques for channel selection in motor imagery EEG applications: A survey. Artif. Intell. Rev. 2020, 53, 1207–1232. [Google Scholar] [CrossRef]
  56. Wang, Z.-M.; Hu, S.-Y.; Song, H. Channel Selection Method for EEG Emotion Recognition Using Normalized Mutual Information. IEEE Access 2019, 7, 143303–143311. [Google Scholar] [CrossRef]
  57. Boonyakitanont, P.; Lek-Uthai, A.; Chomtho, K.; Songsiri, J. A review of feature extraction and performance evaluation in epileptic seizure detection using EEG. Biomed. Signal Process. Control 2020, 57, 101702. [Google Scholar] [CrossRef]
  58. Li, X.; Hu, B.; Sun, S.; Cai, H. EEG-based mild depressive detection using feature selection methods and classifiers. Comput. Methods Programs Biomed. 2016, 136, 151–161. [Google Scholar] [CrossRef]
  59. Liu, Q.; Wang, F.; Xiao, W.; Cui, J. Binary dynamic stochastic search algorithm with support vector regression for feature selection in low-velocity impact localization problem. Eng. Appl. Artif. Intell. 2023, 124, 106554. [Google Scholar] [CrossRef]
  60. Zhuang, Z.; Pan, J.-S.; Li, J.; Chu, S.-C. Parallel binary arithmetic optimization algorithm and its application for feature selection. Knowl. Based Syst. 2023, 275, 110640. [Google Scholar] [CrossRef]
  61. Chen, S.-M.; Chiou, C.-H. Multiattribute Decision Making Based on Interval-Valued Intuitionistic Fuzzy Sets, PSO Techniques, and Evidential Reasoning Methodology. IEEE Trans. Fuzzy Syst. 2014, 23, 1905–1916. [Google Scholar] [CrossRef]
  62. Saremi, S.; Mirjalili, S.; Lewis, A. Grasshopper Optimisation Algorithm: Theory and application. Adv. Eng. Softw. 2017, 105, 30–47. [Google Scholar] [CrossRef]
  63. Moradi, P.; Gholampour, M. A hybrid particle swarm optimization for feature subset selection by integrating a novel local search strategy. Appl. Soft Comput. 2016, 43, 117–130. [Google Scholar] [CrossRef]
  64. Beheshti, Z. UTF: Upgrade transfer function for binary meta-heuristic algorithms. Appl. Soft Comput. 2021, 106, 107346. [Google Scholar] [CrossRef]
  65. Mafarja, M.; Aljarah, I.; Faris, H.; Hammouri, A.; Al-Zoubi, A.; Mirjalili, S. Binary grasshopper optimisation algorithm approaches for feature selection problems. Expert Syst. Appl. 2019, 117, 267–286. [Google Scholar] [CrossRef]
  66. Jiang, Y.; Luo, Q.; Wei, Y.; Zhou, Y. An efficient binary Gradient-based optimizer for feature selection. Math. Biosci. Eng. 2021, 4, 3813–3854. [Google Scholar] [CrossRef]
  67. Wang, H.; He, C.; Li, Z. A new ensemble feature selection approach based on genetic algorithm. Soft Comput 2020, 24, 15811–15820. [Google Scholar] [CrossRef]
  68. Arora, S.; Anand, P. Binary butterfly optimization approaches for feature selection. Expert Syst. Appl. 2019, 116, 147–160. [Google Scholar] [CrossRef]
Figure 1. Time-varying transfer sine function.
Figure 1. Time-varying transfer sine function.
Biomimetics 09 00187 g001
Figure 2. Time-varying V-shaped transfer function mixed with Gaussian function.
Figure 2. Time-varying V-shaped transfer function mixed with Gaussian function.
Biomimetics 09 00187 g002
Figure 3. Flowchart of the proposed algorithm.
Figure 3. Flowchart of the proposed algorithm.
Biomimetics 09 00187 g003
Figure 4. Cancer.
Figure 4. Cancer.
Biomimetics 09 00187 g004
Figure 5. Breast.
Figure 5. Breast.
Biomimetics 09 00187 g005
Figure 6. Wine.
Figure 6. Wine.
Biomimetics 09 00187 g006
Figure 7. Ionosphere.
Figure 7. Ionosphere.
Biomimetics 09 00187 g007
Figure 8. Leukemia.
Figure 8. Leukemia.
Biomimetics 09 00187 g008
Figure 9. Cloud.
Figure 9. Cloud.
Biomimetics 09 00187 g009
Figure 10. Vote.
Figure 10. Vote.
Biomimetics 09 00187 g010
Figure 11. Wall-following robot navigation.
Figure 11. Wall-following robot navigation.
Biomimetics 09 00187 g011
Figure 12. Sonar.
Figure 12. Sonar.
Biomimetics 09 00187 g012
Figure 13. Zoo.
Figure 13. Zoo.
Biomimetics 09 00187 g013
Figure 14. S-shaped.
Figure 14. S-shaped.
Biomimetics 09 00187 g014
Figure 15. V-shaped.
Figure 15. V-shaped.
Biomimetics 09 00187 g015
Figure 16. Cancer.
Figure 16. Cancer.
Biomimetics 09 00187 g016
Figure 17. Breast.
Figure 17. Breast.
Biomimetics 09 00187 g017
Figure 18. Wine.
Figure 18. Wine.
Biomimetics 09 00187 g018
Figure 19. Ionosphere.
Figure 19. Ionosphere.
Biomimetics 09 00187 g019
Figure 20. Leukemia.
Figure 20. Leukemia.
Biomimetics 09 00187 g020
Figure 21. Cloud.
Figure 21. Cloud.
Biomimetics 09 00187 g021
Figure 22. Vote.
Figure 22. Vote.
Biomimetics 09 00187 g022
Figure 23. Wall-following robot navigation.
Figure 23. Wall-following robot navigation.
Biomimetics 09 00187 g023
Figure 24. Sonar.
Figure 24. Sonar.
Biomimetics 09 00187 g024
Figure 25. Zoo.
Figure 25. Zoo.
Biomimetics 09 00187 g025
Figure 26. Classification accuracies of Allfeat.
Figure 26. Classification accuracies of Allfeat.
Biomimetics 09 00187 g026
Figure 27. Classification accuracies of BPSO.
Figure 27. Classification accuracies of BPSO.
Biomimetics 09 00187 g027
Figure 28. Classification accuracies of BGA.
Figure 28. Classification accuracies of BGA.
Biomimetics 09 00187 g028
Figure 29. Classification accuracies of BHHO.
Figure 29. Classification accuracies of BHHO.
Biomimetics 09 00187 g029
Figure 30. Classification accuracies of BWOA.
Figure 30. Classification accuracies of BWOA.
Biomimetics 09 00187 g030
Figure 31. Classification accuracies of BACO.
Figure 31. Classification accuracies of BACO.
Biomimetics 09 00187 g031
Figure 32. Classification accuracies of BGWO.
Figure 32. Classification accuracies of BGWO.
Biomimetics 09 00187 g032
Figure 33. Classification accuracies of BGOA-TVG.
Figure 33. Classification accuracies of BGOA-TVG.
Biomimetics 09 00187 g033
Figure 34. TPR obtained for ten subjects of DEAP dataset.
Figure 34. TPR obtained for ten subjects of DEAP dataset.
Biomimetics 09 00187 g034
Figure 35. PPV obtained for ten subjects of DEAP dataset.
Figure 35. PPV obtained for ten subjects of DEAP dataset.
Biomimetics 09 00187 g035
Figure 36. TNR obtained for ten subjects of DEAP dataset.
Figure 36. TNR obtained for ten subjects of DEAP dataset.
Biomimetics 09 00187 g036
Figure 37. NPV obtained for ten subjects of DEAP dataset.
Figure 37. NPV obtained for ten subjects of DEAP dataset.
Biomimetics 09 00187 g037
Figure 38. Mean ACC obtained for ten subjects of DEAP dataset.
Figure 38. Mean ACC obtained for ten subjects of DEAP dataset.
Biomimetics 09 00187 g038
Figure 39. Mean feature number obtained for ten subjects from the DEAP dataset over forty iterations.
Figure 39. Mean feature number obtained for ten subjects from the DEAP dataset over forty iterations.
Biomimetics 09 00187 g039
Figure 40. Sub-001.
Figure 40. Sub-001.
Biomimetics 09 00187 g040
Figure 41. Sub-002.
Figure 41. Sub-002.
Biomimetics 09 00187 g041
Figure 42. Sub-003.
Figure 42. Sub-003.
Biomimetics 09 00187 g042
Figure 43. Sub-004.
Figure 43. Sub-004.
Biomimetics 09 00187 g043
Figure 44. Sub-005.
Figure 44. Sub-005.
Biomimetics 09 00187 g044
Figure 45. Sub-006.
Figure 45. Sub-006.
Biomimetics 09 00187 g045
Figure 46. Sub-007.
Figure 46. Sub-007.
Biomimetics 09 00187 g046
Figure 47. Sub-008.
Figure 47. Sub-008.
Biomimetics 09 00187 g047
Figure 48. Sub-009.
Figure 48. Sub-009.
Biomimetics 09 00187 g048
Figure 49. Sub-010.
Figure 49. Sub-010.
Biomimetics 09 00187 g049
Figure 50. Sub-011.
Figure 50. Sub-011.
Biomimetics 09 00187 g050
Figure 51. Sub-012.
Figure 51. Sub-012.
Biomimetics 09 00187 g051
Table 1. UCI datasets.
Table 1. UCI datasets.
No.DatasetFeaturesInstances
1Cancer9683
2Breast9277
3Wine13178
4Ionosphere34351
5Leukemia712972
6cloud101024
7Vote18846
8Wall following robot navigation data245456
9sonar60208
10Zoo16101
Table 2. Parameter settings.
Table 2. Parameter settings.
AlgorithmsParameter Values
BDA [23] W m a x = 0.9 ; W m i n = 0.4
BHHO [20] α u [ 2 ,   0 ]
BPSO [11] W m a x = 0.9 ; W m i n = 0.4 ; C 1 = 2 ; C 2 = 2 ; V m a x = 6
BGWO [14] α u [ 2 ,   0 ]
BWOA [17] α u 2 ,   0 ,   α = 0.99 ,   β = 1
BGBO [66]Pr = 0.5
BGOA [65] i t e r v a l 0 ,   2.079 ,   l = 1.5 ,   f = 0.5
BGA [67]px = 0.7, G = 0.95, pm = 0.01
BBOA [68]C = 0.5, α = 0.1, p = 0.7
BGOA-TVG α 0 ,   5 , β 0.05 ,   10 ,   s i g m a 0.01 ,   10 ,   i t e r v a l 0 ,   2.079 ,   l = 1.5 ,   f = 0.5 ,   C m a x = 1 ,   C m i n = 0.00001
Table 3. Test results.
Table 3. Test results.
AlgorithmMeasure
DatasetAverage FitnessAverage AccuracyAverage SizeRank
BDA [23]Cancer0.01060.99187.40002
BHHO [20]0.00471.000014.00001
BPSO [11]0.03240.97376.00006
BGWO [14]0.02900.974310.66675
BGBO [66]0.00171.00004.16671
BGA [67]0.01311.00008.76671
BWOA [17]0.02250.982014.15004
BGOA [65]0.00250.990116.90003
BGOA-TVG0.00161.00005.20001
BDA [23]Breast0.75800.23882.50004
BHHO [20]0.67790.23818.06405
BPSO [11]0.82300.17365.20006
BGWO [14]0.67790.32117.06403
BGBO [66]0.50470.49042.50002
BGA [67]0.87100.12453.16197
BWOA [17]0.84810.14987.59648
BGOA [65]0.83660.16115.22659
BGOA-TVG0.89960.56982.50001
BDA [23]Wine0.00560.99865.45002
BHHO [20]0.00321.00004.12501
BPSO [11]0.06870.98982.66673
BGWO [14]0.00431.00005.65001
BGBO [66]0.00151.00002.06671
BGA [67]0.00341.00004.40001
BWOA [17]0.00260.97645.10004
BGOA [65]0.00391.00005.40001
BGOA-TVG0.00151.00001.90001
BDA [23]Ionosphere0.07290.92919.03335
BHHO [20]0.07190.931413.50004
BPSO [11]0.06420.939112.96673
BGWO [14]0.11720.886215.43338
BGBO [66]0.03260.96813.36671
BGA [67]0.06830.93147.80004
BWOA [17]0.08900.913611.55006
BGOA [65]0.09100.911912.90007
BGOA-TVG0.41120.94414.25602
BDA [23]Leukemia0.00361.00002560.76671
BHHO [20]0.14190.85951997.30002
BPSO [11]0.14630.85713452.63333
BGWO [14]0.00491.00003524.76671
BGBO [66]0.00001.000018.03331
BGA [67]0.00441.00003170.10001
BWOA [17]0.07560.81193498.63334
BGOA [65]0.19150.81193787.06674
BGOA-TVG0.00001.000021.05651
BDA [23]Cloud0.42780.8076467.90008
BHHO [20]0.49370.8302469.66675
BPSO [11]0.63480.8377466.16673
BGWO [14]0.39560.8170650.03337
BGBO [66]0.14610.839353.46674
BGA [67]0.45450.7843429.66679
BWOA [17]0.49730.9145438.10002
BGOA [65]0.49470.8226502.36676
BGOA-TVG0.10650.920049.57601
BDA [23]Vote0.00221.00003.55001
BHHO [20]0.01760.98504.35005
BPSO [11]0.05900.96671.90007
BGWO [14]0.00340.99944.63332
BGBO [66]0.00131.00002.00001
BGA [67]0.05760.94565.96678
BWOA [17]0.01090.99255.55003
BGOA [65]0.02200.98227.10003
BGOA-TVG0.00560.99062.23364
BDA [23]Wall-Following Robot Navigation Data0.02310.80768.25008
BHHO [20]0.01830.83029.5000 s5
BPSO [11]0.15480.83775.13334
BGWO [14]0.04850.817010.36677
BGBO [66]0.00590.83935.00003
BGA [67]0.13580.78437.90009
BWOA [17]0.02060.914523.68422
BGOA [65]0.02310.822612.16673
BGOA-TVG0.00550.91772.00001
BDA [23]Sonar0.05820.945224.06675
BHHO [20]0.08010.923526.00007
BPSO [11]0.15470.933321.60006
BGWO [14]0.05710.946022.00004
BGBO [66]0.05430.947610.22223
BGA [67]0.05230.950821.36672
BWOA [17]0.12060.883330.75009
BGOA [65]0.11820.886534.83338
BGOA-TVG0.05001.00009.52101
BDA [23]Zoo0.00321.00005.10001
BHHO [20]0.00341.00005.40001
BPSO [11]0.12510.94004.43334
BGWO [14]0.00291.00004.66671
BGBO [66]0.00201.00003.10531
BGA [67]0.00221.00003.50001
BWOA [17]0.00350.948125.30003
BGOA [65]0.00890.99679.00002
BGOA-TVG0.00091.00004.00001
The best results are shown in bold.
Table 4. Transfer functions.
Table 4. Transfer functions.
Name (S-Shaped Family)Transfer Function
S1 T x = 1 1 + e x
S2 T x = 1 1 + e 2 x
S3 T x = 1 1 + e x 2
S4 T x = 1 1 + e x 3
Name (V-shaped family)Transfer function
V1 T x = 2 π 0 π 2 x e t 2 d t
V2 T x = tanh ( x )
V3 T x = x 1 + x 2
V4 T x = 2 π arctan ( π 2 x )
Table 5. Characteristics of compared methods.
Table 5. Characteristics of compared methods.
SchemeOptimization AlgorithmIndividual ExpressionObjective FunctionExtracted FeaturesClassifier
AllfeatNo-TPR, PPV, TNR, NPV, ACCOptimizedOptimized
BPSOYesBinaryTPR, PPV, TNR, NPV, ACCOptimizedOptimized
BGAYesBinaryTPR, PPV, TNR, NPV, ACCOptimizedOptimized
BHHO YesBinaryTPR, PPV, TNR, NPV, ACCOptimizedOptimized
BWOAYesBinaryTPR, PPV, TNR, NPV, ACCOptimizedOptimized
BACOYesBinaryTPR, PPV, TNR, NPV, ACCOptimizedOptimized
BGWOYesBinaryTPR, PPV, TNR, NPV, ACCOptimizedOptimized
BGOA-TVGYesBinaryTPR, PPV, TNR, NPV, ACCOptimizedOptimized
Optimized: means that the item value is not constant and chosen by the corresponding optimization algorithm.
Table 6. Best classifiers for ten subjects of DEAP obtained by different algorithms.
Table 6. Best classifiers for ten subjects of DEAP obtained by different algorithms.
Subject No.12345678910
Allfeat [53]RFRFRFDTRFBayesBayesRFRFRF
BPSO [53]DTDTRFDTDTDTDTRFDTDT
BFA [53]DTDTRFDTDTBayesRFRFDTDT
BGARFDTKNNAdaboostDTAdaboostSVMBayesSVMKNN
BHHOBayesKNNBayesKNNDTBayesKNNBayesSVMSVM
BWOADTBayesSVMKNNSVMKNNSVMBayesRFBayes
BACODTSVMDTAdaboostKNNBayesAdaboostSVMBayesBayes
BGWORFAdaboostBayesDTRFKNNSVMBayesSVMAdaboost
BGOA-TVGDTDTDTAdaboostKNNBayesRFRFAdaboostAdaboost
Table 7. Mean values of classification results of algorithms for DEAP data.
Table 7. Mean values of classification results of algorithms for DEAP data.
AlgorithmTPRPPVTNRNPVACCFeature Number
Allfeat [53]0.68480.62090.71570.77860.75802224
BPSO [53]0.84910.78140.79400.86470.84001087
BFA [53]0.87600.93550.85770.81850.921088
BGA0.71140.78620.79530.71160.69981065
BHHO0.85690.79620.56850.6230.861152
BWOA 0.89660.790.84100.8690.93752
BACO0.79520.65590.69740.76650.7451526
BGWO 0.89660.86630.80.81420.896711
BGOA−TVG0.91160.91120.84650.8520.942698
The best results are shown in bold.
Table 8. Test results.
Table 8. Test results.
AlgorithmMeasure (Error Rate)
DatasetAverage ErrorMinimum ErrorRuntimesRank
BGASub-0010.25220.0006302
BHHO [17]0.13870.0064305
BPSO [50]0.09830.0114306
BGWO0.24350.0045304
BBOA 0.07830.0121307
BACO0.16810.0007303
BWOA [46]0.13330.0165308
BGOA−TVG0.04350301
BGASub-0020.52630.0039308
BHHO [17]0.44650.014306
BPSO [50]0.39470.0051305
BGWO0.52630.001303
BBOA 0.42110.0185307
BACO0.42110.0026304
BWOA [46]0.44740.0007302
BGOA−TVG0.27370.0006301
BGASub-0030.28450.0083307
BHHO [17]0.29820.0006304
BPSO [50]0.30170.0074306
BGWO0.31030.0094308
BBOA 0.28450.0001303
BACO0.27590.007305
BWOA [46]0.26720301
BGOA−TVG0.27590301
BGASub-0040.07140.0022304
BHHO [17]0.03330.0183307
BPSO [50]0.04290.0198308
BGWO0.04290.0002302
BBOA 0.02860.0027305
BACO0.04290.0031306
BWOA [46]0.05710301
BGOA−TVG0.00710.0011303
BGASub-0050.41180.0197308
BHHO [17]0.40040.0074305
BPSO [50]0.38210.0171307
BGWO0.38240.0157306
BBOA 0.39710.0033303
BACO0.38240.0024302
BWOA [46]0.42650.0061304
BGOA−TVG0.38210.0010301
BGASub-0060.31940.0074305
BHHO [17]0.32060.0114306
BPSO [50]0.29170.0261308
BGWO0.30560.0003302
BBOA 0.33330.0038303
BACO0.30560.0043304
BWOA [46]0.33330.0177307
BGOA−TVG0.30560301
BGASub-0070.03810.0053304
BHHO [17]0.05290.0258308
BPSO [50]0.23810.0113307
BGWO0.19520.0052306
BBOA 0.04760301
BACO0.03380.0043305
BWOA [46]0.02380.0006303
BGOA−TVG0.01760.0005302
BGASub-0080.85710.0119306
BHHO [17]0.25920.0026303
BPSO [50]0.02860.0083305
BGWO0.05710.0160308
BBOA 0.17140.0122307
BACO0.08570.0028304
BWOA [46]0.17190.0013302
BGOA−TVG0.01430.0006301
BGASub-0090.10000.0256308
BHHO [17]0.20180.0146307
BPSO [50]0.20000.0048303
BGWO0.15000.0142306
BBOA 0.25000.0057304
BACO0.10000.0018301
BWOA [46]0.20000.0099305
BGOA−TVG0.15000.0019302
BGASub-0100.23430.0008304
BHHO [17]0.14180.0061307
BPSO [50]0.16350.0132308
BGWO0.06610.0005303
BBOA 0.07100.0047306
BACO0.15720.0008304
BWOA [46]0.05830.0004302
BGOA−TVG0.00540301
BGASub-0110.03550.0011303
BHHO [17]0.02770.0043306
BPSO [50]0.03550.0114308
BGWO0.07960.0067307
BBOA 0.15190.0014304
BACO0.16230.0027305
BWOA [46]0.00660.0004301
BGOA−TVG0.00840.0006302
BGASub-0120.06790.0569304
BHHO [17]0.07600.0698305
BPSO [50]0.01200.0069303
BGWO0.00250.0005302
BBOA 0.08440.0775306
BACO0.19030.0960307
BWOA [46]0.00880.1336308
BGOA−TVG0.00830301
The best results are shown in bold.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, M.; Luo, Q.; Zhou, Y. BGOA-TVG: Binary Grasshopper Optimization Algorithm with Time-Varying Gaussian Transfer Functions for Feature Selection. Biomimetics 2024, 9, 187. https://doi.org/10.3390/biomimetics9030187

AMA Style

Li M, Luo Q, Zhou Y. BGOA-TVG: Binary Grasshopper Optimization Algorithm with Time-Varying Gaussian Transfer Functions for Feature Selection. Biomimetics. 2024; 9(3):187. https://doi.org/10.3390/biomimetics9030187

Chicago/Turabian Style

Li, Mengjun, Qifang Luo, and Yongquan Zhou. 2024. "BGOA-TVG: Binary Grasshopper Optimization Algorithm with Time-Varying Gaussian Transfer Functions for Feature Selection" Biomimetics 9, no. 3: 187. https://doi.org/10.3390/biomimetics9030187

Article Metrics

Back to TopTop