Next Article in Journal
Machine Learning Strategies for Reconfigurable Intelligent Surface-Assisted Communication Systems—A Review
Previous Article in Journal
SmartDED: A Blockchain- and Smart Contract-Based Digital Electronic Detonator Safety Supervision System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Optimization Techniques in Grammatical Evolution

by
Ioannis G. Tsoulos
*,
Alexandros Tzallas
and
Evangelos Karvounis
Department of Informatics and Telecommunications, University of Ioannina, 47150 Kostaki Artas, Greece
*
Author to whom correspondence should be addressed.
Future Internet 2024, 16(5), 172; https://doi.org/10.3390/fi16050172
Submission received: 2 April 2024 / Revised: 1 May 2024 / Accepted: 13 May 2024 / Published: 16 May 2024

Abstract

:
The Grammatical Evolution technique has been successfully applied to a wide range of problems in various scientific fields. However, in many cases, techniques that make use of Grammatical Evolution become trapped in local minima of the objective problem and fail to reach the optimal solution. One simple method to tackle such situations is the usage of hybrid techniques, where local minimization algorithms are used in conjunction with the main algorithm. However, Grammatical Evolution is an integer optimization problem and, as a consequence, techniques should be formulated that are applicable to it as well. In the current work, a modified version of the Simulated Annealing algorithm is used as a local optimization procedure in Grammatical Evolution. This approach was tested on the Constructed Neural Networks and a remarkable improvement of the experimental results was shown, both in classification data and in data fitting cases.

1. Introduction

Genetic Algorithms belong to the field of evolutionary techniques [1] and were originally formulated by John Hollands and his team [2]. Genetic Algorithms are initiated by generating a series of random candidate solutions to an optimization problem. These candidate solutions are called chromosomes, and they iteratively undergo a series of operations that have their foundation in physical processes, such as selection, crossover, and mutation [3,4,5]. Grammatical Evolution [6] is considered a special case of Genetic Algorithms, where the chromosomes are series of integer numbers. Chromosomes in the Grammatical Evolution process are rules for generating a Backus–Naur form (BNF) grammar [7] and can be used to create functional programs in any programming language.
Grammatical Evolution has been applied on a wide series of real-world applications, such as function approximation [8,9], credit classification [10], network security and prevention of attacks [11], monitoring of water quality [12], modeling glycemia in humans [13], automatic design of Ant algorithms [14], temperature prediction in data centers [15], solving trigonometric equations [16], composing music [17], neural network construction [18,19], producing numeric constants [20], video games [21,22], energy demand estimation [23], combinatorial optimization [24], cryptography [25], evolving of decision trees [26], automatic design of analog electronic circuits [27], etc.
The method of Grammatical Evolution has been extended by various researchers during recent years and some examples of these extensions are Structured Grammatical Evolution [28,29], which applies a one-to-one mapping between the chromosomes and the non-terminal symbols of the grammar; the π Grammatical Evolution method [30], an application of the Particle Swarm Optimization(PSO) [31] to produce programs with Grammatical Evolution denoted as Grammatical Swarm [32,33]; the Probabilistic Grammatical Evolution [34], which introduced a new mapping mechanism for the Grammatical Evolution method, incorporation of parallel programming techniques [35,36], usage of Christiansen grammars [37], etc.
Also, a variety of software has been developed for Grammatical Evolution, such as the GEVA v2.0 [38] that proposes a GUI application for Grammatical Evolution, the gramEvol v2.1-4 software [39] that provides a package in the R programming language, the GeLab v2.0  [40] that implements a Matlab toolbox for Grammatical Evolution, the GenClass v1.0 [41] software used to produce classification programs, the QFc v1.10 software [42] for feature construction, etc.
Although the method of Grammatical Evolution has proven to be extremely efficient, it can often become trapped in the local minimum of the objective problem and its performance may not be as expected. A common method to get out of such situations is the use of local optimization techniques, which have been applied many times in combination with Genetic Algorithms [43,44]. In the case of Grammatical Evolution and due to the integral representation of the candidate solutions, local optimization techniques are not directly applicable, such as, for example, the BFGS method [45] and hence more suitable methods should be adopted. In this paper, an attempt is made to improve upon the excellent results shown by Grammatical Evolution in the past on a variety of problems by applying local optimization techniques, periodically applied to randomly selected chromosomes of Grammatical Evolution. The current work utilizes a modified version of the Simulated Annealing method [46] as a local search procedure. The method of Simulated Annealing has been utilized in various cases, such as image processing [47], protein structure optimization [48], resource allocation [49], convex optimization [50], deep learning [51], etc. To verify the possibility of the proposed methodology to improve the results of Grammatical Evolution, the neural networks construction technique was chosen. This method was initially provided in [52] and it can estimate the topology and the weights of neural networks using the Grammatical Evolution procedure. This method was chosen to test the present methodology because of its many applications. However, the technique proposed in this paper will be able to be applied in the future in other cases of using Grammatical Evolution. To evaluate the performance of the modified technique it was evaluated on a extended series of classification and regression problems found in the relevant literature and the results seem to be promising.
In the same direction, many researchers have published papers in neural network initialization or construction, such as the usage of decision trees to initialize the weights of a neural network [53], initialization of the weights using Cauchy’s inequality [54], and the application of discriminant learning [55]. Also, the issue of constructing the structure of artificial neural networks has been discussed in various papers, such as incorporation of genetic algorithms [56], construction and pruning of the weights [57], application of cellular automata [58], etc.
This paper has the following sections: in Section 2 the Grammatical Evolution procedure is discussed and the proposed modification is described. Section 3 presents the datasets used in the experiments as well as the results from the conducted experiments, and finally Section 4 provides some conclusions and guidelines for future work.

2. The Proposed Method

This section starts with a brief description of the Grammatical Evolution process and the grammar used, then the modified Simulated Annealing method is presented, and finally the artificial neural network construction algorithm will be presented.

2.1. The Grammatical Evolution Method

The chromosomes of Grammatical Evolution stand for production rules of the given BNF grammar. BNF grammars are usually defined as a tuple G = N , T , S , P , where:
  • N is the set of non-terminal symbols.
  • T is the set of terminal symbols.
  • S represents that start symbol of the grammar with S N .
  • P is the production rules of the grammar. Usually these rules are in the form A a or A a B , A , B N , a T .
The BNF grammar is extended by using enumeration in the the production rules. The grammar shown in Figure 1 is used to construct artificial neural networks expressed in the form:
N x , w = i = 1 H w ( d + 2 ) i ( d + 1 ) σ j = 1 d x j w ( d + 2 ) i ( d + 1 ) + j + w ( d + 2 ) i
where the parameter H represents the number of processing units (hidden nodes ) for the neural network and the value d represents the of the vector x . Also, the vector w represents the vector of parameters for the neural network. The function sig(x) represents the sigmoid function σ ( x ) defined as:
σ ( x ) = 1 1 + exp ( x )
The notation <> is used for the non-terminal symbols of the grammar. The sequence numbers of the production rules are inside the parentheses of the extended grammar. The constant d stands for the dimension of the provided dataset. The production of valid expressions initiates from the symbol S and using a series of steps, creates a program by replacing non-terminal symbols with the right hand of the selected production rule. Grammatical Evolution selects production rules using the following scheme:
  • Obtain the next element V from the provided chromosome.
  • The production rule is selected according to the scheme
    Rule = V mod NR
    where NR is the total number of production rules for the current non-terminal symbol.

2.2. The Modified Simulated Annealing Algorithm

The current work utilizes a modified version of the Simulated Annealing algorithm as a local search procedure. Simulated Annealing was chosen as a local search method, since it offers the possibility of representing the considered solutions in an integer form, which is critical for the Grammatical Evolution representation of chromosomes. In addition, this method has been distinguished for its easy adaptation to a multitude of problems but also for its ability to find the total minimum of functions through the point acceptance mechanism at high temperatures. In the present work, the Simulated Annealing will initiate from the current representation of a chromosome and gradually try to find other nearby representations that might lead to lower values of the fitness value. The main steps of this procedure are shown in Algorithm 1.
Algorithm 1 The modified version of the Simulated Annealing algorithm
procedure siman x 0  
  • Set  k = 0 , T 0 > 0 , ϵ > 0 ,   r T > 0 , r T < 1 ..
  • Set  N e p s > 0 , a positive integer indicated the number of samples drawn in every iteration.
  • Set g and R, positive integer values.
  • Set  x b = x 0 , f b = f x b .
  • For  i = 1 N e p s
    (a) 
    Set  y = x k  
    (b) 
    For  j = 1 g
    i. 
    Set  p = rand 1 , size x 0 . The variable p indicates a randomly selected position in y.
    ii. 
    Set  y p = y p + rand R , R  
    (c) 
    EndFor
    (d) 
    If f ( y ) f x k then x k + 1 = y  
    (e) 
    Else Set  x k + 1 = y with probability min 1 , exp f ( y ) f x k T k  
    (f) 
    If  f y < f b  then  x b = y , f b = f ( y ) .
  • EndFor
  • Set  T k + 1 = T k r T  
  • Set  k = k + 1 .
  • If T k ϵ terminate
  • Goto step 5.
  • Return  x b  
end siman
The method accepts the chromosome x 0 as a starting point and, in each iteration, randomly generates chromosomes around it. The parameter g controls the number of changes that will be made to the chromosome and the parameter R controls the radius of these changes. The parameter T represents the temperature of the algorithm. The temperature starts at extremely high values and progressively decreases linearly. In the early stages and at high temperatures, the algorithm may accept points that may have higher values of the fitness function, but as the temperature decreases, this probability also decreases. The used Simulated Annealing variant is also shown graphically as a flow chart in Figure 2.

2.3. The Neural Network Construction Algorithm

The main steps of the algorithm used to construct artificial neural networks with Grammatical Evolution are listed below:
  • Initialization step:
    (a)
    Set  N g as the maximum number of generations allowed.
    (b)
    Set  N c as the number of chromosomes.
    (c)
    Set  p s as the selection rate and p m as the mutation rate.
    (d)
    Define as L i the number of generations that should elapse before applying the local optimization technique.
    (e)
    Define as L c the number of chromosomes involved in the local search procedure.
    (f)
    Initialize the chromosomes. Each chromosome is considered as a series of randomly initialized integers.
    (g)
    Set iter=0.
  • Genetic step:
    (a)
    For i = 1 , , N g do
    • Create for every chromosome a neural network C i using the Grammatical Evolution procedure of Section 2.1 and the associated grammar given in Figure 1.
    • Calculate the fitness f i on the train set of the objective problem as:
      f i = j = 1 M C i x j t j 2
      where the set x j , t j , j = 1 , . . . M is train dataset, with t i being the actual output for the point x i .
    • Perform the selection procedure. Initially, the chromosomes are sorted according to their fitness values. The best 1 p s × N c chromosomes are transferred to the next generation. The remaining chromosomes will be replaced by offspring created during the crossover procedure.
    • Perform the crossover procedure. The crossover procedure produces p s × N c offspring. For every pair of produced offspring z ˜ and w ˜ , there are two offspring ( z , w ) . The selection is performed using tournament selection. The new offspring are produced using the one-point crossover procedure. An example of the one-point crossover procedure is shown in Figure 3.
    • Perform the mutation procedure. For each element of every chromosome, a random number r 0 , 1 is drawn. The corresponding element is altered if r p m .
    (b)
    EndFor
  • Local Search step:
    (a)
    If  iters mod L i = 0 , Then
    • Create a set of randomly L C chosen chromosomes from the genetic population. Denote this set as L S .
    • For every X k in L S apply the modified Simulated Annealing algorithm given in Algorithm 1: X k = siman X k  
  • set iter = iter + 1. If iter > N g  goto Evaluation step else goto Genetic step.
  • Evaluation step:
    (a)
    Obtain the chromosome with the lowest fitness value and create the associated neural network C .
    (b)
    Evaluate the neural network C in the test set of the underlying dataset and report the results.

3. Results

The current work was evaluated by executing a series of experiments on some classification and regression datasets, which are commonly used in the relevant literature. The obtained results were compared with other machine learning techniques. These datasets can be downloaded freely from the following websites:

3.1. Classification Datasets

The following series of classification datasets was used in the conducted experiments:
  • Appendictis a medical dataset, provided in [61].
  • Australian dataset [62], which is related to credit card transactions.
  • Balance dataset [63], a dataset related to psychological experiments.
  • Circular dataset, an artificial dataset that contains 1000 examples.
  • Cleveland dataset, a medical dataset used in a series of papers [64,65].
  • Dermatology dataset [66], which is a medical dataset about dermatological deceases.
  • Ecoli dataset, a dataset about protein localization sites [67].
  • Haberman dataset, related to breast cancer.
  • Heart dataset [68], a medical dataset about heart diseases.
  • Hayes roth dataset [69], which is a human subjects study.
  • HouseVotes dataset [70], related to votes collected from U.S. House of Representatives Congressmen.
  • Ionosphere dataset, which was used in experiments related to the ionosphere [71,72].
  • Liverdisorder dataset [73], a medical dataset related to liver disorders.
  • Mammographic dataset [74], a medical dataset related to breast cancer.
  • Parkinsons dataset, a medical dataset related to Parkinson’s disease (PD) [75].
  • Pima dataset [76], a medical dataset related to the diabetes presence.
  • Popfailures dataset [77], a dataset related to climate measurements.
  • Regions2 dataset, medical dataset related to hepatitis C [78].
  • Saheart dataset [79]. This dataset is used to detect heart diseases.
  • Segment dataset [80], related to image processing.
  • Student dataset [81], related to data collected in Portuguese schools.
  • Transfusion dataset [82], this dataset was taken from the Blood Transfusion Service Center in Hsin-Chu City in Taiwan.
  • Wdbc dataset [83], a medical dataset related cancer detection.
  • Wine dataset. This is a dataset used to detect the quality of a series of wines [84,85].
  • Eeg datasets, a dataset related to EEG measurements [86]. From this dataset the following cases were used: Z_F_S, Z_O_N_F_S, ZO_NF_S, and ZONF_S.
  • Zoo dataset [87], related to animal classification.

3.2. Regression Datasets

The following regression datasets were used in the conducted experiments:
  • Abalone dataset [88], a dataset related to the prediction of age of abalones.
  • Airfoil dataset, a dataset proposed by NASA [89].
  • Baseball dataset, related with the income of baseball players.
  • Concrete dataset [90], which is a civil engineering dataset.
  • Dee dataset. This dataset has measures from the price of electricity.
  • HO dataset, downloaded from the STALIB repository.
  • Housing dataset, mentioned in [91].
  • Laser dataset. This is a dataset related to laser experiments
  • LW dataset, related to risk factors associated with low-weight babies.
  • MORTGAGE dataset, a dataset related to economic measurements from the USA.
  • PL dataset, provided from the STALIB repository.
  • SN dataset, provided from the STALIB repository.
  • Treasury dataset, a dataset related to economic measurements from the USA.
  • TZ dataset, provided from the STALIB repository.

3.3. Experimental Results

For the execution of the experiments, code written in ANSI C++ was used with the help of the programming environment Optimus. The software is freely available from https://github.com/itsoulos/OPTIMUS/ (accessed on 20 March 2024). The experiments were conducted 30 times. In every execution different seed was used for the random number generator and the function drand48() of the C programming language was used. The validation of the results was performed using the technique of 10-fold cross validation. The average classification error is reported for the case of classification datasets and the average regression error for the case of regression datasets. These errors are measured on the corresponding test set. The values for the experimental parameters are displayed in Table 1. The experimental results for the classification datasets are outlined in Table 2 and the results for the regression datasets are shown in Table 3. The following applies to the tables with the experimental results:
  • The column DATASET denotes the used dataset.
  • The column ADAM denotes the application of the ADAM optimization method [92] in an artificial neural network with H = 10 processing nodes.
  • The column NEAT stands for the usage of NEAT method (NeuroEvolution of Augmenting Topologies ) [93].
  • The column MLP stands for the experimental results of an artificial neural network with H = 10 processing nodes. The neural network was trained using a Genetic Algorithm and the BFGS local optimization method [94].
  • The column RBF represents the application of an RBF network with H = 10 processing nodes in each dataset.
  • The column NNC denotes the usage of the original Neural Construction technique, which was constructed with Grammatical Evolution.
  • The column NNC-S denotes the usage of the proposed local optimization procedure in the Neural Construction technique.
  • The line AVERAGE denotes the average classification or regression error.
The above techniques were used in the experiments as they are widespread in machine learning, such as the Adam method [95,96] and because they have a similar complexity to the present technique, such as Genetic Algorithms. As is evident, the proposed modification improves the efficiency of the proposed method in the majority of used datasets. This improvement on some datasets could be as much as an 80% percent error reduction on the test set. In all cases, the simple neural network construction technique outperforms other machine learning techniques and this is evident from the average result (the last row in experimental tables). However, the proposed method significantly reduces the classification error or the error in the data fitting sets in most of the cases where it was used. In fact, in order to show in which cases there was a reduction in the error, bold marking is used in the tables of results.
Figure 4 is a scatter plot that provides a detailed comparative analysis of classification error percentages for six distinct machine learning and optimization algorithms: ADAM, NEAT, GENETIC, RBF, NNC, and NNC-S. Each point on the plot represents the outcome of an individual run of a model, thus showcasing the range of variation in the error rates associated with these classification methods. The vertical dispersion of points for each method reflects the spread of error rates, which is critical for evaluating the reliability of each algorithm. The medians of these error rates are indicated by horizontal lines intersecting the clusters of dots, offering a snapshot of each algorithm’s central tendency in performance. The asterisk-based notation above the clusters denotes statistical significance levels: one asterisk (*) signifies a p-value less than 0.05, two asterisks (**) denote p < 0.01, three asterisks (***) represent p < 0.001, and four asterisks (****) indicate an extremely low p-value of less than 0.0001, suggesting strong evidence against the null hypothesis. Of note is the performance of the NNC-S method, which not only shows a significantly lower median error rate when compared to the NNC method but also displays a notably tighter cluster of data points. This implies that the NNC-S method not only tends to be more accurate on average but also provides greater consistency in its error rates across different runs, underscoring its robustness as a classification tool.
Building on the previous analysis of classification methods, Figure 5 extends the evaluation to regression algorithms, providing a comparative study of regression error rates for the same six methods: ADAM, NEAT, GENETIC, RBF, NNC, and NNC-S. Each data point reflects the regression error from a specific trial, with the array of points for each method revealing the range of performance outcomes. The median error rates are again represented by horizontal lines across the clusters of dots, serving as a summary statistic that facilitates a direct comparison of the method’s central performance trends. The notation of statistical significance is consistent with the previous figure, where asterisks convey the p-value levels, identifying statistically meaningful differences in performance between the methods.
This plot reveals that the NNC-S method maintains its superior performance in the context of regression tasks, demonstrating lower median regression errors compared to the other methods. Significantly, it achieves a markedly lower median regression error than its predecessor, NNC, as indicated by the blue dots and supported by the three asterisks (***). This pattern of results underscores the broader applicability of the NNC-S method’s local optimization enhancements, not only in classification accuracy but also in reducing regression errors.
Additionally, the effectiveness and the robustness of the proposed method was evaluated by performing additional experiments with different values for the critical parameter g of the suggested Simulated Annealing variant. This parameter is used to control the number of random changes performed on any given chromosome. The experimental results in the classification datasets are shown in Table 4. Clearly, no significant variation in the performance of the proposed technique occurs when this critical parameter is varied.
Continuing from the previous analysis, Figure 6 presents the classification error rates for variations of the NNC-S algorithm with differing values of the parameter g. The plot aims to evaluate whether changes in the parameter g led to statistically significant differences in the algorithm’s classification performance. The horizontal bars on the plot indicate the median classification error for each variation of the NNC-S method, providing a clear comparison across the different parameter values. The statistical annotations (“ns” for not significant, “**” for p < 0.01, and “***” for p < 0.001) are used to denote the statistical significance of the differences between the groups. It appears that some variations, particularly between NNC-S ( g = 2 ) and NNC-S ( g = 5 ), as well as between NNC-S ( g = 10 ) and NNC-S ( g = 20 ), do not show significant differences in performance (denoted by “ns”). In contrast, other comparisons do reveal significant differences, suggesting that certain values of g can indeed impact the classification error rates of the NNC-S algorithm.
In addition, in the graph of Figure 7, a comparison is made for the average execution time of each experiment for the classification datasets. The comparison was made between the initial method (denoted as NNC in the graph) as well as the various cases of the proposed technique by changing the parameter g to 5, 10, and 20. As expected, adding the local minimization technique to the Grammatical Evolution method significantly increases the execution time of the method; however, this increase remains almost constant for different values of the critical parameter g. Moreover, by using parallel techniques that appear in the relevant literature this increase in execution time could be reduced.
Furthermore, another experiment was executed by varying the parameter R of the proposed Simulated Annealing variant. This parameter controls the range of changes performed in any given chromosome. The results for this experiment are shown in Table 5. This time the method appeared quite robust in its performance, without large variations in the error as measured in the test set.
As a practical application of the suggested method, we consider the dataset proposed in [97], which relates alcohol consumption and EEG recordings. The proposed method was applied on this dataset and it is compared against the original neural network construction technique; the results are outlined in Figure 8.
Once again, the present method outperforms the original artificial neural network construction technique and the performance appears to be stable as the critical parameter g grows.

4. Conclusions

In the current work, an extension of the original Grammatical Evolution has been proposed, which was applied in the Neural Construction method. In this extension, an application of a modified optimization method was suggested, in order to improve the efficiency of the underlying technique. The proposed optimization algorithm was a variant of the Simulated Annealing method, which was applied on a series of chromosomes that was randomly from the Grammatical Evolution procedure. This method was chosen because of its widespread use in many applications and because of its ability to handle integer representations, since the chromosomes in Grammatical Evolution are vectors of integers. Of course, other local optimization techniques could be incorporated, such as tabu search [98] or hill climbing [99] as feature extensions of the proposed method.
The proposed modification was applied to the neural network construction method and its efficiency was measured on some commonly used classification and regression datasets. Based on the experimental results, it has become clear that the proposed variant significantly improves the performance of the technique both on classification datasets and on data fitting datasets. Also, the effectiveness and the robustness of the proposed modification were measured using experiments with different values of some critical parameters of the current Simulated Annealing variant. These experiments indicated that the method tends to be robust, since the experimental results do not depend on the selection of any critical parameter of the Simulated Annealing variant.
The present modification could be applied to other techniques that use Grammatical Evolution without significant differences and, moreover, it could also be used to optimize functions either without constraints or with constraints in similar techniques that have been proposed in recent years. However, as was seen from the experimental results, the addition of the Simulated Annealing technique significantly increases the required execution time and it is necessary to use techniques that are not particularly demanding in time or to search for termination techniques of the proposed Simulated Annealing variant that take advantage of its particular characteristics. Also, a field of research could be the search for more effective techniques to reduce the critical temperature factor used in the Simulated Annealing.
Future work may include application of the local search procedure in other Grammatical Evolution applications, such as feature construction, creation of classification rules, etc. Also, the proposed process can be significantly accelerated by the use of parallel optimization techniques [100], which take advantage of modern computational structures.

Author Contributions

I.G.T., A.T. and E.K. conceived the idea and methodology and supervised the technical part regarding the software. I.G.T. conducted the experiments, employing several datasets, and provided the comparative experiments. A.T. performed the statistical analysis. E.K. and all other authors prepared the manuscript. E.K. and I.G.T. organized the research team and A.T. supervised the project. All authors have read and agreed to the published version of the manuscript.

Funding

This research has been financed by the European Union: Next Generation EU through the Program Greece 2.0 National Recovery and Resilience Plan, under the call RESEARCH – CREATE – INNOVATE, project name “iCREW: Intelligent small craft simulator for advanced crew training using Virtual Reality techniques” (project code: TAEDK-06195).

Data Availability Statement

The data presented in this study are available in this article and they can be downloaded from the mentioned URL sources.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yusup, N.; Zain, A.M.; Hashim, S.Z.M. Evolutionary techniques in optimizing machining parameters: Review and recent applications (2007–2011). Expert Syst. Appl. 2012, 39, 9909–9927. [Google Scholar] [CrossRef]
  2. Holland, J.H. Genetic algorithms. Sci. Am. 1992, 267, 66–73. [Google Scholar] [CrossRef]
  3. Stender, J. Parallel Genetic Algorithms: Theory & Applications; IOS Press: Amsterdam, The Netherlands, 1993. [Google Scholar]
  4. Goldberg, D. Genetic Algorithms in Search, Optimization and Machine Learning; Addison-Wesley Publishing Company: Reading, MA, USA, 1989. [Google Scholar]
  5. Michaelewicz, Z. Genetic Algorithms + Data Structures = Evolution Programs; Springer: Berlin/Heidelberg, Germany, 1996. [Google Scholar]
  6. O’Neill, M.; Ryan, C. Grammatical evolution. IEEE Trans. Evol. Comput. 2001, 5, 349–358. [Google Scholar] [CrossRef]
  7. Backus, J.W. The Syntax and Semantics of the Proposed International Algebraic Language of the Zurich ACM-GAMM Conference. In Proceedings of the International Conference on Information Processing; UNESCO: Paris, France, 1959; pp. 125–132. [Google Scholar]
  8. Ryan, C.; Collins, J.; O’Neill, M. Grammatical evolution: Evolving programs for an arbitrary language. In Genetic Programming. EuroGP 1998. Lecture Notes in Computer Science; Banzhaf, W., Poli, R., Schoenauer, M., Fogarty, T.C., Eds.; Springer: Berlin/Heidelberg, Germany, 1998; Volume 1391. [Google Scholar]
  9. O’Neill, M.; Ryan, M.C. Evolving Multi-line Compilable C Programs. In Genetic Programming. EuroGP 1999. Lecture Notes in Computer Science; Poli, R., Nordin, P., Langdon, W.B., Fogarty, T.C., Eds.; Springer: Berlin/Heidelberg, Germany, 1999; Volume 1598. [Google Scholar]
  10. Brabazon, A.; O’Neill, M. Credit classification using grammatical evolution. Informatica 2006, 30, 325–335. [Google Scholar]
  11. Şen, S.; Clark, J.A. A grammatical evolution approach to intrusion detection on mobile ad hoc networks. In Proceedings of the Second ACM Conference on Wireless Network Security, Zurich, Switzerland, 16–19 March 2009. [Google Scholar]
  12. Chen, L.; Tan, C.H.; Kao, S.J.; Wang, T.S. Improvement of remote monitoring on water quality in a subtropical reservoir by incorporating grammatical evolution with parallel genetic algorithms into satellite imagery. Water Res. 2008, 42, 296–306. [Google Scholar] [CrossRef] [PubMed]
  13. Hidalgo, J.I.; Colmenar, J.M.; Risco-Martin, J.L.; Cuesta-Infante, A.; Maqueda, E.; Botella, M. Modeling glycemia in humans by means of Grammatical Evolution. Appl. Soft Comput. 2014, 20, 40–53. [Google Scholar] [CrossRef]
  14. Tavares, J.; Pereira, F.B. Automatic Design of Ant Algorithms with Grammatical Evolution. In Genetic Programming. EuroGP 2012. Lecture Notes in Computer Science; Moraglio, A., Silva, S., Krawiec, K., Machado, P., Cotta, C., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; Volume 7244. [Google Scholar]
  15. Zapater, M.; Risco-Martín, J.L.; Arroba, P.; Ayala, J.L.; Moya, J.M.; Hermida, R. Runtime data center temperature prediction using Grammatical Evolution techniques. Appl. Soft Comput. 2016, 49, 94–107. [Google Scholar] [CrossRef]
  16. Ryan, C.; O’Neill, M.; Collins, J.J. Grammatical Evolution: Solving Trigonometric Identities. In Proceedings of the Mendel 1998: 4th International Mendel Conference on Genetic Algorithms, Optimisation Problems, Fuzzy Logic, Neural Networks, Rough Sets, Brno, Czech Republic, 1–2 November 1998. [Google Scholar]
  17. Puente, A.O.; Alfonso, R.S.; Moreno, M.A. Automatic composition of music by means of grammatical evolution. In Proceedings of the APL ’02: Proceedings of the 2002 Conference on APL: Array Processing Languages: Lore, Problems, and Applications, Madrid, Spain, 22–25 July 2002; pp. 148–155. [Google Scholar]
  18. Lídio Mauro Limade Campo, R. Célio Limã Oliveira, Mauro Roisenberg, Optimization of neural networks through grammatical evolution and a genetic algorithm. Expert Syst. Appl. 2016, 56, 368–384. [Google Scholar]
  19. Soltanian, K.; Ebnenasir, A.; Afsharchi, M. Modular Grammatical Evolution for the Generation of Artificial Neural Networks. Evol. Comput. 2022, 30, 291–327. [Google Scholar] [CrossRef] [PubMed]
  20. Dempsey, I.; Neill, M.O.; Brabazon, A. Constant creation in grammatical evolution. Int. J. Innov. Comput. Appl. 2007, 1, 23–38. [Google Scholar] [CrossRef]
  21. Galván-López, E.; Swafford, J.M.; O’Neill, M.; Brabazon, A. Evolving a Ms. PacMan Controller Using Grammatical Evolution. In Applications of Evolutionary Computation. EvoApplications 2010. Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2010; Volume 6024. [Google Scholar]
  22. Shaker, N.; Nicolau, M.; Yannakakis, G.N.; Togelius, J.; O’Neill, M. Evolving levels for Super Mario Bros using grammatical evolution. In Proceedings of the 2012 IEEE Conference on Computational Intelligence and Games (CIG), Granada, Spain, 11–14 September 2012; pp. 304–331. [Google Scholar]
  23. Martínez-Rodríguez, D.; Colmenar, J.M.; Hidalgo, J.I.; Micó, R.J.V.; Salcedo-Sanz, S. Particle swarm grammatical evolution for energy demand estimation. Energy Sci. Eng. 2020, 8, 1068–1079. [Google Scholar] [CrossRef]
  24. Sabar, N.R.; Ayob, M.; Kendall, G.; Qu, R. Grammatical Evolution Hyper-Heuristic for Combinatorial Optimization Problems. IEEE Trans. Evol. Comput. 2013, 17, 840–861. [Google Scholar] [CrossRef]
  25. Ryan, C.; Kshirsagar, M.; Vaidya, G.; Cunningham, A.; Sivaraman, R. Design of a cryptographically secure pseudo random number generator with grammatical evolution. Sci. Rep. 2022, 12, 8602. [Google Scholar] [CrossRef]
  26. Pereira, P.J.; Cortez, P.; Mendes, R. Multi-objective Grammatical Evolution of Decision Trees for Mobile Marketing user conversion prediction. Expert Syst. Appl. 2021, 168, 114287. [Google Scholar] [CrossRef]
  27. Castejón, F.; Carmona, E.J. Automatic design of analog electronic circuits using grammatical evolution. Appl. Soft Comput. 2018, 62, 1003–1018. [Google Scholar] [CrossRef]
  28. Lourenço, N.; Pereira, F.B.; Costa, E. Unveiling the properties of structured grammatical evolution. Genet. Program. Evolvable Mach. 2016, 17, 251–289. [Google Scholar] [CrossRef]
  29. Lourenço, N.; Assunção, F.; Pereira, F.B.; Costa, E.; Machado, P. Structured grammatical evolution: A dynamic approach. In Handbook of Grammatical Evolution; Springer: Cham, Switzerland, 2018; pp. 137–161. [Google Scholar]
  30. O’Neill, M.; Brabazon, A.; Nicolau, M.; Garraghy, S.M.; Keenan, P. πGrammatical Evolution. In Genetic and Evolutionary Computation—GECCO 2004. GECCO 2004. Lecture Notes in Computer Science; Deb, K., Ed.; Springer: Berlin/Heidelberg, Germany, 2004; Volume 3103. [Google Scholar]
  31. Poli, R. James Kennedy kennedy, Tim Blackwell, Particle swarm optimization An Overview. Swarm Intell. 2007, 1, 33–57. [Google Scholar] [CrossRef]
  32. O’Neill, M.; Brabazon, A. Grammatical swarm: The generation of programs by social programming. Nat. Comput. 2006, 5, 443–462. [Google Scholar] [CrossRef]
  33. Ferrante, E.; Duéñez-Guzmán, E.; Turgut, A.E.; Wenseleers, T. GESwarm: Grammatical evolution for the automatic synthesis of collective behaviors in swarm robotics. In Proceedings of the 15th Annual Conference on Genetic and Evolutionary Computation, Amsterdam, The Netherlands, 6–10 July 2013; pp. 17–24. [Google Scholar]
  34. Mégane, J.; Lourenço, N.; Machado, P. Probabilistic Grammatical Evolution. In Genetic Programming. EuroGP 2021. Lecture Notes in Computer Science; Hu, T., Lourenço, N., Medvet, E., Eds.; Springer: Cham, Switzerland, 2021; Volume 12691. [Google Scholar]
  35. Popelka, O.; Osmera, P. Parallel Grammatical Evolution for Circuit Optimization. In Evolvable Systems: From Biology to Hardware. ICES 2008. Lecture Notes in Computer Science; Hornby, G.S., Sekanina, L., Haddow, P.C., Eds.; Springer: Berlin/Heidelberg, Germany, 2008; Volume 5216. [Google Scholar] [CrossRef]
  36. Ošmera, P. Two level parallel grammatical evolution. In Advances in Computational Algorithms and Data Analysis; Springer: Cham, Switzerland, 2009; pp. 509–525. [Google Scholar]
  37. Ortega, A.; Cruz, M.d.; Alfonseca, M. Christiansen Grammar Evolution: Grammatical Evolution with Semantics. IEEE Trans. Evol. Comput. 2007, 11, 77–90. [Google Scholar] [CrossRef]
  38. O’Neill, M.; Hemberg, E.; Gilligan, C.; Bartley, E.; McDermott, J.; Brabazon, A. GEVA: Grammatical evolution in Java. ACM Sigevolution 2008, 3, 17–22. [Google Scholar] [CrossRef]
  39. Noorian, F.; de Silva, A.M.; Leong, P.H.W. gramEvol: Grammatical Evolution in R. J. Stat. Softw. 2016, 71, 1–26. [Google Scholar] [CrossRef]
  40. Raja, M.A.; Ryan, C. GELAB—A Matlab Toolbox for Grammatical Evolution. In Intelligent Data Engineering and Automated Learning—IDEAL 2018. IDEAL 2018. Lecture Notes in Computer Science 2018; Springer: Cham, Switzerland, 2018; Volume 11315. [Google Scholar] [CrossRef]
  41. Anastasopoulos, N.; Tsoulos, I.G.; Tzallas, A. GenClass: A parallel tool for data classification based on Grammatical Evolution. SoftwareX 2021, 16, 100830. [Google Scholar] [CrossRef]
  42. Tsoulos, I.G. QFC: A Parallel Software Tool for Feature Construction, Based on Grammatical Evolution. Algorithms 2022, 15, 295. [Google Scholar] [CrossRef]
  43. Yang, S.; Jat, S.N. Genetic Algorithms with Guided and Local Search Strategies for University Course Timetabling. IEEE Trans. Syst. Man Cybern. Part C Appl. Rev. 2011, 41, 93–106. [Google Scholar] [CrossRef]
  44. Sivaram, M.; Batri, K.; Mohammed, A.S.; Porkodi, V. Exploiting the Local Optima in Genetic Algorithm using Tabu Search. Indian J. Sci. Technol. 2019, 12, 1–13. [Google Scholar] [CrossRef]
  45. Dai, Y.H. Convergence Properties of the BFGS Algoritm. SIAM J. Optim. 2002, 13, 693–701. [Google Scholar] [CrossRef]
  46. Kirkpatrick, S.; Gelatt, C.D.G., Jr.; Vecchi, M.P. Optimization by simulated annealing. Science 1983, 220, 671–680. [Google Scholar] [CrossRef] [PubMed]
  47. Robini, M.C.; Rastello, T.; Magnin, I.E. Simulated annealing; acceleration techniques; image restoration. IEEE Trans. Image Process. 1999, 8, 1374–1387. [Google Scholar] [CrossRef] [PubMed]
  48. Zhang, L.; Ma, H.; Qian, W.; Li, H. Protein structure optimization using improved simulated annealing algorithm on a three-dimensional AB off-lattice model. Comput. Biol. Chem. 2020, 85, 107237. [Google Scholar] [CrossRef]
  49. Aerts, J.C.J.H.; Heuvelink, G.B.M. Using simulated annealing for resource allocation. Int. J. Geogr. Inf. Sci. 2002, 16, 571–587. [Google Scholar] [CrossRef]
  50. Kalai, A.T.; Vempala, S. Simulated annealing for convex optimization. Math. Oper. Res. 2006, 31, 253–266. [Google Scholar] [CrossRef]
  51. Rere, L.M.R.; Fanany, M.I.; Arymurthy, A.M. Simulated Annealing Algorithm for Deep Learning. Procedia Comput. Sci. 2015, 72, 137–144. [Google Scholar] [CrossRef]
  52. Tsoulos, I.G.; Gavrilis, D.; Glavas, E. Neural network construction and training using grammatical evolution. Neurocomputing 2008, 72, 269–277. [Google Scholar] [CrossRef]
  53. Ivanova, I.; Kubat, M. Initialization of neural networks by means of decision trees. Knowl. Based Syst. 1995, 8, 333–344. [Google Scholar] [CrossRef]
  54. Yam, J.Y.F.; Chow, T.W.S. A weight initialization method for improving training speed in feedforward neural network. Neurocomputing 2000, 30, 219–232. [Google Scholar] [CrossRef]
  55. Chumachenko, K.; Iosifidis, A.; Gabbouj, M. Feedforward neural networks initialization based on discriminant learning. Neural Netw. 2022, 146, 220–229. [Google Scholar] [CrossRef]
  56. Leung, F.H.F.; Lam, H.K.; Ling, S.H.; Tam, P.K.S. Tuning of the structure and parameters of a neural network using an improved genetic algorithm. IEEE Trans. Neural Netw. 2003, 14, 79–88. [Google Scholar] [CrossRef]
  57. Han, H.G.; Qiao, J.F. A structure optimisation algorithm for feedforward neural network construction. Neurocomputing 2013, 99, 347–357. [Google Scholar] [CrossRef]
  58. Kim, K.J.; Cho, S.B. Evolved neural networks based on cellular automata for sensory-motor controller. Neurocomputing 2006, 69, 2193–2207. [Google Scholar] [CrossRef]
  59. Kelly, M.; Longjohn, R.; Nottingham, K. The UCI Machine Learning Repository. 2023. Available online: https://archive.ics.uci.edu (accessed on 18 February 2024).
  60. Alcalá-Fdez, J.; Fernandez, A.; Luengo, J.; Derrac, J.; García, S.; Sánchez, L.; Herrera, F. KEEL Data-Mining Software Tool: Data 506 Set Repository, Integration of Algorithms and Experimental Analysis Framework. J. Mult. Valued Log. Soft Comput. 2011, 17, 255–287. [Google Scholar]
  61. Weiss, M.S.; Kulikowski, A.C. Computer Systems That Learn: Classification and Prediction Methods from Statistics, Neural Nets, Machine Learning, and Expert Systems; Morgan Kaufmann Publishers Inc.: Burlington, MA, USA, 1991. [Google Scholar]
  62. Quinlan, J.R. Simplifying Decision Trees. Int. J. Man-Mach. Stud. 1987, 27, 221–234. [Google Scholar] [CrossRef]
  63. Shultz, T.; Mareschal, D.; Schmidt, W. Modeling Cognitive Development on Balance Scale Phenomena. Mach. Learn. 1994, 16, 59–88. [Google Scholar] [CrossRef]
  64. Zhou, Z.H. NeC4.5: Neural ensemble based C4.5. IEEE Trans. Knowl. Data Eng. 2004, 16, 770–773. [Google Scholar] [CrossRef]
  65. Setiono, R.; Leow, W.K. FERNN: An Algorithm for Fast Extraction of Rules from Neural Networks. Appl. Intell. 2000, 12, 15–25. [Google Scholar] [CrossRef]
  66. Demiroz, G.; Govenir, H.A.; Ilter, N. Learning Differential Diagnosis of Eryhemato-Squamous Diseases using Voting Feature Intervals. Artif. Intell. Med. 1998, 13, 147–165. [Google Scholar]
  67. Horton, P.; Nakai, K. A Probabilistic Classification System for Predicting the Cellular Localization Sites of Proteins. Proc. Int. Conf. Intell. Syst. Mol. Biol. 1996, 4, 109–115. [Google Scholar]
  68. Kononenko, I.; Šimec, E.; Robnik-Šikonja, M. Overcoming the Myopia of Inductive Learning Algorithms with RELIEFF. Appl. Intell. 1997, 7, 39–55. [Google Scholar] [CrossRef]
  69. Hayes-Roth, B.; Hayes-Roth, F. Concept learning and the recognition and classification of exemplars. J. Verbal Learn. Verbal Behav. 1977, 16, 321–338. [Google Scholar] [CrossRef]
  70. French, R.M.; Chater, N. Using noise to compute error surfaces in connectionist networks: A novel means of reducing catastrophic forgetting. Neural Comput. 2002, 14, 1755–1769. [Google Scholar] [CrossRef]
  71. Dy, J.G.; Brodley, C.E. Feature Selection for Unsupervised Learning. J. Mach. Learn. Res. 2004, 5, 845–889. [Google Scholar]
  72. Perantonis, S.J.; Virvilis, V. Input Feature Extraction for Multilayered Perceptrons Using Supervised Principal Component Analysis. Neural Process. Lett. 1999, 10, 243–252. [Google Scholar] [CrossRef]
  73. Garcke, J.; Griebel, M. Classification with sparse grids using simplicial basis functions. Intell. Data Anal. 2002, 6, 483–502. [Google Scholar] [CrossRef]
  74. Elter, M.; Schulz-Wendtland, R.; Wittenberg, T. The prediction of breast cancer biopsy outcomes using two CAD approaches that both emphasize an intelligible decision process. Med. Phys. 2007, 34, 4164–4172. [Google Scholar] [CrossRef] [PubMed]
  75. Little, M.A.; McSharry, P.E.; Hunter, E.J.; Spielman, J.; Ramig, L.O. Suitability of dysphonia measurements for telemonitoring of Parkinson’s disease. IEEE Trans. Biomed. Eng. 2009, 56, 1015–1022. [Google Scholar] [CrossRef]
  76. Smith, J.W.; Everhart, J.E.; Dickson, W.C.; Knowler, W.C.; Johannes, R.S. Using the ADAP learning algorithm to forecast the onset of diabetes mellitus. In Proceedings of the Symposium on Computer Applications and Medical Care; IEEE Computer Society Press: New York, NY, USA, 1988; pp. 261–265. [Google Scholar]
  77. Lucas, D.D.; Klein, R.; Tannahill, J.; Ivanova, D.; Brandon, S.; Domyancic, D.; Zhang, Y. Failure analysis of parameter-induced simulation crashes in climate models. Geosci. Model Dev. 2013, 6, 1157–1171. [Google Scholar] [CrossRef]
  78. Giannakeas, N.; Tsipouras, M.G.; Tzallas, A.T.; Kyriakidi, K.; Tsianou, Z.E.; Manousou, P.; Hall, A.; Karvounis, E.C.; Tsianos, V.; Tsianos, E. A clustering based method for collagen proportional area extraction in liver biopsy images. In Proceedings of the 2015 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Milan, Italy, 25–29 August 2015; pp. 3097–3100. [Google Scholar]
  79. Hastie, T.; Tibshirani, R. Non-parametric logistic and proportional odds regression. JRSS-C Appl. Stat. 1987, 36, 260–276. [Google Scholar] [CrossRef]
  80. Dash, M.; Liu, H.; Scheuermann, P.; Tan, K.L. Fast hierarchical clustering and its validation. Data Knowl. Eng. 2003, 44, 109–138. [Google Scholar] [CrossRef]
  81. Cortez, P.; Silva, A.M.G. Using data mining to predict secondary school student performance. In Proceedings of the 5th Future Business Technology Conference (FUBUTEC 2008), Porto, Portugal, 9–11 April 2008; pp. 5–12. [Google Scholar]
  82. Yeh, I.-C.; Yang, K.-J.; Ting, T.-M. Knowledge discovery on RFM model using Bernoulli sequence. Expert Syst. Appl. 2009, 36, 5866–5871. [Google Scholar] [CrossRef]
  83. Wolberg, W.H.; Mangasarian, O.L. Multisurface method of pattern separation for medical diagnosis applied to breast cytology. Proc. Natl. Acad. Sci. USA 1990, 87, 9193–9196. [Google Scholar] [CrossRef]
  84. Raymer, M.; Doom, T.E.; Kuhn, L.A.; Punch, W.F. Knowledge discovery in medical and biological datasets using a hybrid Bayes classifier/evolutionary algorithm. IEEE Trans. Syst. Man Cybern. Part B Cybern. A Publ. IEEE Syst. Man Cybern. Soc. 2003, 33, 802–813. [Google Scholar] [CrossRef]
  85. Zhong, P.; Fukushima, M. Regularized nonsmooth Newton method for multi-class support vector machines. Optim. Methods Softw. 2007, 22, 225–236. [Google Scholar] [CrossRef]
  86. Andrzejak, R.G.; Lehnertz, K.; Mormann, F.; Rieke, C.; David, P.; Elger, C.E. Indications of nonlinear deterministic and finite- dimensional structures in time series of brain electrical activity: Dependence on recording region and brain state. Phys. Rev. E 2001, 64, 1–8. [Google Scholar] [CrossRef] [PubMed]
  87. Koivisto, M.; Sood, K. Exact Bayesian Structure Discovery in Bayesian Networks. J. Mach. Learn. Res. 2004, 5, 549–573. [Google Scholar]
  88. Nash, W.J.; Sellers, T.L.; Talbot, S.R.; Cawthor, A.J.; Ford, W.B. The Population Biology of Abalone (_Haliotis_ species) in Tasmania. I. Blacklip Abalone (_H. rubra_) from the North Coast and Islands of Bass Strait; Sea Fisheries Division, Technical Report 48; Sea Fisheries Division, Department of Primary Industry and Fisheries: Orange, NSW, Australia, 1994. [Google Scholar]
  89. Brooks, T.F.; Pope, D.S.; Marcolini, A.M. Airfoil Self-Noise and Prediction; Technical Report, NASA RP-1218; NASA: Washington, DC, USA, 1989. [Google Scholar]
  90. Yeh, I.C. Modeling of strength of high performance concrete using artificial neural networks. Cem. Concr. Res. 1998, 28, 1797–1808. [Google Scholar] [CrossRef]
  91. Harrison, D.; Rubinfeld, D.L. Hedonic prices and the demand for clean ai. J. Environ. Econ. Manag. 1978, 5, 81–102. [Google Scholar] [CrossRef]
  92. Kingma, D.P.; Ba, J.L. ADAM: A method for stochastic optimization. In Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015), San Diego, CA, USA, 7–9 May 2015; pp. 1–15. [Google Scholar]
  93. Stanley, K.O.; Miikkulainen, R. Evolving Neural Networks through Augmenting Topologies. Evol. Comput. 2002, 10, 99–127. [Google Scholar] [CrossRef]
  94. Powell, M.J.D. A Tolerant Algorithm for Linearly Constrained Optimization Calculations. Math. Program. 1989, 45, 547–566. [Google Scholar] [CrossRef]
  95. Yi, D.; Ahn, J.; Ji, S. An effective optimization method for machine learning based on ADAM. Appl. Sci. 2020, 10, 1073. [Google Scholar] [CrossRef]
  96. Xiao, N.; Hu, X.; Liu, X.; Toh, K.C. Adam-family methods for nonsmooth optimization with convergence guarantees. J. Mach. Learn. Res. 2024, 25, 1–53. [Google Scholar]
  97. Tzimourta, K.D.; Tsoulos, I.; Bilero, T.; Tzallas, A.T.; Tsipouras, M.G.; Giannakeas, N. Direct Assessment of Alcohol Consumption in Mental State Using Brain Computer Interfaces and Grammatical Evolution. Inventions 2018, 3, 51. [Google Scholar] [CrossRef]
  98. Glover, F. Parametric tabu-search for mixed integer programs. Comput. Oper. Res. 2006, 33, 2449–2494. [Google Scholar] [CrossRef]
  99. Lim, A.; Rodrigues, B.; Zhang, X. A simulated annealing and hill-climbing algorithm for the traveling tournament problem. Eur. J. Oper. Res. 2006, 174, 1459–1478. [Google Scholar] [CrossRef]
  100. Bevilacqua, A. A methodological approach to parallel simulated annealing on an SMP system. J. Parallel Distrib. Comput. 2002, 62, 1548–1570. [Google Scholar] [CrossRef]
Figure 1. The used grammar in neural network construction.
Figure 1. The used grammar in neural network construction.
Futureinternet 16 00172 g001
Figure 2. Flowchart of the used Simulated Annealing variant.
Figure 2. Flowchart of the used Simulated Annealing variant.
Futureinternet 16 00172 g002
Figure 3. An example of the one-point crossover method, used in the Grammatical Evolution procedure.
Figure 3. An example of the one-point crossover method, used in the Grammatical Evolution procedure.
Futureinternet 16 00172 g003
Figure 4. Scatter plot illustrating the variability and median classification error rates for six machine learning algorithms, with statistical significance denoted by asterisks. The asterisk-based notation above the clusters denotes statistical significance levels. The NNC-S method demonstrates notably lower error rates, as evidenced by the high statistical significance relative to other methods, which may indicate its superior performance in classification tasks.
Figure 4. Scatter plot illustrating the variability and median classification error rates for six machine learning algorithms, with statistical significance denoted by asterisks. The asterisk-based notation above the clusters denotes statistical significance levels. The NNC-S method demonstrates notably lower error rates, as evidenced by the high statistical significance relative to other methods, which may indicate its superior performance in classification tasks.
Futureinternet 16 00172 g004
Figure 5. Scatter plot of regression error rates for various regression methods, demonstrating the distribution, median error rates, and statistical significance of differences in performance. The asterisk-based notation above the clusters denotes statistical significance levels The NNC-S method, shows statistically significant improvements in accuracy over the NNC method, underlining the efficacy of local optimization enhancements in regression tasks. The y-axis is presented on a logarithmic scale to clearly visualize the range of error magnitudes across methods.
Figure 5. Scatter plot of regression error rates for various regression methods, demonstrating the distribution, median error rates, and statistical significance of differences in performance. The asterisk-based notation above the clusters denotes statistical significance levels The NNC-S method, shows statistically significant improvements in accuracy over the NNC method, underlining the efficacy of local optimization enhancements in regression tasks. The y-axis is presented on a logarithmic scale to clearly visualize the range of error magnitudes across methods.
Futureinternet 16 00172 g005
Figure 6. This figure examines the impact of varying the parameter g on the NNC-S algorithm’s classification errors, revealing that while some parameter adjustments do not significantly affect performance, others result in noticeable differences, as indicated by the statistical significance annotations.
Figure 6. This figure examines the impact of varying the parameter g on the NNC-S algorithm’s classification errors, revealing that while some parameter adjustments do not significantly affect performance, others result in noticeable differences, as indicated by the statistical significance annotations.
Futureinternet 16 00172 g006
Figure 7. Time comparison between the initial method and the proposed modifications.
Figure 7. Time comparison between the initial method and the proposed modifications.
Futureinternet 16 00172 g007
Figure 8. Comparison between the proposed method and the original Neural Construction method for the alcohol consumption dataset.
Figure 8. Comparison between the proposed method and the original Neural Construction method for the alcohol consumption dataset.
Futureinternet 16 00172 g008
Table 1. The values for the parameters used in the conducted experiments.
Table 1. The values for the parameters used in the conducted experiments.
Name PurposeValue
N c Number of chromosomes500
N g Number of generations200
p s Selection rate0.10
p m Mutation rate0.05
gNumber of random changes10
RRange of random changes10
ϵ Small value used in comparisons 10 5
N e p s Number of random samples200
TInitial temperature 10 8
r T Rate of decrease in temperature0.8
Table 2. Experimental results for the series of machine learning methods for the classification datasets. The numbers in cells are the average classification errors as measured in the test set. The bold values it is and indication if the suggested method outperforms the original NNC method.
Table 2. Experimental results for the series of machine learning methods for the classification datasets. The numbers in cells are the average classification errors as measured in the test set. The bold values it is and indication if the suggested method outperforms the original NNC method.
Dataset ADAMNEATGENETICRBFNNCNNC-S
APPENDICITIS16.50%17.20%18.10%12.23%14.40%14.60%
AUSTRALIAN35.65%31.98%32.21%34.89%14.46%14.90%
BALANCE7.87%23.84%8.97%33.42%22.13%7.66%
CIRCULAR3.94%34.07%5.99%6.30%14.26%7.88%
CLEVELAND67.55%53.44%51.60%67.10%49.93%48.59%
DERMATOLOGY26.14%32.43%30.58%62.34%24.80%13.11%
ECOLI64.43%43.24%49.38%59.50%48.82%44.88%
HABERMAN29.00%24.04%28.66%25.10%28.33%28.73%
HAYES ROTH59.70%50.15%56.18%64.36%37.23%28.08%
HEART38.53%39.27%28.34%31.20%15.78%16.00%
HOUSEVOTES7.48%10.89%6.62%6.13%3.52%3.74%
IONOSPHERE16.64%19.67%15.14%16.22%11.86%10.03%
LIVERDISORDER41.53%30.67%31.11%30.84%32.97%32.82%
MAMMOGRAPHIC46.25%22.85%19.88%21.38%18.22%16.58%
PARKINSONS24.06%18.56%18.05%17.41%13.21%12.26%
PIMA34.85%34.51%32.19%25.78%28.47%25.26%
POPFAILURES5.18%7.05%5.94%7.04%6.83%5.52%
REGIONS229.85%33.23%29.39%38.29%25.87%24.47%
SAHEART34.04%34.51%34.86%32.19%30.80%29.52%
SEGMENT49.75%66.72%57.72%59.68%54.89%39.38%
STUDENT5.13%12.50%5.61%7.52%5.70%4.52%
TRANSFUSION25.68%24.87%25.84%27.36%25.30%24.33%
WDBC35.35%12.88%8.56%7.27%7.27%5.59%
WINE29.40%25.43%19.20%31.41%13.53%11.47%
Z_F_S47.81%38.41%10.73%13.16%15.30%7.93%
Z_O_N_F_S78.79%79.08%64.81%60.40%50.48%40.42%
ZO_NF_S47.43%43.75%8.41%9.02%15.22%6.60%
ZONF_S11.99%5.44%2.60%4.03%3.14%2.36%
ZOO14.13%20.27%16.67%21.93%9.10%7.20%
AVERAGE32.23%30.72%24.94%28.74%22.13%18.43%
Table 3. Experimental results as measured on the regression datasets. The numbers in cells denote the average regression errors for the associated machine learning method, as measured on the test set. The bold values it is and indication if the suggested method outperforms the original NNC method.
Table 3. Experimental results as measured on the regression datasets. The numbers in cells denote the average regression errors for the associated machine learning method, as measured on the test set. The bold values it is and indication if the suggested method outperforms the original NNC method.
Dataset ADAMNEATGENETICRBFNNCNNC-S
ABALONE4.309.887.177.375.114.95
AIRFOIL0.0050.0670.0030.270.0030.003
BASEBALL77.90100.39103.6093.0259.4057.30
CONCRETE0.0780.0810.00990.0110.0080.006
DEE0.631.5121.0130.170.260.23
HO0.0350.1672.780.030.0160.012
HOUSING80.2056.4943.2657.6825.5618.82
LASER0.030.0840.590.0240.0260.015
LW0.0280.171.901.140.970.038
MORTGAGE9.2414.112.411.450.290.12
PL0.1170.0970.280.0830.0460.033
SN0.0260.1742.950.900.0260.024
TREASURY11.1615.522.932.020.470.18
TZ0.070.0975.384.100.060.028
AVERAGE13.1214.2012.4512.026.605.84
Table 4. Experiments with the parameter g used in the modified Simulated Annealing method.
Table 4. Experiments with the parameter g used in the modified Simulated Annealing method.
Datatset NNC-S ( g = 2 )  NNC-S ( g = 5 )  NNC-S ( g = 10 )  NNC-S ( g = 20 )  
APPENDICITIS14.90%15.00%14.60%14.50%
AUSTRALIAN14.59%14.85%14.90%15.04%
BALANCE8.53%7.68%7.66%7.56%
CIRCULAR10.49%8.81%7.88%7.50%
CLEVELAND48.41%48.31%48.59%48.10%
DERMATOLOGY15.09%13.80%13.11%13.29%
ECOLI45.30%45.12%44.88%44.36%
HABERMAN27.80%27.97%28.73%27.83%
HAYES ROTH29.85%28.85%28.08%29.15%
HEART16.11%15.04%16.00%14.78%
HOUSEVOTES3.70%4.22%3.74%3.70%
IONOSPHERE10.54%10.09%10.03%10.00%
LIVERDISORDER31.41%33.15%32.82%33.29%
MAMMOGRAPHIC17.16%17.25%16.58%16.99%
PARKINSONS12.32%12.89%12.26%12.11%
PIMA26.12%25.96%25.26%25.92%
POPFAILURES5.58%6.00%5.52%5.68%
REGIONS224.71%24.05%24.47%24.66%
SAHEART30.04%29.67%29.52%29.07%
SEGMENT46.94%42.37%39.38%41.19%
STUDENT4.60%4.73%4.52%4.48%
   TRANSFUSION    24.28%24.34%24.33%24.03%
WDBC6.23%6.22%5.59%5.68%
WINE12.59%11.30%11.47%9.24%
Z_F_S9.57%9.60%7.93%8.10%
Z_O_N_F_S46.04%43.36%40.42%41.54%
ZO_NF_S9.69%8.54%6.60%6.44%
ZONF_S2.58%2.28%2.36%2.36%
ZOO6.90%7.00%7.20%7.70%
AVERAGE 19.38%18.91%18.43%18.42%
Table 5. Experiments with the parameter R of the modified Simulated Annealing algorithm. The parameter g was set to 10.
Table 5. Experiments with the parameter R of the modified Simulated Annealing algorithm. The parameter g was set to 10.
Dataset NNC-S ( R = 2 )  NNC-S ( R = 5 )  NNC-S ( R = 10 )NNC-S ( R = 20 )  
APPENDICITIS14.40%14.90%14.60%15.20%
AUSTRALIAN14.77%14.78%14.90%14.70%
BALANCE7.66%7.74%7.66%7.66%
CIRCULAR8.39%8.29%7.88%7.78%
CLEVELAND49.45%49.28%48.59%47.11%
DERMATOLOGY14.09%12.54%13.11%11.34%
ECOLI44.24%46.30%44.88%44.48%
HABERMAN27.33%28.10%28.73%28.04%
HAYES ROTH29.15%27.92%28.08%26.46%
HEART15.67%15.52%16.00%15.15%
HOUSEVOTES4.00%3.62%3.74%4.52%
IONOSPHERE10.14%10.03%10.03%10.71%
LIVERDISORDER32.80%32.12%32.82%32.29%
MAMMOGRAPHIC17.18%16.78%16.58%16.62%
PARKINSONS12.68%12.16%12.26%11.95%
PIMA25.72%25.11%25.26%26.33%
POPFAILURES5.87%5.72%5.52%5.58%
REGIONS223.55%24.04%24.47%24.08%
SAHEART29.48%28.96%29.52%29.24%
SEGMENT40.32%40.23%39.38%40.82%
STUDENT4.18%4.50%4.52%4.78%
TRANSFUSION24.60%24.12%24.33%24.36%
WDBC6.09%5.68%5.59%5.46%
WINE11.00%10.30%11.47%9.41%
Z_F_S8.33%8.30%7.93%8.50%
Z_O_N_F_S41.70%43.42%40.42%41.44%
ZO_NF_S7.58%7.72%6.60%7.10%
ZONF_S2.50%2.54%2.36%3.00%
ZOO6.80%6.30%7.20%6.10%
AVERAGE 18.61%18.52%18.43%18.28%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tsoulos, I.G.; Tzallas, A.; Karvounis, E. Using Optimization Techniques in Grammatical Evolution. Future Internet 2024, 16, 172. https://doi.org/10.3390/fi16050172

AMA Style

Tsoulos IG, Tzallas A, Karvounis E. Using Optimization Techniques in Grammatical Evolution. Future Internet. 2024; 16(5):172. https://doi.org/10.3390/fi16050172

Chicago/Turabian Style

Tsoulos, Ioannis G., Alexandros Tzallas, and Evangelos Karvounis. 2024. "Using Optimization Techniques in Grammatical Evolution" Future Internet 16, no. 5: 172. https://doi.org/10.3390/fi16050172

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop