Next Article in Journal
Deep Learning and Machine Learning Applications in Biomedicine
Next Article in Special Issue
Enhancing Robustness of Surface Electromyography Pattern Recognition at Different Arm Positions for Transhumeral Amputees Using Deep Adversarial Inception Domain Adaptation
Previous Article in Journal
Adhesion of Varnish Coatings as a Background for Analogue and Digital Printing Technologies
Previous Article in Special Issue
Machine-Learning-Based-Approaches for Sleep Stage Classification Utilising a Combination of Physiological Signals: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Constructing Features for Screening Neurodevelopmental Disorders Using Grammatical Evolution

1
Department of Speech and Language Therapy, School of Health Sciences, University of Ioannina, Panepistimioupoli B’, 45500 Ioannina, Greece
2
Laboratory of New Technologies and Distance Learning, Department of Early Childhood Education, School of Education, University of Ioannina, Panepistimioupoli, 45110 Ioannina, Greece
3
Physics Department, University of Ioannina, 45110 Ioannina, Greece
4
Department of Informatics and Telecommunications, University of Ioannina, 47150 Kostaki Artas, Greece
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(1), 305; https://doi.org/10.3390/app14010305
Submission received: 5 December 2023 / Revised: 26 December 2023 / Accepted: 28 December 2023 / Published: 29 December 2023
(This article belongs to the Special Issue Artificial Intelligence for Healthcare)

Abstract

:

Featured Application

This study is part of an ongoing research project titled “Smart Computing Models, Sensors, and Early diagnostic speech and language deficiencies indicators in Child Communication”, with the acronym “SmartSpeech”. The SmartSpeech project aims to assist clinicians in decision making regarding early diagnosis for children with neurodevelopmental disorders. SmartSpeech employs a serious game designed explicitly by the interdisciplinary team for this project, with activities aiming to evaluate the child’s developmental profile. The game is implemented in a tablet application and utilizes voice, and biomarkers of heart rate and gaze, for additional physiological measurements. A back-end system supports user registration, data collection, data analysis, and decision making. The potential application of this work is to allow the SmartSpeech machine learning model to better capture underlying patterns in the data, determine the most effective feature construction techniques for the given problem, and employ the results of this study to enhance the screening automated prediction results of the machine learning model in neurodevelopmental disorders.

Abstract

Developmental domains refer to different areas of a child’s growth and maturation, including physical, language, cognitive, and social–emotional skills. Understanding these domains helps parents, caregivers, and professionals track a child’s progress and identify potential areas of concern. Nevertheless, due to the high level of heterogeneity and overlap, neurodevelopmental disorders may go undiagnosed in children for a crucial period. Detecting neurodevelopmental disorders at an early stage is fundamental. Digital tools like artificial intelligence can help clinicians with the early detection process. To achieve this, a new method has been proposed that creates artificial features from the original ones derived from the SmartSpeech project, using a feature construction procedure guided by the Grammatical Evolution technique. The new features from a machine learning model are used to predict neurodevelopmental disorders. Comparative experiments demonstrated that using the feature creation method outperformed other machine learning methods for predicting neurodevelopmental disorders. In many cases, the reduction in the test error reaches up to 65% to the next better one.

1. Introduction

Neurodevelopmental disorder (ND) refers to a variety of disorders disturbing neurological development that impact on several domains, including communication, learning, social interaction, behavior, cognitive processes, and emotional functioning. Neurodevelopmental disorders (NDs) typically manifest during childhood [1,2]. Autism Spectrum Disorders (ASD), Attention Deficit Hyperactivity Disorder (ADHD), Intellectual Disability (ID), Specific Learning Disorder (SLD), and Communication Disorders (CD) are among the conditions that fall under the umbrella of neurodevelopmental disorders (NDs) [1].
Specific characteristics associated with each disorder are outlined in the Diagnostic and Statistical Manual of Mental illnesses, Fifth Edition (DSM-5), which offers descriptions for mental illnesses [1,3]. ASD is an ND with clinically severe functional deficits in distinctive, recurring patterns of behavior, interests, or hobbies, along with persistent challenges in social interaction and communication [1,2,4]. ASD can impact individuals’ educational experiences, employment opportunities, and social relationships [5]. ADHD is another ND with hallmarks on inattention, impulsivity, and hyperactivity, with a disruptive impact on daily functioning [1,2,4]. It can affect academic performance, work productivity, and interpersonal relationships [6,7]. ID is an ND characterized by deficiencies in general mental skills that affect adaptive functioning, such as verbal skills, learning aptitude, logical reasoning ability, and practical intelligence (problem-solving) [1,8]. ID can limit individuals’ independence and ability to live a fully inclusive life [9,10]. SLD is characterized by a major impairment in one or more of the following domains: oral expression, listening comprehension, basic reading and/or writing skills, and mathematical calculation and/or problem-solving abilities [1,2,11]. SLD can lead to challenges in academic settings and impact individuals’ self-esteem and confidence [12,13]. CD encompasses a collection of disorders, including speech sound disorder, language disorder, childhood-onset fluency disorder, social (pragmatic) communication disorder, and unspecified communication disorder [1]. These disorders are characterized by ongoing challenges in the acquisition, comprehension, and/or utilization of spoken or written language, resulting in an inability to express themselves, engage in meaningful conversations, and participate fully in social and professional interactions and effective communication [14].
Although NDs are frequently detectable in their early phases, the main obstacle is the lengthy and subjective character of conventional diagnostic techniques [2,15]. Consequently, there exists a minimum waiting period over a year between the initial suspicion and the subsequent confirmation of diagnosis. The process of diagnosis requires a significant amount of time, around 10 h [15]. Moreover, there is a persistent and increasing demand for appointments that surpasses the maximum capacity of pediatric clinics in many countries [16]. Apart from being time-consuming and expensive, conventional diagnostic techniques carry a considerable risk of receiving an incorrect diagnosis. This can lead to unnecessary prolonged pharmaceutical therapy, decreased functionality, and increased vulnerability to further health and social complications [17]. This backlog in diagnostic procedures has led to significant delays in providing timely treatment and intervention for children with suspected developmental disorders. Consequently, many children may go undiagnosed and may be left without the necessary support and resources they require during this critical period of their development. Early detection and intervention of NDs are of the utmost significance, as they contribute significantly to the reduction or mitigation of symptoms, eventually enhancing the individual’s overall quality of life. Nevertheless, due to the temporal intervals between the onset of worry and the establishment of a diagnosis, a significant amount of precious time is squandered while this condition persists undiscovered. Clinicians need to use digital tools, such as artificial intelligence, to aid in efficient early detection. Machine learning techniques possess the potential to not only expedite and enhance the accuracy of assessing the risk for NDs, but also play a crucial role in optimizing the entire diagnostic procedure and facilitating expedited access to vital therapeutic interventions for individuals and affected families [18].
It is well documented that to ensure accuracy and cost-effectiveness, it is necessary to employ swift and sophisticated standards [17,19,20,21]. The study of Alam et al. highlights the use of machine learning (ML) tools and deep learning (DL) techniques, such as convolutional neural network (CNN) and Deep Learning APIs (Application Programming Interface), to detect and treat signs of ADHD and ASD at an early stage [17]. The diagnostic procedures that utilize machine learning (ML) decrease the time required for intervention, enhance accuracy, and also facilitate comprehension of the techniques and algorithms employed for various types of data. Multiple studies have been conducted on ASD [22,23,24,25,26,27,28], ADHD [29,30,31], ID [9,10,32], SLD [33,34], CD [35], and NDs [4,19,20,31,36,37,38], providing evidence that ML algorithms can enhance diagnostic strategies for NDs. Further, more research efforts that seek to investigate ML approaches for early detection and diagnosis of NDs in real-life situations are crucial for ensuring timely intervention and optimizing lifelong outcomes [17,19,20,39,40,41]. This way, early intervention can help children with NDs develop their skills and abilities to the fullest extent possible. Even further, evidence in the literature suggests that SGs offer a flexible and innovative approach to assessing neurodevelopmental issues [3,40,42,43,44]. This is due to their ability to actively engage, adapt, and conduct tests that help to ensure more accurate, reliable, and user-friendly evaluations, eventually enhancing our understanding of NDs and enabling appropriate interventions.
SmartSpeech is an ongoing project intended to support and enhance screening and early detection procedures of NDs, utilizing smart computing models, sensors, and early diagnostic speech and language deficiency indicators [45]. The ultimate goal of this smart project is to improve children’s communication abilities in the context of digital healthcare, thus leading to a positive economic outcome in line with current digital trends. The project includes an online environment for gathering data from parents and physicians as well as a serious game (SG) for the child to interact with. The SG fosters active participation and motivation in individuals with NDs by providing an attractive and stimulating environment with 3D animations that bring life to characters guiding the child in the SG’s adventurous story. The SG also simulates real-world scenarios incorporating various modalities, including visual, auditory, and tactile elements, leading to more accurate observations of behavior and abilities in contexts that closely resemble everyday situations and enabling a more comprehensive assessment of diverse skills, like sensory processing, motor coordination, and social interactions. The SG design is adapted to suit the specific needs and abilities of each individual, ensuring that assessments are tailored to the unique characteristics of those with NDs, providing a more accurate representation of their skills and challenges [40]. Data collection is handled through a mobile application that collects child responses to game activities and biometric data using sensors and timestamps. The game data are processed on a dedicated server back-end service to examine early clinical screening/diagnostic patterns on specified domains or skills towards automated indications. The SmartSpeech ML approach enables automated decision-making based on the child’s communication profile and biometrics.
The primary objective of this study is to contribute to creating new digital tools and procedures aiding clinicians in their decision-making processes. Precisely, this study investigates a new proposed method that creates artificial features from the original ones derived from the SmartSpeech project, using a procedure guided by the Grammatical Evolution technique. The new features from an ML model are used to predict NDs. Comparative experiments are used to identify speech, language, hearing, psychomotor, cognitive, and psychoemotional impairments in both typically developed (TD) children and children with NDs. In a range of neural networks, different optimizers were employed and evaluated using our novel datasets. The objective was to automatically classify individuals in a screening procedure based on neurodevelopmental abilities.
Next, Section 2 “Background information” provides a detailed explanation of the chosen approach’s architecture for feature construction and classification tasks, outlining the algorithms employed to compare the suggested method. In Section 3, the materials and methods are described. Section 4 presents a comparison between the proposed approach and three machine learning methods. Section 5 “Discussion—Conclusions” critically analyzes and evaluates the proposed method and summarizes the findings of this study.

2. Background Information

This section offers a concise overview of the necessary background material and algorithms pertaining to the study.
Learning data are typically categorized into two distinct parts: training data and test data. Learning models adapt their parameters by utilizing the training data as input and, subsequently, undergo evaluation using the test data. The quantity of learning model parameters is directly influenced by the dimensionality of the input problem, namely, the number of features. Large problems require significant memory resources to store and manage learning models with an impact of input problem dimensionality on the efficacy of neural networks. Feature construction refers to the process of creating new features by applying mathematical operations, transformations, or combinations to existing ones. The ultimate goal of this method is to enhance the model by adding more details or connections that may not be immediately apparent in the original features [46].
Feature construction includes various techniques such as polynomial feature creation, interaction term generation, and the application of mathematical transformations, such as logarithmic transformations. To put it simply, feature construction specifically focuses on generating new features by applying mathematical operations or transformations, and it is often used when discussing the enhancement of machine learning model performance. Techniques like PCA, MRMR, and auto-encoder are used to reduce input data dimensionality [46,47].

2.1. The Proposed Method

The proposed method is based on Grammatical Evolution [48] to generate new artificial features from existing ones. Grammatical Evolution is an evolutionary algorithm, used to produce valid programs in any language defined by a BNF grammar, and it has been used in a variety of cases, such as solving trigonometric identities [49], automatic composition of music [50], combinatorial optimization problems [51], etc. The feature construction method was initially proposed by Gavrilis et al. [52] and was applied in many real world problems, such as classification of EEG signals [53], prediction of COVID-19 cases [54], Hemiplegia type detection [55], etc. The feature construction method creates artificial features from the original ones through Grammatical Evolution, and every set of potential features is evaluated on the training set using a machine learning method. For the purpose of this article, the freely available software QFc, version 1.0 [56] was selected and the evaluation machine learning model was a Radial Basis Function (RBF) network [57,58] with H processing nodes. The main steps of the used method are as follows (Algorithm 1):
Algorithm 1. The main steps of the used method
Initialization Step
  • Denote as NC the number of chromosomes in the population, with NG being the maximum number of allowed iterations and NF the number of constructed features.
  • Set the selection rate pS and the mutation rate pM.
  • Initialize the chromosomes as random integer numbers.
  • Set iter = 1
Genetic Step
  • Calculate fitness.
    • For Every chromosome gi, i = 1,…, NC do
    • Create NF artificial features for the gi chromosome, using the Grammatical Evolution procedure.
      • Apply the new features to the training set.
      • Set the fitness value fi as the training error of an RBF network on the modified training set.
    • End For
  • Apply crossover. Firstly, they are sorted according to their fitness values. The first (1-pS)NC chromosomes are copied intact to the next generation. The rest will be replaced by offsprings created in the crossover procedure. Every new offspring is created with one-point crossover from two distinct parents selected by tournament selection.
  • Apply the mutation procedure to the chromosomes with pM rate.
Termination Check
  • Set iter = iter + 1
  • If iter >= NG then terminate else goto Genetic Step.

2.2. Comparative Methods

The following methods were used for comparison.
The RBF neural network is a specific type of neural network that incorporates radial basis functions as activation functions. The Radial Basis Function (RBF) is a mathematical function employed in diverse machine learning techniques, specifically in kernelized approaches like Support Vector Machines (SVMs). The RBF kernel, also called the Gaussian kernel, is utilized to quantify the similarity or distance between data points in a modified feature space [59,60]. RBF neural networks are widely used for various tasks such as classification, regression, and clustering. It has proven to be effective in dealing with problems that involve high-dimensional input spaces and intricate patterns [58,61,62]. Compared with other neural network architectures, the RBF network has numerous advantages, such as its ability to process high-dimensional data, quick training and testing times, and the ability to approximate any continuous function with unrestricted accuracy [58,63]. The RBF network has three layers: input, hidden, and output. The multilayer perceptron train with the BFGS optimization method (MLP BFGS) is also employed. The hidden layer uses radial basis functions as activation functions to convert the input data into a new representation. Subsequently, this representation is used for subsequent analysis in the output layer. The network’s output is calculated by taking the modified inputs and combining them linearly. Therefore, the result is a binary determination expressed as either TRUE or FALSE, representing the two outcomes (NDs is predicted, NDs is not predicted) [40]. The radial basis networks were used in the construction phase of the artificial features since they are distinguished not only for their fast training method but also for their ability to approximate any function if a sufficient number of computing units are available [64].
MLP BFGS is a specific type of artificial neural network trained with the BFGS optimization method. The application of the Broyden–Fletcher–Goldfarb–Shanno (BFGS) optimization algorithm to train a Multilayer Perceptron (MLP) represents a departure from conventional methods. In this approach, the BFGS algorithm, designed for unconstrained optimization, is employed to minimize the MLP’s loss function. The MLP architecture comprises layers of neurons with weighted connections, and the BFGS algorithm iteratively updates the weights by considering the inverse Hessian matrix. This methodology entails a distinct departure from the standard stochastic gradient descent approaches commonly used in neural network training. The study of Hery, Ibrahim, and June [65] illustrated the efficacy and robustness of this unconventional MLP training strategy for small-dimensional test problems. The BFGS algorithm’s capacity for non-convex optimization is leveraged to achieve convergence towards optimal parameter values, showcasing its potential as an alternative in neural network optimization paradigms. It has been utilized in various experimental studies using machine learning, for instance, in automatic EEG epilepsy detection [66], feature extraction for hemiplegia type detection [55], Neural Networks on Biometric Datasets for Screening Speech and Language Deficiencies in Child Communication [41], machine learning for the performance and early drop prediction for higher education students [67], and many more.
MLP PCA is an application of MLP that involves a two-step training process. Initially, Principal Component Analysis (PCA) is employed to reduce the dimensionality of the input data, extracting principal components that capture the essential variance [68]. The resulting transformed features, representing a subset of principal components, serve as inputs for training the MLP. The MLP, with its layered architecture, learns intricate patterns and relationships within the reduced-dimensional data [52]. This approach offers advantages such as mitigating the curse of dimensionality and potentially improving the MLP’s generalization to new, unseen instances [69]. Careful hyperparameter tuning, including selecting the optimal number of principal components and configuring the MLP architecture, is crucial for effective implementation. Overall, the MLP PCA algorithm integrates dimensionality reduction with the capacity of MLPs, presenting a solution for handling complex data with high dimensionality [52].

3. Materials and Methods

This work was part of the project SmartSpeech, with the full title “Smart Computing Models, Sensors, and Early diagnostic speech and language deficiencies indicators in Child Communication” funded by the Region of Epirus and supported by the European Regional Development Fund (ERDF). The participants were recruited through private and public health and education institutions; they were mainly children, and their parents were informed thoroughly regarding the project’s scope and procedures, and asked to provide written consent and details about their child’s developmental and communication profile. Also, the parents were notified about the approval of this study from the University of Ioannina Research Ethics Committee with compliance to the General Data Protection Regulation (GDPR). The child’s active role in this study involved playing the serious game (SG), part of the SmartSpeech system.
The SG consists of children-specific activities with the objective of gathering data regarding the developmental skills of the child and biometric data such as heart rate and gaze responses. These biometric data were collected in order to investigate their role as bio-markers and their potential use for classification purposes.
The interaction of the child with the game involved several activities, each with the goal to actively participate and help to overcome missions/tasks and advance to the next level. All the activities had visual context directed in a way that they were entertaining and attractive, as seen in Figure 1. Nevertheless, behind the scenes activities also served as clinical tools trained toward measuring several speech, language, and developmental skills. The child followed a narrated story and had to solve puzzles through its chapters, select or drag objects on the touchscreen, identify images and shapes, recall names and events, recognize emotions, and even answer questions verbally.
Regarding the child verbal responses, for word recognition, we used the speech-to-text program CMUSphinx, version 5.0.0 [70], which is an open-source project that is free of charge, has cross-platform support for desktop and mobile systems, and can be used offline. Also, there exists a model for Greek language that was created and trained for this software [71].
For the biometric samples, we used a smartwatch with bio-sensors and a software-based eye-tracking module that monitors the gaze of the child while looking at the screen.
The smartwatch the individual wore during the SG activities captured and sent their cardiac rhythm to the SmartSpeech database for analysis. Heart rate variables were computed for each activity, including HRV, which was estimated using heart rate standard deviation and range statistics due to challenges in directly calculating HRV from the wearable device’s heart rate data. Thus, for each activity, we obtained three variables that were the mean, the standard deviation, and the range of the heart rate.
SeeSo software, Unity mobile SDK version 2.4.4, was used for eye tracking [72] to determine where the user’s eyes focused while performing certain activities on the mobile device. It recorded gaze points with X and Y coordinates of the screen at specific time intervals during SG activities. The variables obtained by the software are relevant, with the fixation being the fundamental metric for eye-tracking. A fixation is defined as a cluster of gaze points close to each other in space and time, which means the individual is looking at a specific region. Fixations are the most common measures of visual attention. During the game, certain areas on the screen are predetermined as areas of interest (AOI). An AOI may be, for example, a face, an animal, or a moving object, and it is defined as a rectangular region with specific coordinates and a specific time duration. The software extracted three fundamental variables from the eye movement experiments:
  • Fixation count (FC), i.e., the total number of fixations in an AOI;
  • Time to first fixation (TTFF), i.e., the time needed after an AOI is visible until the first fixation is counted on it;
  • The total duration of fixations (TS), i.e., the total time that an individual spends looking at a specific AOI.
The data collection of the eye-tracking software depends on how the subject reacts to stimuli. The subjects, especially children, move around a lot; thus, the camera fails to take some measurements. Therefore, due to missing values in our datasets, in the final variables selected after clearing the data, there were no variables with TTFF metrics. Eliminating missing values was necessary, and we needed to ensure that all cases were filled with valid data.
After finishing the game, all data were exported and automatically stored as variables in a remote server. The abovementioned variables belong to 3 categories: the game variables (25), which were the scores from the game activities; the eye-tracking variables (16); and the heart rate variables (15) [3]. The data variables for the game score, eye-tracking, and heart rate datasets are illustrated in Figure 2, Figure 3 and Figure 4, respectively, also presenting the dimensionality of each dataset. Table 1 illustrates in more detail the description of the variables for the game dataset.
For the experiments, in total, 435 children aged 8.8 ± 7.4 years participated, of which 224 were boys and 211 were girls. The parents provided written consent along with the child’s neurodevelopmental or medical history. After completion of all the games, data were gathered for each child that belong to the three aforementioned sources—game scores, eye-tracking, and heart rate—forming the three datasets in the study. The eye-tracking dataset had 309 cases, the heart rate dataset had 181 cases, and the game scores had 435 cases. The sample was further divided into two groups according to the existence of a specific neurodevelopmental disorder or not, inserting a new variable in the datasets with the name Disorder. This variable denotes whether an individual has one of the Diagnostic and Statistical Manual of Mental Disorders [1], which defines one or more of the following five disorders: Autism Spectrum Disorder (ASD), Attention Deficit Hyperactivity Disorder (ADHD), Intellectual Disability (ID), Specific Learning Disorder (SLD), and Communication Disorder (CD). The Disorder variable is binary with values true/false, denoting two classes for the classification procedures that follow.

4. Experiments

4.1. Experimental Datasets, Methods, and Parameter Details

This section reports on the assessment of the proposed FC2RBF technique’s efficacy in creating artificial features for feature learning and class prediction using the three datasets from the SmartSpeech project (see Section 3). These issues have been extensively examined by numerous scholars in the pertinent academic discourse, encompassing a diverse array of research domains spanning from economics to health [3,40,54,55,73,74].
The parameters used in the employed algorithms are shown in Table 2. The following methods were used:
  • RBF—an RBF neural network with H processing nodes.
  • MLP BFGS—an artificial neural network with H hidden nodes, trained with the BFGS optimization method.
  • MLP PCA—an artificial neural network with H hidden nodes and trained with the BFGS method. The neural network is applied on two constructed features produced by the PCA method.
  • FC2RBF—an RBF network with 10 processing nodes applied on two artificial features constructed by the proposed method.
To establish a higher level of trust in the outcomes of the experiments that were carried out, the technique of ten-fold cross validation was implemented for each and every experimental dataset. Each experiment was performed a total of thirty times, with a unique seed being input into the random number generator each time. Also, the experiments were executed 30 times using different seeds for the random generator each time. Finally, the same pre-processing and initial manual feature extraction stage was applied in all experiment runs to avoid any bias between the compared methods.
The code utilized was written in ANSI C++ and optimized with the help of the OPTIMUMS programming library. This library, which can be downloaded for free at https://github.com/itsoulos/OPTIMUMS/, was used to implement the code (accessed on 12 September 2023). The software was developed with the ANSI C++ programming language and relies on the freely accessible QT programming library. The software exhibits compatibility with a wide range of operating systems, including mobile platforms such as Android and iOS. The software can be freely downloaded from the official GitHub repository located at https://github.com/itsoulos/QFc (accessed on 1 September 2023).
A visual representation of the overall flowchart and study structure is included in Figure 5 outlining the structure of this study. This visualization includes the steps involved in the feature construction process, the application of machine learning methods, and the comparative analysis framework.
Most prediction models place the data points that they use in one of these four categories:
  • True positive (TP)—the individual in question does in fact have NDs and our prediction was accurate that the individual does have NDs.
  • True negative (TN)—the individual in question does not in fact have NDs and our prediction was accurate that the individual does not have NDs.
  • False positive (FP)—although the individual in question does not in fact have NDs, our prediction was inaccurate that the individual does have NDs. The term for this kind of error is a Type 1 error.
  • False negative (FN)—although the individual in question does in fact have NDs, our prediction was inaccurate that the individual does not have NDs. The term for this kind of error is a Type 2 error.
For the classification of the datasets, the reported error is the average classification error, as measured in the test set. The classification error refers to the percentage of patterns in the test set that were assigned to a class that was not the anticipated one. Error rate is calculated by Equation (1):
E r r o r   r a t e = F P + F N T P + T N + F P + F N
The precision metric quantifies the degree of accuracy of our positive predictions, meaning that it indicates the proportion of projected positive points that really occurred. Equation (2) defines precision:
P r e c i s i o n = T P T P + F P
The recall metric quantifies the proportion of positive instances that our model successfully recognized. In other words, it assesses the accuracy of our model in correctly classifying positive instances out of the total number of instances classified as positive. Recall and sensitivity are synonymous. Next, Equation (3) specifies recall:
R e c a l l = T P T P + F P

4.2. Experimental Results

Table 3 shows the results of the classification experiments for each of the methods in use in terms of error rate percentage. The experiments took place separately for each of the three datasets. Figure 6 presents a visualization of the error rates results.
The precision results of the classification studies for each of the employed methods are presented in Table 4. The experiments were conducted individually for each of the three datasets. The results indicate that the FC2RBF method bested the other three approaches in terms of precision, with great performance on the eye-tracking dataset. A visualization of the precision results is depicted in Figure 7.
The recall findings for each of the methods utilized in the classification experiments are presented in Table 5, designated as sensitivity. The experiments were conducted separately for each of the three datasets. Figure 8 shows a visualization of the recall findings.

5. Discussion—Conclusions

This study contributes to identifying the most accurate and efficient algorithms for a practical application such as SmartSpeech by proposing and comparing ML methodologies on these particular, consistent datasets. It offers valuable analysis of the advantages and disadvantages of various algorithms, enabling informed decision-making in the development and implementation of the SmartSpeech ML solution that may serve as a valuable screening tool for clinicians and other specialists to identify NDs from non-NDs children and, thus, plays a crucial role in their rehabilitation strategy.
In this study, a novel method called FC2RBF has been proposed for feature construction, which aims to improve the machine learning model by incorporating additional details or connections that may not be directly discernible in the original features. Feature construction involves generating new features by applying mathematical operations, transformations, or combinations to existing ones. The primary objective of this method is to enhance the model by introducing more details or connections that may take time to become apparent in the original features. The experimental results presented in Section 4.2 confirm that the optimized features generated by the FC2RBF algorithm in the comparative experiments demonstrated that using this feature creation method outperformed other machine learning methods for predicting neurodevelopmental disorders. The use of the FC2RBF algorithm on the SmartSpeech datasets showed significant results. In the eye-tracking dataset, the test error was reduced by up to 65% compared with the MLP BFGS algorithm, which was the next best-performing algorithm. In the heart rate dataset, the test error was reduced by 6.14% compared with the RBF neural network, the next best-performing algorithm. Similarly, in the game scores dataset, the test error was reduced by 6.78% compared with the RBF neural network, which was again the next best-performing algorithm. Moreover, the proposed technique achieves high success rates using only two artificial features, which are generated as non-linear combinations of the original features in each dataset. Also, the proposed method was applied without any modification in the different SmartSpeech datasets.
The number of learning model parameters is directly related to the dimensionality of the input problem, which is determined by the number of features. Therefore, complex problems, such as the evaluation and screening of neurodevelopmental disorders, require substantial memory resources to accommodate and handle the learning models. Moreover, as the number of parameters within computational models increases, it takes more time to modify those parameters. With a higher dimensionality of the data, a larger number of samples (patterns) is needed to achieve high learning rates. In this study, we managed to reduce the error rate and the dimensionality of the screening NDs features; for instance, in the game scores dataset, from 24 original features to two new artificial features produced from the original ones from non-linear transformations. Hence, only two features are required to obtain a low error rate in each dataset. Based on the results, the proposed technique performs better than the compared ones. Thus, it is faster, more accurate, and has higher sensitivity and specificity, especially for the eye-tracking dataset.
The study’s findings demonstrate the method’s effectiveness with outstanding results in analyzing the SmartSpeech eye-tracking dataset, exhibiting lower error rates and, thus, higher accuracy, together with higher precision and sensitivity. This method can be integrated into the SmartSpeech machine learning model to support automated prediction in neurodevelopmental disorders (NDs), and to further assist clinicians in distinguishing children with NDs from those without during screening procedures.
Future research may focus on addressing challenges and exploring innovations to enhance the efficiency, interpretability, and generalization capabilities of models addressing real-world challenges in NDs.

Author Contributions

Conceptualization, E.I.T. and I.G.T.; methodology, E.I.T. and I.G.T.; software, I.G.T.; validation, E.I.T. and J.P.; formal analysis, E.I.T., G.T. and I.G.T., investigation, G.T.; resources, E.I.T., G.T. and I.G.T.; data curation, G.T. and J.P.; writing—original draft preparation, E.I.T., G.T. and I.G.T.; writing—review and editing, E.I.T., J.P. and I.G.T.; visualization, E.I.T., G.T. and I.G.T.; supervision, E.I.T. and I.G.T.; project administration, E.I.T.; funding acquisition, E.I.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the project titled “Smart Computing Models, Sensors, and Early Diagnostic Speech and Language Deficiencies Indicators in Child Communication”, with code HP1AB-28185 (MIS: 5033088), supported by the European Regional Development Fund (ERDF).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Research Ethics Committee of the UNIVERSITY OF IOANNINA, Greece, (protocol code 18435 on 15 May 2020).

Informed Consent Statement

Informed written consent was obtained from all participating parents after informing them regarding the study’s compliance with GDPR regulations.

Data Availability Statement

Data are contained within the article.

Acknowledgments

We wish to thank all the participants for their valuable contribution in this study as well as administrative and technical support.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. American Psychiatric Association. Diagnostic and Statistical Manual of Mental Disorders, 5th ed.; American Psychiatric Association: Washington, DC, USA, 2013; ISBN 978-0-89042-555-8. [Google Scholar]
  2. Thapar, A.; Cooper, M.; Rutter, M. Neurodevelopmental Disorders. Lancet Psychiatry 2017, 4, 339–346. [Google Scholar] [CrossRef]
  3. Toki, E.I.; Tatsis, G.; Tatsis, V.A.; Plachouras, K.; Pange, J.; Tsoulos, I.G. Applying Neural Networks on Biometric Datasets for Screening Speech and Language Deficiencies in Child Communication. Mathematics 2023, 11, 1643. [Google Scholar] [CrossRef]
  4. Harris, J.C. New Classification for Neurodevelopmental Disorders in DSM-5. Curr. Opin. Psychiatry 2014, 27, 95–97. [Google Scholar] [CrossRef]
  5. Wong, J.; Cohn, E.S.; Coster, W.J.; Orsmond, G.I. “Success Doesn’t Happen in a Traditional Way”: Experiences of School Personnel Who Provide Employment Preparation for Youth with Autism Spectrum Disorder. Res. Autism Spectr. Disord. 2020, 77, 101631. [Google Scholar] [CrossRef]
  6. Young, S.; Adamo, N.; Ásgeirsdóttir, B.B.; Branney, P.; Beckett, M.; Colley, W.; Cubbin, S.; Deeley, Q.; Farrag, E.; Gudjonsson, G. Females with ADHD: An Expert Consensus Statement Taking a Lifespan Approach Providing Guidance for the Identification and Treatment of Attention-Deficit/Hyperactivity Disorder in Girls and Women. BMC Psychiatry 2020, 20, 404. [Google Scholar] [CrossRef]
  7. Abi-Jaoude, E.; Naylor, K.T.; Pignatiello, A. Smartphones, Social Media Use and Youth Mental Health. CMAJ 2020, 192, E136–E141. [Google Scholar] [CrossRef]
  8. American Phychiatic Association. DSM-5 Intellectual Disability Fact Sheet. Am. Psychiatr. Assoc. 2013, 2. Available online: https://www.psychiatry.org/File%20Library/Psychiatrists/Practice/DSM/APA_DSM-5-Intellectual-Disability.pdf (accessed on 4 December 2023).
  9. Merrells, J.; Buchanan, A.; Waters, R. “We Feel Left out”: Experiences of Social Inclusion from the Perspective of Young Adults with Intellectual Disability. J. Intellect. Dev. Disabil. 2019, 44, 13–22. [Google Scholar] [CrossRef]
  10. Chadwick, D.; Ågren, K.A.; Caton, S.; Chiner, E.; Danker, J.; Gómez-Puerta, M.; Heitplatz, V.; Johansson, S.; Normand, C.L.; Murphy, E. Digital Inclusion and Participation of People with Intellectual Disabilities during COVID-19: A Rapid Review and International Bricolage. J. Policy Pract. Intellect. Disabil. 2022, 19, 242–256. [Google Scholar] [CrossRef]
  11. Fletcher, J.M.; Miciak, J. The Identification of Specific Learning Disabilities: A Summary of Research on Best Practices; Meadows Center for Preventing Educational Risk: Austin, TX, USA, 2019; Available online: https://texasldcenter.org/files/resources/SLD–Manual_Final.pdf (accessed on 4 December 2023).
  12. Filippello, P.; Buzzai, C.; Messina, G.; Mafodda, A.V.; Sorrenti, L. School Refusal in Students with Low Academic Performances and Specific Learning Disorder. The Role of Self-Esteem and Perceived Parental Psychological Control. Int. J. Disabil. Dev. Educ. 2020, 67, 592–607. [Google Scholar] [CrossRef]
  13. Haft, S.L.; Greiner de Magalhães, C.; Hoeft, F. A Systematic Review of the Consequences of Stigma and Stereotype Threat for Individuals with Specific Learning Disabilities. J. Learn. Disabil. 2023, 56, 193–209. [Google Scholar] [CrossRef]
  14. Kwame, A.; Petrucka, P.M. A Literature-Based Study of Patient-Centered Care and Communication in Nurse-Patient Interactions: Barriers, Facilitators, and the Way Forward. BMC Nurs. 2021, 20, 158. [Google Scholar] [CrossRef]
  15. Rice, C.E.; Carpenter, L.A.; Morrier, M.J.; Lord, C.; DiRienzo, M.; Boan, A.; Skowyra, C.; Fusco, A.; Baio, J.; Esler, A.; et al. Defining in Detail and Evaluating Reliability of DSM-5 Criteria for Autism Spectrum Disorder (ASD) Among Children. J. Autism Dev. Disord. 2022, 52, 5308–5320. [Google Scholar] [CrossRef]
  16. Kou, J.; Le, J.; Fu, M.; Lan, C.; Chen, Z.; Li, Q.; Zhao, W.; Xu, L.; Becker, B.; Kendrick, K.M. Comparison of Three Different Eye-tracking Tasks for Distinguishing Autistic from Typically Developing Children and Autistic Symptom Severity. Autism Res. 2019, 12, 1529–1540. [Google Scholar] [CrossRef]
  17. Alam, S.; Raja, P.; Gulzar, Y. Investigation of Machine Learning Methods for Early Prediction of Neurodevelopmental Disorders in Children. Wirel. Commun. Mob. Comput. 2022, 2022, 5766386. [Google Scholar] [CrossRef]
  18. Vakadkar, K.; Purkayastha, D.; Krishnan, D. Detection of Autism Spectrum Disorder in Children Using Machine Learning Techniques. SN Comput. Sci. 2021, 2, 386. [Google Scholar] [CrossRef]
  19. Iwauchi, K.; Tanaka, H.; Okazaki, K.; Matsuda, Y.; Uratani, M.; Morimoto, T.; Nakamura, S. Eye-Movement Analysis on Facial Expression for Identifying Children and Adults with Neurodevelopmental Disorders. Front. Digit. Health 2023, 5, 952433. [Google Scholar] [CrossRef]
  20. de Barros, F.R.D.; da Silva, C.N.F.; de Castro Michelassi, G.; Brentani, H.; Nunes, F.L.; Machado-Lima, A. Computer Aided Diagnosis of Neurodevelopmental Disorders and Genetic Syndromes Based on Facial Images-a Systematic Literature Review. Heliyon 2023, 9, e20517. [Google Scholar] [CrossRef]
  21. Andrés-Roqueta, C.; Katsos, N. A Distinction Between Linguistic and Social Pragmatics Helps the Precise Characterization of Pragmatic Challenges in Children With Autism Spectrum Disorders and Developmental Language Disorder. J. Speech Lang. Hear. Res. 2020, 63, 1494–1508. [Google Scholar] [CrossRef]
  22. Schulte-Rüther, M.; Kulvicius, T.; Stroth, S.; Wolff, N.; Roessner, V.; Marschik, P.B.; Kamp-Becker, I.; Poustka, L. Using Machine Learning to Improve Diagnostic Assessment of ASD in the Light of Specific Differential and Co-Occurring Diagnoses. J. Child Psychol. Psychiatry 2023, 64, 16–26. [Google Scholar] [CrossRef]
  23. Manjur, S.M.; Hossain, M.-B.; Constable, P.A.; Thompson, D.A.; Marmolejo-Ramos, F.; Lee, I.O.; Skuse, D.H.; Posada-Quintero, H.F. Detecting Autism Spectrum Disorder Using Spectral Analysis of Electroretinogram and Machine Learning: Preliminary Results. In Proceedings of the 2022 44th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Glasgow, UK, 11–15 July 2022; IEEE: New York, NY, USA, 2022; pp. 3435–3438. [Google Scholar]
  24. Gui, A.; Bussu, G.; Tye, C.; Elsabbagh, M.; Pasco, G.; Charman, T.; Johnson, M.H.; Jones, E.J. Attentive Brain States in Infants with and without Later Autism. Transl. Psychiatry 2021, 11, 196. [Google Scholar] [CrossRef]
  25. Wei, Q.; Xu, X.; Xu, X.; Cheng, Q. Early Identification of Autism Spectrum Disorder by Multi-Instrument Fusion: A Clinically Applicable Machine Learning Approach. Psychiatry Res. 2023, 320, 115050. [Google Scholar] [CrossRef]
  26. Lin, Y.; Yerukala Sathipati, S.; Ho, S.-Y. Predicting the Risk Genes of Autism Spectrum Disorders. Front. Genet. 2021, 12, 665469. [Google Scholar] [CrossRef]
  27. Abdelhamid, N.; Thind, R.; Mohammad, H.; Thabtah, F. Assessing Autistic Traits in Toddlers Using a Data-Driven Approach with DSM-5 Mapping. Bioengineering 2023, 10, 1131. [Google Scholar] [CrossRef]
  28. El Mouatasim, A.; Ikermane, M. Control Learning Rate for Autism Facial Detection via Deep Transfer Learning. Signal Image Video Process. 2023, 17, 3713–3720. [Google Scholar] [CrossRef]
  29. van Rooij, D.; Zhang-James, Y.; Buitelaar, J.; Faraone, S.V.; Reif, A.; Grimm, O. Structural Brain Morphometry as Classifier and Predictor of ADHD and Reward-Related Comorbidities. Front. Psychiatry 2022, 13, 869627. [Google Scholar] [CrossRef]
  30. Chen, T.; Tachmazidis, I.; Batsakis, S.; Adamou, M.; Papadakis, E.; Antoniou, G. Diagnosing Attention-Deficit Hyperactivity Disorder (ADHD) Using Artificial Intelligence: A Clinical Study in the UK. Front. Psychiatry 2023, 14, 1164433. [Google Scholar] [CrossRef]
  31. Ahire, N.; Awale, R.; Wagh, A. Electroencephalogram (EEG) Based Prediction of Attention Deficit Hyperactivity Disorder (ADHD) Using Machine Learning. Appl. Neuropsychol. Adult 2023, 1–12. [Google Scholar] [CrossRef]
  32. Aggarwal, G.; Singh, L. Comparisons of Speech Parameterisation Techniques for Classification of Intellectual Disability Using Machine Learning. In Research Anthology on Physical and Intellectual Disabilities in an Inclusive Society; IGI Global: Hershey, PA, USA, 2022; pp. 828–847. [Google Scholar]
  33. Nilsson Benfatto, M.; Öqvist Seimyr, G.; Ygge, J.; Pansell, T.; Rydberg, A.; Jacobson, C. Screening for Dyslexia Using Eye Tracking during Reading. PLoS ONE 2016, 11, e0165508. [Google Scholar] [CrossRef]
  34. Gran Ekstrand, A.C.; Nilsson Benfatto, M.; Öqvist Seimyr, G. Screening for Reading Difficulties: Comparing Eye Tracking Outcomes to Neuropsychological Assessments. Front. Educ. 2021, 6, 643232. [Google Scholar] [CrossRef]
  35. Chawla, M.; Panda, S.N.; Khullar, V. Assistive Technologies for Individuals with Communication Disorders. In Proceedings of the 2022 10th International Conference on Reliability, Infocom Technologies and Optimization (Trends and Future Directions) (ICRITO), Noida, India, 13–14 October 2022; IEEE: New York, NY, USA, 2022; pp. 1–5. [Google Scholar]
  36. Jacob, S.; Wolff, J.J.; Steinbach, M.S.; Doyle, C.B.; Kumar, V.; Elison, J.T. Neurodevelopmental Heterogeneity and Computational Approaches for Understanding Autism. Transl. Psychiatry 2019, 9, 63. [Google Scholar] [CrossRef]
  37. Bardoloi, P.S. Management of Aggressive Behaviors in Neurodevelopmental Disorders-Role of α 2 Agonists. Acta Sci. Neurol. 2023, 6, 20–25. [Google Scholar] [CrossRef]
  38. Moreau, C.; Deruelle, C.; Auzias, G. Machine Learning for Neurodevelopmental Disorders. In Machine Learning for Brain Disorders; Colliot, O., Ed.; Humana: New York, NY, USA, 2023; ISBN 978-1-07-163194-2. [Google Scholar]
  39. Jacobs, G.R.; Voineskos, A.N.; Hawco, C.; Stefanik, L.; Forde, N.J.; Dickie, E.W.; Lai, M.-C.; Szatmari, P.; Schachar, R.; Crosbie, J.; et al. Integration of Brain and Behavior Measures for Identification of Data-Driven Groups Cutting across Children with ASD, ADHD, or OCD. Neuropsychopharmacology 2021, 46, 643–653. [Google Scholar] [CrossRef]
  40. Toki, E.I.; Tatsis, G.; Tatsis, V.A.; Plachouras, K.; Pange, J.; Tsoulos, I.G. Employing Classification Techniques on SmartSpeech Biometric Data towards Identification of Neurodevelopmental Disorders. Signals 2023, 4, 401–420. [Google Scholar] [CrossRef]
  41. Toki, E.I.; Tatsis, G.; Pange, J.; Plachouras, K.; Christodoulides, P.; Kosma, E.I.; Chronopoulos, S.K.; Zakopoulou, V. Can Eye Tracking Identify Prognostic Markers for Learning Disabilities? A Preliminary Study. In Proceedings of the New Realities, Mobile Systems and Applications; Springer: Cham, Switzerland, 2022; pp. 1032–1039. [Google Scholar] [CrossRef]
  42. Kim, H.H.; An, J.I.; Park, Y.R. A Prediction Model for Detecting Developmental Disabilities in Preschool-Age Children Through Digital Biomarker-Driven Deep Learning in Serious Games: Development Study. JMIR Serious Games 2021, 9, e23130. [Google Scholar] [CrossRef]
  43. Pandria, N.; Petronikolou, V.; Lazaridis, A.; Karapiperis, C.; Kouloumpris, E.; Spachos, D.; Fachantidis, A.; Vasiliou, D.; Vlahavas, I.; Bamidis, P. Information System for Symptom Diagnosis and Improvement of Attention Deficit Hyperactivity Disorder: Protocol for a Nonrandomized Controlled Pilot Study. JMIR Res. Protoc. 2022, 11, e40189. [Google Scholar] [CrossRef]
  44. Rello, L.; Baeza-Yates, R.; Ali, A.; Bigham, J.P.; Serra, M. Predicting Risk of Dyslexia with an Online Gamified Test. PLoS ONE 2020, 15, e0241687. [Google Scholar] [CrossRef]
  45. Toki, E.I.; Zakopoulou, V.; Tatsis, G.; Plachouras, K.; Siafaka, V.; Kosma, E.I.; Chronopoulos, S.K.; Filippidis, D.E.; Nikopoulos, G.; Pange, J.; et al. A Game-Based Smart System Identifying Developmental Speech and Language Disorders in Child Communication: A Protocol Towards Digital Clinical Diagnostic Procedures. In New Realities, Mobile Systems and Applications; Auer, M.E., Tsiatsos, T., Eds.; Lecture Notes in Networks and Systems; Springer International Publishing: Cham, Switzerland, 2022; Volume 411, pp. 559–568. ISBN 978-3-030-96295-1. [Google Scholar]
  46. Ensarioğlu, K.; İnkaya, T.; Emel, E. Remaining Useful Life Estimation of Turbofan Engines with Deep Learning Using Change-Point Detection Based Labeling and Feature Engineering. Appl. Sci. 2023, 13, 11893. [Google Scholar] [CrossRef]
  47. Ding, C.; Peng, H. Minimum Redundancy Feature Selection from Microarray Gene Expression Data. J. Bioinform. Comput. Biol. 2005, 3, 185–205. [Google Scholar] [CrossRef]
  48. O’Neill, M.; Ryan, C. Grammatical Evolution. IEEE Trans. Evol. Comput. 2001, 5, 349–358. [Google Scholar] [CrossRef]
  49. Ryan, C.; O’Neill, M. Grammatical Evolution: Solving Trigonometric Identities. Available online: https://www.semanticscholar.org/paper/Grammatical-Evolution%3A-Solving-Trigonometric-Ryan-O%E2%80%99Neill/7d4bd6c4f2532d17272f36cef016fdf5b59eb4d1 (accessed on 30 November 2023).
  50. de la Puente, A.O.; Alfonso, R.S.; Moreno, M.A. Automatic Composition of Music by Means of Grammatical Evolution. In Proceedings of the 2002 Conference on APL: Array Processing Languages: Lore, Problems, and Applications, Madrid, Spain, 22–25 July 2002; pp. 148–155. [Google Scholar]
  51. Sabar, N.R.; Ayob, M.; Kendall, G.; Qu, R. Grammatical Evolution Hyper-Heuristic for Combinatorial Optimization Problems. IEEE Trans. Evol. Comput. 2013, 17, 840–861. [Google Scholar] [CrossRef]
  52. Gavrilis, D.; Tsoulos, I.G.; Dermatas, E. Selecting and Constructing Features Using Grammatical Evolution. Pattern Recognit. Lett. 2008, 29, 1358–1365. [Google Scholar] [CrossRef]
  53. Tzallas, A.T.; Tsoulos, I.; Tsipouras, M.G.; Giannakeas, N.; Androulidakis, I.; Zaitseva, E. Classification of EEG Signals Using Feature Creation Produced by Grammatical Evolution. In Proceedings of the 2016 24th Telecommunications Forum (TELFOR), Belgrade, Serbia, 22–23 November 2016; IEEE: New York, NY, USA, 2016; pp. 1–4. [Google Scholar]
  54. Tsoulos, I.G.; Tzallas, A.T.; Tsalikakis, D. Prediction of COVID-19 Cases Using Constructed Features by Grammatical Evolution. Symmetry 2022, 14, 2149. [Google Scholar] [CrossRef]
  55. Christou, V.; Tsoulos, I.; Arjmand, A.; Dimopoulos, D.; Varvarousis, D.; Tzallas, A.T.; Gogos, C.; Tsipouras, M.G.; Glavas, E.; Ploumis, A.; et al. Grammatical Evolution-Based Feature Extraction for Hemiplegia Type Detection. Signals 2022, 3, 737–751. [Google Scholar] [CrossRef]
  56. Tsoulos, I.G. QFC: A Parallel Software Tool for Feature Construction, Based on Grammatical Evolution. Algorithms 2022, 15, 295. [Google Scholar] [CrossRef]
  57. Park, J.; Sandberg, I.W. Universal Approximation Using Radial-Basis-Function Networks. Neural Comput. 1991, 3, 246–257. [Google Scholar] [CrossRef]
  58. Yu, H.; Xie, T.; Paszczynski, S.; Wilamowski, B.M. Advantages of Radial Basis Function Networks for Dynamic System Design. IEEE Trans. Ind. Electron. 2011, 58, 5438–5450. [Google Scholar] [CrossRef]
  59. Giveki, D.; Rastegar, H. Designing a New Radial Basis Function Neural Network by Harmony Search for Diabetes Diagnosis. Opt. Mem. Neural Netw. 2019, 28, 321–331. [Google Scholar] [CrossRef]
  60. Karimi, N.; Kazem, S.; Ahmadian, D.; Adibi, H.; Ballestra, L.V. On a Generalized Gaussian Radial Basis Function: Analysis and Applications. Eng. Anal. Bound. Elem. 2020, 112, 46–57. [Google Scholar] [CrossRef]
  61. Tsoulos, I.G.; Tzallas, A.; Tsalikakis, D. Use RBF as a Sampling Method in Multistart Global Optimization Method. Signals 2022, 3, 857–874. [Google Scholar] [CrossRef]
  62. Haykin, S.S.; Haykin, S.S. Neural Networks and Learning Machines, 3rd ed.; Prentice Hall: New York, NY, USA, 2009; ISBN 978-0-13-147139-9. [Google Scholar]
  63. Bishop, C. Neural Networks for Pattern Recognition; Oxford University Press: Oxford, UK, 1995. [Google Scholar]
  64. Liao, Y.; Fang, S.-C.; Nuttle, H.L.W. Relaxed Conditions for Radial-Basis Function Networks to Be Universal Approximators. Neural Netw. 2003, 16, 1019–1028. [Google Scholar] [CrossRef]
  65. Hery, M.A.; Ibrahim, M.; June, L.W. BFGS Method: A New Search Direction. Sains Malays. 2014, 43, 1591–1597. [Google Scholar]
  66. Christou, V.; Miltiadous, A.; Tsoulos, I.; Karvounis, E.; Tzimourta, K.D.; Tsipouras, M.G.; Anastasopoulos, N.; Tzallas, A.T.; Giannakeas, N. Evaluating the Window Size’s Role in Automatic EEG Epilepsy Detection. Sensors 2022, 22, 9233. [Google Scholar] [CrossRef]
  67. Christou, V.; Tsoulos, I.; Loupas, V.; Tzallas, A.T.; Gogos, C.; Karvelis, P.S.; Antoniadis, N.; Glavas, E.; Giannakeas, N. Performance and Early Drop Prediction for Higher Education Students Using Machine Learning. Expert Syst. Appl. 2023, 225, 120079. [Google Scholar] [CrossRef]
  68. Toki, E.I.; Zakopoulou, V.; Tatsis, G.; Pange, J. Exploring the Main Principles for Automated Detec????on of Neurodevelopmental Disorders: Findings from Typically Developed Children. J. Med. Internet Res. 2023, preprint. [Google Scholar] [CrossRef]
  69. Georgoulas, G.; Gavrilis, D.; Tsoulos, I.G.; Stylios, C.; Bernardes, J.; Groumpos, P.P. Novel Approach for Fetal Heart Rate Classification Introducing Grammatical Evolution. Biomed. Signal Process. Control 2007, 2, 69–79. [Google Scholar] [CrossRef]
  70. CMUSphinx 2022. Available online: https://cmusphinx.github.io/2022/10/release/ (accessed on 4 December 2023).
  71. Pantazoglou, F.K.; Papadakis, N.K.; Kladis, G.P. Implementation of the Generic Greek Model for CMU Sphinx Speech Recognition Toolkit. In Proceedings of the International Scientific Conference eRA-12, Piraeus, Greece, 24–26 October 2017. [Google Scholar]
  72. SeeSo: Eye Tracking Software 2022. Available online: https://www.seeso.io/ (accessed on 4 December 2023).
  73. Guo, H.-W.; Huang, Y.-S.; Lin, C.-H.; Chien, J.-C.; Haraikawa, K.; Shieh, J.-S. Heart Rate Variability Signal Features for Emotion Recognition by Using Principal Component Analysis and Support Vectors Machine. In Proceedings of the 2016 IEEE 16th International Conference on Bioinformatics and Bioengineering (BIBE), Taichung, Taiwan, 31 October–2 November 2016; IEEE: New York, NY, USA, 2016; pp. 274–277. [Google Scholar]
  74. Rahman, M.M.; Usman, O.L.; Muniyandi, R.C.; Sahran, S.; Mohamed, S.; Razak, R.A. A Review of Machine Learning Methods of Feature Selection and Classification for Autism Spectrum Disorder. Brain Sci. 2020, 10, 949. [Google Scholar] [CrossRef]
Figure 1. Screenshots from the in-house SG.
Figure 1. Screenshots from the in-house SG.
Applsci 14 00305 g001
Figure 2. Variables comprising the game scores dataset.
Figure 2. Variables comprising the game scores dataset.
Applsci 14 00305 g002
Figure 3. Variables comprising the eye-tracking dataset.
Figure 3. Variables comprising the eye-tracking dataset.
Applsci 14 00305 g003
Figure 4. Variables comprising the heart rate dataset.
Figure 4. Variables comprising the heart rate dataset.
Applsci 14 00305 g004
Figure 5. Overall flowchart and study structure (g* stands for the best chromosome of the population).
Figure 5. Overall flowchart and study structure (g* stands for the best chromosome of the population).
Applsci 14 00305 g005
Figure 6. Error rate results visualization.
Figure 6. Error rate results visualization.
Applsci 14 00305 g006
Figure 7. Precision results visualization.
Figure 7. Precision results visualization.
Applsci 14 00305 g007
Figure 8. Recall results visualization.
Figure 8. Recall results visualization.
Applsci 14 00305 g008
Table 1. Description of variables.
Table 1. Description of variables.
Variable TypeDescription—Examples
Objects recognitionIdentification of shadow (shape)
Identification of object by acoustic stimuli
Categorization (i.e., distinguishing fruits from vegetables)
Time sequences (i.e., setting pictures to correct order)
Click on objectsBurst balloons
Color sequences (i.e., fill bridge gap with colored boards)
Pre-writing skills (i.e., move a teleferic with hand)
Cognitive flexibility (i.e., lead character out of a maze)
Sustained attention (i.e., catch thrown fruits in basket)
Fine motor skills (i.e., solve classic puzzle with pieces)
Sequences for size (arrange boards according to size)
Vocal intensityAvoid clouds using voice intensity in a flying game
Verbal responsesRepeat a vocalization (word)
Naming objects
Answer questions
Naming feelings
Memory tasksRecall names of characters
Remember object’s position in a grid
Emotion recognitionColor sequences (i.e., fill bridge gap with colored boards)
Table 2. Experimental settings parameters.
Table 2. Experimental settings parameters.
Parameter NameValueParameter
NC500Chromosomes
NF2Number of constructed features
NG200Maximum number of generations
H10Processing nodes
pS0.10Selection rate
pM0.05Mutation rate
Table 3. Error rates (%) of the methods applied for the classification procedures.
Table 3. Error rates (%) of the methods applied for the classification procedures.
Method
FC2RBFRBFMLP BFGSMLP PCADATASET
5.41%15.48%14.45%27.16%Eye-tracking
21.85%23.28%35.19%28.58%Heart rate
20.33%21.81%27.20%25.04%Game scores
Table 4. Precision of the methods applied for the classification procedures.
Table 4. Precision of the methods applied for the classification procedures.
Method
FC2RBFRBFMLP BFGSMLP PCADATASET
0.91250.68870.76440.5558Eye-tracking
0.57480.52640.50670.5006Heart rate
0.56040.53440.55740.5344Game scores
Table 5. Recall of the methods applied for the classification procedures.
Table 5. Recall of the methods applied for the classification procedures.
Method
FC2RBFRBFMLP BFGSMLP PCADATASET
0.93710.89060.82310.6004Eye-tracking
0.76390.80760.54050.5545Heart rate
0.70650.69050.58720.5598Game scores
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Toki, E.I.; Tatsis, G.; Pange, J.; Tsoulos, I.G. Constructing Features for Screening Neurodevelopmental Disorders Using Grammatical Evolution. Appl. Sci. 2024, 14, 305. https://doi.org/10.3390/app14010305

AMA Style

Toki EI, Tatsis G, Pange J, Tsoulos IG. Constructing Features for Screening Neurodevelopmental Disorders Using Grammatical Evolution. Applied Sciences. 2024; 14(1):305. https://doi.org/10.3390/app14010305

Chicago/Turabian Style

Toki, Eugenia I., Giorgos Tatsis, Jenny Pange, and Ioannis G. Tsoulos. 2024. "Constructing Features for Screening Neurodevelopmental Disorders Using Grammatical Evolution" Applied Sciences 14, no. 1: 305. https://doi.org/10.3390/app14010305

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop