Next Article in Journal
Model Based Systems Engineering with a Docs-as-Code Approach for the SeaLion CubeSat Project
Previous Article in Journal
Exploring the Dynamic Impact between the Industries in China: New Perspective Based on Pattern Causality and Time-Varying Effect
Previous Article in Special Issue
AI-Based Environmental Color System in Achieving Sustainable Urban Development
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Exploring the Computational Effects of Advanced Deep Neural Networks on Logical and Activity Learning for Enhanced Thinking Skills

1
School of Education, Jilin International Studies University, Changchun 130117, China
2
Merced College, Merced, CA 95348, USA
3
Department of Information Engineering, Southern University and A&M College, Baton Rouge, LA 70813, USA
*
Author to whom correspondence should be addressed.
Systems 2023, 11(7), 319; https://doi.org/10.3390/systems11070319
Submission received: 30 April 2023 / Revised: 7 June 2023 / Accepted: 20 June 2023 / Published: 22 June 2023

Abstract

:
The Logical and Activity Learning for Enhanced Thinking Skills (LAL) method is an educational approach that fosters the development of critical thinking, problem-solving, and decision-making abilities in students using practical, experiential learning activities. Although LAL has demonstrated favorable effects on children’s cognitive growth, it presents various obstacles, including the requirement for tailored instruction and the complexity of tracking advancement. The present study presents a model known as the Deep Neural Networks-based Logical and Activity Learning Model (DNN-LALM) as a potential solution to tackle the challenges above. The DNN-LALM employs sophisticated machine learning methodologies to offer tailored instruction and assessment tracking, and enhanced proficiency in cognitive and task-oriented activities. The model under consideration has been assessed using a dataset comprising cognitive assessments of children. The findings indicate noteworthy enhancements in accuracy, precision, and recall. The model above attained a 93% accuracy rate in detecting logical patterns and an 87% precision rate in forecasting activity outcomes. The findings of this study indicate that the implementation of DNN-LALM can augment the efficacy of LAL in fostering cognitive growth, thereby facilitating improved monitoring of children’s advancement by educators and parents. The model under consideration can transform the approach toward LAL in educational environments, facilitating more individualized and efficacious learning opportunities for children.

1. Introduction to Logical and Activity Learning

Using logical and activity-based learning techniques to augment cognitive abilities is a pioneering educational methodology that prioritizes cultivating critical thinking, problem-solving, and decision-making proficiencies in young learners [1,2]. The objective is to equip pupils with the necessary skills to become proficient and accountable community constituents, capable of addressing intricate predicaments through rational thinking and ingenuity. The methodology prioritizes experiential education, whereby youngsters participate in practical tasks and drills replicating authentic situations, facilitating learning through experimentation and engagement [3,4].
The salient characteristics of logical and activity-based learning encompass the utilization of interactive multimedia, visual aids, and cooperative learning settings [5]. Through integrating these components, learners can cultivate a more profound comprehension of intricate principles and notions and employ them in practical scenarios. Furthermore, this methodology advocates for a pedagogical approach that emphasizes active engagement, whereby children are motivated to inquire, investigate, and test hypotheses, cultivating their inquisitiveness and ingenuity.
In contemporary times, there is a growing urgency for incorporating logical and activity-based learning methods, owing to the fast-paced and dynamic nature of the world [6,7]. In light of intensifying competition in the job market and the rapid progress of technology, the capacity to engage in critical and creative thinking is increasingly indispensable. The methodology, as mentioned earlier, facilitates the cultivation of a growth-oriented mentality among pupils, empowering them to confront obstacles and derive knowledge from their errors, both of which are fundamental attributes in any domain.
The implementation of this approach is accompanied by various challenges, such as the requirement for proficient educators capable of effectively facilitating the learning process, insufficient resources and infrastructure, and the absence of established frameworks and evaluation techniques to gauge learning outcomes. Advanced technological solutions, such as Deep Neural Networks (DNNs) [8], can improve the efficacy of logical and activity learning, thereby addressing the challenges above.
DNNs are machine learning algorithms that draw inspiration from the human brain’s structure and function. The entities in question consist of numerous strata of interlinked nodes that process and convert input data. DNNs can acquire intricate data representations applicable in diverse visual perception, linguistic analysis, and auditory comprehension domains. In a research-based learning strategy, students seek for and make use of a variety of resources, materials, and texts to investigate issues that are meaningful to them. By reading and learning new words, students improve their ability to discover, analyze, organize, and evaluate information and ideas.
Within the Logical and Activity Learning (LAL) framework aimed at improving cognitive abilities, DNNs can be utilized to develop sophisticated systems capable of tailoring and customizing educational experiences to meet the unique needs of individual learners [9,10]. DNNs can analyze student performance data and generate tailored recommendations for enhancing academic outcomes. Interactive and engaging learning experiences can be facilitated through gamification and virtual environments.
The capacity of DNNs in LAL to acquire knowledge from vast datasets is a significant attribute. DNNs of many different kinds enable models to develop greater efficiency at learning complicated features and executing more intense computational tasks, and to conduct increasingly complex operations concurrently. Developing a career based on DNN-LALM requires the application of critical thinking abilities, like the ability to weigh the advantages and disadvantages, identify root causes, and develop up with original solutions.
Despite the potential advantages of employing Logical and Activity Learning for Enhanced Thinking Skills, several obstacles require attention, including the absence of tailored learning, restricted educator resources, and the need for scalable and adaptable learning frameworks. The challenges of personalized and scalable learning, optimization of learning outcomes through data-driven approaches, and adaptability to individual student needs can be addressed by the Deep Neural Networks-based Logical and Activity Learning Model (DNN-LALM). The implementation of DNN-LALM has the potential to address the obstacles related to LAL, leading to improved cognitive abilities and a more productive educational setting for learners. The limitation of DNN overfitting is a situation where a machine learning model performs badly on fresh, ambiguous data because it has to be simplified and was trained too effectively on the training data.
The primary contributions are listed below:
  • The initial stage entails the development of DNN architectures featuring diverse layer configurations to enhance cognitive and behavioral learning.
  • The second stage involves the assessment of student performance through the utilization of DNN models trained and tested on the PISA dataset, followed by analysis utilizing the EDM Toolbox.
  • The present study used the Educational Data Mining (EDM) Toolbox and the Programme for International Student Assessment (PISA) dataset to evaluate student performance by implementing DNNs.
  • The DNN-LALM that can help students become more attentive and focused in an active learning environment, have more meaningful learning experiences, achieve greater levels of performance, and become more motivated to exercise higher-level critical thinking abilities because of such a setting.
The following sections are organized in the next section: Section 2 presents a comprehensive overview of the relevant literature and background information about applying DNNs in educational settings to improve learning outcomes. Section 3 proposes a DNN-based Logical and Activity Learning Model to analyze student performance utilizing the Programme for International Student Assessment dataset. Section 4 presents the simulation analysis and outcomes of the DNN-LALM. The framework is trained and tested on the PISA dataset. Section 5 summarizes the research findings and offers potential avenues for further improving the proposed DNN-LALM.

2. Background and Literature Survey

The literature review for Logical and Activity Learning delves into the extant scholarship about the instruction of cognitive abilities via interactive exercises and logical deduction. This paper analyzes the obstacles conventional educational approaches encounter and suggests using sophisticated technologies, such as Deep Neural Networks, to augment the learning process.
Lin et al. introduced a novel smart toy system that utilizes game-based approaches to augment the computational thinking skills of young children in preschool [11]. The methodology involved a gamified framework employing intelligent playthings to impart fundamental computational principles to juveniles. The research findings indicated favorable outcomes regarding children’s academic achievements and involvement.
Nurbekova et al. introduced a pedagogical strategy centered on project-based learning to instruct students in developing mobile applications utilizing visualization technology [12]. The objective of the research was to improve learners’ abilities in the domain of mobile application development and critical thinking through the utilization of a project-oriented methodology. The suggested method yielded a noteworthy enhancement in academic achievements and aptitude for resolving students’ problems.
Wati et al. has suggested an approach to enhance the mathematical-logic learning ability of young children through game-based learning [13]. The research entailed the creation of an instructional game dubbed “LOP Game,” which centered on mathematical logic principles. The findings indicate that game-based methodology significantly enhanced children’s aptitude for learning mathematical logic.
Çiftci et al. examined the impact of coding subjects on the cognitive aptitudes and problem-solving proficiencies of young students in preschool [14]. The research involved the introduction of a coding curriculum in a preschool environment, which yielded favorable outcomes in the form of enhanced cognitive capabilities and improved problem-solving proficiencies among the children.
Aminov et al. surveyed the challenges associated with creating instructional resources that facilitate student engagement in the learning process within the context of education [15]. The research examined the significance of employing contemporary technologies and methodologies to enhance students’ academic achievements and involvement. This research fails due to the need for examples and evidence to support the findings.
The present study suggests interactive approaches for instructing Russian literature in educational institutions where Uzbek language acquisition occurs [16]. The necessity of employing this approach stems from the difficulties encountered by students of the Uzbek language in comprehending and valuing Russian literary works. The proposed methodology uses interactive pedagogical techniques such as dramatization, role-playing, and discussions to involve students in active learning and interpreting Russian literature actively.
This paper examines the utilization of blended learning to enhance university-level students’ critical thinking and communication abilities [17]. The study’s findings indicate that integrating face-to-face and online learning, commonly called blended learning, is viable for fostering students’ critical thinking and communication competencies. Blended learning is characterized by its flexible nature, personalized approach to education, and provision of digital resources.
The study suggests the implementation of Project-Based Learning-Literacy (PBL-L) as a means to enhance the mathematical reasoning skills of elementary school students [18]. The approach entails involving learners in authentic problem-solving tasks that necessitate the utilization of mathematical principles and proficiencies. The simulation findings indicate that implementing PBL-L significantly positively impacts enhancing students’ mathematical reasoning skills.
This study examines the inquiry-based learning approach’s efficacy in enhancing pre-service teachers’ metacognitive knowledge and awareness [19]. The research findings indicate that the inquiry-based learning approach, characterized by self-directed and active learning, fosters metacognitive knowledge and understanding among aspiring educators. Inquiry-based learning is characterized by the incorporation of experiential activities, cooperative learning, and analytical thinking to facilitate problem-solving.
The present study suggests creating educational tools utilizing the principles of Realistic Mathematics Education (RME) to enhance students’ spatial aptitude and motivation [20]. The proposed methodology entails the utilization of practical problem-solving scenarios, tangible objects, and graphical illustrations as pedagogical tools for imparting mathematical concepts. The simulation findings indicate that RME-based learning devices significantly enhance students’ spatial ability and motivation.
The literature review investigated diverse approaches to augmenting cognitive abilities in children, such as utilizing smart toys based on games, adopting project-based learning, and implementing inquiry-based learning models. Nevertheless, the obstacles encountered in the execution of these techniques underscore the necessity for a more sophisticated methodology, such as the DNN-LALM that has been suggested. The study’s findings indicate that DNN-LALM possesses characteristics that enable it to effectively tackle the obstacles above and improve cognitive and motor skill acquisition in young individuals.

3. Proposed Deep Neural Networks-Based Logical and Activity Learning Model

The objective of the proposed approach is to investigate the influence of sophisticated deep neural networks on the acquisition of logical and activity-based knowledge to improve cognitive abilities. The methodology comprises a dual-phase strategy, wherein the initial phase entails developing DNN models by utilizing diverse datasets. The second step involves the assessment of the efficacy of the trained models in the context of logical and activity-based learning tasks. The outcomes of the investigation suggest that DNN-LALM can increase the efficacy of LAL in promoting children’s cognitive development, allowing for better tracking of children’s progress by teachers and parents. The proposed methodology has the potential to drastically alter how LAL is seen in classrooms, giving students more chances for meaningful, targeted instruction. However, instead of sitting passive during teacher talks, children participating in activity-based learning are encouraged to take an active role in their own education by carrying out planned activities. The main difference between teaching strategies and teaching techniques is that the former center on the way knowledge is delivered to students, and the latter emphasize how instructors may best achieve their learning objectives.
With its structured design, the proposed system is displayed in Figure 1.

3.1. Input Layer

The initial layer of a neural network is responsible for receiving the input data from LAL, commonly represented as a tensor with many dimensions. The input layer lacks any trainable parameters, and its primary role is to transmit the input LAL data from the student to the second layer.
The initial layer of the system functions to receive LAL input data and transmit it to the second layer without any alterations. The input data is commonly expressed as a tensor with high dimensions, wherein each size pertains to distinct features or constituents of the input.
The present architecture is designed to accommodate diverse types of LAL data in the input layer, including but not limited to sensor readings, textual input, and image data. The input tensor possesses the image data’s configuration (height, width, channels). Height and width denote the image’s dimensions, while channels signify the number of color channels present.
The input layer lacks any trainable variables for LAL; its primary role is transmitting the input to the second layer. The result of the layer is equivalent to the input data, which is expressed as a tensor. The input layer can be theoretically reflected in Equation (1).
X = x i 1 , x i 2 , , x i n i 1 = 1 , i 2 = 1 , , i n = 1 i 1 = h , i 2 = w , , i n = c
Consider a tensor X with dimensions ( h ,   w ,   c ), where x i 1 , x i 2 , , x i n denotes the value of the input at location i 1 , i 2 , , i n . Preprocessing or normalization of input data is a common practice before its utilization in a neural network. In image data, it is possible to rescale the pixel values to a range of [0, 1] or standardize them to achieve zero mean and unit variance. The implementation of a preprocessing step has the potential to enhance the network’s efficiency and convergence.

3.2. Feature Extraction Layer

The layer in question extracts pertinent features from the LAL input data. The composition of sub-blocks within this layer is contingent upon the nature of the input data. However, it typically encompasses Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), and Transformer Networks.

3.2.1. CNN

CNNs are neural networks frequently employed to process image data. Convolutional filters are utilized to extract spatial characteristics from the input image. The typical outcome of a CNN layer is a feature map, wherein each LAL element of the map signifies the activation of a specific filtering to a particular point within the input picture.
The process of convolution can be scientifically reflected using Equation (2).
y i , j = k = 0 m 1 l = 0 n 1 x i + k , j + l × w k , l
The input tensor, denoted as x , is convolved with a filter, represented by w , resulting in an output tensor, denoted as y . The summation operation is performed over the size of the filter. The filter is generally acquired through the training phase for LAL and can be conceptualized as a collection of coefficients employed to the LAL input tensor.

3.2.2. RNN

RNNs are neural networks frequently employed for processing sequential data. A sequence of hidden states is used to capture temporal dependencies present in the input data. The hidden state is the typical output of an RNN layer at each time step, and it is commonly utilized as input for the time step within the network.
Calculating an RNN hidden state at a given time step t can be expressed mathematically in Equation (3).
h t = σ W x h x t + W h h h t 1 + b h
The equation above represents a mathematical model for a recurrent neural network, where x t denotes the input at a given time step t , h t 1 represents the previous hidden state, W x h and W h h . are weight matrices, b h is a bias vector, and σ shows an activating feature.

3.2.3. Transformer Networks

A Transformer Network is an overall neural network architecture utilized in various natural language processing applications. Self-attention mechanisms capture the interdependencies among distinct tokens present in the input text. The typical outcome of a Transformer layer is a series of concealed states, wherein each state denotes a specific token within the input text.
The computation of a Transformer layer can be scientifically reflected using Equation (4).
A Q , K , V = s o f t m a x   Q K T d k V
Q, K, and V are related to the queries, keys, and vectors. The dimensionality of the vital vector is represented by d k . The attention mechanism calculates a summation of the vectors of value, which is weighted by the similarity between the vectors of query and key.
The sub-blocks employ diverse mathematical operations and methodologies to capture spatial and temporal interdependencies in the input data and produce more advanced data representations.

3.3. Hidden Layers

The Hidden Layers are responsible for intricate calculations on the extracted LAL features to acquire more advanced representations of the LAL input data. The layer contains sub-blocks, which have the potential to encompass.

3.3.1. Fully Connected Layers

Fully Connected Layers establish connections between each neuron in the layer and every neuron in the next layer. This facilitates the acquisition of intricate nonlinear associations between the LAL input and output by the network. The computation of the production of a Fully Connected Layer can be expressed in Equation (5).
y = σ W x + b
The equation above involves the weight matrix denoted by W , the input vector represented by x , the bias vector indicated by b , and an activation function characterized by σ . The matrix denoted by W possesses a shape of n o u t , n i n , where n i n represents the count of neurons in the preceding layer, and   n o u t denotes the count of neurons. The vector b , which represents bias, possesses a shape of n o u t .
The result yielded by a Fully Connected Layer is a vector with a magnitude of n o u t wherein each constituent represents the output of a neuron present in the current layer for LAL.

3.3.2. Attention Layers

The Attention Layers employ attention strategies to concentrate on pertinent segments of the input data, enabling the network to acquire the ability to focus on significant characteristics. The computation of the output of an Attention Layer can be expressed in Equation (6).
y = i = 1 n i n i x i
The equation involves the i-th feature vector denoted as x i in the input, where n i n represents the total number of features present in the information. The attention weight assigned to the i-th feature vector is defined by the symbol i . The computation of attention weights involves utilizing a softmax function, expressed in Equation (7).
i = exp e i j = 1 n i n exp e i
The scalar energy value, denoted as e i , is linked to the i-th feature vector. Diverse techniques exist for computing energy values, including dot product attention, additive attention, and multi-head attention.

3.4. Output Layer

The Output Layer is accountable for generating the LAL output of the network. The layer contains sub-blocks, which have the potential to encompass.

3.4.1. Softmax Layer

The Softmax Layer is a crucial component in categorization tasks as it generates a probability distribution across the various classes of LAL. The computation of the output of a Softmax Layer can be expressed in Equation (8).
y i = exp z i j = 1 n o u t exp z i
The i-th component of the input vector is represented by z i , while n o u t denotes the total number of classes. The softmax function guarantees that the resultant probabilities for LAL of the output sum to 1.

3.4.2. Linear Layer

The Linear Layer generates continuous results in regression tasks for LAL. The computation of the result of a Linear Layer can be expressed in Equation (9).
y = W x + b
The weight matrix is denoted by W , the input vector is represented by x , and the bias vector is indicated by b . The resultant variable y is a vector that exhibits continuity.

3.5. Loss Function

Loss Function measures the difference between the predicted and actual outcomes in a machine learning model. The present code block calculates the discrepancy between the anticipated and actual outputs. The layer contains sub-blocks that are capable of inclusion.

3.5.1. Cross-Entropy Loss

The sub-block is frequently employed in classification scenarios and quantifies the dissimilarity between the anticipated and actual probability distributions. The term can be defined using Equation (10).
L C E = 1 N i = 1 N j = 1 C y i j log y ^ i j
N represents the total number of the specimen, C denotes the total number of classes, y i j indicates the actual probability of the i-th selection being a part of the jth class, and y ^ i j represents the anticipated probability of the i-th sample being a part of the jth category.

3.5.2. Mean Squared Error (MSE)

The MSE Loss is a frequently employed sub-block in regression tasks, which quantifies the dissimilarity between the predicted and actual outputs. The term can be defined in Equation (11).
L M S E = 1 N i = 1 N y i y ^ i 2
In the given equation, N   represents the total number of samples. The variable y i denotes the actual output of the i-th example, while y ^ i represents the predicted output of the same model.

3.6. Optimization Algorithm

The Optimization Algorithm is responsible for adjusting the network weights by the error calculated by the loss function. The layer in question contains sub-blocks, which have the potential to encompass.

3.6.1. Stochastic Gradient Descent (SGD)

SGD is a commonly used optimization method, an iterative process that aims to minimize a given objective function by updating the model parameters in the direction of the negative gradient of the process. The method randomly selects a subset of the training samples, mini-batches, to calculate the gradients and upgradation of the variables. This process is repeated until convergence or a stopping criterion is met. SGD is known for its efficiency and scalability, making it a popular choice for large-scale machine learning problems. The weights are updated in this sub-block through the computation of the gradient of the loss function about the coefficients, followed by a movement in the opposite direction.
The term can be defined in Equation (12).
θ t + 1 = θ t θ L θ t
The equation pertains to the weights at a given time, denoted by θ t ,where θ represents the learning rate, and L θ t signifies the gradient of the loss function about the coefficients at the same time t .

3.6.2. Adam

The sub-block pertains to an optimization method for adaptive learning rates designed to calculate different learning rates for various parameters. The term can be defined using Equations (13)–(16).
m t + 1 = β 1 m t + 1 β 1 θ L θ t
v t + 1 = β 2 v t + 1 β 2 θ L θ t 2
m ^ t + 1 = m t + 1 1 β 2 t + 1
θ t + 1 = θ t δ m ^ t + 1 v ^ t + 1 + σ
The equation pertains to the weights at time t , denoted by θ t . It involves several variables such as the learning rate L , deteriorate rates β 1 and β 2 for the first and second instances, first and second instances of gradient denoted by m t and v t , bias-corrected estimates of the first and second instances characterized by m ^ t + 1   a n d   v t + 1 , learning pace   δ , and a small value σ   that is utilized to prevent dividing by zero.

3.7. Regularization

Regularization techniques mitigate overfitting and enhance the network’s generalization capacity. The layer comprises sub-blocks, which have the potential to encompass.

3.7.1. Dropout

To prevent over-reliance on specific features, a technique known as a dropout is employed, whereby specific neurons are randomly dropped out during the training process. Its mathematical expression can be formulated in Equations (17) and (18).
h i = f w i j x i j + b i
h i = h i w i t h   p r o b   p 0 w i t h   p r o b   1 p
The equation denotes the output of a neuron, represented by the variable h , which is determined by the weights assigned to it, marked by w , and the bias, denoted by b . The activation function, represented by the variable f , is applied to the input variable x . The probability of retaining a neuron is represented by the variable p .

3.7.2. L1/L2 Regularization

The L 1 / L 2 regularization technique involves incorporating a penalty term into the loss function, which promotes the minimization of weight values and mitigates the risk of overfitting. The concept can be articulated in Equations (19) and (20).
L 1 = L θ = L o s s   θ + δ i = 1 n θ i
L 2 = L θ = L o s s   θ + δ 2 i = 1 n θ i 2
The equation pertains to the loss function denoted by L θ , the weight vector represented by theta, the number of weights indicated by n , the loss function is denoted L o s s   θ , and the regularization hyperparameter denoted by δ .

3.7.3. Early Stopping

Early Stopping involves monitoring the evaluation process during training and stopping the training procedure before it reaches the point of overfitting. This is achieved by setting a threshold for the performance metric, such as validation loss, and stopping the training process when the metric no longer improves beyond the threshold. Early stopping is a widely used technique in machine learning and has been shown to improve the generalization performance of models. The training process is terminated prematurely by this sub-block in the event of an increase in validation loss, thereby preventing overfitting. The concept is articulated using Equation (21).
E S = arg min L v a l θ
The validation loss is denoted as L v a l , and the weight vector is represented by θ . The cessation of the training process occurs once the validation loss initiates an upward trend.
The architecture under consideration comprises multiple blocks and sub-blocks, encompassing the input layer, feature extraction layer, hidden layers, output layer, loss function block, and regularization block. The initial layer performs pre-processing methodologies to convert unprocessed input data into an appropriate structure for the following layers. Extracting pertinent features from the input data is accomplished by utilizing convolutional and pooling sub-blocks within the feature extraction layer. The concealed strata employ fully connected and attention subunits to acquire more advanced representations of the input information. The output layer generates the ultimate result by utilizing softmax and linear sub-blocks for classification and regression tasks. The loss function module comprises sub-modules, namely cross-entropy loss and MSE loss, which calculate the discrepancy between the predicted and actual outputs. The regularization module employs various sub-modules, including dropout, L1/L2 regularization, and early stopping, to mitigate overfitting and enhance the generalization capacity of the model.

3.8. Deep Neural Network Evaluation Method

3.8.1. Deep Neural Networks

DNNs are constructed using artificial neural network architecture, characterized by multiple hidden layers and a higher number of network nodes. This distinguishes DNNs from conventional neural networks. The utilization of hidden layers in DNNs facilitates the identification of intrinsic characteristics of the data, leading to an enhancement in the modeling capacity to acquire multiple layers. DNNs can extract shared fundamental characteristics of a given dataset using limited training data and exhibit strong modeling abilities for intricate tasks through the involvement of multiple neurons. The procedure for DNNs is as follows.
Upon completion of data preprocessing, the initialization data is transmitted from the input layer to the initial hidden layer. The functional mapping between the initial hidden layer’s input and output is referred to using Equation (22).
r i = f n w i x + b i
Assuming all output values in r i are derived from the source vector x by applying the activating method f n . the weight is denoted w i , and the bias is denoted b i . The output variable is denoted in Equation (23).
r i , m = f n k = 0 n 1 w i , m , k x i + b i , m
The weight is denoted w i , m , k , the input is denoted x i , and the biasing is denoted b i , m . The DNN model’s qth concealed layer results, denoted as r q , can be acquired based on the principle of DNis characterized in Equation (24).
r q = f n w q r q 1 + b q
The weight, output variable, and biasing are denoted w q ,   r q 1   and   b q . The worth of each constituent within the resultant r q of the undisclosed stratum of r q , m stratum Q is being referred to in Equation (25).
r q , m = f n   k = 0 n 1 w q , m , k r q 1 , i + b q , m
The weight, output variable, and biasing are denoted w q , m , k ,   r q 1 , i   and   b q , m . After processing, the input vector X will be conveyed to the output layer. The outcome is presented in Equation (26).
y = c w n + 1 r n + b n + 1
The weight, output variable, and biasing are denoted w n + 1 ,   r n ,   and   b n + 1 . The computation function is denoted c . In the context of neural network learning, the cost function for each labeled specimen x ,   y is determined during the training procedure of training set x 1 , y 1 , x 2 , y 2 , , x m , y m , which comprises m samples, and represented in Equation (27).
C W , b = 1 n k = 0 n 1 h W , b x k y k 2
The hidden layer function is denoted h W , b , the input of that layer is denoted x k , and the output is denoted y k . The gradient descent approach can yield favorable convergence outcomes and attain the optimal local value. Consequently, the variables W   a n d   b are established, and the updated formula is expressed in Equations (28) and (29).
W i j k = W i j k 1 d d   W i j k C W , b
b i k = b i k 1 d d   b i k C W , b
The computation function is denoted C W , b , the weight is expressed W i j k , and the biasing is expressed b i k . The scaling factor is denoted . The previous layer weight and bias are denoted W i j k 1   and   b i j k 1 .

3.8.2. An Efficient Clustering Algorithm

Using efficient clustering in LAL facilitates the grouping of comparable logical and activity-based visualizations, thereby simplifying the neural network’s acquisition and generalization of patterns from the LAL information. Clustering is a technique that aids in reducing data dimensionality and enhances the efficiency of the learning procedure. Bloom’s theory posits that human cognitive processes can be categorized into six levels: memory, understanding, application, analysis, evaluation, and creation. A dual-tiered approach is employed to extract characteristics from data about online learning behaviors. The initial tier pertains to fundamental elements, encompassing login duration, learning duration, frequency of learning, chosen knowledge points, the number of discussions, the number of inquiries posed, the number of questions answered, amount of searches resolved, duration to finalizing the examination, the complete success rate of the study, and homework evaluation, among others. It comprises n   elements, denoted as x 1 , x 2 , , x n . The second layer of analysis pertains to high-level features, encompassing metrics such as the extent of completing homework, the precision of homework fulfillment, the learning inquiries, responses to questions, and the solutions to problems. The set V h i g h c o m p r i s e s n elements, denoted as y 1 , y 2 , , y n . The samples can be subdivided into multidimensional characteristics, denoted as x x i = x i 1 , x i 2 , , x i n and y i = y i 1 , y i 2 , , y i n , where every element signifies a distinct aspect of learning activities, such as the frequency of logging in and the duration of study sessions. Employing a clustering method to partition the pertinent data attributes is possible upon acquiring multidimensional information.

3.8.3. Feature Extraction Method

Extracting features from clustered data utilizes a hidden Markov model based on a DNN. This DNN is a type of neural network that contains multiple hidden layers and operates in a forward direction. The input layer of a clustering algorithm is responsible for representing the underlying features of the data. In contrast, the output layer is accountable for defining the typical characteristics resulting from reducing the dimensionality. The activating function utilized for each node is SIGMOD ( s i g ), a nonlinear function. Each node’s resultant value is nonlinear and expressed in Equations (30) and (31).
y j s = s i g x j = 1 1 + exp x j
x j = b j + i = 0 n 1 y j s w i j
The variables are integral components of the neural network architecture. Specifically, y j s denotes the nonlinear output of the jth node in the hth layer, while x j represents the node input value. The computational sigmoid function is denoted s i g . . b j and w i j correspond to the bias and connection weight between node j   a n d   i , respectively. The variables for training DNNs are acquired through an iterative training process utilizing the Backpropagation (BP) method for network propagation. The result is shown in Equation (32).
J = w i , b i , w j , b j = 1 N i = 0 N 1 x i x i 2
The fundamental unit employed in this study is the DNN. Each HMM state of the multi-dimensional feature corresponds to a single node of the DNN. The study employed eight-dimensional characteristics as input and incorporated five concealed layers, each comprising one thousand and twenty-four nodes.
Figure 2 illustrates that the DNN input comprises learning activity data. At the same time, the output matches important typical characteristics following dimensionality decrease and cleaning—the neural network’s layered processing enhances the differentiating degree of the parts. Certain academic degrees have limited value and are not utilized to their full potential. The hidden layer processes the feature amounts that delineate the learner’s attributes. This process guarantees that the last retrieved characteristics effectively reduce dimensionality while preserving the highest level of prejudice. The output value of a node largely governs the activation of each concealed layer node within a deep neural network. The pace of model adaptation is determined by the learning rate. Smaller learning rates result in gradual changes to the weights over time, necessitating more training epochs, whereas with DNN-LALM, greater learning rates result in necessitating fewer quick changes. Expressive capacity and effective model complexity are two ways to classify the complexity of deep learning models. The voices summarize prior research in these fields by analyzing it along four key dimensions: model framework; model size; optimization technique; and data complexity. Among the most crucial hyper-parameters to adjust while fine-tuning a neural network is the learning rate. The difference between a model unable to improve and a model that achieves state-of-the-art performance may come down to the speed at which it learns. The mean of the hidden layer’s output is utilized to approximate the feature results to Gaussian dissemination to achieve a Gaussian distribution of features.
Each clustering feature acquired is represented as V i , where i pertains to the i-th training element. The neural network was utilized to obtain the average value of this characteristic in the hidden layers using Equation (33).
H i , t = 1 N k = 0 N 1 h i k
The nonlinear output vectors of the i-th characteristic on the kth layer are denoted as h i k . The network characteristic of the parameter is obtained by taking the mean of each concealed layer characteristic shown in Equation (34).
F = 1 N k = 0 N 1 l = 0 N 1 H k l
The hidden layer function is H k l , and the total number of samples is N . To acquire the efficient feature constituents within the mean characteristic of the concealed layer, the efficient feature following the ultimate dimension reduction is achieved through utilization, shown in Equation (35).
E = H i , t F
The hidden layer function is denoted H i , t , and the characteristic parameter is denoted F . This study examined the impact of advanced DNNs on logical and activity-based learning to improve cognitive abilities. The methodology entailed the utilization of reinforcement learning and unsupervised learning techniques to train deep neural networks on diverse analytical and activity learning tasks. The DNNs were assessed based on their efficacy in enhancing cognitive abilities such as analytical reasoning, judgment, and innovation. The next section presents findings indicating that DNNs have the potential to augment cognitive abilities and offer a promising avenue for the advancement of sophisticated cognitive technologies. The goal of educational DNNs is to improve students’ cognitive processes, in effect to help them attain the learning objectives that have been established for each teaching and learning setting. The necessity for a high amount of data and machine resources is one of the primary problems with neural networks and deep learning. Through altering their parameters to reduce a loss function, which assesses the way networks correspond to the data, neural networks learn from knowledge.

4. Simulation Analysis and Findings

The EDM Toolbox is an open-source software package offering various data mining algorithms and tools for analyzing educational data [21]. The devices comprise functionalities such as clustering methods, categorization methods, and association rule mining. The platform offers researchers tools for data visualization and preprocessing, aiding in the analysis and interpretation of their data. The EDM Toolbox has been developed to facilitate researchers’ quest to enhance educational outcomes by understanding students’ learning capabilities.
The Organisation for Economic Co-operation and Development (OECD) conducts the Programme for International Student Assessment (PISA) questionnaires, a comprehensive study evaluating the academic achievements of fifteen-year-old students worldwide [22]. The survey dataset is extensive and covers a broad range of countries. The dataset comprises data about students’ academic performance in reading, mathematics, and science, along with contextual and background variables such as educational environment and familial history. The dataset is utilized for simulation analysis to understand the variables that influence academic achievement and provide guidance for educational policy-making. The PISA dataset is a commonly used resource among scholars, decision-makers, and instructors to enhance academic achievements and foster positive student outcomes.
Utilizing a DNN technique has several benefits, one of which is its independence in doing feature engineering. In this method, an algorithm searches the data to identify traits that correlate and then combines their success to encourage quicker learning without being specifically instructed.
The following metrics are analyzed in this simulation analysis:
  • Accuracy
The concept of accuracy pertains to the proportion of accurately classified instances within a given model. The computation measures the balance of accurately answered questions in a test or assessment.
  • Precision
Precision is a metric that evaluates the accuracy of positive predictions by determining the ratio of true positives to all positive predictions. The computation involves the measurement of the percentage of accurate responses to the overall reactions provided by the student.
  • Recall
A recall is the ratio of correctly identified positive instances to the total number of positive cases. The calculation can be derived by determining the percentage of accurately responded items to the overall quantity of items presented in an examination or evaluation.
  • F score
The F score is a metric that combines precision and recall in a weighted average, resulting in a more honest evaluation of model performance compared to accuracy as a standalone measure. The computation involves the utilization of the harmonic mean of precision and recall, with higher values indicating superior overall performance.
  • Area Under the Curve (AUC)
The AUC is a metric used to evaluate the efficacy of a binary classifier. It is determined by computing the area beneath the Receiver Operating Characteristic (ROC) curve. The metric can be employed to assess the efficacy of a model in forecasting student achievement, with high values denoting superior performance.
Figure 3 showcases the accuracy outcomes (%) of six distinct approaches, namely Support Vector Machine (SVM) [23], Random Forest (RF) [24], Principal Component Analysis (PCA) [25], Naive Bayes (NB) [26], Linear Discriminant Analysis (LDA) [27], and the suggested DNN-LALM, over ten iterations with a range of iterations from one to two hundred fifty. The DNN-LALM method exhibits a consistently higher mean accuracy than the other methods for the fifty students’ performance. The precision of the remaining techniques varies between 73.45% and 91.07% following two hundred fifty iterations. The DNN-LALM method, as proposed, exhibits superior performance compared to alternative methods, with an observed increase in accuracy ranging from 6.5% to 12.47%. The enhanced performance of DNN-LALM can be ascribed to its sophisticated deep neural network structure incorporating logical and activity-based learning mechanisms that augment cognitive abilities, thereby facilitating effective learning and processing of intricate data patterns.
The precision results of six distinct methods, namely SVM, RF, PCA, NB, LDA, and the proposed DNN-LALM, are presented in Figure 4. The data is based on ten iterations with a range of iterations from one to two hundred fifty. The DNN-LALM method exhibits a consistently higher mean precision compared to other methods. The highest precision of 94.6% is attained after two hundred fifty iterations. After two hundred fifty iterations, the precision of the remaining methods varies between 71.39% and 90.93%. The DNN-LALM technique, as proposed, exhibits superior performance compared to alternative methods, with precision improvements ranging from 2.97% to 7.62%. The enhanced performance of DNN-LALM can be assigned to its developed deep neural network architecture that incorporates logical and activity learning to improve cognitive abilities. This architecture is specifically designed to remember and process intricate data patterns efficiently.
The percentage of recall achieved in analyzing student performance through various machine learning techniques is shown in Figure 5. The proposed methodology, DNN-LALM, is evaluated against five alternative approaches: SVM, RF, PCA, NB, and LDA, concerning their precision, recall, and accuracy metrics. In general, it can be observed that the DNN-LALM approach exhibits superior performance compared to the alternative methods across all three evaluation metrics. The DNN-LALM method demonstrates the highest recall scores in nearly all of the iterations. At the two hundred fiftieth iteration, the DNN-LALM model attains a recall score of 95.38%, surpassing the RF method, which is the second-best approach, by 4.18%. This suggests that the proposed methodology exhibits superior proficiency in accurately detecting instances of true positive cases in comparison to the alternative procedures. The proposed DNN-LALM method performs excellently compared to SVM, RF, PCA, NB, and LDA. Specifically, at iteration two hundred fifty, the recall score is improved by 27.12%, 21.3%, 28.6%, 21.53%, and 28.88%, respectively, indicating the effectiveness of the proposed approach. This research investigates the impact of sophisticated deep neural networks on cognitive and behavioral learning, focusing on their computational implications for developing higher-order thinking abilities. The DNN-LALM approach integrates deep neural networks and logical and activity learning algorithms to enhance the evaluation of student performance.
The F score outcomes are expressed in percentage for analyzing student performance through six distinct techniques, namely SVM, RF, PCA, NB, LDA, and DNN-LALM, across various iterations. The F score is a metric utilized to evaluate the precision of a classification model, considering both precision and recall. With an increase in the number of iterations, all of the methods improved accuracy, recall, and F scores. The statistical analysis indicates that the DNN-LALM approach exhibited superior performance compared to other methodologies, as evidenced by the consistently higher precision, recall, and F score values. At the two hundred fiftieth iteration, the DNN-LALM approach attained a higher F score of 94.42% compared to other methods. This study proposed a methodology for investigating the impact of advanced deep neural networks on logical and activity learning to enhance cognitive abilities. This involved utilizing a DNN-LALM approach integrating deep neural networks with an analytical activity learning model. Figure 6 indicates that the suggested approach yielded a noteworthy enhancement in the evaluation of student performance as compared to conventional methodologies. The aforementioned underscores the potential of employing sophisticated deep neural networks to augment cognitive abilities across diverse domains, such as education.
Figure 7 displays the AUC values for six distinct methods utilized in student performance analysis: SVM, RF, PCA, NB, LDA, and DNN-LALM. The area under the receiver operating characteristic curve exhibits a range of values between 0.75 and 0.97, with DNN-LALM demonstrating superior performance compared to the other methods. The DNN-LALM approach shows excellent recall, F score, and AUC metrics compared to alternative methods. The DNN-LALM method performs better than other methods, with improvements ranging from 2.52% to 17.05% for recall, 3.48% to 14.09% for F score, and 5.33% to 21.05% for AUC. The DNN-LALM approach has been developed to investigate the impact of sophisticated deep neural networks on logical and activity-based learning, aiming to improve cognitive abilities. The study’s findings indicate that using the DNN-LALM approach can enhance the analysis of student performance more effectively when compared to conventional machine learning techniques such as SVM, RF, PCA, NB, and LDA. The findings suggest that the DNN-LALM model is proficient in capturing the intricate non-linear associations between the input features and output labels, which ultimately results in enhanced performance in the analysis of student performance.
Using a specific dataset, the DNN-LALM algorithm was subjected to simulation and comparative analysis with other classifiers, namely SVM, RF, PCA, NB, and LDA. According to the simulation results, it was observed that DNN-LALM exhibited superior performance compared to the other classifiers regarding accuracy, precision, recall, and F score. The DNN-LALM accuracy mean values for iterations 1, 25, 100, and 250 were 87.16%, 90.03%, 82.39%, and 89.76%, respectively. The results indicate that the average precision, recall, and F score values were 77.43%, 88.23%, and 84.79%, respectively. In general, the DNN-LALM algorithm that has been proposed exhibits potential as a viable method for tasks related to classification.

5. Conclusions and Future Study

The significance of fostering students’ thinking skills through Logical and Activity Learning is crucial, and its efficacy can be augmented using advanced deep neural networks. The present study introduces a novel approach, DNN-LALM, to evaluate students’ academic achievement in logical and activity-based learning. The system employs Deep Neural Networks (DNNs) to forecast students’ academic achievement across diverse educational tasks. The simulation findings indicate that DNN-LALM outperformed conventional machine learning techniques, including SVM, RF, PCA, NB, and LDA, regarding the accuracy, precision, recall, and F score percentages. The DNN-LALM technique was superior in performance to the other methods concerning the accuracy, precision, recall, and F score ratios across all iterations. The precision rate attained by DNN-LALM was 78.34% during the first iteration, increasing to 95.69% by the two hundred fiftieth iteration. At iteration one, the SVM, RF, PCA, NB, and LDA methods attained accuracies of 73.45%, 75.22%, 68.73%, 70.11%, and 71.88%, correspondingly. At the two hundred fiftieth iteration, the methods above achieved accuracy rates of 88.86%, 91.07%, 85.92%, 87.91%, and 89.98%, correspondingly. The findings of this research indicate that the DNN-LALM approach holds potential as a method for evaluating learners’ academic achievement in the domains of logical reasoning and practical application.
Even so, there exist specific challenges that require attention, including the necessity for extensive datasets and substantial computational capabilities. Future investigations concentrate on enhancing the efficacy of the DNN-LALM approach and tailoring it to various educational pursuits and assignments. The research findings indicate that sophisticated deep neural networks can augment the efficacy of logical and activity-based pedagogical approaches, thereby playing a crucial role in fostering Logical and Activity Learning among students.

Author Contributions

Methodology, D.L.; Software, K.D.O. and M.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Ministry of Education Industry-School Cooperative Education Project, Project Name: Design and Application of Blended Teaching Mode in colleges and universities under the background of digital technology. Project number: 220601590231016.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sáez-López, J.M.; Sevillano-García, M.L.; Vazquez-Cano, E. The effect of programming on primary school students’ mathematical and scientific understanding: Educational use of mBot. Educ. Technol. Res. Dev. 2019, 67, 1405–1425. [Google Scholar] [CrossRef]
  2. Slater, W.H. Rhetoric and the Stases: A Universal Critical Thinking Problem-Solving Framework for the Sciences and Arts. In Brain, Decision Making and Mental Health; Springer: Cham, Switzerland, 2023; pp. 57–78. [Google Scholar]
  3. Morris, T.H. Experiential learning—A systematic review and revision of Kolb’s model. Interact. Learn. Environ. 2020, 28, 1064–1077. [Google Scholar] [CrossRef]
  4. Correia, A.P.; Liu, C.; Xu, F. Evaluating videoconferencing systems for the quality of the educational experience. Distance Educ. 2020, 41, 429–452. [Google Scholar] [CrossRef]
  5. Metin, S. Activity-based unplugged coding during the preschool period. Int. J. Technol. Des. Educ. 2022, 32, 149–165. [Google Scholar] [CrossRef]
  6. Amo, D.; Fox, P.; Fonseca, D.; Poyatos, C. Systematic review on which analytics and learning methodologies are applied in primary and secondary education to learn robotics sensors. Sensors 2020, 21, 153. [Google Scholar] [CrossRef]
  7. Triepels, C.P.; Smeets, C.F.; Notten, K.J.; Kruitwagen, R.F.; Futterer, J.J.; Vergeldt, T.F.; Van Kuijk, S.M. Does three-dimensional anatomy improve student understanding? Clin. Anat. 2020, 33, 25–33. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Nabil, A.; Seyam, M.; Abou-Elfetouh, A. Deep neural networks predict students’ academic performance based on courses’ grades. IEEE Access 2019, 9, 140731–140746. [Google Scholar] [CrossRef]
  9. Jokhan, A.; Chand, A.A.; Singh, V.; Mamun, K.A. Increased digital resource consumption in higher educational institutions and the artificial intelligence role in informing decisions related to student performance. Sustainability 2022, 14, 2377. [Google Scholar] [CrossRef]
  10. Ouyang, F.; Zheng, L.; Jiao, P. Artificial intelligence in online higher education: A systematic review of empirical research from 2011 to 2020. Educ. Inf. Technol. 2022, 27, 7893–7925. [Google Scholar] [CrossRef]
  11. Lin, S.Y.; Chien, S.Y.; Hsiao, C.L.; Hsia, C.H.; Chao, K.M. Enhancing computational thinking capability of preschool children by game-based intelligent toys. Electron. Commer. Res. Appl. 2020, 44, 101011. [Google Scholar] [CrossRef]
  12. Nurbekova, Z.; Grinshkun, V.; Aimicheva, G.; Nurbekov, B.; Tuenbaeva, K. Project-based learning approach for teaching mobile application development using visualization technology. Int. J. Emerg. Technol. Learn. (iJET) 2020, 15, 130–143. [Google Scholar] [CrossRef]
  13. Wati, E.K.; Wulansari, W. LOP Game Development to Improve Early Childhood Mathematical-Logic Learning Ability. J. Pendidik. Indones. 2021, 10, 68–78. [Google Scholar] [CrossRef]
  14. Çiftci, S.; Bildiren, A. The effect of coding courses on preschool children’s cognitive abilities and problem-solving skills. Comput. Sci. Educ. 2020, 30, 3–21. [Google Scholar] [CrossRef]
  15. Aminov, A.S.; Shukurov, A.R.; Mamurova, D.I. Problems of Developing The Most Important Didactic Tool For Activating Students’ Learning Process in the Educational Process. Int. J. Progress. Sci. Technol. 2021, 25, 156–159. [Google Scholar] [CrossRef]
  16. Vitalyevna, C.Y. Interactive methods of teaching Russian literature in schools with Uzbek language learning. Orient. Renaiss. Innov. Educ. Nat. Soc. Sci. 2021, 1, 1169–1174. [Google Scholar]
  17. Hasanah, H.; Malik, M.N. Blended learning in improving students’ critical thinking and communication skills at University. Cypriot J. Educ. Sci. 2020, 15, 1295–1306. [Google Scholar] [CrossRef]
  18. Abidin, Z.; Utomo, A.C.; Pratiwi, V.; Farokhah, L.; Jakarta, U.B.; Jakarta, U.M. Project-Based Learning-Literacy n Improving Students’ Mathematical Reasoning Abilities in Elementary Schools. JMIE J. Madrasah Ibtidaiyah Educ. 2020, 4, 39–52. [Google Scholar] [CrossRef]
  19. Asy’ari, M.; Ikhsan, M. The Effectiveness of Inquiry Learning Model in Improving Prospective Teachers’ Metacognition Knowledge and Awareness. Int. J. Instr. 2019, 12, 455–470. [Google Scholar] [CrossRef]
  20. Putri, S.K.; Syahputra, E. Development of Learning Devices Based on Realistic Mathematics Education to Improve Students’ Spatial Ability and Motivation. Int. Electron. J. Math. Educ. 2019, 14, 393–400. [Google Scholar] [CrossRef] [Green Version]
  21. Available online: https://educationaldatamining.org/resources/ (accessed on 19 June 2023).
  22. Available online: https://www.oecd.org/pisa/data/ (accessed on 19 June 2023).
  23. Naicker, N.; Adeliyi, T.; Wing, J. Linear support vector machines for prediction of student performance in school-based education. Math. Probl. Eng. 2020, 2020, 4761468. [Google Scholar] [CrossRef]
  24. Dass, S.; Gary, K.; Cunningham, J. Predicting student dropout in self-paced MOOC course using random forest model. Information 2021, 12, 476. [Google Scholar] [CrossRef]
  25. Charitaki, G.; Soulis, S.G.; Tyropoli, R. Academic self-regulation in autism spectrum disorder: A principal components analysis. Int. J. Disabil. Dev. Educ. 2021, 68, 26–45. [Google Scholar] [CrossRef]
  26. Wickramasinghe, I.; Kalutarage, H. Naive Bayes: Applications, variations and vulnerabilities: A literature review with code snippets for implementation. Soft Comput. 2021, 25, 2277–2293. [Google Scholar] [CrossRef]
  27. Sood, S.; Saini, M. Hybridization of cluster-based LDA and ANN for student performance prediction and comments evaluation. Educ. Inf. Technol. 2021, 26, 2863–2878. [Google Scholar] [CrossRef]
Figure 1. The proposed DNN-LALM design.
Figure 1. The proposed DNN-LALM design.
Systems 11 00319 g001
Figure 2. DNN architecture for the proposed DNN-LALM.
Figure 2. DNN architecture for the proposed DNN-LALM.
Systems 11 00319 g002
Figure 3. Accuracy analysis of the DNN-LALM.
Figure 3. Accuracy analysis of the DNN-LALM.
Systems 11 00319 g003
Figure 4. Precision analysis of the DNN-LALM.
Figure 4. Precision analysis of the DNN-LALM.
Systems 11 00319 g004
Figure 5. Recall analysis of DNN-LALM.
Figure 5. Recall analysis of DNN-LALM.
Systems 11 00319 g005
Figure 6. F score analysis of the DNN-LALM.
Figure 6. F score analysis of the DNN-LALM.
Systems 11 00319 g006
Figure 7. AUC analysis of the DNN-LALM.
Figure 7. AUC analysis of the DNN-LALM.
Systems 11 00319 g007
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, D.; Ortegas, K.D.; White, M. Exploring the Computational Effects of Advanced Deep Neural Networks on Logical and Activity Learning for Enhanced Thinking Skills. Systems 2023, 11, 319. https://doi.org/10.3390/systems11070319

AMA Style

Li D, Ortegas KD, White M. Exploring the Computational Effects of Advanced Deep Neural Networks on Logical and Activity Learning for Enhanced Thinking Skills. Systems. 2023; 11(7):319. https://doi.org/10.3390/systems11070319

Chicago/Turabian Style

Li, Deming, Kellyt D. Ortegas, and Marvin White. 2023. "Exploring the Computational Effects of Advanced Deep Neural Networks on Logical and Activity Learning for Enhanced Thinking Skills" Systems 11, no. 7: 319. https://doi.org/10.3390/systems11070319

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop