Next Article in Journal
Design and Implementation of Simulation System for Multi-UAV Mission
Next Article in Special Issue
A Domain-Driven Framework to Analyze Learning Dynamics in MOOCs through Event Abstraction
Previous Article in Journal
A Preliminary Feasibility Study of Electric Taxi Promotion in Hong Kong—Behavior Modelling of Driving Patterns and Preferences
Previous Article in Special Issue
All-Year Dropout Prediction Modeling and Analysis for University Students
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Two-Phase Ensemble-Based Method for Predicting Learners’ Grade in MOOCs

Department of Computer Science, College of Computing, Khon Kaen University, Khon Kaen 40002, Thailand
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(3), 1492; https://doi.org/10.3390/app13031492
Submission received: 22 December 2022 / Revised: 18 January 2023 / Accepted: 19 January 2023 / Published: 23 January 2023
(This article belongs to the Special Issue Artificial Intelligence in Online Higher Educational Data Mining)

Abstract

:
MOOCs are online learning environments which many students use, but the success rate of online learning is low. Machine learning can be used to predict learning success based on how people learn in MOOCs. Predicting the learning performance can promote learning through various methods, such as identifying low-performance students or by grouping students together. Recent machine learning has enabled the development of predictive models, and the ensemble method can assist in reducing the variance and bias errors associated with single-machine learning. This study uses a two-phase classification model with an ensemble technique to predict the learners’ grades. In the first phase, binary classification is used, and the non-majority class is then sent to the second phase, which is multi-class classification. The new features are computed based on the distance from the class’s center. The distance between the data and the center of an overlapping cluster is calculated using silhouette score-based feature selection. Lastly, Bayesian optimization boosts the performance by fine tuning the optimal parameter set. Using data from the HMPC- and the CNPC datasets, the experiment results demonstrate that the proposed design, the two-phase ensemble-based method, outperforms the state-of-the-art machine learning algorithms.

1. Introduction

Massive open online courses (MOOCs) provide an environment in which many learners can enhance their knowledge or skills and reach their learning goals on a global scale [1]. Even though their features are different, these and learning management systems (LMSs) are seen as innovations for learning. For example, LMSs are often used in traditional classes with a set schedule, while MOOCs can be used in many different ways. Similar to LMSs, MOOCs can be set up for learning by a program. They can accommodate enrollment and are easily used as an educational tool for short courses. One of the advantages of MOOCs over traditional classrooms is that the system can collect data on a broader range of learning aspects than that which can be achieved in traditional classrooms [2]. Typically, teachers cannot monitor their students’ behavior outside the classroom. In MOOCs, however, all of the students’ actions are collected. The system can identify courses in which learners engage in numerous actions and with videos that motivate them to watch and pay attention [3,4]. Moreover, due to a large number of learners in MOOCs, these various actions and their effects on the learners’ behaviors have become an attractive topic for researchers in the education field.
MOOCs are full of information about the learners’ behaviors, making them an attractive data source for educators. Learner behavior data are utilized in the learning process or by the system. To analyze the students’ behavior, Hogo [5] used log files from an e-learning system to investigate learners’ behaviors. Learners with similar learning patterns were clustered by applying the kernelized fuzzy C-means (KFCM) technique. The learners in the substandard performance cluster were found, and their teachers were told about it so they could plan the next steps in their education. Not only can data be used to put students into groups, but the students’ actions can also be used to figure out which course each student should take. The system can use this information to suggest courses to students with similar methods of learning [6]. The data from MOOCs are used to train simple deep learning, recommending the best course for each learner.
Some learning systems require the learners to pass assessments or submit assignments to receive grades for evaluating their performance. Nevertheless, how a learner learns will affect how well they perform at every stage of the learning process. Consequently, various studies concentrate on predicting the performance of learners. By predicting their performance, the learners can figure out their learning state and improve or change how they learn [7]. In addition, instructors or course owners might apply the prediction results to enhance their teaching or learning materials. In addition, performance information can be combined with other learning behavior characteristics in a learning analysis.
Even though many students try to gain enhanced knowledge and skills through MOOCs, the success rate is low [8]. Thus, when MOOCs discuss statistics regarding the success of their students, imbalanced data are usually one of the most prominent features. Clustering Synthetic Minority Oversampling Technique (SMOTE) oversampling is broadly applied to determine the imbalance problem for learner grade prediction [9]. However, it is important to note that resampling data will cause some biases, while undersampling will lead to information loss. In contrast, oversampling increases the risk of overfitting [10], and it cannot overcome the data distributing problem [11].
This study shows how to deal with unbalanced data using a hybrid strategy that combines a two-phase ensemble-based architecture with a centroid distance feature. Following these contributions, the proposed architecture is used to estimate the students’ grades based on two imbalanced MOOC datasets. The experiment in this research will collect data without oversampling or undersampling, which could lead to an unbalanced set of data. The new distance features will be chosen based on their silhouette score before they are added to the model. To discover the optimal set of classification parameters, Bayesian optimization is applied. These are the contributions of this research:
  • The two-phase architecture of the grade prediction model is constructed using the ensemble approach. AdaBoost is utilized in the first phase as a binary classifier for categorizing class ‘c0’ and non-class ‘c0’. The remainder of the data will then be categorized using multi-class classification. In this phase, One-versus-One will collaborate with XGBoost to predict all of the grades. Due to the imbalanced dataset, this experiment’s data will not be over- or undersampled.
  • This research presents new features that compute the distance between the data and the centroid of each grade class to determine how far the data points are from the center of each grade class.
  • Adding many training features to a prediction model may diminish its performance. In this research, a silhouette coefficient-based feature selection is utilized for selecting only the data associated with the overlap of the grade clusters.
  • The proposed architecture employs the Bayesian-based optimization algorithm to tune the ensemble methods’ hyperparameters.
The remainder of this paper’s content is organized as follows. Section 2 outlines the related works. Section 3 presents the research approach, which includes the data preparation procedure, the engineering of the features, and the processing steps for the two-phase architecture. Next, Section 4 describes the experimental setting, results, and discussion. Lastly, Section 5 is a conclusion.

2. Related Works

There are many proposed methods that are used for predicting grades. Mueller and Weber proposed using the RF Regressor to predict students’ grades [12]. The model was used to determine the predicted scores of the learners. From this study, one remarkable finding is that learners’ backgrounds are ineffective in forecasting their grades. In addition, Bidirectional Long Short-Term Memory (Bi-LSTM) with knowledge distillation was proposed by Kang et al. [8] as the predictor of the learner’s success rate. The model was taught to replicate the behavior of a larger model. The results were then compared to those of traditional machine learning models, including Random Forest (RF), Logical Regression (LR), and Support Vector Machine (SVM) ones, in terms of precision and F1 scores. The learners’ grade prediction model in [13] employed a de-noising auto-encoder that had been previously trained. It was utilized as the initial value for the neural network model, after which this base predictor was incorporated into an ensemble predictor whose weight was determined by the weight of each student’s year. Xing and Du [14] used deep learning to estimate the student dropout rates from the MOOC data, then they used the prediction to personalize and prioritize interventions for at-risk students. However, to apply deep learning to a prediction model, the amount of data is one of the factors that must be considered.
To handle the classification in an imbalanced dataset, many techniques for sampling training data have been proposed and applied to the predictive model, which is solving at the data level, whereas some researchers have tried to improve the classification techniques for solving the problem at an algorithm level. Using the methods at both the data and algorithm levels for classifying imbalanced data is a hybrid-level technique. However, a large portion of the studies about grade prediction using imbalanced data propose using a solving method at the data level, such as using SMOTE for oversampling the training data. The algorithm and hybrid levels are generally lacking [9]. Ashraf et al. [15] advocated for using a naïve Bayes with boosting technique and SMOTE to forecast the grades of learners. Ayienda et al. [16] attempted to identify influential features from school reports and questionnaire data, and then they used them to predict the learners’ performance with three grades (fair, good, and excellent) using a hybrid algorithm of a weighted voting classifier with logistic regression. Yang et al. presented an improved RF, one of the ensemble techniques for predicting the final grades [17]. Due to the problem of unbalanced data, the SMOTE is used to oversample the data during the training process. After the training dataset is ready, the RF model will be trained using the features that passed a threshold value in a feature selection procedure. The K-Means algorithm weights the important features in the prior task’s feature selection [18]. Deepika et al. [19] also proposed a model which applied SMOTE with RF and used Grey Wolf optimization in the feature selection process.

3. Methodology

This research classifies the learners’ predicted grades based on two well-known available datasets. In this section, we describe the details of each dataset and how they were prepared as features of the prediction model. Next, feature engineering, including how the new features are formed, are described, followed by the architecture of the proposed classification model.

3.1. Dataset

This research focuses on two open datasets that collect MOOC data. Both of the datasets include the same collection of attributes with a similar structure [20]. They have been frequently utilized in a variety of papers relating to learning analytics of MOOC data [6,8,12,17,18]. Detailed descriptions of each dataset are provided below.

3.1.1. HarvardX Person-Course Academic Year 2013 De-Identified Dataset (HMPC)

The HMPC dataset contains de-identified data from MITx and HarvardX courses offered on the edX platform during the 2013 academic year. Each record indicates a single student’s participation in a single edX course. The activity of the learners in 13 courses spanning a variety of subject areas is included in the HMPC datasets. This dataset contains 641,138 records with 16 administrative attributes that are system-provided data, three user-provided attributes, and one attribute that is a mix of system and user-provided data [21]. Figure 1 depicts an illustration of a data record.

3.1.2. Canvas Network Person-Course (1/2014–9/2015) De-Identified Dataset (CNPC)

The CNPC dataset consists of de-identified information from open Canvas Network courses held between January 2014 and September 2015. Similar to the HMPC dataset, the data set shows an individual’s participation in a single course for each record. Even the format of the attributes in the dataset is designed after HMPC’s dataset; however, the data are not associated with those of HMPC. The dataset contains 355,199 records. Seventeen administrative attributes and eight user-provided data points from two hundred and thirty-eight open courses make up CNPC’s attributes [20]. Figure 2 provides an illustration of CNPC dataset records.

3.2. Data Pre-Processing

According to Section 3.1, there are differences between the properties of the HMPC- and CNPC-collected datasets. Following each attribute in Table 1, the steps of preprocessing data are detailed in this section.
From Table 1, the attributes of two datasets are modified and translated into a similar format. In the data pre-processing step, attributes ‘user_id’ and ‘nplay_video’, etc., that only appear in one dataset and have a specific meaning are ignored. The remaining nine attributes will be categorized as eight features and one label attribute. According to Equation (1), the ‘grade’ attribute is classified into the ‘grade_code’ attribute, which includes five classes [17].
g r a d e _ c o d e = C 0 ,     0.0 g r a d e < 0.2 C 1 ,     0.2 g r a d e < 0.4 C 2 ,     0.4 g r a d e < 0.6 C 3 ,     0.6 g r a d e < 0.8 C 4 ,     0.8 g r a d e 1.0
Following Equation (1), C0 through C4 represent five grades of the learners. The records that “fail” the conditional value in Equation (1) are eliminated. However, there are still data components that need to be cleaned. Records with missing values for the user-supplied attributes of ‘age’ and ‘edu’ are discarded. The missing values for the system-provided data for ‘nevents’, ‘ndays act’, ‘nforum post’, and ‘ncharpters’, assuming there is no activity, are replaced with zero [6,22], but the missing values from ‘grade’ are rejected. Unfortunately, out of 325,199 total records in CNPC, 243,197 of them contain null values for the ‘grade’ attribute. This is a significant reason why the number of records in the CNPC dataset reduces substantially, while the missing data in the ‘grade’ attribute of HMPC comprise only 57,406 records, which is a small portion when it is compared with the total number of records. For the purpose of transforming the data into a format suitable for machine learning, all of the categorical features are encoded as numbers, and then the entire dataset’s features are normalized using Min–Max normalization. Finally, the outliers are removed from the dataset. In this step, the interquartile range (IQR) is utilized [23]. Table 2 displays the number of records in each dataset after removing the outliers.

3.3. Centroid Distance Features and Selection Method

Following data preprocessing, eight features are utilized to train the model for the grade prediction. The features consist of two categorical variables, ‘edu’ and ‘age’, and six numerical features that were created by the system. Based on the number of ‘grade’ classes, five extra features are produced during the feature engineering process. The distance features are derived from the distance between each data point and the centroid point of each ‘grade’ class cluster. The distances are computed utilizing the Euclidean distance, with all of the training data being placed in the vector space, and Equation (2) is as follows:
d p , q = i = 1 n q i p i 2
where d p , q is the distance between the data points p and q . n is the number of dimensions of venter space. p i and q i are the value of each datum at the i t h dimension.
Finally, each dataset contains thirteen features from eight selected original dataset and five new distance features. The example of the CNPC dataset is illustrated in Figure 3.
However, the results of the practical experiment indicate that including all of the new features as training features reduces the prediction model’s effectiveness. The method of feature selection is utilized to solve this problem. A silhouette coefficient is used to choose the appropriate features. The silhouette coefficient is derived using the average intra-class distance and the average nearest cluster distance. Typically, it is employed to assess the quality of the clustering [24,25]. The formula provided in Equation (3) is used to compute a dataset’s silhouette score of clusters.
s i l h o u e t t e   c o e f f i c i e n t = b a max a ,   b
where a is the mean cluster centroid distance. b is the average nearest cluster distance for every sample.
For the purpose of understanding the silhouette score, numbers approaching zero imply overlapping clusters, but negative numbers indicate that a different cluster is more similar. The silhouette score will be closer to one the more clearly the clusters are recognized. After calculating the silhouette score based on the data from the HMPC and CNPC datasets, the results are displayed in Table 3.
To utilize the silhouette coefficient in the feature selection procedure, each pair of grade clusters is analyzed. Because a low silhouette score represents overlapping data between the examined clusters, the grade clusters with the lowest scores will be utilized as additional training features for the prediction model. In addition, the ‘cen_c0’ attribute will always be used to support binary classification. Consequently, the HMPC dataset’s new features are ‘cen_c0’, ‘cen_c3’, and ‘cen_c4’, whereas the CNPC dataset’s new features are ‘cen_c0’, ‘cen_c1’, and ‘cen_c3’.

3.4. Machine Learning Architecture

Ensemble learning is a machine learning technique that entails combining the predictions made by training data and multiple models to improve the performance. Ensemble algorithms ensure diversity by resampling the data or modifying the individual learners’ structure. Individual learners must perform distinct tasks to learn, and they are expected to obtain errors that differ from those of the other learners, while basic learners aim for high accuracy [26].
To predict the grades of the students based on the MOOCs data, a two-phase concatenation architecture was established as the foundation for classifying the data of the majority class, followed by the remaining classes. Consequently, the first phase will respond to specific data that should be in class ‘c0’ before pushing the remaining data to the subsequent phase. The second phase refers to the multi-class classification that the model must learn to predict data for five grade levels (from ‘c0’ to ‘c4’). Figure 4 depicts an overview of the proposed architecture: the two-phase ensemble-based grade prediction model.
A.
Dataset
To use the dataset for the building of a prediction model, it will be divided into training- and testing sets with an equal proportion of each grade class. Eighty percent of the data is used to train the model, while the remaining twenty percent is used for testing. In the first phase—binary classification—the training set’s labels will be translated into binary classes. The ‘grade code’ attribute provides five classes for usage in the second phase: multi-class classification.
The ‘grade code’ label will be binned to become a binary one in the beginning stages. The first label is ‘C0’, which corresponds to the dataset’s majority class, while ‘Non-C0’ is allocated to the other classes. The number of binary class labels is listed in Table 4.
B.
Binary Classification
To specify which ensemble technique should be placed in the binary classification, one boosting method called Adaboost is used [27]. In contrast to the RF approach, which is an ensemble bagging approach that trains the model using a random subset of the data, Adaboost trains its individual models by learning from the previous model’s mistakes. Based on the preliminary test, Adaboost performed better than RF did in classifying binary grades in both of the datasets. Its basis learners are decision trees with a single split. After that, the sample weights are modified based on the predictions made by the classifier, and the samples that have been modified are the ones that will be used to train the subsequent classifier. When there is an instance of samples being misclassified, the weights that are allocated to them are increased. In contrast, the correctly identified samples are given a reduced weight. Finally, the strong classifier is constructed by employing Equation (4).
H x = s i g n t = 1 T α t h t x
where t is the number of iterations, h x is the base classifier, and α is the weight of classifier h x . s i g n ( ) is defined by s i g n x = 1 if x 0 and s i g n x = 0 otherwise. Adaboost is widely employed in the binary classification of imbalanced data, such as fraud detection, hence, this boosting method is chosen for binary classification in the starting phase.
C.
Multi-Class Classification
After the model generates the results of the binary class classification, the data labeled as ‘Non-C0’ are forwarded to the multi-class classifier. Based on the preliminary test between the One-versus-Rest (OvR) and One-versus-One (OvO) strategies, the latter one can overcome the results of the OvR. In addition, other research [28,29] shows that the combination of OvO with the other machine is better than OVR is. Hence, the OvO strategy is chosen to cooperate with the XGBoost ensemble method. In this step, the ‘Non-C0’ class will be classified as being from grade C0 to C4 in this phase.
An OvO technique can be applied to handle the imbalanced datasets [30]. It addresses the multi-class classification problem by decomposing the data from the original dataset to classify them using a binary criterion. In the second phase of the proposed architecture, the label 5 grade classes will be converted into the binary classification, as shown in Figure 5. After receiving the multiple binary classification task from OvO, the boosting ensample technique XGBoost is implemented. The XGBoost technique is a decision-tree-based ensemble that uses the gradient boosting framework, requires minimum feature engineering, and is capable of handling large datasets [27]. Each tree in XGBoost attempts to learn from the previous tree’s errors until there are no false residuals remaining. The loss function of XGBoost tries to prevent overfitting by using Equation (5).
L M = i = 1 T i I j g i ω j + 1 2 i I j h i + λ ω j 2 + γ T
where g i and h i represent the first and second derivatives of the loss function, respectively. The samples in leaf node j are denoted by I j . These will be computed for the total number of leaf nodes T , while γ represents the complexity parameter, λ is the penalty parameter, and ω j is the outputs of the j t h leaf nodes. In the preliminary experiment, the performances of XGBoost (boosting) and RF (bagging) were also evaluated to find the optimal strategy for OvO implementation. The results indicated that XGBoost performed better than RF did, which is consistent with the conclusions in [31,32] regarding using XGBoost and RF on imbalanced data. Consequently, the prediction of OvO-XGBoost will be chosen by the majority vote of the XGBoost results.
D.
Bayesian Optimization
Using a collection of appropriate parameters can aid in achieving a better performance. Due to the limitations of the boosting machine technique, they contain a large number of difficult-to-tune parameters. Bayesian optimization is utilized to find the optimal solution. The Bayesian optimization technique discovers how the values of various hyperparameters of machine learning influence the target outcomes by iteratively traversing the results based on different hyperparameter values until they converge on the highest value of the target function [33]. The following sampling point is found by optimizing the acquisition function over a Gaussian process by using Equation (6).
x t = argmax x u ( x | D 1 : t 1 )
D 1 : t 1 = x 1 , y 1 , ,   x t 1 , y t 1
where u is an acquisition function that directs the next sampling point ( x t ) to targets where an improvement over the current best observation is possible. Additionally, D 1 : t 1 represents the t 1 samples drawn from the objective function ( f x t ). After calculating the possible noisy sample from the objective function ( y t = f x t + ε ), where ε is a Gaussian noise. It is added to the previous samples as D 1 : t = D 1 : t 1 , x t , y t and used to update the Gaussian process.
E.
Results Integration
After obtaining the predicted grade of the testing dataset, the records identified as ‘c0’ grade by the binary classification will be combined with the results of the multi-class classification. The metrics described in Section 4 will be used to evaluate these consolidated results.

4. Performance Evaluation

In accordance with the procedure proposed in the previous section, the experiments were configured so that we could use both the HMPC and CNPC datasets. The experiment was mainly implemented using the Python programming language and the Scikit-learn library. To avoid the problems caused by a too-small minority class in the test set, the experiments were processed using five-fold cross-validation, in which 80 percent of the total dataset was designated as the training set and the remaining 20 percent was the testing set. The training and testing process were iterated five times, and the training and testing datasets were different each time. In each iteration of the experiment, the data in each class were divided evenly depending on their respective ratios.
To evaluate the performance of a prediction model with an imbalanced dataset, the weighted-precision, weighted-recall, and weighted-average F1 scores are often considered as the standard evaluation measures [34,35]. Equations (8)–(13) explain how to compute these evaluation scores.
p r e c i s i o n = i = 0 n t p i t p i + f p i
r e c a l l = i = 0 n t p i t p i + f n i
w e i g h t e d   p r e c i s i o n = i = 0 n p r e c i s i o n i × N i i = 0 n N i
w e i g h t e d   r e c a l l = i = 0 n r e c a l l i × N i i = 0 n N i
F 1   s c o r e = 2 × p r e c i s i o n × r e c a l l p r e c i s i o n + r e c a l l
w e i g h t e d   a v e r a g e   F 1   s c o r e = i = 0 n ( F 1 s c o r e i × N i ) i = 0 n N i
where t p i is the number of true positive instances for label i . t n i is the number of true negative instances for label i . f p i is the number of false positive instances for label i . f n i is the number of false negative instances for label i . n is the number of class labels in the dataset. N i is the number of instances or the support value for label i .
The precision value typically represents the proportion of valid positive predictions, whereas the recall value indicates the proportion of positive cases that have been correctly categorized. We also calculated the F1 score, which is a harmonic mean of the precision and recall. To apply these metrics for evaluating the prediction results of an imbalanced dataset, the scores will be divided by the number of true instances of each class. Table 5 represents the results of evaluation metrics that are average values from five-fold cross-validation.
To evaluate the performance of the proposed two-phase ensemble-based method, the state-of-the-art machine learning RF method is compared to the improved RF method [17], which is an enhanced RF that was achieved by utilizing clustering SMOTE and a hybrid indicator with a decision mechanism to select the appropriate features. However, to compare it to the proposed method, the resampling data step is left out. The comparison also includes the RF Regressor [12], which is used to predict the learners’ grades based on their exam results. Following the same conditions as in Equation (1), the outcomes of the prediction as a floating point score are binned to become a ‘grade code’. The deep learning proposed in [13], which implemented a forward-backward algorithm and minimized the cost function by the gradient descent, is also examined.
Table 5 illustrates the average weighted precision, weighted recall, and weighted average F1 scores from the five-fold cross-validation configuration of the experiment. To compare the weighted-precision and weighted-recall means, the weighted-average F1 score (weighted F1 score) is utilized. The optimization of the proposed model with the selected distance feature produced the highest weighted F1 score of the two datasets, whereas deep learning produced the lowest score. In the case of CNPC, the IRF without SMOTE was successful in overcoming the traditional RF and RF regressions, while the efficiency trend from the HMPC dataset is distinct. The outcome of the RF regression for HMPC is worse than the those for RF and IRF. However, even the proposed model is deployed without the selected distance features. For both of the datasets, its performance is better than those of RF, IRF, RF regression, and deep learning. It was compared with the tasks that have been mentioned before because all of them learn and predict the learners’ grade using the multi-class classification, which uses the imbalanced data directly. The results obtained from them cannot reach the same level as the proposed method can for all of the classes, especially for class ‘c0’, which is the majority class.
Focusing on the two-phase ensemble-based method, the proposed model’s base outperforms the previous techniques. To improve it, the centroid distance features are added, and then filtered by the lowest silhouette score between the clusters of grades in the training set. In this step, the CNPC performance improves in all of the measures, except for HMPC. In HMPC, only the weighted F1 score increases, which the confusion matrix will analyze in depth. Figure 6 shows the confusion matrix of five-fold cross-validation from the CNPC dataset experiments to support a detailed discussion. Figure 7 displays the enhancement achieved by deploying the optimization and feature selection to the proposed model’s core.
Only ‘cen_c0’, ‘cen_c1’, and ‘cen_c3’ are integrated with CNPC dataset features when feature selection is employed. The silhouette scores of all of the clusters within the dataset denote these important features. The model is then enhanced via Bayesian optimization. Adding the selected distance characteristics improves the model’s ability to predict the samples from classes ‘c1’ through to ‘c3’ compared to that of the originally proposed model. However, the number of accurate predictions from ‘c0’ to ‘c4’, which represent the majority class and the second largest in the dataset, decreases marginally.
The HMPC datasets receive ‘cen_c0’, ‘cen_c3’, and ‘cen_c4’ as additional features, and optimization is performed. However, the enhanced performance of this modified model is not very significant. Figure 8 demonstrates the confusion matrix of five-fold cross-validation for predicting the learners’ grades using the two-phase ensemble-based model that was applied to the HMPC dataset. In addition, Figure 9 displays the outcomes of applying optimization to the model and including the selected centroid distance features in the dataset.
From the confusion matrix, the first phase of the proposed model’s binary classification efficiently identifies ‘C0’ and ‘Non-C0’. This may be because class ‘c0’ of the HMPC has 13 times more data than the other classes do. The prediction of class ‘c0′ influences the quality of the grade prediction in the HMPC dataset. The accuracy values of ‘c2’ and ‘c3’ improve Figure 9’s prediction findings, whereas all of the others except for ‘c0’ decrease. This demonstrates that the specified distance features can help the model to clear the ‘c3’ feature, and they also influence ‘c2’, which has the same poor silhouette score as that of ‘c3’. This adjustment is in response to the decrease in weight precision and weighted recall and the rise in the weighted F1 scores.
To improve the classification performance, Bayesian optimization is enabled for binary and multi-class classification. The optimizer learns by iteratively traversing all of the possible sets of hyperparameters until the highest value of the given evaluation metric is obtained. In parameter spaces, only the parameters ‘n_estimators’ and ‘learning_rate’ are present for the binary classifier’s AdaBoost, while XGBoost’s parameter spaces include seven parameters: ‘n_estimators’, ‘max_depth’, ‘reg_alpha’, ‘reg_lambda’, ‘min_child_weight’, ‘num_ boost_round’, and ‘gamma’. In this step, the optimization that is represented in Table 6 can improve the results from both of the datasets.
Compared to the prior discussion, the CNPC dataset is a better fit for the proposed grade prediction model than the HMPC dataset is. It is possible that the number of samples in each class in the CNPC dataset is comparable to that of the remaining classes, which differs from the HMPC dataset. According to Table 4, the ‘Non-C0’ class of CNPC receives approximately 52.94 percent of all of the records. While the ‘Non-C0’ category of HMPC accounts for only 7.25 percent of all of the records, it does represent a relatively small portion of the dataset. Changing the architecture of the prediction model from a single-phase model into a two-phase model dramatically improves the ‘c0’ class prediction accuracy, which can be further improved through the optimization procedure.
Even if the system can track and collect all of the student behaviors when it is studying or interacting with the system, the focus should be on the characteristics of the data when we are applying MOOC learner behavior data. The data from the HMPC and CNPC datasets demonstrate how imbalanced data can cause classification model difficulties. Class ‘c0’, which constitutes the majority of the dataset, influences the prediction model, whilst the remaining data are insufficient for accurately training the model. In this situation, the two-phase ensemble-based technique can combine binary and multi-class classification. Additionally, it can improve the performance by learning the applicable distance features and optimizing the associated hyperparameters.

5. Conclusions

MOOCs are a great source of data regarding the behaviors of students. Using a prediction of their performance, the learners can evaluate their learning behavior. Instructors or course administrators could utilize the forecast to improve their teaching or learning materials. In addition, the performance data may be combined with other features of learning behavior for learning analyses.
The proposed design, a two-phase ensemble-based architecture with a centroid distance feature, is utilized to estimate students’ grades based on two imbalanced MOOC datasets without using over- and undersampling methods. The centroid distance features will be chosen according to their silhouette score. In the first phase, the ensembled method selected for the binary classification is Adaboost, while the cooperation of One-versus-One and XGBoost is deployed in the second phase as the multi-class classification. Bayesian optimization is used to determine the best classification parameters for performance improvements. The experiment findings reveal that the presented two-phase ensemble-based method design outperforms the state-of-the-art machine learning algorithms in both of the HMPC and CNPC datasets.
However, making an improvement is possible by overlaying the data for each grade level in the training dataset. In the future, a new learning approach or more appropriate training elements will be required to solve the identification samples in the overlapping clusters. From the learning analysis viewpoint, it is equally interesting to analyze the behavior of the learners from each subject’s perspective. In addition, the open datasets in this work are collected at the end of the course. Using partial learning behavior data, such as the data from the sub-sections of a course, is an interesting potential topic. It can improve the prediction performance for detecting learner performance earlier, and teachers can support their students in a timely manner.

Author Contributions

Conceptualization; methodology; software; validation; formal analysis; investigation; resources; data curation; writing—original draft preparation; visualization; project administration; funding acquisition, W.W.; Supervision; writing review and editing; C.S.-I.; Conceptualization; writing review and editing; P.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research project was financial supported by the young researcher development project of Khon Kaen University.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Pursel, B.K.; Zhang, L.; Jablokow, K.W.; Choi, G.W.; Velegol, D. Understanding MOOC students: Motivations and behaviours indicative of MOOC completion. J. Comput. Assist. Learn. 2016, 32, 202–217. [Google Scholar] [CrossRef]
  2. Pailai, J.; Wunnasri, W. Learning Behavior Visualization of an Online Lecture Support. ICIC Express Lett. Part B Appl. 2022, 13, 1155–1164. [Google Scholar]
  3. Abe, K.; Tanaka, T.; Matsumoto, K. Lecture support system using digital textbook for filling in blanks to visualize student learning behavior. Int. J. Educ. Learn. Syst. 2018, 3, 138–144. [Google Scholar]
  4. Kuosa, K.; Distante, D.; Tervakari, A.; Cerulo, L.; Fernandez, A.; Koro, J.; Kailanto, M. Interactive visualization tools to improve learning and teaching in online learning environments. Int. J. Distance Educ. Technol. 2016, 14, 21. [Google Scholar] [CrossRef]
  5. Hogo, M.A. Evaluation of e-learning systems based on fuzzy clustering models and statistical tools. Expert Syst. Appl. 2010, 37, 6891–6903. [Google Scholar] [CrossRef]
  6. Sakboonyarat, S.; Tantatsanawong, P. Massive open online courses (MOOCs) recommendation modeling using deep learning. In Proceedings of the 23rd International Computer Science and Engineering Conference, Phuket, Thailand, 30 October–1 November 2019. [Google Scholar]
  7. Albreiki, B.; Zaki, N.; Alashwal, H. A systematic literature review of student’performance prediction using machine learning techniques. Educ. Sci. 2021, 11, 552. [Google Scholar] [CrossRef]
  8. Kang, T.; Wei, Z.; Huang, J.; Yao, Z. MOOC student success prediction using knowledge distillation. In Proceedings of the Computer Information and Big Data Applications, Guiyang, China, 17–19 April 2020. [Google Scholar]
  9. Bujang, S.D.A.; Selamat, A.; Krejcar, O.; Mohamed, F.; Cheng, L.K.; Chiu, P.C.; Fujita, H. Imbalanced Classification Methods for Student Grade Prediction: A Systematic Literature Review. IEEE Access 2022, 11, 1970–1989. [Google Scholar]
  10. Douzas, G.; Bacao, F.; Fonseca, J.; Khudinyan, M. Imbalanced learning in land cover classification: Improving minority classes’ prediction accuracy using the geometric SMOTE algorithm. Remote Sens. 2019, 11, 3040. [Google Scholar] [CrossRef] [Green Version]
  11. Liang, X.W.; Jiang, A.P.; Li, T.; Xue, Y.Y.; Wang, G.T. LR-SMOTE—An improved unbalanced data set oversampling based on K-means and SVM. Knowl. -Based Syst. 2020, 196, 105845. [Google Scholar] [CrossRef]
  12. Mueller, M.; Weber, G. Machine Learning Regression Analysis of EDX 2012-13 Data for Identify The Auditors Use Case. Int. J. Integr. Technol. Educ. 2017, 6, 14. [Google Scholar] [CrossRef]
  13. Kuo, J.Y.; Chung, H.T.; Wang, P.F.; Baiying, L.E. Building Student Course Performance Prediction Model Based on Deep Learning. J. Inf. Sci. Eng. 2021, 37, 243–257. [Google Scholar]
  14. Xing, W.; Du, D. Dropout prediction in MOOCs: Using deep learning for personalized intervention. J. Educ. Comput. Res. 2019, 57, 547–570. [Google Scholar] [CrossRef]
  15. Ashraf, M.; Zaman, M.; Ahmed, M. An intelligent prediction system for educational data mining based on ensemble and filtering approaches. Procedia Comput. Sci. 2020, 167, 1471–1483. [Google Scholar] [CrossRef]
  16. Ayienda, R.; Rimiru, R.; Cheruiyot, W. Predicting Students Academic Performance using a Hybrid of Machine Learning Algorithms. In Proceedings of the 2021 IEEE AFRICON, Arusha, Tanzania, 13–15 September 2021. [Google Scholar]
  17. Yang, Y.; Fu, P.; Yang, X.; Hong, H.; Zhou, D. MOOC learner’s final grade prediction based on an improved random forests method. Comput. Mater. Contin. 2020, 65, 2413–2423. [Google Scholar] [CrossRef]
  18. Yang, Y.; Zhou, D.; Yang, X. A multi-feature weighting based K-means algorithm for MOOC learner classification. Comput. Mater. Contin. 2019, 59, 625–633. [Google Scholar] [CrossRef]
  19. Deepika, K.; Sathyanarayana, N. Hybrid model for improving student academic performance. Int. J. Adv. Res. Eng. Technol. 2020, 11, 768–779. [Google Scholar]
  20. Canvas Network Person-Course (1/2014–9/2015) De-Identified Open Dataset. Available online: https://doi.org/10.7910/DVN/1XORAL (accessed on 23 December 2022). [CrossRef]
  21. HarvardX Person-Course Academic Year 2013 De-Identified Dataset, Version 3.0. Available online: https://doi.org/10.7910/DVN/26147 (accessed on 23 December 2022). [CrossRef]
  22. Musil, C.M.; Warner, C.B.; Yobas, P.K.; Jones, S.L. A comparison of imputation techniques for handling missing data. West. J. Nurs. Res. 2002, 24, 815–829. [Google Scholar] [CrossRef]
  23. Sainis, N.; Srivastava, D.; Singh, R. Feature classification and outlier detection to increased accuracy in intrusion detection system. Int. J. Appl. Eng. Res. 2018, 13, 7249–7255. [Google Scholar]
  24. Yuan, C.; Yang, H. Research on K-value selection method of K-means clustering algorithm. J 2019, 2, 16. [Google Scholar] [CrossRef] [Green Version]
  25. Han, J.; Pei, J.; Tong, H. Data mining: Concepts and Techniques, 3rd ed.; Morgan Kaufmann: Burlington, MA, USA, 2012; pp. 489–490. [Google Scholar]
  26. Sun, Y.; Li, Z.; Li, X.; Zhang, J. Classifier selection and ensemble model for multi-class imbalance learning in education grants prediction. Appl. Artif. Intell. 2021, 35, 290–303. [Google Scholar] [CrossRef]
  27. Mienye, I.D.; Sun, Y. A Survey of Ensemble Learning: Concepts, Algorithms, Applications, and Prospects. IEEE Access 2022, 10, 99129–99149. [Google Scholar] [CrossRef]
  28. Yan, Z.; Chen, H.; Dong, X.; Zhou, K.; Xu, Z. Research on prediction of multi-class theft crimes by an optimized decomposition and fusion method based on XGBoost. Expert Syst. Appl. 2022, 207, 117943. [Google Scholar] [CrossRef]
  29. Sun, J.; Fujita, H.; Zheng, Y.; Ai, W. Multi-class financial distress prediction based on support vector machines integrated with the decomposition and fusion methods. Inf. Sci. 2021, 559, 153–170. [Google Scholar] [CrossRef]
  30. Song, Y.; Zhang, J.; Yan, H.; Li, Q. Multi-class imbalanced learning with one-versus-one decomposition: An empirical study. In Proceedings of the Cloud Computing and Security, Haikou, China, 8–10 June 2018. [Google Scholar]
  31. Le, T.T.H.; Oktian, Y.E.; Kim, H. XGBoost for imbalanced multiclass classification-based industrial internet of things intrusion detection systems. Sustainability 2022, 14, 8707. [Google Scholar] [CrossRef]
  32. Mardiansyah, H.; Sembiring, R.W.; Efendi, S. Handling problems of credit data for imbalanced classes using SMOTEXGBoost. J. Phys. Conf. Ser. 2021, 1830, 012011. [Google Scholar] [CrossRef]
  33. Snoek, J.; Larochelle, H.; Adams, R.P. Practical bayesian optimization of machine learning algorithms. Adv. Neural Inf. Process. Syst. 2012, 25, 9. [Google Scholar] [CrossRef]
  34. Mandl, T.; Modha, S.; Majumder, P.; Patel, D.; Dave, M.; Mandlia, C.; Patel, A. Overview of the hasoc track at fire 2019: Hate speech and offensive content identification in indo-european languages. In Proceedings of the 11th Forum for Information Retrieval Evaluation, Kolkata, India, 12–15 December 2019. [Google Scholar]
  35. Wawer, A.; Nielek, R.; Wierzbicki, A. Predicting webpage credibility using linguistic features. In Proceedings of the 23rd International Conference on World Wide Web, Seoul, Republic of Korea, 7 April 2014. [Google Scholar]
Figure 1. Examples of data records from the HMPC dataset.
Figure 1. Examples of data records from the HMPC dataset.
Applsci 13 01492 g001
Figure 2. Examples of data records from the CNPC dataset.
Figure 2. Examples of data records from the CNPC dataset.
Applsci 13 01492 g002
Figure 3. Examples of data records after pre-processing.
Figure 3. Examples of data records after pre-processing.
Applsci 13 01492 g003
Figure 4. Overview of two-phase ensemble-based method architecture.
Figure 4. Overview of two-phase ensemble-based method architecture.
Applsci 13 01492 g004
Figure 5. The process of the OvO-XGBoost algorithm.
Figure 5. The process of the OvO-XGBoost algorithm.
Applsci 13 01492 g005
Figure 6. Confusion matrix of the proposed model on the CNPC dataset.
Figure 6. Confusion matrix of the proposed model on the CNPC dataset.
Applsci 13 01492 g006
Figure 7. Confusion matrix of the optimization on the proposed model with selected distance features on the CNPC dataset.
Figure 7. Confusion matrix of the optimization on the proposed model with selected distance features on the CNPC dataset.
Applsci 13 01492 g007
Figure 8. Confusion matrix of the proposed model on the HMPC dataset.
Figure 8. Confusion matrix of the proposed model on the HMPC dataset.
Applsci 13 01492 g008
Figure 9. Confusion matrix of the optimization on the proposed model with selected distance features on the HMPC dataset.
Figure 9. Confusion matrix of the optimization on the proposed model with selected distance features on the HMPC dataset.
Applsci 13 01492 g009
Table 1. A comparison of the attributes of the HMPC and CNPC datasets, and the attributes after pre-processing.
Table 1. A comparison of the attributes of the HMPC and CNPC datasets, and the attributes after pre-processing.
HMPCCNPCAfter Pre-Processing
AttributeAttributeAttributeExample of Value
course_idcourse_id_DI--
userid_DIuserid_DI--
registeredregistered--
viewedviewed--
exploredexploredexplored1
certifiedcompleted_%completed1
-course_reqs--
gradegradegrade0.75
-grade_reqs--
-primary_reason--
final_cc_cname_DIfinal_cc_cname_DI--
-primary_reason--
-learner_type--
-expected_hours_week--
LoELoEedu“Bachelor’s”
YoBage_DIage“{19–34}”
gendergender--
start_time_DIstart_time_DI--
-course_start--
-course_end--
last_event_DIlast_event_DI--
neventsneventsnevents502
ndays_actndays_actndays_act16
nchaptersnchaptersnchapters52
nforum_postsnforum_postsnforum_posts8
nplay_video---
-course_length--
roles---
inconsistent_flag---
Table 2. The number of data records in each dataset.
Table 2. The number of data records in each dataset.
ClassHMPCCNPC
C0201,87412,818
C137143918
C220253701
C335171544
C465265254
Total217,65627,235
Table 3. The silhouette scores between each class in two datasets.
Table 3. The silhouette scores between each class in two datasets.
Silhouette Score between Each Class in the DatasetHMPCCNPC
C0 and C10.32990.1607
C0 and C20.47720.3124
C0 and C30.58840.2302
C0 and C40.59480.3527
C1 and C20.06760.0624
C1 and C30.34820.0261
C1 and C40.42900.1247
C2 and C30.22630.0324
C2 and C40.32810.0703
C3 and C40.03250.1509
Table 4. The number of binary class labels in each dataset.
Table 4. The number of binary class labels in each dataset.
ClassHMPCCNPC
C0201,87412,818
Non-C015,78214,417
Table 5. Results of the experiments on two datasets.
Table 5. Results of the experiments on two datasets.
MethodHMPCCNPC
Weighted
Precision
Weighted
Recall
Weighted Average F1Weighted
Precision
Weighted
Recall
Weighted Average F1
RF0.97150.97340.97200.69750.72740.7067
RF + Selected distance feature0.97150.97340.97210.70400.73280.7127
IRF [17] (without SMOTE)0.97200.97470.97210.69850.72720.7074
RF Regression [12]0.97170.97470.97080.68930.67470.6779
Deep Learning [13]0.97030.97050.96770.66490.70180.6712
Proposed model0.97410.97640.97270.71490.74760.7110
Proposed model + Selected distance features0.97320.97530.97340.71620.74860.7125
Proposed model + Selected distance features + Bayesian Optimization0.97350.97560.97350.72700.75580.7236
Table 6. The optimal values of hyperparameters in each dataset.
Table 6. The optimal values of hyperparameters in each dataset.
Machine LearningHMPCCNPC
AdaBoostn_estimators = 10
learning_rate = 0.1
n_estimators = 500
learning_rate = 1.0
OvO + XGBoostgamma = 10
max_depth = 40
min_child_weight = 1
n_estimators = 10
num_boost_round = 100
reg_alpha = 0.1
reg_lambda = 0.0
gamma = 10
max_depth = 40
min_child_weight = 1
n_estimators = 100
num_boost_round = 1000
reg_alpha = 0.0
reg_lambda = 0.0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wunnasri, W.; Musikawan, P.; So-In, C. A Two-Phase Ensemble-Based Method for Predicting Learners’ Grade in MOOCs. Appl. Sci. 2023, 13, 1492. https://doi.org/10.3390/app13031492

AMA Style

Wunnasri W, Musikawan P, So-In C. A Two-Phase Ensemble-Based Method for Predicting Learners’ Grade in MOOCs. Applied Sciences. 2023; 13(3):1492. https://doi.org/10.3390/app13031492

Chicago/Turabian Style

Wunnasri, Warunya, Pakarat Musikawan, and Chakchai So-In. 2023. "A Two-Phase Ensemble-Based Method for Predicting Learners’ Grade in MOOCs" Applied Sciences 13, no. 3: 1492. https://doi.org/10.3390/app13031492

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop