Next Article in Journal
Secondary School Apprenticeship Research Experience: Scientific Dispositions and Mentor-Student Interaction
Next Article in Special Issue
Connections between Online International Learning and Inclusion of Intercultural and International Elements in the Curriculum—The Perspective of Slovene Academics
Previous Article in Journal
Using Big Data for Educational Decisions: Lessons from the Literature for Developing Nations
Previous Article in Special Issue
Teacher Professional Development in Higher Education: The Impact of Pedagogical Training Perceived by Teachers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Predictive Model That Aligns Admission Offers with Student Enrollment Probability

1
Department of Statistics, Feng-Chia University, Taichung 407102, Taiwan
2
Department of Risk Management and Insurance, Feng-Chia University, Taichung 407102, Taiwan
*
Author to whom correspondence should be addressed.
Educ. Sci. 2023, 13(5), 440; https://doi.org/10.3390/educsci13050440
Submission received: 1 March 2023 / Revised: 21 April 2023 / Accepted: 21 April 2023 / Published: 25 April 2023
(This article belongs to the Special Issue Challenges and Trends for Modern Higher Education)

Abstract

:
This study develops a process that helps admission committees of higher education institutions select interested and qualified students. This enables institutions to maintain their financial viability by reaching the quota given by the Education Administration of Taiwan. We aimed to predict the decision-making behavior of students in terms of enrollment. A logistic regression analysis was conducted on publicly and inexpensively accessible data; the selection criteria of the model are based on metrics from a confusion matrix comprising predicted and observed data. The results indicate a matching rate of close to 80% between the training data of a target university from 2018 to 2020 and the testing data from 2021. This system outputs a probability that the student will enroll and thus helps admission committees more effectively select students.

1. Introduction

From 1994 to 2005, the number of universities and colleges in Taiwan approximately doubled; in addition, the total fertility rate in Taiwan declined from approximately 1.8 in 1995 to 1.2 in 2015, one of the fastest falling rates in the world [1]. These two factors have caused serious societal problems, one of which is a decreasing rate of enrollment in universities. As enrollments have decreased, competition between universities to recruit students has increased; universities have become commercialized to attract students, which leads to less stringent admission standards. Thus, in this study, we designed a process that helps admissions committees in higher education.
The student selection process is a critical success factor for higher education institutions because every university wants to offer admission to qualified students who are willing to enroll in the university they receive an offer from. On the one hand, admitting unqualified or uninterested students is likely to cause problems in the future. Universities are often forced to do this because failing to achieve enrollment quotas threatens their financial viability; this is an especially serious problem for private institutions. On the other hand, when students are accepted to or are on the waiting list for more than one university, aspects such as the reputation of the institution, department specialties, and faculty strength influence their final decision on which university to attend. Educational institutions can thus benefit from a better understanding of the decision-making behavior of students; we attempt to explain this behavior in this paper.
The reputation or ranking of a department or university is often conflated with the academic achievement of its students. In general, such evaluations are necessarily based on some element of subjectivity. For example, the Department of Accounting cannot be objectively compared with the Department of Finance. Moreover, comparing the achievements of the top two students from two different high schools without using private data, such as their transcripts, is challenging. The goal of our research is to address these problems using objective and publicly available data.
One of the main reasons for public data is the budget issue. In the realm of higher education, the recruitment process, encompassing a range of activities such as campaigning, requires the prudent allocation of limited financial resources. In light of these constraints, utilizing publicly available data emerges as a compelling requirement if they are to use any modeling for assisting their recruitment. The feasibility of obtaining source data from public repositories obviates the need for significant financial investments, thereby enabling academic institutions to channel their resources toward other crucial domains. By adopting our suggested process, they can capitalize on the wealth of information accessible through public data sources and leverage it to inform the admission committee’s decision-making processes. The process we developed aims to predict the likelihood that a prospective student will accept an admission offer from a given university or that a student on a waitlist will wait for an offer; we applied our process to publicly accessible data. We constructed metrics using logistic regression to predict the probability that a student will enroll. This process can also evaluate the decisions of admission committees in terms of the distribution of willingness to enroll among the students who receive an admission offer; developing such an evaluation is the primary goal of our research.
Each individual can apply to only a limited number of departments, and the students can only choose to enroll in one department after receiving admission offers. Several studies on this topic have been conducted. One study investigated the predictors of college enrollment across different age groups and genders [2]. Another article examined the influence of the Ronald E. McNair Post-Baccalaureate Achievement Program on graduate school enrollment for students from disadvantaged backgrounds [3]. One study analyzed the datasets of different universities with machine learning to predict the chance of a student being admitted to a specific university [4]. Another study reviewed and compared various machine learning techniques used for university admission predictions [5]. A fuzzy logic for an intelligent and automated decision support system to assess a student’s eligibility for admission to a specific university was proposed in another paper [6], which compared the testing results and received a 96% approval rate.
Studies on admission problems have primarily taken the perspectives of students to predict how competitive a student is (i.e., how likely they are to be accepted) based on various factors; studies have often used a student’s academic profile to predict the probability of admission. However, we seek to predict the probability of enrollment among admitted students. Other studies have focused on students who have applied to certain departments or universities and used the academic profiles of students in their analysis; by contrast, our research focuses on students who enrolled and do not use academic profiles.
Few studies have taken the perspective of admission committees possibly because the supply for higher education has only recently outstripped demand in Taiwan. Our work addresses the current situation in higher education in Taiwan, and our approach is primarily based on empirical experience from one university rather than any given theory. The findings of this study may thus be inapplicable to other universities. Unfortunately, due to the competition between universities, other universities will probably not release their data, and generalizing our findings to other universities is difficult. Moreover, the evolution of the education system may make our process obsolete in the future. Despite this, we hope our pioneering work will be applied to admission processes to help universities balance candidate quality and financial viability.

2. Materials and Methods

Students can apply to universities in Taiwan through three channels. One of the channels is personal applications, which accounted for approximately 60% of applications in 2021; this percentage that has remained steady for years. Personal application requires students to take the General Scholastic Ability Test (GSAT), and students can submit applications to several universities for various departments based on their scores; in this paper, a student’s portfolio was defined as their GSAT scores. Each student can apply to at most 11 departments, referred to as the choice list (CL). Students do not rank their CL; however, the admission committee has access to the CL of those students who have applied to its department.
After reviewing the portfolio in the first phase and going through the interview process in the second, the admission committee decides whether each student is accepted (A), wait-listed (W), or rejected (R). The students are then informed of the committee’s decision and order their preference for the departments they are qualified for and would like to enroll in. If the status of a student’s first choice is A, the student can enroll at that institution. If the status is W, the student will be placed on a waiting list; only when enough other students reject the admission offer and a space becomes available would the student be moved up to the acceptance list.
The textual information of the CL was transformed into 38 feature values as the major inputs of logistic regression. Unlike other studies, which have included student scores for specific exams, for example, Graduate Record Examination (GRE) score or grade point average (GPA) [7], this study did not use nor had access to the score of GSAT, which made the cost of research relatively low.
We separated the 38 feature values into three categories; they are listed in Table 1.
Differences in explanatory power between feature values were not evident. However, department-specific values were likely to be more heavily weighted because departments differed in their admission criteria and preferences. Moreover, the weight of specific values may change every year. Thus, our goal was not to identify the feature values with the highest weight; instead, we aimed to design a process that worked for all departments. We achieved this goal by combining the basic feature values along with either interuniversity or intrauniversity values, or both, to determine the most optimal one for every department, and then select the one with the best performance; the definition of best performance is discussed in Section 3. We planned to explore disparate combinations of feature values in future studies. However, this study used the following three combinations of feature values in our logistic regression model: (A) basic and interuniversity; (B) basic and intrauniversity; and (C) basic, interuniversity, and intrauniversity.
The geographic location where a student takes the exam GSAT was used as the proxy for the student’s hometown; this was a variable that may be included in the logistic regression.
The modeling data were from the academic year lasting from fall 2018 to spring 2019 of the target university, and we used 2020 data as the criteria data to select from a pool of systems by imposing specific metrics. We then applied our model to test our system by predicting the enrollment outcome of 2021, where the students were enrolled in the university after the summer of 2021. Different from the traditional method of using separate training and testing data from the same pool of data [9], our method avoids the problem of the influence of the separation method on the model [10].
The modeling data were categorized into the following three groups: (1) 2018, (2) 2019, and (3) 2018 and 2019. Table 2 explains how the data were applied.
Table 3 displays the total number of applications from the students who chose at least one department of the target university. For example, if a student applied to two departments in the academic year 2018, their data was counted twice in the total number of applications.
We processed the inputs using the following two approaches: (a) scaling, which involves normalizing the inputs to range from 0 to 1, and (b) principal component analysis (PCA), which is applied to transform the feature values to a new set of variables.
Table 4 includes a summary of the possible designs of the system. Combining the three possible inputs of modeling data as in Table 2 results in 3 types of modeling data × 3 combinations of feature values × 2 types of location inputs (to include or not to include) × 3 processing methods × 2 predicted targets (acceptance or enrollment) = 108 candidate systems (Table 4).
In addition to using the feature values, we used 36 for each combination of feature values, totaling 108 systems. We applied these candidate systems to each of the three combinations of feature values and select the system with the highest metric for each department.
For simplicity, we used A, B, and C, as specified in the second column of Table 4, to represent the aforementioned combinations of feature values. We used logistic regression to estimate the acceptance and enrollment probability [11] in the following equation in which  x i  is the feature value:
p x = 1 1 + e β 0 + i = 1 k β i x i
The traditional outcome of logistic regression [12] is a probability value, with 0.5 as the cutoff point that determines which category an observation belongs to. We ranked each student by the probability of acceptance and enrollment predicted by the model. Because we knew the quota of each department, the model output a positive prediction if and only if an individual’s ranking was high enough to be included in the quota.
By applying this predictive process to the admission and enrollment behavior of students and comparing its predictions with the actual data, we could compose four confusion matrices. The confusion matrix has four possible outcomes: true positive (TP), false positive (FP), false negative (FN), and true negative (TN).
The observed number of enrolled students is displayed in the first column of Table 5 (ETP + EFP), the enrollment matrix, and in the first column of Table 6 (AETP + AEFP), the admission-enrollment matrix. This sum should be less than or equal to the admission quota of the respective department. The sum of the first column of Table 7 and Table 8 should equal this quota. For example, in Table 7, the sum of the two numbers (ATP + AFP) in the first column should equal the department quota because the admission committee offers admission to enough students to satisfy the quota. The sum of the first row of these four tables should equal to this quota too since our prediction should be congruent with this fact.
For each matrix, we can compose the following two metrics: (A) accuracy, which equals (TP + TN)/(TP + FN + FP + FN) and measures the performance of the model on how accurately it correctly predicts the outcome, and (B) sensitivity, which equals TP/(TP + FN) and measures the ability of the model to predict the proportion of positive results correctly. With four confusion matrices and two metrics, we have eight metrics to test our system performance. We denoted them using a matrix–metric format; for example, AE–accuracy is the accuracy in the Admission–Enrollment Matrix.
To ensure the quality of our system, we used the receiver operating characteristic curve (ROC) and only included the system in our analysis when the area under the ROC (AUC) was greater than 0.65; the systems with AUC 0.65 or less were deleted. For the goodness of fit of the logistic regression, we used the Hosmer–Lemeshow test and included the results with a p value greater than 0.05, which indicated statistical significance [13].

3. Results

3.1. Optimal System for Each Metric of the Three Combinations

For each department, the system with the highest value of the respective metric was selected as the optimal system. With combination A and metric A-accuracy taken as an example, the system with the highest value of this metric is illustrated in Figure 1.
With the same department used to demonstrate the result of the three combinations in Table 9, Table 10 and Table 11, the first column is the target metric, and the second column is the system with the highest value among the 36 different systems. Starting from the third column is the description of the system with this highest value; AC stands for acceptance and EN stands for enrollment.
These results matched our expectation that the A matrix and AE matrix best-predicted admission offers whereas the other two matrices best-predicted enrollment. For each combination, eight systems had the highest value of the respective metric. We then used principal components analysis to integrate these eight systems and took only the first principal component, PC1 [14], in each combination; the PCA has different weightings for different combinations.

3.2. Matching Level

The sequential step is better explained by using one specific department as an example. This department had a quota of 82 students, and the quota for its waiting list was 164. Therefore, 246 students were selected by the committee. These students were then ranked from highest to lowest by their PC1 value and divided into five buckets with the set cutoff points of (49.2, 98.4, 147.6, and 196.8); the bucket size was chosen by dividing the total number of students by the number of groups (246/5). The bucket a student was in was represented by a metric called matching level (ML); bucket 5 contained the highest ranking students (1 to 49), bucket 4 contained 50 to 98 students, and bucket 1 contained 197 to 246 students. Students in the highest ranking bucket (bucket 5) thus all had an ML of 5, and the ones in the lowest-ranking bucket had an ML of 1. A higher ML indicated a higher probability of being enrolled.
We drew a histogram of these five MLs using the data of the 82 observed enrolled students for the academic year 2020 (Figure 2). Each bar in the histogram corresponds to the number of students in each ML bucket but only accounts for students that were enrolled. The left-skewed histogram represents a closer match between the enrolled students and the committee’s choice than histograms with other distributions.

3.3. Matching Performance

To quantify this degree of matching, we defined a new metric called matching performance (MP), which equals the ratio of the number of students in buckets 4 and 5 to the number of students in buckets 1 and 2. For instance, in combination A (Figure 2), the total number of students in buckets 4 and 5 was 54 (23 + 31); we divided this by the total number of students in buckets 1 and 2, which was 15 (3 + 12) resulting in an MP of 3.60 (54/15). The higher the MP of the selected system for a department was, the higher the proportion of students predicted to be enrolled.
The combination with the highest MP was then selected as the best-performing model, which was combination A of this example department. We then analyzed the 2021 enrollment data with combination A to compare the results with those of 2020. Figure 3 illustrates the box plot of the MP values from 34 departments; each box goes from the first quartile to the third quartile; the middle line represents the median. The two upper boxes represent the current model applied to the data from 2020 and 2021, which we call the full model (FM); the middle and the bottom ones represent the following two other indicators used to select feature values, namely the Akaike information criterion (AIC) and Bayesian information criterion (BIC), respectively. The average MP values of the 2020 data were all well above one because we used 2020 data as the criteria data. The data from 2021 had a slighter decline in MP values, but the averages all remained above one. Several outliers were present on the right of the plot, which indicated the ML was relatively high; no outliers were present on the left of the plot.
Our process is summarized in Figure 4.

4. Discussion

To understand how well our system can predict the acceptance and enrollment data for 2021, we arbitrarily chose an MP value of 1 + 0.25 × (standard deviation) in 2021 as a cutoff point. If the MP value was greater than this number, we considered the prediction of the department applying the respective model to the data to be trending up, meaning most of the actually enrolled students were predicted by the model to have a high probability of enrollment for the year 2021. If this metric was below 1 − 0.25 × (standard deviation), then we considered it to be trending down; otherwise, we considered it to be indistinguishable. This information is displayed in Table 12.
We then determined whether the trends for 2020 and 2021 matched. Specifically, we determined whether they both trended upward (UU; left graph in Figure 5), whether they both trended downward (DD), or whether they were both indistinguishable (II). By calculating the number of departments with matching trends and including II but not DD, we obtained a matching trend rate. Because the DD example had the highest MP values among the three combinations, the other two combinations necessarily had a downward trend.
The results of the three models are presented in Table 13.
Future studies could combine more feature values, which we categorized into only three combinations, to discover combinations that are more effective. We used three combinations for practical reasons; for example, the combined effects of that are groups similar in nature to each other tend to be easier to analyze.
Another point that warrants further research is the maximum number of departments each student is allowed to apply to. Our system has no built-in constraints that limit this number; however, we do not know if our system would be as effective and robust if this number increases. If possible, we aim to apply our approach to the application of universities in countries such as the United States, where students can apply to as many universities as they wish.
We also applied the same system to the data from the academic year 2022 using the modeling data from 2019 and 2020 and the criteria data from 2021. We provided the admission committees of each department with the relative probability that each student that applies to the target university will choose to enroll; its use as a tool was strongly recommended by the director of the admission office. Nonetheless, each committee may choose whether to use this data and how they use it. The fact that the enrollment rate of the target university has improved amid falling enrollment rates of most universities in Taiwan provides strong evidence that our system is of practical use.

5. Conclusions

Decades ago, when admission to university was difficult in Taiwan due to the low number of universities, researchers focused mostly on predicting acceptance by the committee based on student academic performance. However, with an aging population, universities must develop strategies to maintain their financial viability; however, this topic has not received much attention from researchers. In this study, we designed a system to help the admission committee of the target university select students for admission; our goal is to offer admission to qualified students who have a higher probability of accepting an admission offer and enrolling.
Our system predicts the enrollment probability of students using only publicly accessible data about the potential majors students choose. This system could be of great value to the admission committee because it can increase the enrollment rates of each department. After transforming the textual information into 38 feature values, we used them as our inputs in logistic regression. With three combinations of data from various years, three combinations of feature values, with or without imposing location as input, three different processing methods, and two predicted targets; summing these combinations leads to 108 candidate systems. To choose the most suitable system, four confusion matrices and two metrics for each matrix were created, resulting in eight metrics. The system with the highest metric value was picked. By PCA, the eight most suitable systems from the eight metrics were integrated and the first component, PC1, was selected. A summary metric called MP was designed to determine the best-performing combination of feature values. We then used AIC and BIC to repeat the process. The trending up and matching trend rates were imposed to evaluate the predictive power among the Full, AIC, and BIC models. Ultimately, the FM performed better than the AIC and BIC.

Author Contributions

Conceptualization, J.-P.W. and M.-S.L.; methodology J.-P.W.; programming, J.-P.W. and C.-L.T.; validation, M.-S.L.; formal analysis, J.-P.W.; investigation, J.-P.W., M.-S.L. and C.-L.T.; resources, J.-P.W.; data curation, C.-L.T.; writing—original draft preparation, M.-S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data is extracted from the following public website, http://www.com.tw. Given that it is an open website with huge amount of information, labor work is required for data curation.

Acknowledgments

The authors would like to acknowledge our dear friend, Bao-Ling Lee, for her support and trust in our project.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jones, G.W. Ultra-low fertility in East Asia: Policy responses and challenges. Asian Popul. Stud. 2019, 15, 131–149. [Google Scholar] [CrossRef]
  2. Monaghan, D.B. Predictors of College Enrollment across the Life Course: Heterogeneity by Age and Gender. Educ. Sci. 2021, 11, 344. [Google Scholar] [CrossRef]
  3. Renbarger, R.; Beaujean, A. A meta-analysis of graduate school enrollment from students in the Ronald E. McNair post-baccalaureate program. Educ. Sci. 2020, 10, 16. [Google Scholar] [CrossRef]
  4. Raghavendran, C.V.; Pavan Venkata Vamsi, C.; Veerraju, T.; Veluri, R.K. Predicting student admissions rate into university using machine learning models. In Machine Intelligence and Soft Computing: Proceedings of ICMISC 2020; Springer: Singapore, 2021; pp. 151–162. [Google Scholar]
  5. Golden, P.; Mojesh, K.; Devarapalli, L.M.; Reddy PN, S.; Rajesh, S.; Chawla, A.A. Comparative Study on University Admission Predictions Using Machine Learning Techniques. Int. J. Sci. Res. Comput. Sci. Eng. Inf. Technol. 2021, 7, 537–548. [Google Scholar] [CrossRef]
  6. Yudono MA, S.; Faris, R.M.; De Wibowo, A.; Sidik, M.; Sembiring, F.; Aji, S.F. Fuzzy Decision Support System for ABC University Student Admission Selection. In International Conference on Economics, Management and Accounting (ICEMAC 2021); Atlantis Press: Amsterdam, The Netherlands, 2022; pp. 230–237. [Google Scholar]
  7. Fathiya, H.; Sadath, L. University Admissions Predictor Using Logistic Regression. In Proceedings of the 2021 International Conference on Computational Intelligence and Knowledge Economy (ICCIKE), Dubai, United Arab Emirates, 17–18 March 2021; pp. 46–51. [Google Scholar]
  8. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  9. Hyndman, R.J. Athanasopoulos G. In Forecasting: Principles and Practice; Otexts: Columbia, MD, USA, 2018. [Google Scholar]
  10. Fenlon, C.; O’Grady, L.; Doherty, M.L.; Dunnion, J. A discussion of calibration techniques for evaluating binary and categorical predictive models. Prev. Vet. Med. 2018, 149, 107–114. [Google Scholar] [CrossRef] [PubMed]
  11. Basu, K.; Basu, T.; Buckmire, R.; Lal, N. Predictive models of student college commitment decisions using machine learning. Data 2019, 4, 65. [Google Scholar] [CrossRef]
  12. Wright, R.E. Logistic regression. In Reading and Understanding Multivariate Statistics; Grimm, L.G., Yarnold, P.R., Eds.; American Psychological Association: Washington, DC, USA, 1995; p. 3. [Google Scholar]
  13. Hosmer, D.W., Jr.; Lemeshow, S.; Sturdivant, R.X. Applied Logistic Regression; John Wiley & Sons: Hoboken, NJ, USA, 2013. [Google Scholar]
  14. Maciejowska, K.; Uniejewski, B.; Serafin, T. PCA forecast averaging—Predicting day-ahead and intraday electricity prices. Energies 2020, 13, 3530. [Google Scholar] [CrossRef]
Figure 1. System with the highest value of A-accuracy of combination A.
Figure 1. System with the highest value of A-accuracy of combination A.
Education 13 00440 g001
Figure 2. Number of students in each bucket.
Figure 2. Number of students in each bucket.
Education 13 00440 g002
Figure 3. MP values of 2020 and 2021 using three selection strategies.
Figure 3. MP values of 2020 and 2021 using three selection strategies.
Education 13 00440 g003
Figure 4. Summary of the process of the predictive system.
Figure 4. Summary of the process of the predictive system.
Education 13 00440 g004
Figure 5. Matching trend example of the three combinations.
Figure 5. Matching trend example of the three combinations.
Education 13 00440 g005
Table 1. The 38 feature values.
Table 1. The 38 feature values.
Basic
1The number of departments chosen in the CL. This value ranged from 1, 2, …, 11 (see Point 4 for explanation of why 11 departments are chosen)
Interuniversity
2The number of different universities.
3The number of private universities.
4The type of university: general or vocational. Among the 11 maximum number of departments, 6 were from general universities, and five were from vocational universities, per the regulations of the education authorities; this value was either 1 or 2.
5The number of different general universities.
6The number of different colleges chosen, which gives a general idea of the type of major the student might be interested.
7The number of different domains chosen. Eleven domains are present under the category of educational administration. Thus, this value ranged 1 to 11.
8The number of different subdomains chosen. The category of educational administration had 27 subdomains. This ranged from 1 to 11 because the maximum number of departments chosen was 11.
9The ratio of private university departments over the total number of departments in the CL.
10The diversity between universities, which was calculated in terms of entropy [8] using the following equation, where  p i  is the ratio of the ith university out of the total number of universities.
Hu = −∑  p i  Log  p i  (1)
11The diversity between colleges. This is calculated similar to Equation (1) where  p i  is the ratio of the  i t h  college.
Hc = −∑  p i  Log  p i  (2)
12The diversity between domains, calculated similar to Equation (1) and  p i  is the ratio of the  i t h  domain.
13The diversity between subdomains, calculated similar to Equation (1) and  p i  is the ratio of the  i t h  subdomain.
14The diversity between geographic locations of the university, calculated similar to Equation (1) and  p i  is the ratio of the  i t h  county or city.
Intrauniversity
15The number of departments in the target university. Because we acquired the CL of those students who applied to at least one department in the target university that was in the “general” category, this feature value ranges from 1 to 6.
16The ratio of the departments in the target university over the total number of departments in the CL.
17The number of colleges in the target university.
18The number of domains in the target university.
19The number of subdomains in the target university.
20The diversity between the colleges of the target university, calculated in a similar manner as for diversity between universities.
21The diversity between the domains of the target university, calculated in a similar manner as for diversity between universities.
22The diversity between the subdomains of the target university, calculated in a similar manner as for diversity between universities.
23~30The number of departments in a specific college of the target university; for instance, eight colleges corresponded to eight feature values.
31~38The ratio of departments in a specific college of the target university over the total number of departments of the target university. For example, if the target university had four departments and two of them were in college X, this ratio would be 1:2.
Table 2. Year of modeling and criteria data.
Table 2. Year of modeling and criteria data.
Year of Modeling DataYear of Criteria Data
20182020
20192020
2018 + 20192020
Table 3. Yearly data on the number of applications to the target university.
Table 3. Yearly data on the number of applications to the target university.
YearNumber of Application
20187367
20197979
20208163
20217951
Table 4. Candidate systems: feature values, locations, processing methods, and predicted targets.
Table 4. Candidate systems: feature values, locations, processing methods, and predicted targets.
Year of Modeling DataFeature ValuesLocation InputProcessing MethodPredicted Target
2018A: Basic + Inter-universityImposeNothingAccepted (AC)
2019B: Basic + Intra-universityDo not imposeScalingEnrolled (EN)
2018 + 2019C: Basic + Inter-university + Intra-universityPCA
Table 5. Enrollment matrix.
Table 5. Enrollment matrix.
Outcome of Enrollment Matrix (E)Observed
EnrolledNot Enrolled
PredictedEnrolledETPEFN
Not enrolledEFPETN
Table 6. Admission–enrollment matrix.
Table 6. Admission–enrollment matrix.
Outcome of Admission-Enrollment Matrix (AE)Observed
EnrolledNot Enrolled
Predicted AcceptedAETPAEFN
Not acceptedAEFPAETN
Table 7. Admission matrix.
Table 7. Admission matrix.
Outcome of Admission Matrix (R)Observed
AcceptedNot Accepted
PredictedAcceptedATPAFN
Not acceptedAFPATN
Table 8. Enrollment–admission matrix.
Table 8. Enrollment–admission matrix.
Outcome of Enrollment-Admission Matrix (EA)Observed
AcceptedNot Accepted
Predicted EnrolledEATPEAFN
Not enrolledEAFPEATN
Table 9. A: basic + interuniversity.
Table 9. A: basic + interuniversity.
Metric Metric ValueTraining DataImpose LocationProcessing MethodPredicted Target
A-accuracy0.652018NoNothingAC
A-sensitivity0.452018NoNothingAC
E-accuracy0.732018NoPCAEN
E-sensitivity0.572018NoPCAEN
AE-accuracy0.692018YesNothingAC
AE-sensitivity0.522018NoScaleAC
EA-accuracy0.622019YesPCAEN
EA-sensitivity0.42019YesPCAEN
Table 10. B: basic + intrauniversity.
Table 10. B: basic + intrauniversity.
Metric Metric ValueTraining DataImpose LocationProcessing MethodPredicted Target
A-accuracy0.642018NoNothingAC
A-sensitivity0.442018NoNothingAC
E-accuracy0.682019NoNothingEN
E-sensitivity0.502019NoNothingEN
AE-accuracy0.732018NoNothingAC
AE-sensitivity0.592018NoNothingAC
EA-accuracy0.622019NoNothingEN
EA-sensitivity0.42019NoNothingEN
Table 11. C: basic + interuniversity + intrauniversity.
Table 11. C: basic + interuniversity + intrauniversity.
Metric Metric ValueTraining DataImpose LocationProcessing MethodPredicted Target
A-accuracy0.652018NoScaleAC
A-sensitivity0.462018NoScaleAC
E-accuracy0.692018NoNothingEN
E-sensitivity0.512018NoNothingEN
AE-accuracy0.722018NoNothingAC
AE-sensitivity0.562018NoNothingAC
EA-accuracy0.622018NoScaleEN
EA-sensitivity0.42018NoScaleEN
Table 12. Trending criteria of data for the year 2021 by the range of MP values.
Table 12. Trending criteria of data for the year 2021 by the range of MP values.
Range of MPTrend
MP > 1 + 0.25 × (Standard Deviation)Trending Up (U)
1 − 0.25 × (Standard Deviation) < MP < 1 + 0.25 × (Standard Deviation)Indistinguishable (I)
MP < 1 − 0.25 × (Standard Deviation)Trending Down (D)
Table 13. Trending up rate and matching rate.
Table 13. Trending up rate and matching rate.
ModelTrending up Rate of 2021Matching Trend Rate between 2020 and 2021
Full model73.5%79.4%
AIC67.6%64.7%
BIC52.9%50.0%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, J.-P.; Lin, M.-S.; Tsai, C.-L. A Predictive Model That Aligns Admission Offers with Student Enrollment Probability. Educ. Sci. 2023, 13, 440. https://doi.org/10.3390/educsci13050440

AMA Style

Wu J-P, Lin M-S, Tsai C-L. A Predictive Model That Aligns Admission Offers with Student Enrollment Probability. Education Sciences. 2023; 13(5):440. https://doi.org/10.3390/educsci13050440

Chicago/Turabian Style

Wu, Jung-Pin, Ming-Shr Lin, and Chi-Lun Tsai. 2023. "A Predictive Model That Aligns Admission Offers with Student Enrollment Probability" Education Sciences 13, no. 5: 440. https://doi.org/10.3390/educsci13050440

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop