Next Article in Journal
Characterization of All Graphs with a Failed Skew Zero Forcing Number of 1
Next Article in Special Issue
Common Fixed Point Results in Bicomplex Valued Metric Spaces with Application
Previous Article in Journal
A New Slack Lyapunov Functional for Dynamical System with Time Delay
Previous Article in Special Issue
A Modified Inertial Parallel Viscosity-Type Algorithm for a Finite Family of Nonexpansive Mappings and Its Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Convergent Algorithm for Equilibrium Problem to Predict Prospective Mathematics Teachers’ Technology Integrated Competency

by
Nipa Jun-on
1,
Watcharaporn Cholamjiak
2 and
Raweerote Suparatulatorn
3,*
1
Department of Mathematics, Faculty of Science, Lampang Rajabhat University, Lampang 52100, Thailand
2
School of Science, University of Phayao, Phayao 56000, Thailand
3
Department of Mathematics, Faculty of Science, Chiang Mai University, Chiang Mai 50200, Thailand
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(23), 4464; https://doi.org/10.3390/math10234464
Submission received: 23 October 2022 / Revised: 20 November 2022 / Accepted: 23 November 2022 / Published: 26 November 2022

Abstract

:
Educational data classification has become an effective tool for exploring the hidden pattern or relationship in educational data and predicting students’ performance or teachers’ competency. This study proposes a new method based on machine learning algorithms to predict the technology-integrated competency of pre-service mathematics teachers. In this paper, we modified the inertial subgradient extragradient algorithm for pseudomonotone equilibrium and proved the weak convergence theorem under some suitable conditions in Hilbert spaces. We then applied to solve data classification by extreme learning machine using the dataset comprised of the technology-integrated competency of 954 pre-service mathematics teachers in a university in northern Thailand, longitudinally collected for five years. The flexibility of our algorithm was shown by comparisons of the choice of different parameters. The performance was calculated and compared with the existing algorithms to be implemented for prediction. The results show that the proposed method achieved a classification accuracy of 81.06%. The predictions were implemented using ten attributes, including demographic information, skills, and knowledge relating to technology developed throughout the teacher education program. Such data driven studies are significant for establishing a prospective teacher competency analysis framework in teacher education and contributing to decision-making for policy design.

1. Introduction

The National Council of Teachers of Mathematics (NCTM) [1] reports that teaching with technology to support conceptual development has been a focus of mathematics education for decades. Utilizing multiple technologies to teach mathematics is significantly more difficult than using technology in everyday life. It included teachers’ perspectives, recognizing the significance of technology in teaching mathematics and their confidence in using technology when constructing relevant technology-integrated mathematics classrooms. Therefore, the technology-integrated competency of teachers was defined in this study as the competency to design mathematics lessons that enable students to work on challenging mathematics problems through technology [2,3,4,5].
To unleash the benefits of technology in the mathematics classroom [6,7], teachers require extensive preparation and support. Accordingly, the integration of technology in mathematics education to produce competent prospective mathematics teachers has been incorporated into the most recent mathematics teacher preparation standards [8]. Utilizing validated classification to evaluate prospective mathematics teachers’ technology-integrated competency provides a solid foundation for entry into the profession [9,10].
In recent years, there has been significant interest in the application of data classification approaches to the field of education. Classification is a method for identifying the class of data points provided, referred to as targets/labels or categories. It can be the study of discovering new and potentially helpful information or meaningful outcomes from data. It also seeks to obtain new trends and patterns from datasets by employing various categorization techniques. Particularly, data classification in the education field is now an effective technique for identifying hidden patterns in educational data, predicting students’ academic performance, determining teachers’ competency or enhancing the learning and teaching policy plan. Thus, in this study, we focused on prospective mathematics teachers’ information as our educational data, longitudinally collected over five years, for classification to identify hidden patterns in their technology-integrated competency development.
First of all, we studied the equilibrium problem (EP), initially introduced by Muu and Oettli [11]. The EP is to find an element z * in a nonempty closed convex subset C of a real Hilbert space H such that
f ( z * , y ) 0 , y C ,
where f : H × H R is a bifunction with f ( x , x ) = 0 for all x C , and E P ( C , f ) is denoted for a solution set of the EP (1). The EP (1) generalizes various mathematical problems in optimization analysis such as variational inequalities, minimization problems, linear programming problems, and Nash equilibrium problems, among others, see in [12,13,14,15].
In 2008, Tran et al. [16] presented the two-step extragradient method (TSEM) for solving the EP (1), which is inspired by the concept of solving the variational inequalities of Korpelevich [17]. The iterative scheme is formulated as follows: x 1 C and
y n = arg min y C λ f ( x n , y ) + 1 2 x n y 2 , x n + 1 = arg min y C λ f ( y n , y ) + 1 2 x n y 2 ,
where λ is some constant depending on the interval that makes the bifunction f satisfy the Lipschitz condition. However, as remarked by the authors of [18], two projections on C of the two-step extragradient algorithm, which was introduced by Korpelevich [17], are very difficult to use and can affect the efficiency of the method if C has a complex structure.
In 2019, Rehman et al. [19] modified the subgradient explicit iterative algorithm to solve the problem of pseudomonotone equilibrium problems. The weak convergence of these algorithms is well established under stepsize, which are updated at each iteration without the Lipschitz-type condition. The algorithm is generated by arbitrary elements x 0 H , λ 0 > 0 and μ ( 0 , 1 ) ,
y n = arg min y C λ n f ( x n , y ) + 1 2 x n y 2 , x n + 1 = arg min y H n λ n f ( y n , y ) + 1 2 x n y 2 ,
where w n 2 f ( x n , y n ) satisfy x n λ n w n y n N C ( y n ) with a half-space
H n = { z H : x n λ n w n y n , z y n 0 } ,
and
λ n + 1 = λ n , μ x n y n 2 + x n + 1 y n 2 2 [ f ( x n , x n + 1 ) f ( x n , y n ) f ( y n , x n + 1 ) ] .
Very recently, Rehman et al. [20] focused on improving the stepsize of the subgradient extragradient method to find a solution to the problems of pseudo-monotone equilibrium in a real Hilbert space. The inertial technique term which was first proposed by Polyak [21] was added to speed up the convergence of the algorithm. The weak convergence of the method is well-established based on the standard assumptions on a bifunction. This algorithm is generated by arbitrary elements x 0 , x 1 H . Choose ϱ ( 0 , 1 ) , σ < min 1 3 δ ( 1 δ ) 2 , 1 2 c 1 , 1 2 c 2 , μ ( 0 , σ ) , λ 1 > 0 and non-decreasing sequence 0 δ n δ [ 0 , 1 3 ) ;
ρ n = x n + δ n ( x n x n 1 ) y n = arg min y C λ n f ( ρ n , y ) + 1 2 ρ n y 2 , x n + 1 = arg min y H n λ n f ( y n , y ) + 1 2 ρ n y 2 ,
where w n 2 f ( ρ n , y n ) satisfy ρ n λ n w n y n N C ( y n ) and construct a half-space
H n = { z H : ρ n λ n w n y n , z y n 0 } ,
and
λ n + 1 = σ , μ f ( y n , x n + 1 ) f ( ρ n , x n + 1 ) f ( ρ n , y n ) c 1 ρ n y n 2 c 2 x n + 1 y n 2 + 1 .
Inspired by the above research, we introduce a new modified inertial subgradient extragradient method for obtaining weak convergence to a solution of the set E P ( f , C ) and try to relax the update stepsize { λ n } that can be chosen in many ways. In applications, we apply our algorithm to solve classification problems in machine learning and show the performance of our algorithm by comparing it with existing algorithms to predict prospective mathematics teachers’ technology-integrated competency.

2. Preliminaries

In what follows, recall that H is a real Hilbert space. Let C be a nonempty, closed and convex subset of H . We denote ⇀ and → as weak and strong convergence, respectively. We next collect some necessary definitions and lemmas for proving our main results. Assume u , v H ,
a u + ( 1 a ) v 2 = a u 2 + ( 1 a ) v 2 a ( 1 a ) u v 2
for any a [ 0 , 1 ] .
A normal cone of C at x C is defined by
N C ( x ) = { z H : z , y x 0 , for all y C } .
Let g : C R be a convex function and subdifferential of g at x C defined by
g ( x ) = { z H : g ( y ) g ( x ) z , y x , for all y C } .
A bifunction f : H × H R on C is stated to be
( i ) pseudomonotone if f ( u , v ) 0 f ( v , u ) 0 , for all u , v C ;
( i i ) satisfies the Lipschitz-like criteria for some c 1 , c 2 > 0 , the following inequality is satisfied
f ( u , w ) f ( u , v ) + f ( v , w ) + c 1 u v 2 + c 2 v w 2 , for all u , v , w C .
Lemma 1
([22]). Let g : C R be a subdifferentiable, convex and lower semi-continuous function on C. An element x C is a minimizer of a function g if and only if
0 g ( x ) + N C ( x ) ,
where g ( x ) stands for the subdifferential of g at x C and N C ( x ) the normal cone of C at x.
Lemma 2
([23]). Let { a n } and { b n } be nonnegative sequences of real numbers satisfying n = 1 b n < and a n + 1 a n + b n . Then, { a n } is a convergent sequence.
Lemma 3
([24], Opial). Let Φ be a nonempty set of H and { x n } be a sequence in H . Suppose the following assertions hold.
( i ) For every x Φ , the sequence x n x converges;
( i i ) Every weak sequential cluster point of { x n } belongs to Φ.
Then { x n } converges weakly to a point in Φ.

3. Convergence Theorem

To study the convergence analysis, consider the following conditions.
( C 1 ) The solution set E P ( f , C ) is nonempty and f is pseudomonotone on C;
( C 2 ) f meets the Lipschitz-like condition on H through c 1 > 0 and c 2 > 0 ;
( C 3 ) f ( z , · ) is subdifferentiable and convex on H for each fixed z H ;
( C 4 ) lim sup n f ( z n , y ) f ( z * , y ) for each y C and { z n } C satisfies z n z * .
Lemma 4.
Let ρ n = y n in Algorithm 1, then ρ n E P ( f , C ) .
Algorithm 1 Modified inertial subgradient extragradient Mann Algorithm 1
  • Initialization: Select arbitrary elements x 0 , x 1 H .
  • Iterative Steps: Construct { x n } by using the following steps:
    Step 1. Set ρ n = x n + δ n ( x n x n 1 ) , where { δ n } [ 0 , ) , and compute
    y n = arg min y C λ n f ( ρ n , y ) + 1 2 ρ n y 2 ,
    where 0 < λ n λ < min 1 2 c 1 , 1 2 c 2 . If ρ n = y n , then stop. Otherwise,
  • Step 2. Compute
    u n = arg min y H n λ n f ( y n , y ) + 1 2 ρ n y 2 ,
    where w n 2 f ( ρ n , y n ) satisfying ρ n λ n w n y n N C ( y n ) and construct a half-space
    H n = { z H : ρ n λ n w n y n , z y n 0 } .
    Step 3. Compute
    x n + 1 = θ n ρ n + ( 1 θ n ) u n ,
    where { θ n } ( 0 , 1 ) . Replace n with n + 1 and then repeat Step 1.
Proof. 
By the definition of y n with Lemma 1, we have
0 2 λ n f ( ρ n , · ) + 1 2 ρ n · 2 ( y n ) + N C ( y n ) .
Thus, we can write λ n w ˜ n + y n ρ n + w ¯ n = 0 , where w ˜ n 2 f ( ρ n , y n ) and w ¯ n N C ( y n ) . Due to ρ n = y n implies that λ n w ˜ n + w ¯ n = 0 . Thus, we have
λ n w ˜ n , y y n + w ¯ n , y y n = 0
for all y C . By w ¯ n N C ( y n ) implies w ¯ n , y y n 0 for all y C and through above expression, we obtain
λ n w ˜ n , y y n 0
for all y C . Due to w ˜ n 2 f ( ρ n , y n ) and using the subdifferential definition, we obtain
w ˜ n , y y n f ( ρ n , y ) f ( ρ n , y n )
for all y C . From the inequalities (6) and (7) with 0 < λ n λ implies that f ( ρ n , y ) 0 for all y C , that is, ρ n E P ( f , C ) .    □
Lemma 5.
Suppose that f : H × H R meet the items ( C 1 ) ( C 3 ) , we have
u n ξ ¯ 2 + ( 1 2 c 1 λ n ) ρ n y n 2 + ( 1 2 c 2 λ n ) y n u n 2 ρ n ξ ¯ 2
for all ξ ¯ E P ( f , C ) .
Proof. 
Let ξ ¯ E P ( f , C ) , then by using Lemma 1, we have
0 2 λ n f ( y n , · ) + 1 2 ρ n · 2 ( u n ) + N H n ( u n ) .
Thus, we can write λ n w ˜ n + u n ρ n + w ¯ n = 0 , where w ˜ n 2 f ( y n , u n ) and w ¯ n N H n ( u n ) . This implies that
ρ n u n , y u n = λ n w ˜ n , y u n + w ¯ n , y u n
for all y H n . Given that w ¯ n N H n ( u n ) then w ¯ n , y u n 0 for all y H n . Therefore, we have
ρ n u n , y u n λ n w ˜ n , y u n
for all y H n . Since w ˜ n 2 f ( y n , u n ) , we have
w ˜ n , y u n f ( y n , y ) f ( y n , u n )
for all y H . From (9) and (10), we obtain
ρ n u n , y u n λ n f ( y n , y ) λ n f ( y n , u n )
for all y H n . Substituting y = ξ ¯ in (9), we obtain
ρ n u n , ξ ¯ u n λ n f ( y n , ξ ¯ ) λ n f ( y n , u n ) .
Given ξ ¯ E P ( f , C ) imply that f ( ξ ¯ , y n ) 0 and owing to the item ( C 1 ) gives that f ( y n , ξ ¯ ) 0 . Thus, we obtain
ρ n u n , u n ξ ¯ λ n f ( y n , u n ) .
Following the condition ( C 2 ) , we have
f ( y n , u n ) f ( ρ n , u n ) f ( ρ n , y n ) c 1 ρ n y n 2 c 2 y n u n 2 .
Combining (13) and (14), we obtain
ρ n u n , u n ξ ¯ λ n f ( ρ n , u n ) λ n f ( ρ n , y n ) c 1 λ n ρ n y n 2 c 2 λ n y n u n 2 .
By using the half-space definition, we have ρ n λ n w n y n , u n y n 0 , which implies that
ρ n y n , u n y n λ n w n , u n y n .
Since w n 2 f ( ρ n , y n ) , we obtain
w n , y y n f ( ρ n , y ) f ( ρ n , y n )
for all y H . By replacing y = u n , we obtain
w n , u n y n f ( ρ n , u n ) f ( ρ n , y n ) .
It follows from inequalities (16) and (17) that
ρ n y n , u n y n λ n f ( ρ n , u n ) λ n f ( ρ n , y n ) .
From (15) and (18), we have
ρ n u n , u n ξ ¯ ρ n y n , u n y n c 1 λ n ρ n y n 2 c 2 λ n y n u n 2 .
Now, we obtain the following equalities:
ρ n ξ ¯ 2 u n ρ n 2 u n ξ ¯ 2 = 2 ρ n u n , u n ξ ¯
and
ρ n y n 2 + u n y n 2 ρ n u n 2 = 2 ρ n y n , u n y n .
Combining the above equalities with expression (19) finalizes the proof.    □
Lemma 6.
Assume that the items ( C 1 ) ( C 4 ) hold. If there is a subsequence { ρ n k } of { ρ n } such that ρ n k x * H and
lim k ρ n k y n k = lim k ρ n k u n k = lim k u n k y n k = 0 ,
then x * E P ( f , C ) .
Proof. 
From y n C , ρ n k x * and lim k ρ n k y n k = 0 , we get y n k x * C . This follows from lim k u n k y n k = 0 that the subsequence { u n k } is bounded. For any y H n , using (11), (14) and (18), we have
λ n k f ( y n k , y ) λ n k f ( y n k , u n k ) + ρ n k u n k , y u n k λ n k f ( ρ n k , u n k ) λ n k f ( ρ n k , y n k ) c 1 λ n k ρ n k y n k 2 c 2 λ n k y n k u n k 2         + ρ n k u n k , y u n k ρ n k y n k , u n k y n k + ρ n k u n k , y u n k c 1 λ n k ρ n k y n k 2 c 2 λ n k y n k u n k 2 .
This implies by (20) and the boundedness of { u n k } that the right hand side tends to zero. Due to 0 < λ n k λ < min 1 2 c 1 , 1 2 c 2 , the condition ( C 4 ) , and y n k x * , we obtain 0     lim sup k f ( y n k , y ) f ( x * , y ) for all y H n . Since C H n , we get f ( x * , y ) 0 for all y C , that is, x * E P ( f , C ) .    □
With the above results, we are now ready for the main convergence theorem.
Theorem 1.
Suppose that n = 1 δ n x n x n 1 < , lim inf n θ n ( 1 θ n ) > 0 and the items ( C 1 ) ( C 4 ) are satisfied. Then, the sequence { x n } generated due to Algorithm 1 converges weakly to a point in E P ( f , C ) .
Proof. 
Let ξ ¯ E P ( f , C ) . Since 0 < λ n λ < min 1 2 c 1 , 1 2 c 2 with expression (8) implies that
u n ξ ¯     ρ n ξ ¯ .
By the definition of ρ n and n = 1 δ n x n x n 1 < , we get
lim n ρ n x n   = lim n δ n x n x n 1 = 0 .
Next, from the definitions of ρ n and x n + 1 , and using (21), the following relation is obtained:
x n + 1 ξ ¯ θ n ρ n ξ ¯   +   ( 1 θ n ) u n ξ ¯ θ n ρ n ξ ¯   +   ( 1 θ n ) ρ n ξ ¯ = ρ n ξ ¯ x n ξ ¯   +   δ n x n x n 1 .
Applying this to n = 1 δ n x n x n 1 < with Lemma 2, we can conclude that the sequence x n ξ ¯ converges. It follows from (22) and (23) that
lim n x n ξ ¯   =   lim n ρ n ξ ¯ .
Next, applying the definition of x n + 1 with (5) and (21), we have
x n + 1 ξ ¯ 2 θ n ρ n ξ ¯ 2 + ( 1 θ n ) u n ξ ¯ 2 θ n ( 1 θ n ) ρ n u n 2 ρ n ξ ¯ 2 θ n ( 1 θ n ) ρ n u n 2 ,
which means that
θ n ( 1 θ n ) ρ n u n 2 ρ n ξ ¯ 2 x n + 1 ξ ¯ 2 .
It is implied by expression (24) and lim inf n θ n ( 1 θ n ) > 0 that
lim n ρ n u n = 0 .
By the inequality (8), we obtain
( 1 2 c 1 λ n ) ρ n y n 2 + ( 1 2 c 2 λ n ) y n u n 2   ρ n ξ ¯ 2     u n ξ ¯ 2 ρ n ξ ¯   +   u n ξ ¯ ρ n ξ ¯     u n ξ ¯ M ¯ ρ n u n
for some M ¯ > 0 . Using (25) with (26) and 0 < λ n λ < min 1 2 c 1 , 1 2 c 2 , we infer that
lim n ρ n y n   = lim n y n u n = 0 .
Finally, let x * H such that x n k x * as k for some subsequence { x n k } of { x n } . By (22), we get ρ n k x * as k . Then, Lemma 6 together with (25) and (27) implies that x * E P ( f , C ) . Using Opial’s lemma (Lemma 3), we can conclude that { x n } converges weakly to a point in E P ( f , C ) .    □
We next show that we can construct stepsizes under the Lipschitz-like condition of the parameter λ n for obtaining new algorithms in many ways. This means that our algorithm is flexible to use. Inspired by the stepsize idea of Rehman et al. [19,20], we can modify stepsizes { λ n } for our Algorithm 1 which satisfy the condition that 0 < λ n λ < min 1 2 c 1 , 1 2 c 2 ; then, we obtain the following algorithms.

4. Application to Data Classification Problem of Educational Dataset

The educational dataset shown in this classification is prospective mathematics teachers’ technology-integrated competency level identified as A, B, C, and D. According to Niess et al. [25], the levels to which prospective mathematics teachers integrate technology into their teaching were classified in this study: Exploring (A), Adapting (B), Accepting (C), and Recognizing (D).
At level D, prospective mathematics teachers recognize technology usage in the classroom as distinct from pedagogical content knowledge. At level C, prospective mathematics teachers desire to integrate technology into their classrooms but may struggle to find ways to connect it to specific topics. At level B, to determine the use of technology in their classrooms, prospective mathematics teachers begin to make noticeably different adjustments in their pedagogy. Lastly, at level A, prospective mathematics teachers begin seeking more ways to integrate technology throughout the curriculum as another learning tool.
This study is a part of a longitudinal study; the research team planned the research design before collecting the data, according to the educational theory of mathematics education. The observation and analysis were validated by three experts in mathematics education and were proved reliable by three researchers to find consensus from all experts and researchers.
In the very first phase of this research, only four attributes were considered as factors. However, other unobserved factors emerged. Those factors were analyzed to determine whether they affected the competency by statistical analysis and affirmed by inspecting pieces of literature. This analysis was provided in another part of the research and was in-proceeding in other contributions. Moreover, all inputs (prospective mathematics teachers) entered the program under the same selective examination and were controlled by the exit examination.
Consequently, in data training, 954 instances were used, containing ten attributes, including major, gender, GPA, IT for learning competency, innovative skill, technology knowledge for a specific subject, number of supplementation, curriculum pattern, selective technology courses and competency level. The statistical overview of the data is illustrated in Table 1.
Denoted CV as coefficient of variation (%) and SD as standard deviation.
Before starting our work, we will provide a brief concept of an extreme learning machine (ELM) [26] for data classification problems. Let U : = { ( x n , b n ) : x n R K , b n R G , n = 1 , 2 , . . . , N } be a training set of N distinct samples where x n is input training data and b n is a target. The output function of ELM for single-hidden layer feed forward neural networks (SLFNs) with H hidden nodes and activation function L is
O n = i = 1 H w i L ( c i x n + e i ) ,
where c i and e i are parameters of weight and finally the bias, respectively. To find the optimal output weight w i at the i-th hidden node, then the hidden layer output matrix L is generated as follows:
L = L ( c 1 x 1 + e 1 ) L ( c H x 1 + e H ) L ( c 1 x N + e 1 ) L ( c H x N + e H ) .
To solve ELM is to find optimal output weight w = [ w 1 T , . . . , w H T ] T such that [ O 1 T , . . . , O N T ] T = L w = B , where B = [ b 1 T , . . . , b N T ] T is the training target data. The least square problem is considered for finding the solution in the cases of the Moore–Penrose generalized inverse of L and may be difficult to find when the matrix L does not exist.
To avoid overfitting in machine learning, we use least square regularization. This problem can be determined as the following convex minimization problem:
min w R K { L w B 2 2 + λ w 1 } ,
where λ is a regularization parameter. This problem is called the least absolute shrinkage and selection operator (LASSO) [27]. For applying our algorithms we set the bifunction f ( x , y ) = L T ( L x B ) , y x for all x , y R H
We use four evaluation metrics: Accuracy, Precision, Recall, and F1-score [28] as explained below for comparing the performance of the classification algorithms.
A c c u r a c y ( % ) = T P + T N T P + F P + T N + F N × 100 % .
P r e c i s i o n ( % ) = T P T P + F P × 100 % .
R e c a l l ( % ) = T P T N + F N × 100 % .
F 1 - s c o r e ( % ) = 2 × ( P r e c i s i o n × R e c a l l ) P r e c i s i o n + R e c a l l ,
where these matrices gave True Negative ( T N ), False Positive ( F P ), False Negative ( F N ), and True Positive ( T P ) results. The multi-class cross entropy loss is used in multi-class classification by the form:
L o s s = i = 1 N y k log y ^ k ,
where y k is 0 or 1, indicating whether class label k is the correct classification and y ^ k is a probability of class y k and N is the number of scalar values in the model output.
To start our computation, we set the activation function as sigmoid, hidden nodes K = 300 , and regularization parameter λ = 10 . Setting δ n = 1 3 n and α n = 5 n 10 n + 1 for Algorithms 1–3, μ = 0.99 for Algorithm 2 and μ = λ 1 2 for Algorithm 3. The stopping criteria is the best accuracy of the training process (81.06%). The comparison of all algorithms with different parameters λ n of Algorithm 1 and λ 1 of Algorithms 2 and 3 is presented in Table 2:
Where K = max ( e i g e n v a l u e ( L T L ) ) , we can see that λ 1 = 0.999 / max ( e i g e n v a l u e ( L T L ) ) of Algorithms 1–3 receives less training time and iteration number. This means that it highly improves the performance of the algorithm. Next, we consider the differences of the parameters δ n and α n of Algorithms 1–3 in Table 3 and Table 4, respectively, when λ 1 = 0.999 / max ( e i g e n v a l u e ( L T L ) ) . Setting α n = 5 n 10 n + 1 , then we obtain the following numerical results of different parameters δ n .
Using the best parameter δ n = 0.99 for Algorithms 1 and 2 and δ n = 1 / 3 n for Algorithm 3 in Table 3 and setting λ 1 = 0.999 / max ( e i g e n v a l u e ( L T L ) ) , μ = λ 1 / 2 , we obtain the following numerical results of different parameters α n .
Algorithm 2 Modified inertial subgradient extragradient Mann Algorithm 2
  • Initialization: Select arbitrary elements x 0 , x 1 H , 0 < λ 1 < min 1 2 c 1 , 1 2 c 2 and μ ( 0 , 1 ) .
  • Iterative Steps: Construct { x n } by using the following steps:
    Step 1. Set ρ n = x n + δ n ( x n x n 1 ) , where { δ n } [ 0 , ) , and compute
    y n = arg min y C λ n f ( ρ n , y ) + 1 2 ρ n y 2 .
    If ρ n = y n , then stop. Otherwise
  • Step 2. Compute
    u n = arg min y H n λ n f ( y n , y ) + 1 2 ρ n y 2 ,
    where w n 2 f ( ρ n , y n ) satisfying ρ n λ n w n y n N C ( y n ) and construct a half-space
    H n = { z H : ρ n λ n w n y n , z y n 0 } .
    Step 3. Compute
    x n + 1 = θ n ρ n + ( 1 θ n ) u n ,
    where { θ n } ( 0 , 1 ) and λ n + 1 = λ n , μ z n y n 2 + x n + 1 z n 2 2 [ f ( y n , x n + 1 ) f ( y n , z n ) f ( z n , x n + 1 ) ] + .
  • Replace n with n + 1 and then repeat Step 1.
Algorithm 3 Modified inertial subgradient extragradient Mann Algorithm 3
  • Initialization: Select arbitrary elements x 0 , x 1 H , 0 < λ 1 < min 1 2 c 1 , 1 2 c 2 and μ ( 0 , 1 ) .
  • Iterative Steps: Construct { x n } by using the following steps:
    Step 1. Set ρ n = x n + δ n ( x n x n 1 ) , where { δ n } [ 0 , ) , and compute
    y n = arg min y C λ n f ( ρ n , y ) + 1 2 ρ n y 2 ,
    If ρ n = y n , then stop. Otherwise,
  • Step 2. Compute
    u n = arg min y H n λ n f ( y n , y ) + 1 2 ρ n y 2 ,
    where w n 2 f ( ρ n , y n ) satisfy ρ n λ n w n y n N C ( y n ) and construct a half-space
    H n = { z H : ρ n λ n w n y n , z y n 0 } .
    Step 3. Compute
    x n + 1 = θ n ρ n + ( 1 θ n ) u n ,
    where { θ n } ( 0 , 1 ) and λ n + 1 = λ n , μ f ( z n , x n + 1 ) f ( y n , x n + 1 ) f ( y n , z n ) c 1 y n z n 2 c 2 x n + 1 z n 2 + 1 .
  • Replace n by n + 1 and then repeat Step 1.
From Table 4, we see that α n = 9 n / ( 10 n + 1 ) highly improves the performance of the Algorithms 1 and 2 and α n = 3 n / ( 10 n + 1 ) highly improves the performance of Algorithm 3. We next show the performance of our Algorithms 1–3 compared with the other existing Algorithms (2)–(4).
Table 5 demonstrates that our algorithm is among those with the highest precision, recall, F1-score, and accuracy efficiency. Additionally, it has the lowest number of iterations. Although it may slightly reduce training time, compared to previous examinations, it has the best probability of correctly categorizing prospective mathematics teachers’ technology-integrated competency level. Moreover, we deliver the training and validation loss with the accuracy of training to show that our algorithm has no overfitting in the training dataset.
From Figure 1 and Figure 2, we see that our model from Algorithm 3 with the suitable parameters in Table 2, Table 3 and Table 4 obtains a good fitting model that is the measure of a machine learning model and generalizes well to similar data. Based on Figure 1 and Figure 2, the over fitting problem can be controlled by finding the best parameters of our algorithms to solve the least square regularization problem (28).
We implemented an inertial subgradient extragradient method for the equilibrium problem on an educational dataset of 954 instances containing ten attributes, including major, gender, GPA, IT for learning competency, innovative skill, technology knowledge for a specific subject, number of supplementation, curriculum pattern, selective technology courses and competency level. The accuracy of classification achieved by the proposed machine learning algorithm was evaluated and 81.06% of the dataset was classified accurately with fewer iterations compared to other methods.

5. Conclusions and Discussion

This study proposes a new method based on machine learning algorithms to predict the technology-integrated competency level of prospective mathematics teachers, taking their data related to different aspects as the source data. Performances of an extragradient method for equilibrium problem were calculated and compared to predict the technology-integrated competency. This study emphasized two focuses. The first one was the prediction of competency based on the skills and knowledge developed throughout teacher education programs. The second focus was the comparison of the performances of machine learning algorithm.
The results show that the proposed method achieved a classification accuracy of 81.06%. Accordingly, it can be said that major, gender, GPA, IT for learning competency, innovative skill, technology knowledge for a specific subject, number of supplementation, curriculum pattern, and selective technology courses are significant predictors to be used for predicting their technology-integrated competency.
Even though this study focused on technology-integrated competency in mathematics classrooms, it was noticed that the major of prospective teachers was one of the predictors. Because there is a large number of prospective teachers in Thailand who may teach out of their field upon entering the profession [29], the major of teacher education programs was also analyzed to determine if their technology-integrated competency differs when they are required to teach mathematics in the future [30].
Comparing the results of this study to other studies on technology integration by mathematics teachers, it was discovered that gender is one of the best predictors of teachers’ intentions to implement technology in their classes [31,32]. In addition, general technological skills and knowledge, such as IT for learning competency and innovative skill, are without a doubt effective predictors of technology-integrated competency [25,33,34]. Additionally, the integration of technology knowledge and content knowledge, also known as Technological Content Knowledge (TCK), was represented by the attribute of technology knowledge for a specific subject, which is a predictor for predicting technology-integrated competency [25,34,35].
The curriculum pattern, according to the educational dataset analyzed in this study, is the new finding that distinguishes this study from others. Since Pattern 1, Pattern 2, and Pattern 3 are included in this study—the attributes of the curriculum pattern—there are three sorts of pattern. The patterns that prospective teachers study the most in courses of pedagogical knowledge, content knowledge, and technological knowledge are Pattern 1, Pattern 2, and Pattern 3, respectively. This finding indicates that when prospective teachers were trained in a variety of knowledge patterns, their technology-integrated competency also performed differently [36,37].
Using this approach, it is possible to anticipate future technology-integrated competency based on these findings. By projecting prospective teachers’ technology-integrated competency in the future, pre-service teachers can examine and improve their working techniques and proficiency. Given that there are about four years between teacher education programs, it is easier to comprehend the significance of the proposed strategy.
The practical achievement of this study is a curriculum revision policy for university-level mathematics education programs. Particularly, the program should offer additional TCK courses, and the redesigned curriculum should place a greater emphasis on technology knowledge for mathematics. In addition, the result specifies the concept of the required curriculum pattern in terms of weighing pedagogical knowledge, content knowledge, and technology knowledge courses.
In the comparison of performances of a machine learning algorithm with other methods, it was found that our Algorithm 3 uses fewer iterations than the existing algorithms with the same highest precision, recall, F1-score, and accuracy efficiency and has the same number of iterations compared with Algorithm 3, although it takes slightly less time to train the data. This means that we can choose to use both algorithms to work with these data.
The results demonstrate that machine learning techniques can be used to predict the technology-integrated competency of prospective mathematics teachers. The results of this study can assist educators in identifying pre-service teachers with below or above average technology integration. In addition, such data driven studies are very significant for establishing a prospective teacher competency analysis framework in teacher education and contributing to decision-making for policy design.
Future research can be undertaken by incorporating additional input attributes and machine learning methods into the modeling procedure. In addition, it is crucial to leverage the efficacy of an extra gradient method in order to analyze the learning patterns of individuals, address their issues, enhance the educational environment, and enable data driven decision-making for the policy design of teacher education in Thailand.

Author Contributions

Conceptualization, W.C. and N.J.-o.; methodology, W.C.; software, W.C. and R.S.; validation, R.S. and N.J.-o.; formal analysis, W.C. and N.J.-o.; investigation, R.S.; resources, N.J.-o.; data curation, N.J.-o. and W.C.; writing—original draft preparation, N.J.-o.; writing—review and editing, R.S.; supervision, W.C.; project administration, N.J.-o. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

Watcharaporn Cholamjiak would like to thank National Research Council of Thailand and University of Phayao (N42A650334), and Thailand Science Research and Innovation, University of Phayao (FF66-UoE). The authors would like to thank Lampang Rajabhat University for supportting the raw dataset to be analyzed in the main result of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. National Council of Teachers of Mathematics. Principles to Actions: Ensuring Mathematical Success for All; National Council of Teachers of Mathematics: Reston, VA, USA, 2014. [Google Scholar]
  2. Graham, C.R.; Burgoyne, N.; Cantrell, P.P.; Smith, L.M.; Clair, L.S.; Harris, R. TPACK development in science teaching: Measuring the TPACK confidence of inservice science teachers. TechTrends 2009, 53, 70–79. [Google Scholar]
  3. Roshelle, J.; Leinwand, S. Improving student achievement by systematically integrating effective technology. NCSM J. Math. Educ. Leadersh. 2011, 13, 3–9. [Google Scholar]
  4. Niess, M.L.; Roschelle, J. Transforming Teachers’ Knowledge for Teaching Mathematics with Technologies through Online Knowledge-Building Communities. In Proceedings of the 40th Annual Meeting of the North American Chapter of the International Group for the Psychology of Mathematics Education, Greenville, SC, USA, 15–18 November 2018; pp. 44–62. [Google Scholar]
  5. Ahshan, R. A framework of implementing strategies for active student engagement in remote/online teaching and learning during the COVID-19 pandemic. Educ. Sci. 2021, 11, 483. [Google Scholar] [CrossRef]
  6. Hill, H.C.; Rowan, B.; Ball, D.L. Effects of teachers’ mathematical knowledge for teaching on student achievement. Am. Educ. Res. J. 2005, 42, 371–406. [Google Scholar] [CrossRef] [Green Version]
  7. Barlovits, S.; Jablonski, S.; Lázaro, C.; Ludwig, M.; Recio, T. Teaching from A Distance—Math Lessons during COVID-19 in Germany and Spain. Educ. Sci. 2021, 11, 406. [Google Scholar] [CrossRef]
  8. National Council of Teachers of Mathematics. Catalyzing Change in Middle School Mathematics: Initiating Critical Conversations; National Council of Teachers of Mathematics: Reston, VA, USA, 2020. [Google Scholar]
  9. Adipat, S. Developing Technological Pedagogical Content Knowledge (TPACK) through Technology-enhanced Content and Language-Integrated Learning (T-CLIL) instruction. Educ. Inf. Technol. 2021, 26, 6461–6477. [Google Scholar] [CrossRef]
  10. Thomas, M.O.J.; Hong, Y.Y. Teacher integration of technology into mathematics learning. Int. J. Technol. Math. Educ. 2013, 20, 69–84. [Google Scholar]
  11. Muu, L.D.; Oettli, W. Convergence of an adaptive penalty scheme for finding constrained equilibria. Nonlinear Anal. Theory Methods Appl. 1992, 18, 1159–1166. [Google Scholar] [CrossRef]
  12. Blum, E.; Oettli, W. From optimization and variational inequalities to equilibrium problems. Math. Stud. 1994, 63, 123–145. [Google Scholar]
  13. Tan, B.; Cho, S.Y. Strong convergence of inertial forward-backward methods for solving monotone inclusions. Appl. Anal. 2021, 101, 1–29. [Google Scholar] [CrossRef]
  14. Rehman, H.U.; Kumam, W.; Sombut, K. Inertial modification using self-adaptive subgradient extragradient techniques for equilibrium programming applied to variational inequalities and fixed-point problems. Mathematics 2022, 10, 1751. [Google Scholar] [CrossRef]
  15. Muangchoo, K. Three novel two-step proximal-like methods for solving equilibrium and fixed point problems in real Hilbert spaces. Comp. Appl. Math. 2022, 41, 374. [Google Scholar] [CrossRef]
  16. Tran, D.Q.; Dung, M.L.; Nguyen, V.H. Extragradient algorithms extended to equilibrium problems. Optimization 2008, 57, 749–776. [Google Scholar] [CrossRef]
  17. Korpelevich, G. The extragradient method for finding saddle points and other problems. Matecon 1976, 12, 747–756. [Google Scholar]
  18. Censor, Y.; Gibali, A.; Reich, S. Algorithms for the split variational inequality problem. Numer. Algorithms 2012, 59, 301–323. [Google Scholar] [CrossRef]
  19. Rehman, H.U.; Kumam, P.; Cho, Y.J.; Yordsorn, P. Weak convergence of explicit extragradient algorithms for solving equilibirum problems. J. Inequalities Appl. 2019, 2019, 1–25. [Google Scholar] [CrossRef]
  20. Rehman, H.U.; Kumam, P.; Kumam, W.; Shutaywi, M.; Jirakitpuwapat, W. The inertial sub-gradient extra-gradient method for a class of pseudo-monotone equilibrium problems. Symmetry 2020, 12, 463. [Google Scholar] [CrossRef] [Green Version]
  21. Polyak, B.T. Some methods of speeding up the convergence of iteration methods. USSR Comput. Math. Math. Phys. 1964, 4, 1–17. [Google Scholar] [CrossRef]
  22. Tiel, J.V. Convex Analysis: An Introductory Text, 1st ed.; Wiley: New York, NY, USA, 1984. [Google Scholar]
  23. Auslender, A.; Teboulle, M.; Ben-Tiba, S. A logarithmic-quadratic proximal method for variational inequalities. Comput. Optim. Appl. 1999, 12, 31–40. [Google Scholar] [CrossRef]
  24. Bauschke, H.H.; Combettes, P.L. Convex Analysis and Monotone Operator Theory in Hilbert Spaces, 2nd ed.; CMS Books in Mathematics; Springer: Cham, Switzerland, 2017. [Google Scholar]
  25. Niess, M.L.; Ronau, R.N.; Shafer, K.G.; Driskell, S.O.; Harper, S.R.; Johnston, C.; Browning, C.; Aslizgn-Koca, S.; Kersaint, G. Mathematics teacher TPACK standards and development model. Contemp. Issues Technol. Teach. Educ. 2009, 9, 4–24. [Google Scholar]
  26. Huang, G.-B.; Zhu, Q.-Y.; Siew, C.-K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  27. Tibshirani, R. Regression shrinkage and selection via the Lasso. J. R. Stat. Soc. Ser. B Stat. Methodol. 1996, 58, 267–288. [Google Scholar] [CrossRef]
  28. Han, J.; Kamber, M.; Pei, J. Data Mining: Concepts and Techniques, 3rd ed.; Morgan Kaufman Publishers: Waltham, MA, USA, 2012; p. 978. [Google Scholar]
  29. Ingersoll, R. A comparative study of teacher preparation and qualifications in six nations. CPRE Policy Briefs 2007, 47, 1–16. [Google Scholar]
  30. Ndlovu, M.; Ramdhany, V.; Spangenberg, E.D.; Govender, R. Preservice teachers’ beliefs and intentions about integrating mathematics teaching and learning ICTs in their classrooms. ZDM 2020, 52, 1365–1380. [Google Scholar] [CrossRef]
  31. Anderson, S.E.; Maninger, R.M. Preservice teachers abilities, beliefs, and intentions regarding technology integration. J. Educ. Comput. Res. 2007, 37, 151–172. [Google Scholar] [CrossRef]
  32. Raman, A.; Thannimalai, R. Importance of technology leadership for technology integration: Gender and professional development perspective. SAGE Open 2019, 9, 1–13. [Google Scholar] [CrossRef] [Green Version]
  33. Niess, M.L. Investigating TPACK: Knowledge growth in teaching with technology. J. Educ. Comput. Res. 2011, 44, 299–317. [Google Scholar] [CrossRef]
  34. Bonafini, F.C.; Lee, Y. Investigating prospective teachers TPACK and their use of mathematical action technologies as they create screencast video lessons on iPads. Techtrends 2021, 65, 303–319. [Google Scholar] [CrossRef]
  35. Mouza, C.; Karchmer-Klein, R.; Nandakumar, R.; Ozden, S.Y.; Hu, L. Investigating the impact of an integrated approach to the development of preservice teachers’ Technological Pedagogical Content Knowledge (TPACK). Comput. Educ. 2014, 71, 206–221. [Google Scholar] [CrossRef]
  36. Durak, H.Y. Preparing pre-Service teachers to integrate teaching technologies into their classrooms: Examining the effects of teaching environments based on open-ended, hands-on and authentic tasks. Educ. Inf. Technol. 2021, 26, 5365–5387. [Google Scholar] [CrossRef]
  37. Ratnayake, I.; Thomas, M.; Kensington-Miller, B. Professional development for digital technology task design by secondary mathematics teachers. ZDM 2020, 52, 1423–1437. [Google Scholar] [CrossRef]
Figure 1. Accuracy plots of the iteration of Algorithm 3 from Table 5.
Figure 1. Accuracy plots of the iteration of Algorithm 3 from Table 5.
Mathematics 10 04464 g001
Figure 2. Loss plots of the iteration of Algorithm 3 from Table 5.
Figure 2. Loss plots of the iteration of Algorithm 3 from Table 5.
Mathematics 10 04464 g002
Table 1. Overview of all attributes used to train the models.
Table 1. Overview of all attributes used to train the models.
AttributesMeanSDCVMinMax
Major2.541.0842.7214
Gender1.770.4223.7112
GPA3.550.6819.2425
IT for learning competency3.810.9926.0515
Innovative skill3.970.8822.2515
Technology knowledge for a specific subject4.380.9221.0015
Number of suplementation1.160.4135.7614
Curriculum pattern2.090.7335.2713
Selective technology courses7.342.1228.91213
Table 2. Numerical results of different parameters λ n , λ 1 .
Table 2. Numerical results of different parameters λ n , λ 1 .
λ n , λ 1 Algorithm 1 Algorithm 2 Algorithm 3
Training Time (s)Iter.Training Time (s)Iter.Training Time (s)Iter.
0.999 / L 2 0.1329440.1508440.06362
0.999 / K 0.1226440.1464440.04982
0.9999 / L 2 0.1497480.1629480.06782
0.9999 / K 0.1363480.1585480.05192
1 / L 2 0.1465490.1660490.05962
1 / K 0.1364490.1599490.05072
Table 3. Numerical results of different parameters δ n .
Table 3. Numerical results of different parameters δ n .
δ n Algorithm 1 Algorithm 2 Algorithm 3
Training Time (s)Iter.Training Time (s)Iter.Training Time (s)Iter.
0.20.1114370.1902370.05932
0.990.0411110.0420110.07142
1 / 3 n 0.1552440.1800440.04792
1 / 10 n 0.1746450.2034450.05722
1 n 2 x n x n 1 + 1 0.0430140.0675140.05732
Table 4. Numerical results of different parameters α n .
Table 4. Numerical results of different parameters α n .
α n Algorithm 1 Algorithm 2 Algorithm 3
Training Time (s)Iter.Training Time (s)Iter.Training Time (s)Iter.
n / ( 10 n + 1 ) 0.1210260.1099260.05242
3 n / ( 10 n + 1 ) 0.0543150.0569150.04792
5 n / ( 10 n + 1 ) 0.0411110.0420110.04982
7 n / ( 10 n + 1 ) 0.041890.037390.04872
9 n / ( 10 n + 1 ) 0.029880.032480.05052
Table 5. The performance of our Algorithm 3 compared with the other existing algorithms.
Table 5. The performance of our Algorithm 3 compared with the other existing algorithms.
Iter. No.Training Time (s)Precision (%)Recall (%)F1-Score (%)Accuracy (%)
Algorithm 180.029881.0610089.5481.06
Algorithm 280.032481.0610089.5481.06
Algorithm 320.047981.0610089.5481.06
Algorithm (2)230.071181.0610089.5481.06
Algorithm (3)20.048081.0610089.5481.06
Algorithm (4)230.078781.0610089.5481.06
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jun-on, N.; Cholamjiak, W.; Suparatulatorn, R. A Convergent Algorithm for Equilibrium Problem to Predict Prospective Mathematics Teachers’ Technology Integrated Competency. Mathematics 2022, 10, 4464. https://doi.org/10.3390/math10234464

AMA Style

Jun-on N, Cholamjiak W, Suparatulatorn R. A Convergent Algorithm for Equilibrium Problem to Predict Prospective Mathematics Teachers’ Technology Integrated Competency. Mathematics. 2022; 10(23):4464. https://doi.org/10.3390/math10234464

Chicago/Turabian Style

Jun-on, Nipa, Watcharaporn Cholamjiak, and Raweerote Suparatulatorn. 2022. "A Convergent Algorithm for Equilibrium Problem to Predict Prospective Mathematics Teachers’ Technology Integrated Competency" Mathematics 10, no. 23: 4464. https://doi.org/10.3390/math10234464

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop