Next Article in Journal
Valuing in the Agrifood System: The Case of Fresh Grain Legumes in Denmark
Next Article in Special Issue
A Study on Teachers’ Continuance Intention to Use Technology in English Instruction in Western China Junior Secondary Schools
Previous Article in Journal
Short Term Power Load Forecasting Based on PSVMD-CGA Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

A Meta-Analysis and Systematic Review of the Effect of Chatbot Technology Use in Sustainable Education

Faculty of Foreign Studies, Beijing Language and Culture University, Beijing 100083, China
*
Author to whom correspondence should be addressed.
Sustainability 2023, 15(4), 2940; https://doi.org/10.3390/su15042940
Submission received: 29 December 2022 / Revised: 3 February 2023 / Accepted: 3 February 2023 / Published: 6 February 2023
(This article belongs to the Special Issue The Use of Digital Technology for Sustainable Teaching and Learning)

Abstract

:
The development of artificial intelligence in recent years has promoted the use of chatbot technology in sustainable education. Many studies examined the effect of chatbots on learning outcomes. However, scant studies summarized the effectiveness of chatbots in education. The aim of the study is to investigate the effect of chatbot-assisted learning on various components and how different moderator variables influenced the effectiveness. This study, through a meta-analysis, reviewed 32 empirical studies with 2201 participants published between 2010 and 2022. The results showed that chatbot technology exerted a medium-to-high effect on overall learning outcomes regardless of moderator variables, i.e., intervention duration, chatbot roles, and learning content. In terms of learning components, chatbots could significantly improve explicit reasoning, learning achievement, knowledge retention, and learning interest despite negative findings in critical thinking, learning engagement, and motivation. Future research could expand chatbot research by including different learning components.

1. Introduction

As the popularity of information technologies grows, chatbots have caught the increasing attention of stakeholders in the educational context. A chatbot is a conversational program that can process input and accordingly provide information through verbal or written interactions [1]. Researchers and practitioners could even design chatbots by themselves based on Dialogflow (e.g., [2,3]). Dialogflow is a natural language understanding platform integrating conversational interfaces into various devices, applications, and bots [4]. During the post-pandemic period when online learning still plays an important role, chatbot integration alleviates teachers’ workload to provide individual support for students with limited resources and personalizes students’ pace of learning [5]. Unlike teachers, educational chatbots could answer students’ questions anytime and anywhere. They have the ability to handle several questions at the same time.
However, the use of chatbot technology also brings challenges. They can be described as issues in ethics, evaluation, users’ attitudes, programming, supervision, and maintenance [6]. Problems have also included technological limitations and training side effects [7]. Simultaneously, there comes another issue. The novelty effect may appear when students are introduced to new technology. The improvement of learning outcomes might result from students’ newness to chatbot technology [8]. Given the benefits and concerns of educational chatbot use, many studies have measured the effectiveness of chatbots but have obtained inconsistent findings.
Although some review studies focused on chatbot-based education, few of them synthesized previous studies to identify the overall effect of chatbots. Recent review articles either provided basic information on chatbot research through visualization (e.g., [9]), or summarized chatbot roles, evaluation methods, application fields, affordances, and challenges through content analysis (Table 1). In terms of the effectiveness of chatbots, there were two meta-analyses calculating the overall effect size of chatbots on language learning. Lee and Hwang [10] limited their research to English as a foreign language education, and Bibauw et al. [11] also used language learning as the research topic.
However, some previous studies only investigated a particular context. Lee and Hwang [10] focused on the effect of chatbots in the Korean context. In recent years, the Chinese Ministry of Education has been advocating technology-enhanced education. Chatbots could improve Chinese students’ thinking ability and facilitate interactive learning [13]. The meta-analysis of chatbot technology in education across the world remains sparse. This study thus aims to examine the effect of chatbot use on various components and whether the effectiveness would be influenced by different variables. This study would shed light on the effect of chatbot-assisted learning not only in China but also in other countries and regions. It could also provide a reference for sustainable education and developing certain abilities and affective domains.

2. Literature Review

Chatbots perform three roles in education, i.e., teaching assistants, learning partners, and personal tutors. Inspired by Li and Yu [14], the authors summarized the role of chatbots in Figure 1. Operating as a teaching assistant, the chatbot mechanism provided professional knowledge and formative feedback [15] and scaffolded students’ online learning [16]. Chatbots, as learning partners, chatted and interacted with students through either texts or voices. The tutorial role required chatbots to offer questions and answers, guided students to start their learning [17], and give quizzes [18]. The three educational roles of chatbots are intertwined with each other, contributing to effective teaching and learning [14,19]. Given the function of chatbots, it is most likely that chatbot-based education would positively and significantly influence critical thinking, explicit reasoning, learning achievement, knowledge retention, engagement, motivation, and interest.

2.1. Critical Thinking

Critical thinking refers to the thinking process of forming self-regulatory and reflective judgments which could determine one’s beliefs and behaviors [20]. As one of the 21st-century skills, critical thinking has become increasingly pivotal in education [21]. Students are encouraged to express their viewpoints based on critical analysis and reasoning [22]. Therefore, recent years have witnessed many studies on the cultivation of critical thinking, especially with the assistance of artificial intelligence and information communication technology tools [23]. The use of emerging technologies such as chatbot systems could guide and inspire students to think over and make judgments, thus gradually developing the habit of critical thinking. The artificial intelligence-integrated chatbot was proved effective to enhance students’ thinking ability and expectations [13]. Accordingly, the authors proposed the following null hypothesis.
H1. 
The use of chatbot technology could not significantly improve critical thinking at the 0.05 level.

2.2. Explicit Reasoning

Explicit reasoning is derived from the academically productive talk framework which emphasizes social interaction and reasoned participation [24]. The explicitness of reasoning is an important feature of students’ effective communication. Peer dialogues should involve learning community, accurate knowledge, and rigorous thinking [25]. Specifically, students are expected to learn from each other, explicate their reasoning, and construct logical arguments in open-ended discussions and collaborative activities. It was found that chatbots could trigger and scaffold students’ discussions, stimulating explicit reasoning processes [26]. Moreover, explicit reasoning could increase collaboration practices, improve learning outcomes, and promote conceptual knowledge acquisition [27]. The authors thus raised the following null hypothesis.
H2. 
The use of chatbot technology could not significantly improve explicit reasoning at the 0.05 level.

2.3. Learning Achievement and Knowledge Retention

Learning achievement is the measurement of students’ academic success in a given period of time [28]. Most of the existing research on chatbot-based learning investigated learning achievement, including gains in second language speaking proficiency [29], aircraft engine maintenance test scores [30], vaccine knowledge scores [31], numerical system conversion test results [18], and transfer ability [32]. Through the pretest–posttest design, Ghanaian undergraduate students in the chatbot group better performed academically than those interacting with instructors [33]. However, chatbots may not significantly improve secondary school students’ academic performance [3].
Knowledge retention, also known as learning retention, is defined as the ability to capture information and transfer it from short-term to long-term memory [34]. Many researchers mainly examined this ability via retention tests, e.g., delayed posttests for vocabulary knowledge [35] and multiple-choice question tests for programming knowledge [32]. The dynamic assessment of chatbots was effective in enhancing vocabulary retention and providing detailed information about personal learning [2]. The chatbot also facilitated the retention of Python programming knowledge [32]. Therefore, the authors proposed the following null hypotheses.
H3. 
The use of chatbot technology could not significantly improve learning achievement at the 0.05 level.
H4. 
The use of chatbot technology could not significantly improve knowledge retention at the 0.05 level.

2.4. Learners’ Engagement, Motivation, and Interest

Many previous studies have examined the effect of chatbots on students’ engagement. Engagement is the extent to which students actively involve or participate in learning activities [36]. A narrative-based learning system equipped with chatbot feedback significantly improved users’ engagement [37]. Likewise, a mobile chatbot-based learning approach enabled nursing students to believe that chatbots could promote their learning engagement [31]. However, the interaction with chatbots failed to make a statistically significant difference in engagement in the extensive reading activity [38]. Considering inconsistent findings, the following null hypothesis was determined.
H5. 
The use of chatbot technology could not significantly enhance learning engagement at the 0.05 level.
Researchers also investigated the effect of chatbots on learning motivation. Motivation, unlike engagement, refers to the possibility of engaging in learning tasks and maintaining learning behavior [39]. Essentially, engagement highlighted action, while motivation emphasized intent [40]. Motivation can be categorized into intrinsic and extrinsic motivations [41]. Intrinsic motivation focuses on learners’ inner satisfaction, whereas extrinsic motivation is defined as the behavior for external and separable results [42]. The voice-based chatbots positively influenced middle school students’ motivation [43]. Furthermore, chatbot-assisted instructional videos and micro-learning systems could also effectively promote intrinsic motivation [18,44]. Thus, the authors presented the following null hypothesis.
H6. 
The use of chatbot technology could not significantly enhance learning motivation at the 0.05 level.
Previous studies also focused on the effect of chatbot technology on learning interest. Interest indicates the individual readiness or predisposition to engage in a given learning task with effort [45]. Chatbots improved Korean students’ interest in foreign language learning [43]. Similarly, nursing college students using chatbots experienced a higher level of learning interest than their peers in the control group [46]. Chatbots also increased college students’ interest in English vocabulary learning [47]. The authors thus raised the following null hypothesis.
H7. 
The use of chatbot technology could not significantly enhance learning interest at the 0.05 level.

2.5. Intervention Duration

Previous studies experimented with different durations and obtained different results. The study [18] whose intervention lasted for only 40 min found no significant differences in learning performance between the chatbot and control groups. Fifteen instructional sessions carried out over three weeks facilitated students’ comprehension of English adjectival constructions but failed to help learners generate prepositional constructions [48]. However, speaking test scores in the experimental group were significantly higher than those of the control group after the four-month experiment [49]. In the field of educational technology, some meta-analyses have investigated the influence of implementation duration. For example, Chen et al. [50] examined the effects of mobile devices on language learning across different intervention durations. The authors therefore developed the following null hypothesis.
H8. 
Intervention duration could not influence the effect of chatbot-assisted learning.

2.6. Chatbot Roles

Researchers assigned different roles to chatbots in their experiments. University students created conversations with the chatbot Elbot ranging from school life to movies. After eight weeks, the experimental group better acquired vocabulary knowledge than the control group [47]. The AsasaraBot, acting as a tutor, provided questions, encouragement, and interactions, aiming to support students’ language learning [15]. However, the experimental group performed worse than students equipped with other technological tools such as Google Forms. Using chatbots as teaching assistants, learners significantly outperformed those in traditional classroom settings in terms of projected-based learning performance [51]. The meta-analysis of robot-assisted language learning [52] examined the influence of robot roles on the effectiveness of social robots. Therefore, the authors formulated the following null hypothesis.
H9. 
Chatbot roles could not influence the effect of chatbot-assisted learning.

2.7. Learning Content

Chatbot technology has been applied to many disciplines; thus, participants’ learning content has varied among studies. In nursing education, students used chatbot systems to learn courses about the physical examination, effectively enhancing students’ academic performance [53]. The LINE Bot used in military science could provide procedures for engine fan module decomposition, which improved trainee performance and reduced training costs [30]. In the field of second language learning, students interacted with chatbots to practice their speaking skills [29]. However, the chatbot was ineffective when students developed computing knowledge, e.g., conversion of numerical systems [18]. The authors thus proposed the following null hypothesis.
H10. 
Learning content could not influence the effect of chatbot-assisted learning.

3. Materials and Methods

The researchers conducted a meta-analysis by collecting studies, coding included studies, and calculating the effect sizes. They strictly followed the Preferred Reporting Items for Systematic reviews and Meta-Analyses (PRISMA) guidelines [54]. It was not necessary to pre-register this systematic review in a designated public repository such as Prospero since this study did not involve the health of animals and human beings.

3.1. Literature Search

At the beginning, the authors determined search keywords by clustering the literature. They obtained 741 results in the Web of Science by keying in the research themes “chatbot” AND “learning”. The results were output in the form of plain texts and imported into VOSviewer. The type of analysis was co-occurrence, and the unit of analysis was all keywords. The minimum number of occurrences of a keyword was set at three. Of the 2295 keywords, 217 met the threshold and were categorized into 11 clusters (Figure 2). The researchers obtained the top 10 frequently-occurring keywords: chatbot (N = 340), artificial intelligence (N = 96), machine learning (N = 90), chatbots (N = 89), deep learning (N = 81), natural language processing (N = 73), conversational agents (N = 36), conversational agent (N = 28), education (N = 24), and technology (N = 23).
The authors obtained 2322 studies from online databases on 3 September 2022. The major databases included Web of Science, Wiley Online Library, Springer Link, Taylor & Francis Online, Elsevier ScienceDirect, and Google Scholar (Figure 3). Considering the above findings of bibliometric analysis and the aim of this study, the researchers retrieved 228 results by entering the topic terms: (chatbot OR “conversational agent”) AND (education OR learn* OR teach*) AND (“control group” OR experim* OR experient*) from Web of Science Core Collection. They also obtained 56 results from Wiley and 54 results from Taylor & Francis by keying in “chatbot OR conversational agent” AND “education OR learning OR teaching OR control OR experiment” in the Abstract. Researchers obtained 20 records from Springer and 1860 results from Google Scholar by entering in “chatbot” in where the title contains AND “control group” in with the exact phrase AND “education OR learn OR teach” in with at least one of the words. Researchers also retrieved 104 studies by keying in “chatbot OR conversational agent” in the Title AND “education OR learning OR teaching OR control OR experiment” in Title, abstract, keywords from Elsevier.

3.2. Inclusion and Exclusion Criteria

The researchers followed the inclusion and exclusion criteria to select the literature. The identified studies should (1) determine the effect of chatbot technology on educational outcomes; (2) include the experimental group that adopted chatbot-based learning and the control group that used traditional learning approaches; (3) report sufficient statistics, i.e., sample sizes, means, and standard deviations, to calculate effect sizes; (4) ensure homogeneity between the control group and the experimental group, i.e., students’ prior learning outcomes should be equivalent; and (5) be written in English and be published from 2010 to 2022. The studies were excluded if they (1) were irrelevant to educational use of chatbots; (2) lacked control groups; (3) did not provide adequate information for effect-size calculations; and (4) were not written in an acceptable English language.
Based on the inclusion and exclusion criteria, the researchers included 25 studies in this meta-analysis after the first round of screening. To avoid the case that some studies may be excluded by mistake, the researchers conducted another two rounds of screening. They found that four studies, which met all the inclusion criteria, were accidentally excluded in the first round of screening. Therefore, given three studies from previous reviews, there were altogether 32 studies included in this meta-analysis (Figure 3).

3.3. Coding Procedures

The researchers developed a coding scheme consisting of comparable features of chatbot-based learning. First, basic information included the author’s last name, publication year, and sample sizes. Second, regarding the instruction duration, the researchers at first coded it as a continuous variable. However, inspired by Chen et al. [50], the researchers decided to recode this variable as a categorical one for further analysis. Considering the data characteristics, the researchers divided the duration variable into five categories: less than 1 week, less than 5 weeks, less than 10 weeks, more than 10 weeks, and not specified. Third, the chatbot role variable was classified into teaching assistants, tutors, and learning partners. Fourth, the researchers categorized the educational outcomes into critical thinking, explicit reasoning, learning achievement, knowledge retention, learning engagement, learning motivation, and learning interest. The defining terms of the above categories were, respectively, critical thinking scores [53], the frequency of explicit position and explicit argument [24], test scores [29], retention test scores [2], engagement scale scores [38], motivation questionnaire scores [18], and interest scale scores [47]. Fifth, in terms of learning content, the researchers divided it into five categories, i.e., computer science, instructional technology, language, medicine, and others. Table 2 shows the detailed information of included studies.

3.4. Data Analysis

This study used Stata MP/14.0 to carry out the meta-analysis. Cohen’s d, responsible for the measurement of effect sizes, is calculated by dividing the mean difference between the experimental and control groups by the pooled standard deviation [67]. The calculation formula is as follows. The researchers calculated 76 effect sizes with a total sample size of 2201 in 32 identified studies.
Cohen s   d = M E     M C N E     1   S E 2 + N C 1   S C 2   N E 1 + N C 1
The researchers examined publication bias using both visual and mathematical tests, including the funnel plot, Begg’s test, and Egger’s test. Through the shape of the funnel plot, researchers could preliminarily assess publication bias. The dots will be symmetrically distributed along the no-effect line if there is no publication bias, while the plot will be asymmetric if there is a presence of publication bias. Statistically, both Begg’s and Egger’s tests reported p-value, which determines the presence or absence of publication bias.
The researchers also measured heterogeneity using I2 test and conducted sensitivity analysis. The heterogeneity will be considered low if I2 is less than 25%, moderate if I2 falls between 25% and 75%, and substantial if I2 is greater than 75% [68]. Generally, heterogeneity is significant when I2 is larger than 50% and a random effects model should be adopted accordingly. Otherwise, heterogeneity is insignificant, and the fixed effects model should be used.

4. Results

4.1. Analyses of Publication Bias

Both Egger’s and Begg’s tests (Table 3) checked whether the identified studies were influenced by publication bias. The results of both tests indicated the absence of publication bias in critical thinking (t = 14.17, p = 0.111; z = 1.47, p = 0.142), explicit reasoning (t = 1.90, p = 0.424; z = 1.32, p = 0.188), learning achievement (t = 4.36, p = 0.083; z = 0.53, p = 0.598), knowledge retention (t = 10.18, p = 0.061; z = 1.47, p = 0.142), learning engagement (t = −0.14, p = 0.946; z = 1.35, p = 0.176), and learning motivation (t = 8.17, p = 0.216; z = 1.23, p = 0.217). The funnel plot also confirmed test results (Figure 4). Using learning motivation as an example, the funnel graph was obviously symmetrical, revealing no publication bias. However, Begg’s test showed a presence of publication bias in learning interest (z = 2.04, p = 0.042), which was different from the result of Egger’s test (t = 15.30, p = 0.056).

4.2. Results of Sensitivity Analysis

The researchers implemented a sensitivity analysis to examine the stability of meta-analytical results. As shown in Figure 5, all estimates range from the lower confidence interval limit (95% CI = 0.55) to the upper confidence interval limit (95% CI = 0.68). It indicates that none of the included studies could influence the pooled effect size when a specific study is omitted, confirming that the meta-analytical results are robust.

4.3. Results of Moderator Analyses

Table 4 presents the overall effect sizes of each outcome domain. Since I2 statistics revealed that the effect size in critical thinking (Q = 55.89, I2 = 92.8%, p < 0.001) was of considerable heterogeneity, the researchers adopted the random effects model. The effect size did not show a significant difference in critical thinking (d = 0.690, 95% CI [−0.235, 1.615], p = 0.144) between the chatbot-assisted and conventional learning methods. Thus, the authors accepted hypothesis 1.
However, the effect size in explicit reasoning (Q = 1.88, I2 = 0.0%, p = 0.865) was not significantly heterogeneous. The fixed effects model was thus used. Table 4 shows a significant difference in explicit reasoning (d = 1.190, 95% CI [0.868, 1.512], p < 0.001) between the experimental and control groups. The authors thus rejected hypothesis 2.
Effect sizes in both learning achievement (Q = 221.14, I2 = 85.5%, p < 0.001) and knowledge retention (Q = 22.96, I2 = 82.6%, p < 0.001) were considered heterogeneous. The authors accordingly employed the random effects model to pool the data. It was revealed that learning achievement (d = 1.033, 95% CI [0.743, 1.322], p < 0.001) and retention (d = 0.691, 95% CI [0.101, 1.281], p = 0.022) in the experimental group significantly improved compared with the control group. Therefore, the authors rejected hypotheses 3 and 4.
Because of I2 = 0.0% in learning engagement (Q = 3.22), the effect sizes were of insignificant heterogeneity at the 0.05 level (p = 0.780). The authors thus chose the fixed effects model to pool the effect sizes. No significant differences in learning engagement (d = 0.147, 95% CI [−0.068, 0.363], p = 0.114) were found between chatbot-assisted and traditional learning methods. The authors, therefore, accepted hypothesis 5.
The effect sizes in learning motivation (Q = 161.61, I2 = 93.2%, p < 0.001) and interest (Q = 23.39, I2 = 87.2%, p < 0.001) were heterogeneous. Thus, the authors adopted the random effects model when conducting the meta-analysis regarding both educational outcomes. Meta-analytical results (Table 4) showed that compared with the control group, the experimental group maintained significantly higher levels of learning interest (d = 0.842, 95% CI [0.034, 1.650], p = 0.041). However, learning motivation in both groups was not significantly different (d = 0.409, 95% CI [−0.099, 0.916], p = 0.114). The authors thus accepted hypothesis 6 and rejected hypothesis 7.
The overall effect size of using chatbots in education was 0.789 (p < 0.001), with a 95% confidence interval between 0.593 and 0.985 (Table 4). According to Cohen’s [69] effect-size criteria, which identified 0.2, 0.5, and 0.8 as small, medium, and large effect sizes respectively. The overall effect size in this study was thus considered as a medium-to-large level, indicating that the use of chatbot technology exerted a positive effect on learning outcomes. Additionally, the heterogeneity test showed that the overall effect size (Q = 582.09, I2 = 87.8%, p < 0.001) was of considerable heterogeneity, which deserved further analysis to explore potentially critical moderator variables.
The researchers implemented a meta-regression analysis for three moderator variables (Table 5). Regarding intervention duration, there were medium-sized effects on chatbot-assisted learning for less than one week (d = 0.775, p < 0.001), less than 10 weeks (d = 0. 561, p < 0.001), and more than 10 weeks (d = 0.601, p < 0.01) and large-sized effects for less than 5 weeks (d = 1.060, p < 0.05) and not specified (d = 1.844, p < 0.001). However, there were no significant differences between effect sizes of the different intervention durations (p > 0.05). Thus, hypothesis 8 was accepted.
The results for the remaining two variables presented similar patterns. Regarding chatbot roles, a large effect size was reported for using chatbots as teaching assistants (d = 1.631, p < 0.001), while the role of tutors (d = 0.748, p < 0.001) and partners (d = 0.712, p < 0.001) yielded medium effect sizes. Nevertheless, no significant differences existed between effect sizes of three roles of chatbots (p > 0.05). The authors thus accepted hypothesis 9.
In terms of learning content, computer science (d = 0.695, p < 0.01) and language (d = 0.749, p < 0.01) showed medium effect sizes, and large effect size estimates came from instructional technology (d = 0.928, p < 0.001) and medicine (d = 1.588, p < 0.05). The categories of math and military science were merged into one category (i.e., others) due to the limited studies, with no significant effect sizes (d = 0.519, p > 0.05). The meta-regression results revealed no statistically significant differences between the abovementioned values. Therefore, the authors accepted hypothesis 10. The results of the hypothesis testing are summarized in Table 6.

5. Discussion

This study investigated the effect of chatbot-assisted learning on various components and how different moderator variables influenced the effectiveness. There were no significant changes in critical thinking through the use of chatbots, which was inconsistent with the findings of Li et al. [13]. Although limited studies focused on critical thinking, there were still contradictory findings, probably because of the elusive property and different measurement instruments. Critical thinking is difficult to measure. Some studies (e.g., [53]) used a newly developed scale consisting of five items. However, other studies (e.g., [61]) adopted the four-factor inventory developed through exploratory factor analysis. Chatbots may only exert influence on some dimensions of critical thinking, such as inquiring mindset.
The use of chatbot technology significantly enhanced explicit reasoning, which was also underexplored. The existing studies mainly developed students’ explicit reasoning in collaborative activities since one display of explicit reasoning could associate with another one, which was the core of transactivity theory [70]. Chatbot interventions could effectively stimulate group discussions and help students utter their thoughts [55]. Chatbots also asked for clear and convincing statements, motivating students to find strong evidence. Thus, the conversational behaviors in the chatbot group were considered more transactive, i.e., with more explicit arguments, than those in the control group.
Chatbot technology also significantly improved learning achievement and retention. This finding was corroborated by the previous studies that confirmed the positive influence of chatbots on linguistic competence [10] and programming course achievement [33]. Chatbots can easily notice learners’ knowledge gaps and accordingly make responses in order to create meaningful interactions. Review activities before presenting new information activate students’ prior knowledge, facilitating the integration of the new and old knowledge [57]. Regarding knowledge retention, chatbots could randomly generate multiple-choice questions for declarative knowledge testing and open questions for procedural knowledge testing. Students in this way could timely recall their newly acquired information.
However, significant differences in learning engagement and motivation were not found between the chatbot-based condition and the control condition. One possible reason was that some students preferred to finish learning tasks in their own ways and paid little attention to chatbots [38], leading to a decrease in learning engagement. Another reason may be that factors such as peer feedback could influence motivation and that influencing factors varied with learning environments. Specifically, pressure was a significant predictor of motivation in the chatbot-assisted learning context, whereas perceived competency was an influencing factor in the traditional context [18].
With chatbot technology, students experienced more learning interest than those without it, supporting previous studies [10,43,46]. The enhancement of learning interest can be attributed to the flexibility of learning and affective feedback. Chatbot systems allowed users to learn based on individual needs and pace, which avoided frustration and learned helplessness for slow learners. Chatbots were designed to give encouraging messages if students failed to correctly answer questions [30]. They also gave human-like utterances such as uh-huh and yeah.
Three types of chatbot roles revealed no significant differences in learning outcomes. This finding echoed Huang et al.’s [1] suggestion for future research on determining how chatbots can be utilized to best achieve learning outcomes. Students can benefit from chatbot technology regardless of its roles. Chatbots were qualified as teaching assistants and learning partners since human teachers still took the leading role. Interestingly, chatbots could also be employed as tireless personal tutors. The intelligent systems guided and monitored students’ personalized learning, while teachers had the opportunity and time to discover learners’ potential problems [53], thus jointly promoting learning outcomes.
Intervention duration failed to influence the effectiveness of chatbot-based learning. The result did not indicate the novelty effect that learning outcomes may improve in the short term but ultimately decrease over time, which was inconsistent with the study [8]. It was possibly because a growing number of studies (e.g., [44,55]) have attempted to mitigate the novelty effect by introducing and familiarizing students with chatbot technology prior to intervention. Students in the information age can readily reach different technological innovations. Therefore, they were most likely to familiarize themselves with chatbots within several minutes.
Learning content was also not a significant variable. Moderator analysis suggested a more positive result for computer science, instructional technology, language, and medicine than the “others” category. Computer science and language were the most targeted fields in chatbot-based education, while engineering and mathematics received less attention [12]. Due to the small number of studies on other domains, educational fields such as military science and mathematics were subsumed into one category in this study. Therefore, it cannot be concluded that chatbots were more effective for certain learning content than for other categories.

6. Conclusions

6.1. Major Contributions

Methodologically, this study included major databases and conducted a meta-analysis under the PRISMA guidelines to examine the effectiveness of chatbot technology on educational outcomes. Theoretically, the results showed a medium-to-high overall effect size of chatbots on educational outcomes regardless of intervention duration, chatbot roles, and learning content. Chatbot technology exerted a significant and positive influence on explicit reasoning, learning achievement, knowledge retention, and learning interest. However, chatbots did not significantly improve critical thinking, learning engagement, and motivation. Practically, teachers and instructors could adopt appropriate teaching approaches to facilitate sustainable education.

6.2. Limitations

It should be noted that there are some limitations to the present study. First, the researchers only included studies written in English. Some publications, especially in Korean, could not be understood by the researchers. Second, this study only included three moderator variables that did not significantly influence the effectiveness of chatbots. Third, the results may still be influenced by unpublished studies with insignificant results despite the absence of publication bias.

6.3. Implications for Future Research

The findings may shed light on future research directions and propose suggestions for practitioners. First, the results revealed that chatbot-based learning was more effective than traditional learning in terms of explicit reasoning, learning achievement, knowledge retention, and learning interest. Therefore, future research could explore more components of educational outcomes, e.g., learning confidence, self-efficacy, social media use, and cognitive load. Experimental studies with large sample sizes are also expected. Teachers and instructors could integrate chatbot technology with different activities to meet learners’ needs. Educational institutions could also provide training to improve teachers’ and students’ digital literacy and knowledge about artificial intelligence [71].
Second, future research could further explore users’ attitudes towards chatbot technology and students’ learning attitudes. Since the control group in the included studies did not obtain access to chatbots, it was difficult to compare users’ attitudes between the control and intervention groups. Researchers could employ such models as the technology acceptance model and task-technology fit model to analyze the influencing factors of users’ attitudes. On the other hand, learning attitudes could be compared. Only a few studies, however, focused on this aspect [43,57]. Future research could also expand chatbot research by using interdisciplinary research methods.
Third, intervention duration, chatbot roles, and learning content did not influence learning effectiveness. Thus, researchers could include more potential moderator variables for future meta-analyses, e.g., educational levels and interaction types. Future studies could also consider chatbot integration in other underexplored disciplines, e.g., arts, mathematics, and psychology [12]. Teachers could feel free to adopt chatbot-integrated teaching. They could introduce chatbot technology at any stage of the semester and assign any role to chatbots according to teaching needs. Developers and designers could introduce intriguing elements by learning natural language processing and improve chatbots’ performance based on machine learning algorithms.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/su15042940/s1, File S1: Supporting data; File S2: PRISMA checklist.

Author Contributions

Conceptualization, X.D. and Z.Y.; methodology, X.D.; software, X.D. and Z.Y.; validation, X.D. and Z.Y.; formal analysis, X.D. and Z.Y.; investigation, X.D.; resources, X.D.; data curation, X.D.; writing—original draft preparation, X.D.; writing—review and editing, Z.Y.; visualization, X.D.; supervision, Z.Y.; project administration, Z.Y.; funding acquisition, X.D. and Z.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the 2019 MOOC of Beijing Language and Culture University (MOOC201902) (important) “Introduction to Linguistics”; “Introduction to Linguistics” of online and offline mixed courses in Beijing Language and Culture University in 2020; the Special fund of Beijing Co-construction Project Research and reform of the “Undergraduate Teaching Reform and Innovation Project” of Beijing higher education in 2020-innovative “multilingual +” excellent talent training system (202010032003); The research project of Graduate Students of Beijing Language and Culture University “Xi Jinping: The Governance of China” (SJTS202108); The Fundamental Research Funds for the Central Universities, and the Research Funds of Beijing Language and Culture University (22YCX017).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in File S1 in Supplementary Materials.

Acknowledgments

The authors would like to extend their gratitude to the people who help this study and the projects which financially support this study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Huang, W.J.; Hew, K.F.; Fryer, L.K. Chatbots for language learning-Are they really useful? A systematic review of chatbot-supported language learning. J. Comput. Assist. Learn. 2022, 38, 237–257. [Google Scholar] [CrossRef]
  2. Chatbot-Assisted Dynamic Assessment (CA-DA) for L2 Vocabulary Learning and Diagnosis. Available online: https://doi.org/10.1080/09588221.2021.1987272 (accessed on 8 September 2022).
  3. Topal, A.D.; Eren, C.D.; Gecer, A.K. Chatbot application in a 5th grade science course. Educ. Inf. Technol. 2021, 26, 6241–6265. [Google Scholar] [CrossRef] [PubMed]
  4. Dialogflow ES Documentation. Available online: https://cloud.google.com/dialogflow/es/docs (accessed on 31 January 2023).
  5. Cai, W.L.; Grossman, J.; Lin, Z.J.; Sheng, H.; Wei, J.T.Z.; Williams, J.J.; Goel, S. Bandit algorithms to personalize educational chatbots. Mach. Learn. 2021, 110, 2389–2418. [Google Scholar] [CrossRef]
  6. Okonkwo, C.W.; Ade-Ibijola, A. Chatbots applications in education: A systematic review. Comput. Educ. Artif. Intell. 2021, 2, 100033. [Google Scholar] [CrossRef]
  7. Perez, J.Q.; Daradoumis, T.; Puig, J.M.M. Rediscovering the use of chatbots in education: A systematic literature review. Comput. Appl. Eng. Educ. 2020, 28, 1549–1565. [Google Scholar] [CrossRef]
  8. Fryer, L.K.; Ainley, M.; Thompson, A.; Gibson, A.; Sherlock, Z. Stimulating and sustaining interest in a language course: An experimental comparison of Chatbot and Human task partners. Comput. Hum. Behav. 2017, 75, 461–468. [Google Scholar] [CrossRef]
  9. A Review of Opportunities and Challenges of Chatbots in Education. Available online: https://doi.org/10.1080/10494820.2021.1952615 (accessed on 8 September 2022).
  10. Lee, J.-Y.; Hwang, Y. A meta-analysis of the effects of using AI chatbot in Korean EFL education. Stud. Engl. Lang. Lit. 2022, 48, 213–243. [Google Scholar] [CrossRef]
  11. Bibauw, S.; Van den Noortgate, W.; Francois, T.; Desmet, P. Dialogue systems for language learning: A meta-analysis. Lang. Learn. Technol. 2022, 26, 1–24. Available online: https://hdl.handle.net/10125/73488 (accessed on 27 September 2022).
  12. Kuhail, M.A.; Alturki, N.; Alramlawi, S.; Alhejori, K. Interacting with educational chatbots: A systematic review. Educ. Inf. Technol. 2023, 28, 973–1018. [Google Scholar] [CrossRef]
  13. Li, L.; Subbareddy, R.; Raghavendra, C.G. Al intelligence chatbot to improve students learning in the higher education platform. J. Interconnect. Netw. 2022, 22, 2143032. [Google Scholar] [CrossRef]
  14. Li, M.; Yu, Z.G. Teachers’ satisfaction, role, and digital literacy during the COVID-19 pandemic. Sustainability 2022, 14, 1121. [Google Scholar] [CrossRef]
  15. Mageira, K.; Pittou, D.; Papasalouros, A.; Kotis, K.; Zangogianni, P.; Daradoumis, A. Educational AI chatbots for content and language integrated learning. Appl. Sci.-Basel 2022, 12, 3239. [Google Scholar] [CrossRef]
  16. Song, D.; Kim, D. Effects of self-regulation scaffolding on online participation and learning outcomes. J. Res. Technol. Educ. 2021, 53, 249–263. [Google Scholar] [CrossRef]
  17. Chen, H.L.; Widarso, G.V.; Sutrisno, H. A chatbot for learning Chinese: Learning achievement and technology acceptance. J. Educ. Comput. Res. 2020, 58, 1161–1189. [Google Scholar] [CrossRef]
  18. Yin, J.Q.; Goh, T.T.; Yang, B.; Xiaobin, Y. Conversation technology with micro-learning: The impact of chatbot-based learning on students’ learning motivation and performance. J. Educ. Comput. Res. 2021, 59, 154–177. [Google Scholar] [CrossRef]
  19. Jung, H.; Lee, J.; Park, C. Deriving design principles for educational chatbots from empirical studies on human-chatbot interaction. J. Digit. Content. Soc. 2020, 21, 487–493. [Google Scholar] [CrossRef]
  20. Norris, S.P.; Ennis, R.H. Evaluating Critical Thinking; Midwest Publications: Pacific Grove, CA, USA, 1989. [Google Scholar]
  21. Tang, T.; Vezzani, V.; Eriksson, V. Developing critical thinking, collective creativity skills and problem solving through playful design jams. Think. Skills. Creat. 2020, 37, 100696. [Google Scholar] [CrossRef]
  22. Bag, H.K.; Gursoy, E. The effect of critical thinking embedded English course design to the improvement of critical thinking skills of secondary school learners. Think. Skills Creat. 2021, 41, 100910. [Google Scholar] [CrossRef]
  23. Critical Thinking Cultivation in TESOL with ICT Tools: A Systematic Review. Available online: https://doi.org/10.1080/09588221.2022.2033788 (accessed on 8 September 2022).
  24. Tegos, S.; Demetriadis, S.; Papadopoulos, P.M.; Weinberger, A. Conversational agents for academically productive talk: A comparison of directed and undirected agent interventions. Int. J. Comp.-Supp. Coll. Learn. 2016, 11, 417–440. [Google Scholar] [CrossRef]
  25. Resnick, L.B.; Michaels, S.; O’Connor, C. How (well structured) talk builds the mind. In Innovations in Educational Psychology: Perspectives on Learning, Teaching and Human Development; Sternberg, R.J., Preiss, D.D., Eds.; Springer: New York, NY, USA, 2010; pp. 163–194. [Google Scholar]
  26. Tegos, S.; Demetriadis, S. Conversational agents improve peer learning through building on prior knowledge. Educ. Technol. Soc. 2017, 20, 99–111. [Google Scholar]
  27. Papadopoulos, P.M.; Demetriadis, S.N.; Weinberger, A. ‘Make it explicit!’: Improving collaboration through increase of script coercion. J. Comput. Assist. Learn. 2013, 29, 383–398. [Google Scholar] [CrossRef]
  28. A Meta-Analysis of the Effect of Virtual Reality Technology Use in Education. Available online: https://doi.org/10.1080/10494820.2021.1989466 (accessed on 8 September 2022).
  29. Dizon, G. Evaluating intelligent personal assistants for L2 listening and speaking development. Lang. Learn. Technol. 2020, 24, 16–26. Available online: https://hdl.handle.net/10125/44705 (accessed on 20 September 2022).
  30. Development of Mobile Interactive Courses Based on an Artificial Intelligence Chatbot on the Communication Software LINE. Available online: https://doi.org/10.1080/10494820.2021.1937230 (accessed on 8 September 2022).
  31. Chang, C.Y.; Hwang, G.J.; Gau, M.L. Promoting students’ learning achievement and self-efficacy: A mobile chatbot approach for nursing training. Brit. J. Educ. Technol. 2022, 53, 171–188. [Google Scholar] [CrossRef]
  32. Winkler, R.; Hobert, S.; Salovaara, A.; Söllner, M.; Leimeister, J.M. Sara, the Lecturer: Improving Learning in Online Education with a Scaffolding-Based Conversational Agent. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, 25−30 April 2020. [Google Scholar] [CrossRef]
  33. Essel, H.B.; Vlachopoulos, D.; Tachie-Menson, A.; Johnson, E.E.; Baah, P.K. The impact of a virtual teaching assistant (chatbot) on students’ learning in Ghanaian higher education. Int. J. Educ. Technol. High. Educ. 2022, 19, 57. [Google Scholar] [CrossRef]
  34. Al Mulhim, E.N. Flipped learning, self-regulated learning and learning retention of students with internal/external locus of control. Int. J. Instr. 2021, 14, 827–846. [Google Scholar] [CrossRef]
  35. Shi, Y.S.; Tsai, C.Y. Fostering vocabulary learning: Mind mapping app enhances performances of EFL learners. Comput. Assist. Lang. Learn. 2022. [Google Scholar] [CrossRef]
  36. Cole, P.G.; Chan, L.K.S. Teaching Principles and Practice; Prentice Hall: Haboken, NJ, USA, 1994. [Google Scholar]
  37. Ruan, S.; He, J.Y.; Ying, R.; Burkle, J.; Hakim, D.; Wang, A.; Yin, Y.F.; Zhou, L.; Xu, Q.Y.; AbuHashem, A.; et al. Supporting Children’s Math Learning with Feedback-Augmented Narrative Technology. In Proceedings of the ACM Interaction Design and Children (IDC) Conference, London, UK, 21–24 June 2020. [Google Scholar]
  38. Liu, C.C.; Liao, M.G.; Chang, C.H.; Lin, H.M. An analysis of children’ Interaction with an AI chatbot and its impact on their interest in reading. Comput. Educ. 2022, 189, 104576. [Google Scholar] [CrossRef]
  39. Motivation, Learning Strategies, and Outcomes in Mobile English Language Learning. Available online: https://doi.org/10.1007/s40299-022-00675-0 (accessed on 8 September 2022).
  40. Lyu, B.N.; Lai, C. Learners’ engagement on a social networking platform: An ecological analysis. Lang. Learn. Technol. 2022, 26, 1–22. [Google Scholar] [CrossRef]
  41. Vallerand, R.J. Toward a hierarchical model of intrinsic and extrinsic motivation. Adv. Exp. Soc. Psychol. 1997, 29, 271–360. [Google Scholar] [CrossRef]
  42. Ryan, R.M.; Deci, E.L. Intrinsic and extrinsic motivations: Classic definitions and new directions. Contemp. Educ. Psychol. 2000, 25, 54–67. [Google Scholar] [CrossRef]
  43. Han, D.E. The effects of voice-based AI chatbots on Korean EFL middle school students’ speaking competence and affective domains. Asia-Pac. J. Converg. Res. Interchange 2020, 6, 71–80. [Google Scholar] [CrossRef]
  44. Fidan, M.; Gencel, N. Supporting the instructional videos with chatbot and peer feedback mechanisms in online learning: The effects on learning performance and intrinsic motivation. J. Educ. Comput. Res. 2022, 60, 1716–1741. [Google Scholar] [CrossRef]
  45. Krapp, A. Interest, motivation and learning: An educational-psychological perspective. Eur. J. Psychol. Educ. 1999, 14, 23–40. [Google Scholar] [CrossRef]
  46. Han, J.W.; Park, J.; Lee, H.N. Analysis of the effect of an artificial intelligence chatbot educational program on non-face-to-face classes: A quasi-experimental study. BMC Med. Educ. 2022, 22, 830. [Google Scholar] [CrossRef] [PubMed]
  47. Kim, N.-Y. Chatbots and Korean EFL students’ English vocabulary learning. J. Digit. Converg. 2018, 16, 1–7. [Google Scholar] [CrossRef]
  48. Kim, R. Effects of AI Chatbot-Based Instruction on the Learning of English Adjectival Transitive Resultative Construction by Korean High School Students. Ph.D. Thesis, Seoul National University, Seoul, South Korea, 2022. Available online: https://hdl.handle.net/10371/181242 (accessed on 21 September 2022).
  49. Proposing a Task-Oriented Chatbot System for EFL Learners Speaking Practice. Available online: https://doi.org/10.1080/10494820.2021.1960864 (accessed on 8 September 2022).
  50. Chen, Z.Z.; Chen, W.C.; Jia, J.Y.; An, H.L. The effects of using mobile devices on language learning: A meta-analysis. Educ. Technol. Res. Dev. 2020, 68, 1769–1789. [Google Scholar] [CrossRef]
  51. Kumar, J.A. Educational chatbots for project-based learning: Investigating learning outcomes for a team-based design course. Int. J. Educ. Technol. High. Educ. 2021, 18, 65. [Google Scholar] [CrossRef]
  52. Lee, H.; Lee, J.H. The effects of robot-assisted language learning: A meta-analysis. Educ. Res. Rev. 2022, 35, 100425. [Google Scholar] [CrossRef]
  53. Chang, C.Y.; Kuo, S.Y.; Hwang, G.H. Chatbot-facilitated nursing education: Incorporating a knowledge-based chatbot system into a nursing training program. Educ. Technol. Soc. 2022, 25, 15–27. [Google Scholar]
  54. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. Br. Med. J. 2021, 372, n71. [Google Scholar] [CrossRef]
  55. Tegos, S.; Demetriadis, S.; Karakostas, A. Promoting academically productive talk with conversational agent interventions in collaborative learning settings. Comput. Educ. 2015, 87, 309–325. [Google Scholar] [CrossRef]
  56. Vazquez-Cano, E.; Mengual-Andres, S.; Lopez-Meneses, E. Chatbot to improve learning punctuation in Spanish and to enhance open and flexible learning environments. Int. J. Educ. Technol. High. Educ. 2021, 18, 33. [Google Scholar] [CrossRef]
  57. Lee, Y.F.; Hwang, G.J.; Chen, P.Y. Impacts of an AI-based chatbot on college students’ after-class review, academic performance, self-efficacy, learning attitude, and motivation. Educ. Technol. Res. Dev. 2022, 70, 1843–1865. [Google Scholar] [CrossRef]
  58. Kim, N.-Y. A study on chatbots for developing Korean college students’ English listening and reading skills. J. Digit. Converg. 2018, 16, 19–26. [Google Scholar] [CrossRef]
  59. Kim, H.S.; Kim, N.Y.; Cha, Y. Is it beneficial to use AI chatbots to improve learners’ speaking performance? J. Asia Tefl 2021, 18, 161–178. [Google Scholar] [CrossRef]
  60. Farah, J.C.; Spaenlehauer, B.; Sharma, V.; Rodriguez-Triana, M.J.; Ingram, S.; Gillet, D. Impersonating chatbots in a code review exercise to teach software engineering best practices. In Proceedings of the 2022 IEEE Global Engineering Education Conference (Educon 2022), Gammarth, Tunisia, 28–31 March 2022. [Google Scholar]
  61. Goda, Y.; Yamada, M.; Matsukawa, H.; Hata, K.; Yasunami, S. Conversation with a chatbot before an online EFL group discussion and the effects on critical thinking. J. Inf. Syst. Educ. 2014, 13, 1–7. [Google Scholar] [CrossRef]
  62. Wambsganss, T.; Kueng, T.; Soellner, M.; Leimeister, J.M. ArgueTutor: An adaptive dialog-based learning system for argumentation skills. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, Yokohama, Japan, 8–13 May 2021. [Google Scholar] [CrossRef]
  63. Abbasi, S.; Kazi, H.; Hussaini, N.N. Effect of chatbot systems on student’s learning outcomes. Sylwan 2019, 163, 49–63. Available online: https://www.researchgate.net/publication/336373880 (accessed on 8 September 2022).
  64. Abbasi, S.; Kazi, H. Measuring effectiveness of learning chatbot systems on student’s learning outcome and memory retention. Asian J. Appl. Sci. Eng. 2014, 3, 57–66. [Google Scholar] [CrossRef]
  65. Lin, M.P.-C.; Chang, D. Enhancing post-secondary writers’ writing skills with a chatbot: A mixed-method classroom study. J. Educ. Technol. Soc. 2020, 23, 78–92. Available online: https://www.jstor.org/stable/26915408 (accessed on 12 September 2022).
  66. Na-Young, K. A study on the use of artificial intelligence chatbots for improving English grammar skills. J. Digit. Converg. 2019, 17, 37–46. [Google Scholar] [CrossRef]
  67. Sedgwick, P.; Marston, L. Meta-analyses: Standardised mean differences. Br. Med. J. 2013, 347, f7257. [Google Scholar] [CrossRef]
  68. Higgins, J.P.; Thompson, S.G.; Deeks, J.J.; Altman, D.G. Measuring inconsistency in meta-analyses. Br. Med. J. 2003, 327, 557–560. [Google Scholar] [CrossRef] [PubMed]
  69. Cohen, J. Statistical Power Analysis for the Behavioral Sciences, 2nd ed.; Routledge: New York, NY, USA, 1988. [Google Scholar]
  70. Sionti, M.; Ai, H.; Rosé, C.P.; Resnick, L. A framework for analyzing development of argumentation through classroom discussions. In Educational Technologies for Teaching Argumentation Skills; Pinkwart, N., McLaren, B., Eds.; Bentham Science Publishers: New York, NY, USA, 2012; Rosé, C.P.; pp. 28–55. [Google Scholar] [CrossRef]
  71. Yu, Z.G.; Xu, W.; Yu, L.H. Constructing an online sustainable educational model in COVID-19 pandemic environments. Sustainability 2022, 14, 3598. [Google Scholar] [CrossRef]
Figure 1. Three roles of chatbots in education.
Figure 1. Three roles of chatbots in education.
Sustainability 15 02940 g001
Figure 2. Clustering keywords related to chatbot-assisted learning.
Figure 2. Clustering keywords related to chatbot-assisted learning.
Sustainability 15 02940 g002
Figure 3. The flowchart of study inclusion.
Figure 3. The flowchart of study inclusion.
Sustainability 15 02940 g003
Figure 4. A funnel plot of publication bias of learning motivation.
Figure 4. A funnel plot of publication bias of learning motivation.
Sustainability 15 02940 g004
Figure 5. A plot of sensitivity analysis.
Figure 5. A plot of sensitivity analysis.
Sustainability 15 02940 g005
Table 1. The research focus of relevant review studies.
Table 1. The research focus of relevant review studies.
NFocusRelevant Review Study
1Research authors, journals, and countriesHwang and Chang [9]
2Advantages and challenges of chatbot useHuang et al. [1], Okonkwo, and Ade-Ibijola [6], Perez et al. [7]
3Chatbot rolesKuhail et al. [9]
4Evaluation methodsPerez et al. [4], Kuhail et al. [12]
5Application fieldsOkonkwo and Ade-Ibijola [6], Hwang and Chang [9], Kuhail et al. [12]
6The overall effect of chatbots on language learningLee & Hwang [10], Bibauw et al. [11]
7The overall effect of chatbots on educationThis study
Table 2. The coding results of included studies.
Table 2. The coding results of included studies.
NAuthor (Year)Sample Size (Experimental Group/Control Group)Instruction DurationChatbot RoleLearning ContentEducational Outcomes
1Tegos and Demetriadis (2017) [26]72 (38/34)40 minTutorComputer scienceExplicit reasoning and learning achievement
2Winkler et al. (2020) [32]72 (37/35)30 minTutorComputer scienceLearning achievement and retention
3Song and Kim (2021) [16]56 (27/29)15 weeksPartnerInstructional technologyLearning achievement
4Chang et al. (2022) [53]32 (16/16)100 minTutorMedicineCritical thinking and learning achievement
5Tegos et al. (2015) [55]43 (21/22)70 minTutorComputer scienceExplicit reasoning and learning achievement
6Tegos et al. (2016) [24]64 (32/32)40 minTutorComputer scienceExplicit reasoning and learning achievement
7Fidan and Gencel (2022) [44]94 (54/40)4 weeksTutorInstructional technologyLearning motivation and achievement
8H. L. Chen et al. (2020) [17]58 (19/29)4 weeksTutorLanguageLearning achievement and retention
9Yuan et al. (2021) [30]40 (20/20)2–3 weeksTutorOthersLearning achievement
10Mageira et al. (2022) [15]35 (18/17)2 daysTutorLanguageLearning achievement
11Vazquez-Cano et al. (2021) [56]103 (52/51)2 weeksTeaching assistantLanguageLearning achievement
12Chang, Hwang, et al. (2022) [31]36 (18/18)100 minTutorMedicineLearning achievement
13Yin et al. (2021) [18]99 (51/48)40 minTutorComputer scienceLearning motivation and achievement
14Lee et al. (2022) [57]38 (18/20)200 minTutorMedicineLearning motivation and achievement
15Ruan et al. (2020) [37]36 (18/18)25.75 minTutorOthersLearning engagement and achievement
16Kim (2018) [47]47 (24/23)8 weeksPartnerLanguageLearning interest, motivation, and achievement
17Han (2020) [43]44 (22/22)10 weeksPartnerLanguageLearning interest, motivation, and achievement
18Kim (2018) [58]46 (24/22)16 weeksPartnerLanguageLearning achievement
19Kumar (2021) [51]60 (30/30)10 weeksTeaching assistantInstructional technologyLearning achievement
20Kim et al. (2021) [59]75 (37/38)1 semesterPartnerLanguageLearning achievement
21Jeon (2021) [2]35 (18/17)25 minTutorLanguageLearning achievement and retention
22Farah et al. (2022) [60]20 (11/9)30 minTutorComputer scienceLearning engagement and achievement
23Goda et al. (2014) [61]63 (31/32)30 minPartnerLanguageCritical thinking and learning achievement
67 (32/35)30 minPartnerLanguageCritical thinking
24Wambsganss et al. (2021) [62]55 (31/24)15 minTeaching assistantLanguageLearning achievement
25Kim (2022) [48]64 (32/32)3 weeksTutorLanguageLearning achievement and retention
26Dizon (2020) [29]28 (13/15)10 weeksTutorLanguageLearning achievement
27Abbasi et al. (2019) [63]110 (55/55)Not availableTeaching assistantComputer scienceLearning achievement
28Abbasi and Kazi (2014) [64]72 (36/36)Not availableTeaching assistantComputer scienceLearning achievement
29Lin and Chang (2020) [65]357 (167/190)2 weeksTeaching assistantLanguageLearning achievement
30Liu et al. (2022) [38]62 (41/21)6 weeksPartnerLanguageLearning engagement and interest
31Hsu et al. (2021) [49]48 (24/24)4 monthsTutorLanguageLearning achievement
32Na-Young (2019) [66]70 (36/34)10 sessionsPartnerLanguageLearning achievement
Table 3. Test results of publication bias.
Table 3. Test results of publication bias.
NOutcomenEgger’s TestBegg’s TestContinuity Corrected
BiaspScoresdzpzp
1Critical thinking514.170.11164.081.470.1421.220.221
2Explicit reasoning61.900.42475.321.320.1881.130.260
3Learning achievement334.360.0833464.540.530.5980.510.609
4Knowledge retention510.180.06164.081.470.1421.220.221
5Learning engagement7−0.140.94696.661.350.1761.200.230
6Learning motivation128.170.2161814.581.230.2171.170.244
7Learning interest415.300.05662.942.040.0421.700.089
Table 4. The effect sizes of seven educational outcomes.
Table 4. The effect sizes of seven educational outcomes.
NOutcomeEffect SizeHeterogeneity Test
d95% CIzpQ-ValuedfpI2
1Critical thinking0.690−0.235, 1.6151.460.14455.8940.00092.8%
2Explicit reasoning1.1900.868, 1.5127.250.0001.8850.8650.0%
3Learning achievement1.0330.743, 1.3226.990.000221.14320.00085.5%
4Knowledge retention0.6910.101, 1.2812.290.02222.9640.00082.6%
5Learning engagement0.147−0.068, 0.3631.340.1803.2260.7800.0%
6Learning motivation0.409−0.099, 0.9161.580.114161.61110.00093.2%
7Learning interest0.8420.034, 1.6502.040.04123.3930.00087.2%
8Overall0.7890.593, 0.9857.900.000582.09710.00087.8%
Table 5. The effect sizes of moderator variables.
Table 5. The effect sizes of moderator variables.
NModerator VariableEffect Size Heterogeneity Test
nd95% CIzpCoefficient95% CItp
1Intervention duration 0.026−0.194, 0.2470.240.814
Less than 1 week390.7750.561, 0.9907.090.000
Less than 5 weeks111.0600.218, 1.9022.470.014
Less than 10 weeks180.5610.318, 0.8054.520.000
More than 10 weeks20.6010.203, 0.9982.960.003
Not specified21.8441.496, 2.19310.370.000
2Chatbot roles −0.012−0.271, 0.248−0.090.930
Teaching assistant41.6311.012, 2.2505.170.000
Tutor460.7480.491, 1.0055.710.000
Partner220.7120.438, 0.9865.090.000
3Learning content 0.061−0.147, 0.2700.590.559
Computer science240.6950.267, 1.1233.180.001
Instructional technology80.9280.642, 1.2156.350.000
Language330.7490.546, 0.9517.250.000
Medicine41.5880.363, 2.8122.540.011
Others30.519−0.422, 1.4611.080.279
Table 6. The results of hypothesis testing.
Table 6. The results of hypothesis testing.
NNull HypothesesResults
1The use of chatbot technology could not significantly improve critical thinking at the 0.05 level.Accepted
2The use of chatbot technology could not significantly improve explicit reasoning at the 0.05 level.Rejected
3The use of chatbot technology could not significantly improve learning achievement at the 0.05 level.Rejected
4The use of chatbot technology could not significantly improve knowledge retention at the 0.05 level.Rejected
5The use of chatbot technology could not significantly enhance learning engagement at the 0.05 level.Accepted
6The use of chatbot technology could not significantly enhance learning motivation at the 0.05 level.Accepted
7The use of chatbot technology could not significantly enhance learning interest at the 0.05 level.Rejected
8Intervention duration could not influence the effect of chatbot-assisted learning.Accepted
9Chatbot roles could not influence the effect of chatbot-assisted learning.Accepted
10Learning content could not influence the effect of chatbot-assisted learning.Accepted
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Deng, X.; Yu, Z. A Meta-Analysis and Systematic Review of the Effect of Chatbot Technology Use in Sustainable Education. Sustainability 2023, 15, 2940. https://doi.org/10.3390/su15042940

AMA Style

Deng X, Yu Z. A Meta-Analysis and Systematic Review of the Effect of Chatbot Technology Use in Sustainable Education. Sustainability. 2023; 15(4):2940. https://doi.org/10.3390/su15042940

Chicago/Turabian Style

Deng, Xinjie, and Zhonggen Yu. 2023. "A Meta-Analysis and Systematic Review of the Effect of Chatbot Technology Use in Sustainable Education" Sustainability 15, no. 4: 2940. https://doi.org/10.3390/su15042940

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop