Your new experience awaits. Try the new design now and help us make it even better

SYSTEMATIC REVIEW article

Front. Psychol., 31 October 2025

Sec. Educational Psychology

Volume 16 - 2025 | https://doi.org/10.3389/fpsyg.2025.1691414

This article is part of the Research TopicInnovative ICT Strategies for Inclusive Education: Enhancing Teacher Competencies and Student EngagementView all 12 articles

Unlocking the potential of artificial intelligence in improving learning achievement in blended learning: a meta-analysis

  • 1Faculty of Education, City University of Macau, Taipa, Macau SAR, China
  • 2Smart Learning Institute of Beijing Normal University, Beijing, China
  • 3An-Najah National University, Nablus, Palestine
  • 4Department of Educational Studies, American University in Cairo, New Cairo, Egypt
  • 5School of Computing, University of Eastern Finland, Kuopio, Finland

Introduction: While several studies investigated the effect of blended learning on students’ learning achievement, scant information exists on whether using artificial intelligence (AI) in blended learning could further contribute to the obtained effect.

Methods: To effectively address the challenges and opportunities presented by blended learning and AI, the present study conducts a meta-analysis to systematically examine the impact of AI-enhanced blended learning on students’ learning achievement, considering the significant role of multiple variables in shaping this achievement, including the type of AI technology, instruction duration, research design, and sample size as well as across different educational levels and subject areas. Specifically, 21 studies (N = 2,873 participants) were meta-analyzed.

Results: The obtained results revealed that AI has a medium effect (g = 0.5) on students’ learning achievement in blended learning. Particularly, personalized systems in blended learning had the highest effect (i.e., large) compared to chatbots and intelligent tutoring systems. Finally, it is seen that the educational context (grade level and educational subject), as well as the experiment type (research design, intervention duration, and sample size), moderate the effect of AI on students’ learning achievement in blended learning.

Discussion: The findings of this study can help researchers and practitioners better understand the effects of AI in blended learning, thereby contributing to a better design of teaching and learning experiences accordingly.

1 Introduction

The concept of Blended Learning (BL) has evolved significantly over the past few decades, building upon traditional distance education and face-to-face instruction (Mizza and Rubio, 2020). As technology has advanced, so too has the potential of BL to enhance the learning experience. BL seamlessly integrates the benefits of both online and in-person learning (Mizza and Rubio, 2020). It provides opportunities for personalized feedback, collaborative activities, and asynchronous communication, making it particularly suitable for all students, with diverse learning profiles and those in remote locations (Garrison and Kanuka, 2004; Graham and Dziuban, 2007; Watson, 2008; Wanner and Palmer, 2015).

With the advent of the Fourth Industrial Revolution, Artificial Intelligence (AI) has emerged as a transformative technology in educational technology ecosystems (Holmes et al., 2019). AI-powered adaptive learning systems and intelligent tutoring systems are now enabling educators to create more sophisticated and personalized BL environments through advanced algorithmic approaches. These advanced capabilities are offered through machine learning-driven content recommendation engines, dynamic assessment adaptation mechanisms, and intelligent learner support systems. In doing so, AI demonstrated remarkable pedagogical personalization capabilities that were hardly realizable at the scale before (Kumar et al., 2024).

Nevertheless, while AI-enhanced BL holds promise in fostering personalization and enhancing students’ learning outcomes, research is needed to fully understand its potential impact (Kumar et al., 2024). A significant challenge in evaluating BL practices has been the lack of consistent, comprehensive evaluation criteria (Yan and Chen, 2021; Zhang et al., 2022). Han (2023) has highlighted the importance of considering multiple factors when assessing BL’s impact on student learning outcomes. More robust literature synthesis is crucial to avoid inflating outcomes or obfuscating results. The present study addresses this gap through a comprehensive meta-analysis that considers a wide range of variables, including AI technology type, intervention duration, research design, sample size, and educational contexts across different grade levels and subject areas. The findings will discuss effective integration strategies to help educators make informed, equitable decisions about AI-enhanced BL solutions.

2 Literature review

2.1 Blended learning and AI impact

While no standard definition of impact exists, this study defines impact as the measurable outcomes of a student’s achievement in both the online and in-person components of BL (Spanjers et al., 2015). Positively influencing students’ academic achievement is a fundamental objective shared by educators, institutions, and societies at large. Nevertheless, research on the impact of BL has thus far shown mixed results, emphasizing the need to understand what makes teaching approaches effective and how they work in different contexts.

Since the early 2000s, particularly in higher education in the United States, there has been evidence of a positive impact when comparing BL approaches to traditional learning methods. Numerous studies have documented improvements across various educational dimensions. Researchers have linked BL to enhanced student achievement and increased motivation (Vaughan, 2014), better support mechanisms (Lim et al., 2019), improved access to comprehensive learning materials (Kim et al., 2014), and greater overall engagement and achievement (Owston et al., 2013; Bernard et al., 2014; Halverson et al., 2014; Spanjers et al., 2015; Boelens et al., 2017; Asarta and Schmidt, 2020; Rasheed et al., 2020). This trend extends across both higher and secondary education, offering personalized learning experiences and broader educational opportunities (Picciano, 2012; Hilliard, 2015).

However, other studies have found no significant contribution of BL to student achievement and test scores compared to traditional learning environments (Means et al., 2013, Müller et al., 2018). These conflicting findings can be attributed to a range of complex and interconnected factors. BL design features, including technology quality, online tools, and face-to-face support, interact closely with student characteristics such as technological proficiency (Alfadda and Mahdi, 2021), individual attitudes, and self-regulation capabilities. Contextual factors further complicate the assessment, with instructors’ expertise, subject matter, and specific course goals playing crucial roles in determining learning outcomes.

With the emergence of AI, many researchers sought to explore its potential and embed its capabilities in different applications. Due to the recency and the rapid development of AI, researchers sought to answer the question of impact by exploring the potential of AI to enhance BL environments. A systematic review by Park and Doo (2024) examining AI applications in BL from January 2007 to October 2023 revealed several promising developments in educational technology. AI-powered tools and platforms may create flexible and personalized learning experiences. These AI technologies demonstrate remarkable capabilities in tailoring content to individual needs, providing personalized instruction and scaffolding (Liao and Wu, 2022; Phillips et al., 2020), adapting assessments to student achievement, and offering personalized guidance and feedback (Jovanović et al., 2017; Liao and Wu, 2022). Nevertheless, despite these promising developments, the research on AI’s impact on BL remains inconclusive. While some studies show potential benefits, AI integration does not consistently lead to improved learning outcomes across different educational levels. Challenges persist, including data dependency, the ongoing need for human expertise, and the lack of standardized measurement tools. Future investigations should focus on systematically measuring AI’s effect on student learning achievement and understanding the intricate interplay of factors influencing its effectiveness.

2.2 Previous meta-analyses on blended learning

To address the challenges of comprehensively evaluating the integration of BL and AI, which impedes systematic understanding of their effectiveness and impact on educational outcomes, meta-analysis offers a powerful analytical approach. Meta-analyses systematically aggregate data from multiple studies, providing valuable insights that inform the development of new or revised BL criteria and contribute to establishing more robust and standardized evaluation methods (Cohn and Becker, 2003).

Unlike individual studies that may yield varying results due to differences in methodology, sample size, or other contextual factors, meta-analysis can reconcile these discrepancies by aggregating data, thereby increasing sample size and statistical power. This methodological approach makes it more likely to detect significant effects and help identify moderating factors that influence BL effectiveness, such as the specific blend of face-to-face and online instruction, student characteristics, or instructor expertise.

Several meta-analyses have examined the effectiveness of BL. While earlier research supported the perspective that BL can result in better learning outcomes for higher education students (Means et al., 2010, 2013; Bernard et al., 2014), recent meta-analyses have yielded mixed results, highlighting the complexities of implementing BL in various educational settings. Vo et al. (2017) conducted a cross-regional meta-analysis with 51 studies in higher education and they found that the effect of BL on student achievement is small.

A more recent study by Yu et al. (2022) analyzed 30 peer-reviewed studies encompassing 70 effect sizes, exploring BL’s impact on student outcomes and attitudes. The findings indicated that BL significantly outperforms traditional instruction, with a medium effect size. Students in BL environments demonstrated higher academic achievement and more positive learning attitudes. However, the researchers emphasized that BL effectiveness varies depending on the implementation model, student characteristics, and the quality of instructional materials and technologies.

Another significant meta-analysis by Xu L. et al. (2022) and Xu Z. et al. (2022) explored the potential of self-regulated learning (SRL) interventions in online and BL environments. This research examined the efficacy of self-regulated learning on academic achievement through the moderators of article type, subject type, learning context, educational level, SRL strategy, SRL phase, SRL scaffolds, and intensity and duration of intervention. The study revealed a moderate effect of SRL intervention on academic achievement in online and blended environments.

Cao's (2023) meta-analysis provided an additional perspective by evaluating BL effectiveness across different countries. While BL generally demonstrated a positive moderate impact on student achievement, attitudes, and achievement, engagement levels varied significantly.

Despite the limited number of meta-analyses on blended learning, the findings from these studies vary to some extent. While the effect size reported by Vo et al. (2017) is small, it was found to be medium in Yu et al. (2022), Xu L. et al. (2022) and Xu Z. et al. (2022). Additionally, the focus of each meta-analysis differed. For example, Cao (2023) concentrated on comparing effect sizes across several countries, while Vo et al. (2017) focused solely on higher education, and Xu L. et al. (2022) and Xu Z. et al. (2022) centered on self-regulated learning as the main factor. It is clear that the initial meta-analyses identified several gaps to be explored, such as the various applications of AI in BL and the different educational levels. Additionally, the existing meta-analyses further highlight the complexity of BL on learning achievement and the need for context-specific strategies, particularly when integrating AI. They emphasize the importance of ongoing research to understand the many factors that contribute to creating effective BL environments.

3 Research gap and study objectives

As mentioned in the previous section, while several meta-analyses have examined the effectiveness of BL, no research, to the best of the authors’ knowledge, has specifically investigated the impact of AI on student achievement within BL environments. However, while there are some systematic reviews of AI on blended learning (Park and Doo, 2024; Al-Maroof et al., 2022), all of them were qualitative and did not provide quantitative evidence on how AI would impact students’ learning achievement in blended learning. Therefore, more rigorous quantitative studies are needed to establish a definitive causal link between AI and improved students’ learning achievement. In other words, fragmented evidence was found related to the impact of AI on learning achievement in blended learning. Thus, to better understand this effect, a meta-analysis and synthesis is needed. Meta-analyses are effective in this context as they synthesize results from multiple studies and sources to provide an overall effect size, offering a more comprehensive understanding of AI effect. To address this research gap, the present study quantitatively measures the effect of AI on learning achievement through a systematic review and meta-analysis.

In addition, it is seen that the effect of AI in education is moderated by several variables, including educational subject and level (Vo et al., 2017), intervention duration, geographical distribution, and learning context (Xu L. et al., 2022; Xu Z. et al., 2022). Therefore, to fully understand the effectiveness of AI in BL environments, it is essential to consider various factors, such as the specific implementation of the BL model, the characteristics of the students, the used instruction and technologies, and the broader educational context. Thus, the impact of AI in BL needs to be further unpacked to better understand its effect on learning achievement. To do so, the present study takes one step forward and investigates what might moderate the effect of AI in blended learning. Specifically, it explores the impact of its integration whose magnitude is influenced by a wide range of critical variables. Therefore, the research question is: ‘To what extent do AI-enhanced BL environments improve student learning achievement, and what critical variables influence the magnitude of these effects?’

To address this research question, the present study identifies certain factors that were not adequately explored in previous research and must be prioritized in this meta-analysis to contribute to a more comprehensive understanding. It expands the scope to examine BL by considering specific types of applications across various grade levels and different subject areas. Additionally, this study focuses on specific factors such as the intervention duration, the research design, and the sample size. Accordingly, the following two main research questions were further developed and guided this research:

RQ1. What is the overall effect of AI on student learning achievement in blended learning environments?

RQ2. How does the effectiveness of AI in blended learning vary across different moderators, including grade level, educational subject, instruction duration, study design, and sample size?

The findings of this study can contribute to the ongoing debate related to the effectiveness of AI in education generally and in blended learning particularly. It advances understanding of AI’s role in pedagogical frameworks and the different variables to be considered when developing AI-based interventions.

4 Methodology

4.1 Search and data retrieval

A search was performed in the electronic databases of Science Direct, IEEE Xplore, Taylor & Francis, Scopus, and Web of Science, as these databases are familiar in the field of AI and blended learning, and include several of the most important journals. The search strings were adapted from several AI and blended learning reviews in the literature (e.g., Park and Doo, 2024), and are as follows: (Artificial intelligence substring) AND (blended learning) AND (education substring), where:

Artificial intelligence substring: “artificial intelligence” OR AI OR “machine intelligence” “OR “machine learning” OR “natural language processing” OR “deep learning” OR robotic.

Blended learning substring: “blended learning” OR “hybrid learning” OR “flipped learning” OR “integrated learning” OR “multi-method learning.”

Education substring: “learning achievement” OR “learning performance” OR “academic achievement” OR “academic performance.”

The included articles were peer-reviewed to ensure the quality of the meta-analysis. Moreover, the search period was set starting from 2011, since this year was considered as the year where AI applications became more mature and AI assisted technology was booming (Wang et al., 2023).

The last search was performed on June 01, 2024, and the overall process yielded 538 potential studies. A 307 potential studies were removed due to duplication, then the inclusion/exclusion criteria were applied. A study was included if it: (1) was in English; (2) was an empirical research; (3) used AI for blended learning; (4) was not qualitative or review research since these studies do not have the needed statistical data (e.g., mean effect sizes; standard errors and confidence intervals; and the samples sizes) to conduct a meta-analysis; (5) provided sufficient information (e.g., mean, and standard error) to compute the effect size; or (6) included a control condition. It should be noted that no study was excluded based on the used AI application and the obtained set of AI applications in this study were the result of the aforementioned inclusion/exclusion criteria. Finally, 21 studies (2,873 participants in total) were considered for this meta-analysis. There were 16 studies followed true experimental design and 5 studies followed quasi experimental design. Figure 1 presents the data selection process.

Figure 1
Flowchart of the selection process for articles. Identification: 538 records identified, 307 duplicates removed. Screening: 231 records screened, 116 excluded. Eligibility: 115 full-text articles assessed, 94 excluded due to insufficient information or missing control condition. Inclusion: 21 articles included in synthesis.

Figure 1. PRISMA chart.

4.2 Meta-analysis

The software of Comprehensive Meta-Analysis V.4 was used to conduct this meta-analysis. Hedges’ g was used to calculate the effect sizes (Hedges, 1981). The motivation behind using Hedges’ g instead of Cohen’s d effect size is to provide a less biased estimate of the effect size (Tlili et al., 2023). Hedges’ g incorporates a small sample correction factor that reduces the upward bias that can occur in Cohen’s d with limited data (Hedges and Olkin, 1985). Nineteen studies followed the pretest–posttest-control (PPC) research design. In the current research design, students are assigned to experimental and control interventions and are evaluated before and after the intervention (i.e., the learning process). As stated by Morris (2008), the PPC design provides reliable and precise values of effect sizes and minimizes the threats to internal validity. The other two papers followed the design of posttest only with control (POWC), where participants are assigned to experimental and control interventions and evaluated just after the intervention (i.e., learning process).

Four methods were used to assess publication bias. The first method is the trim-and-fill with the focus of identifying publication bias by means of a funnel plot wherein the papers are represented by dots. It is assumed that there is no publication bias when the dots are distributed on both sides of a vertical line representing the average effect size (Borenstein et al., 2010). The second method was Rosenthal’s (1979) fail-safe number which aims to bring the meta-analytic mean effect size down to a statistically insignificant level. A fail-safe number larger than 5 k + 10 (where k is the original number of studies included in the meta-analysis) is robust. It suggests that publication bias is unlikely to significantly affect the overall results (Borenstein et al., 2021). The third method was Egger’s regression test where a significant intercept suggests publication bias. The fourth method was p-curve analysis, which assesses whether the distribution of statistically significant p-values in the included studies demonstrates evidential value or is indicative of selective reporting practices (Simonsohn et al., 2014).

5 Results

5.1 Effect of AI in blended learning

The overall pooled effect size of the 21 studies shows a medium effect size (Hedges’ g = 0.50), with a 95% confidence interval of [0.27, 0.74] (Table 1). The z-value is 4.17, and the effect is statistically significant (p = 0.001). There was substantial variability across the included studies manifested in high heterogeneity values (I2 = 88.97, τ2 = 0.25). As such the prediction interval (Figure 2) ranged from −0.57 to 1.57 indicating that future replication of the positive results of the meta-analysis may be unwarranted.

Table 1
www.frontiersin.org

Table 1. Effect of AI on blended learning environment.

Figure 2
Forest plot displaying the results of multiple studies comparing control (Not AI) and treatment (AI) using Hedges's g. Each study lists Hedges's g, a 95% confidence interval, and weight. The summary effect is indicated by a red diamond, with a prediction interval below it. The heterogeneity is marked as I squared equals eighty-nine percent, tau squared equals 0.2467, p is less than 0.01.

Figure 2. Forest plot.

Looking at the differences by the type of AI implemented (Table 1), it is evident that studies that used a personalized AI system (k = 4) reported a large effect size (Hedges’ g = 0.88), which was statistically significant (p = 0.001, 95% CI [0.48, 1.27]). The heterogeneity was moderate (I2 = 60.89, τ2 = 0.10) and lower compared to other categories, indicating more consistent results for the personalized approach. Intelligent Tutoring Systems (ITS) (k = 16) studies reported a small-to-medium and statistically significant effect size (Hedges’ g = 0.42, 95% CI [0.15, 0.69], p = 0.002), although heterogeneity was high (I2 = 90.19, τ2 = 0.24). Lastly, only one study used a chatbot system and reported a small effect size (Hedges’ g = 0.34), which was non-significant (95% CI [−0.17, 0.84], p = 0.188). The I2 statistic showed that 88.97% of variance resulted from between-study factors, implying that other variables might moderate the effect size of AI in blended learning.

5.2 Moderating effect of grade level

There were marked variations in grade levels (Table 2) where the highest effect sizes were obtained in higher education and the lowest in early childhood studies. In that, studies in higher education (k = 9) showed a large effect size (Hedges’ g = 0.7, 95% CI [0.03, 1.02], p = 0.001), although with considerable heterogeneity (I2 = 70.41%, τ2 = 0.15). Primary education (k = 2) followed, with a medium effect size (Hedge’s g = 0.55), although the results are not statistically significant (95% CI [−0.14, 1.25], p = 0.12) and showed considerable heterogeneity. Secondary education studies (k = 9) yielded a small-to-medium effect size (Hedges’ g = 0.40) which was statistically significant, although the studies presented very high heterogeneity (I2 = 93.07%, τ2 = 0.26). Lastly, the analysis of early childhood education included only one study which reported a small negative effect size (Hedges’ g = −0.31). The negative effect was statistically significant (95% [−0.58, −0.04], p = 0.02).

Table 2
www.frontiersin.org

Table 2. Effect of AI on blended across various grade levels on learning achievement.

5.3 Moderating effect of educational subject

The meta-analysis sheds light on the varied impact of AI-enhanced blended learning across educational subject areas (Table 3). Notably, in teacher training, AI demonstrated a substantial positive effect, with a very large effect size (Hedges’s g = 0.99), although only one study was conducted in this subject. Similarly, a large effect size was observed in the five studies in ICT and Engineering (Hedges’s g = 0.88). However, the results in other subjects were not decisive. In Languages, although there was a moderate effect size of 0.56, the lack of statistical significance and high variability among studies (I2 = 95.26%) suggest unclear and inconsistent results. Mathematics presented a modest outcome, with a negligible effect size of 0.11. In Science and Social Sciences, AI had a moderate effect size of 0.44 with statistically insignificant effect size and moderate heterogeneity (I2 = 67.75%).

Table 3
www.frontiersin.org

Table 3. Effect of AI on blended learning across various educational subject areas on learning achievement.

5.4 Moderating effect of instruction duration

The effectiveness of AI-enhanced blended learning varies significantly depending on the duration of the instruction (Table 4). The strongest impact was seen in short-term duration (k = 3, 1 week to 1 month), with a very large effect size of 1.14, indicating AI’s substantial immediate influence. This effect was consistent across studies. In longer durations (k = 14, between 1 month and 1 semester), the effect size was moderate 0.53, significant but with very high variability across studies (I2 = 90.97). In long-term durations (k = 4, 1 semester to 1 year), the effect size was almost negligible and non-significant 0.08. These variations may be explained—at least partially—by the immediacy effect where closer time to assessment may result in a more positive effect as students are more likely to retain what they have learned.

Table 4
www.frontiersin.org

Table 4. Effect of AI on blended with different instruction durations on learning achievement.

5.5 Moderating effect of study design

Regarding study design, studies using an experimental design reported a statistically significant medium effect size of 0.52 (Table 5). This effect was inconsistent across the studies giving rise to high levels of heterogeneity (I2 = 91.35%). In contrast, studies employing a quasi-experimental design reported a slightly lower and statistically significant effect size of 0.42. Yet, the effect was more consistent across the studies with heterogeneity measures of (I2 = 49.16%).

Table 5
www.frontiersin.org

Table 5. Effect of AI on blended learning in different study designs on learning achievement.

5.6 Moderating effect of sample size

Table 6 reveals that studies with smaller sample sizes (≤250 participants) showed a moderate effect size of 0.57, which was statistically significant (p = 0.001). On the other hand, studies with larger sample sizes (>250 participants) showed a much smaller effect size of 0.24, which was not statistically significant (p = 0.121). Both groups show a very high heterogeneity (I2 > 80%).

Table 6
www.frontiersin.org

Table 6. Effect of AI on blended learning in different sample size on learning achievement.

5.7 Meta-regression

By and large, the results of meta-regression (Table 7) reflect the results of the subgroup analysis. Regarding the grade level, the analysis shows that AI is less effective in Early Childhood (EC) and Primary Education (PE), with both having negative coefficients of −0.45. The impact on the overall effect size was statistically significant (p = 0.04) in Early Childhood, though negative. In contrast, Secondary Education (SE) shows a significant positive coefficient of 0.87 (p = 0.001), suggesting that AI interventions may be more likely to be effective for older students. Regarding educational subjects, the field of teacher training shows a positive coefficient of 0.80, with marginal significance (p = 0.05). On the contrary, a significant negative impact was reported in Mathematics, with a coefficient of −0.38 (p = 0.001). This suggests that AI applications or tools may be less ready for Mathematics education. Regarding instruction durations, shorter durations (1 week to 1 month) show a non-significant positive effect (coefficient = 0.14), implying that the benefits of AI are more apparent in the short term. In contrast, the analysis shows that longer durations (1 semester to 1 year) are associated with a significantly negative coefficient of −1.28 (p = 0.001), suggesting that the positive effects of AI be lost or—better said—not retained over time. Put another way, AI might show strong initial results, but sustaining these benefits over longer periods can be challenging—which is not only specific to AI education. Regarding research design, compared to experimental design, quasi-experimental designs show a significantly lower effect size, as indicated by a negative coefficient of −0.58 (p = 0.02). Similarly, compared with studies with smaller sizes, studies with larger sample sizes (>250 participants) have a significantly negative coefficient of −1.15 (p = 0.001).

Table 7
www.frontiersin.org

Table 7. Meta-regression results for the learning achievement of response from grade level, educational subject, instruction duration, study design and sample size.

5.8 Publication bias

The funnel plot (Figure 3) displays the effect sizes on the x-axis against standard errors on the y-axis, with the larger studies (smaller standard errors) clustering near the top and the smaller studies (larger standard errors) spreading out towards the bottom. Further, the plot shows a slight asymmetry, with more studies skewed to the right, especially among the smaller studies at the bottom, suggesting that smaller studies reporting larger positive effects are overrepresented.

Figure 3
Funnel plot showing standard error on the y-axis against estimated effect size on the x-axis. Points are scattered symmetrically around the vertical line at zero, forming an inverted funnel shape bounded by dashed lines.

Figure 3. Funnel plot.

These results are confirmed by a Rank Correlation Test for Funnel Plot Asymmetry which was statistically significant and positive (Kendall’s tau: 0.4571, p-value: 0.0033) suggesting the presence of publication bias. Furthermore, a Regression Test for Funnel Plot Asymmetry was also statistically significant with a z-value of 3.3348 and a p-value of 0.0009. The estimated effect size as the standard error approaches zero was b = −0.2091, with a confidence interval of −0.6749 to 0.2567, indicating potential publication bias where smaller studies may report larger effect sizes.

To assess potential publication bias, Duval and Tweedie’s (2000) Trim-and-Fill method using the L-estimator to correct for funnel plot asymmetry was applied. The analysis was conducted in two steps: (1) identifying and removing extreme outliers, and (2) applying Trim-and-Fill to estimate and adjust for missing studies. To reduce the influence of extreme values, six outliers were identified and removed from the dataset, reducing the number of studies from k = 24 to k = 18. Studies are defined as outliers when their 95% confidence interval lies outside the 95% confidence interval of the pooled effect. Following outlier removal, Trim-and-Fill imputed three potentially missing studies on the left side of the funnel plot. After removing six outliers, the Trim-and-Fill adjusted estimate increased to g = 0.3181, with a narrower confidence interval [0.1672, 0.4690], which remained statistically significant (p < 0.0001). The funnel plot became more symmetric after outlier removal, suggesting that the initial asymmetry was at least partially driven by extreme values rather than systematic publication bias.

On the other hand, the p-curve analysis (Figure A1) strongly indicates that the analyzed studies have substantial evidential value, are likely to reflect true effects, and are not just a result of selective reporting or p-hacking. The extremely low p-values for the right-skewness test and high-power estimate further reinforce this conclusion.

6 Discussion and conclusions

Discussions around AI are often accompanied by words like “potential,” “promise” and “hope.” While optimism is needed, it often risks overshadowing the need for tangible, measurable outcomes. It is important to not forget that AI is no stranger to hype and in fact, it has known several waves of hype and winters before. Such optimism will only be justified if the technology begins delivering real-world impact that addresses current challenges. Therefore, the present study aimed to examine in a systematic way what AI has achieved in terms of impact on students’ outcomes in BL environments through a meta-analysis of 21 studies.

6.1 Effect of AI on students’ learning achievement in blended learning

The present study revealed that AI has an overall medium effect size (g = 0.50) on learning achievement in blended learning, which is higher than the small effect found by Vo et al. (2017) when investigating the effect of blended learning without the AI component on students’ learning achievement. This implies that AI has improved students’ learning achievement in blended learning. This improvement could be attributed to the various features —and advances— provided by AI applications that can play the same role as teachers or instructors, students, or peers in blended learning (Park and Doo, 2024). AI also supports blended learning by providing personalized learning experiences and optimizing course delivery (Lee et al., 2020). Specifically, personalized systems had the highest effect on blended learning, where it achieved a large effect (g = 0.88), which could be explained as personalized systems could free up the teachers from the routine and repetitive tasks of preparing lesson plans and grading answer sheets. The effect of ITS on blended learning was found to be medium (g = 0.42). This result could be attributed to the effectiveness of ITS in a blended environment, by enabling teachers to consider a variety of models for how they might integrate ITS into their lessons to accomplish usage and blended learning guidelines (Phillips et al., 2020). However, this decrease of ITS effect compared to other AI applications (e.g., personalized learning and chatbots) reveal the importance of investigating instructional approaches used within AI applications and the effective implementation of human-machine collaboration in education.

However, it is worth noting that the effectiveness of AI in blended learning was not very high, i.e., only medium effect size, raising questions about what might hinder AI in reaching its full potential when it comes to enhancing students’ learning achievement in blended learning. This calls for more investigation in this regard to unpack both how AI might be implemented in blended learning and what might hinder its effect.

6.2 Educational context moderates the AI effect size in blended learning

The study found that AI had the highest effect, large (g = 0.70), in higher education. This could be explained by students in higher education are using AI applications for improving their problem-solving skills, conceptual understanding, and overall learning outcomes, as well as to enhance their engagement, promoting personalized learning experiences, and revolutionizing assessment strategies (Shi et al., 2023), which is greatly increasing students’ motivation to learn (Wang and Jan, 2022). However, the findings revealed that AI is less effective in early educational levels. This might be because children do not understand how AI works and the knowledge behind it. Additionally, there is not much known about how teachers can enhance children’s learning with AI applications with a methodological and appropriate approach. Moreover, there is a shortage of research on AI education for children who have no prior knowledge of computer programming and robotics. In terms of educational subjects, the current study showed that AI on blended learning has a large effect size (g = 0.88) in ICT and engineering. This could be explained by the potential of AI to enhance individual projects, as well as creative problem-solving abilities using technological skills such as computational thinking and data-driven reasoning. In fact, AI-generated teaching practices have proven to increase learning outcomes among engineering students (Chiang, 2021). Such knowledge suggests that AI integration strategies should be tailored to specific educational contexts in blended learning, avoiding a one-size-fits-all approach.

6.3 Experiment type moderates the AI effect size in blended learning

The present study revealed that the AI conducted experiment moderates the effect size on learning achievement in blended learning. Specifically, the findings indicated that AI on blended learning has a very large effect size (g = 1.14) on short-term interventions (1 week to 1 month). This result could be explained as this duration is considered sufficient to promote learners’ familiarization and stimulate motivation for proper engagement and achievement (Merilampi et al., 2014). It is noted that the AI effect size in long-term interventions (1 semester to 1 year) was negligible (g = 0.08) and this could be clarified as students lose interest and motivation in a given technology as time passes (Tlili, 2024). Among methodologists, there was significant discussion about determining the appropriate intervention duration for technological implementations. For instance, Slavin (1986) suggested longer intervention duration when using a given technology to draw solid evidence about that technology effect beyond the temporary technology novelty effect. Others, on the other hand, highlighted concerns related to that longer intervention duration might lead to a dip in performance and confidence (Cung et al., 2019). Particularly, longer AI interventions might lead to a poorer implementation fidelity, hence causing a shift from the initial experiment goals and achieving low effects (Tlili et al., 2025; Wang et al., 2023).

In terms of experiment design, the present study revealed that AI on blended learning has medium effect sizes when it is used as quasi-experimental design (g = 0.42) and true experimental design (g = 0.52). This similarity between both experiment designs can be explained with both designs being based on testing and validating hypotheses and understanding AI in multiple contexts and subjects (Ofosu-Ampong, 2024). Additionally, most experimental studies improved learning through the use of AI applications. In terms of sample size, the current study pointed out that AI on blended learning has a medium size effect (g = 0.57) in smaller sample sizes (≤250 participants), while the effect size was small (g = 0.24) in larger sample sizes (>250 participants). These results could be attributed to the nature of small sample sizes, as personalized and interactive learning approaches can be more feasible, allowing for supportive feedback and individualized support. On the other hand, large sample sizes may require more collaborative learning platforms, which can affect the depth of engagement (Tlili et al., 2024).

6.4 Conclusion

Through a rigorous meta-analysis of 21 studies, this research provides compelling evidence for the positive impact of AI on student learning achievement in BL environments. The findings indicate that AI-powered tools and platforms can significantly enhance learning achievement. However, it is important to acknowledge that the effectiveness of AI in BL is influenced by several factors, including the specific AI technology, the instructional design, and the characteristics of the learners. To maximize the benefits of AI, it is crucial to carefully consider these factors and implement AI-enhanced learning strategies in a thoughtful and systematic manner.

Future research should explore the long-term impacts of AI on student learning, particularly in terms of critical thinking, problem-solving, and creativity, as well as the optimal integration of AI tools into various learning contexts. Moreover, study designs aimed at assessing learning outcomes (such as pre-posttest, experimental, or quasi-experimental settings) focus on taking snapshots of students’ learning before and after a specific intervention (e.g., using AI). Research that focuses on the process rather than only on the outcomes is needed to understand whether students are making use of AI in an effective way.

There is paucity of studies in underperforming contexts (e.g., primary education) and evidence from these areas is rather thin. Additionally, there is an ongoing tension about whether AI should be used in primary education or not. For instance, the President von der Leyen stated in the State of the European Union 2025 “I strongly believe that parents, not algorithms, should be raising our children.” Nevertheless, research is needed to examine such context given the potential benefits of AI, the evidence from other contexts and the fact that AI is expected to be an important player in everyday life. It is crucial to understand if and how AI can enhance learning in underperforming primary education settings, address challenges like teacher readiness, infrastructure, and safety. Research should address if and to what extent AI can enhance students’ engagement, support and personalization. Such research is needed to optimize learning and most importantly to inform evidence-based strategies for integrating AI effectively into the classrooms. The same applies to other areas and disciplines where AI has been rarely applied. However, extending AI to schools requires digital infrastructure, teacher training, and considering ethical considerations. Some schools lack stable internet access, adequate devices, and most importantly, the technical and practical expertise to effectively and safely integrate AI. Furthermore, AI must be implemented in ways that uphold data privacy, avoid biases that could further disadvantage certain student groups.

The present meta-analysis demonstrates AI’s potential to improve learning achievement in blended environments, particularly through personalized and adaptive functionalities. This potential is highly synergistic with inclusive education goals by enabling unprecedented levels of differentiation, accessibility support, and early intervention tailored to individual learner needs. It also highlights critical moderating variables that determine whether this potential AI efficacy translates into genuinely inclusive benefits. Ultimately, AI in blended learning can be a powerful lever for inclusive education, but only if its deployment is intentionally designed, critically evaluated, and continuously refined with equity and human agency at its core.

6.5 Future research

The agenda for future research on AI can be long given that AI is extending to all areas of research and practice with little research on its long-term impact. While the present study has shown primary evidence of the impact of AI in some contexts, research is needed to address areas where evidence is lacking like early school years. Most importantly, the knowledge about the long-term impact of AI is still limited, requiring longitudinal studies that track its effects over time. Research should explore how AI influences cognitive development, learning outcomes, and social interactions in young children. Furthermore, it is important to investigate how AI affects students’ motivation, dependence on technology and students’ soft skills. Future research should also investigate the unintended consequences of AI deployment, such as bias, misinformation, and environmental sustainability. Alignment between AI and students’ needs has never been more important to ensure that AI-driven educational solutions support rather than hinder student development. Unfortunately, research on alignment is lacking and is therefore badly needed.

Finally, blended learning has always meant blending online and face-to-face learning. With the emergence of AI, it may expand beyond this traditional definition, introducing a dynamic interaction between human teachers, students, and intelligent systems. However, this shift raises critical questions about the evolving role of teachers in AI-enhanced classrooms which calls for research on safe implementations as well as monitoring of impact.

6.6 Limitations

While the reliability of this meta-analysis has been investigated through several methods, including bias assessment, it still has several limitations that should be acknowledged. For instance, the findings of the present study might be limited by the search keywords or databases. Additionally, it covered only journal papers written in English. Therefore, future researchers could complement this study by including papers written in other languages as well as covering more electronic databases and search keywords.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author/s.

Author contributions

JW: Writing – original draft, Writing – review & editing. AT: Methodology, Writing – review & editing, Writing – original draft. SS: Formal analysis, Writing – original draft, Writing – review & editing. DM: Conceptualization, Writing – original draft, Writing – review & editing. MS: Formal analysis, Writing – original draft, Investigation, Software, Writing – review & editing. SL-P: Writing – original draft, Software, Visualization, Writing – review & editing, Formal analysis, Validation. RH: Project administration, Writing – original draft, Conceptualization, Writing – review & editing, Resources.

Funding

The author(s) declare that no financial support was received for the research and/or publication of this article.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

The handling editor FA declared a past co-authorship with the authors AT, SS, and RH.

The author(s) declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.

Generative AI statement

The authors declare that no Gen AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Alfadda, H. A., and Mahdi, H. S. (2021). Measuring students' use of zoom application in language course based on the technology acceptance model (TAM). J. Psycholinguist. Res. 50, 883–900. doi: 10.1007/s10936-020-09752-1

PubMed Abstract | Crossref Full Text | Google Scholar

Al-Maroof, R., Al-Qaysi, N., Salloum, S. A., and Al-Emran, M. (2022). Blended learning acceptance: a systematic review of information systems models. Technol. Knowl. Learn. 27, 891–926. doi: 10.1007/s10758-021-09519-0

Crossref Full Text | Google Scholar

Asarta, C., and Schmidt, J. R. (2020). The effects of online and blended experience on outcomes in a blended learning environment. Internet High. Educ. 44:100708. doi: 10.1016/j.iheduc.2019.100708

Crossref Full Text | Google Scholar

Bernard, R. M., Borokhovski, E., Schmid, R. F., Tamim, R. M., and Abrami, P. C. (2014). A meta-analysis of blended learning and technology use in higher education: from the general to the applied. J. Comput. High. Educ. 26, 87–122. doi: 10.1007/s12528-013-9077-3

Crossref Full Text | Google Scholar

Boelens, R., De Wever, B., and Voet, M. (2017). Four key challenges to the design of blended learning: a systematic literature review. Educ. Res. Rev. 22, 1–18. doi: 10.1016/j.edurev.2017.05.001

Crossref Full Text | Google Scholar

Borenstein, M., Hedges, L. V., Higgins, J. P., and Rothstein, H. R. (2010). A basic introduction to fixed-effect and random-effects models for meta-analysis. Res. Synth. Methods, 1, 97–111.

Google Scholar

Borenstein, M., Hedges, L. V., Higgins, J. P. T., and Rothstein, H. R. (2021). Introduction to meta-analysis (2nd ed.). John Wiley & Sons. doi: 10.1016/b978-0-12-209005-9.50005-9

Crossref Full Text | Google Scholar

Cao, W. (2023). A meta-analysis of effects of blended learning on performance, attitude, achievement, and engagement across different countries. Front. Psychol. 14:1212056. doi: 10.3389/fpsyg.2023.1212056

PubMed Abstract | Crossref Full Text | Google Scholar

Chiang, T. (2021). Estimating the artificial intelligence learning efficiency for civil engineer education: a case study in Taiwan. Sustainability 13:11910. doi: 10.3390/su132111910

Crossref Full Text | Google Scholar

Cohn, L. D., and Becker, B. J. (2003). How meta-analysis increases statistical power. Psychol. Methods 8, 243–253. doi: 10.1037/1082-989X.8.3.243

PubMed Abstract | Crossref Full Text | Google Scholar

Cung, B., Xu, D., Eichhorn, S., and Warschauer, M. (2019). Getting academically underprepared students ready through college developmental education: does the course delivery format matter? Am. J. Dist. Educ. 33, 178–194. doi: 10.1080/08923647.2019.1582404

Crossref Full Text | Google Scholar

Duval, S., and Tweedie, R. (2000). Trim and fill: A simple funnel-plot-based method of testing and adjusting for publication bias in meta-analysis. Biometrics, 56, 455–463. doi: 10.1111/j.0006-341x.2000.00455.x

Crossref Full Text | Google Scholar

Garrison, D. R., and Kanuka, H. (2004). Blended learning: uncovering its transformative potential in higher education. Internet High. Educ. 7, 95–105. doi: 10.1016/j.iheduc.2004.06.001

Crossref Full Text | Google Scholar

Graham, C., and Dziuban, C. (2007). Blended learning environments. In J. M. Spector, M. D. Merrill, J. J. G. Merrienboervan, and M. P. Driscoll (Eds.), Handbook of research on educational communications and technology (4th ed., pp. 1–22). Mahwah: Lawrence Erlbaum Associates.

Google Scholar

Halverson, L. R., Graham, C. R., Spring, K. J., Drysdale, J. S., and Henrie, C. R. (2014). A thematic analysis of the most highly cited scholarship in the first decade of blended learning research. Internet High. Educ. 20, 20–34. doi: 10.1016/j.iheduc.2013.10.004

Crossref Full Text | Google Scholar

Han, X. (2023). Evaluating blended learning effectiveness: an empirical study from undergraduates' perspectives using structural equation modeling. Front. Psychol. 14:1059282. doi: 10.3389/fpsyg.2023.1059282

PubMed Abstract | Crossref Full Text | Google Scholar

Hedges, L., and Olkin, I. (1985). Statistical methods for meta-analysis. Academic Press.

Google Scholar

Hedges, L. V. (1981). Distribution theory for Glass’s estimator of effect size and related estimators. J. Educ. Stat, 6, 107–128. doi: 10.3102/10769986006002107

Crossref Full Text | Google Scholar

Hilliard, A. T. (2015). Global blended learning practices for teaching and learning, leadership and professional development. J. Int. Educ. Res. 11, 179–188. doi: 10.19030/jier.v11i3.9366

Crossref Full Text | Google Scholar

Holmes, W., Bialik, M., and Fadel, C. (2019). Artificial intelligence in education: promise and implications for teaching and learning. Boston: Center for Curriculum Redesign.

Google Scholar

Jovanović, J., Gašević, D., Dawson, S., Pardo, A., and Mirriahi, N. (2017). Learning analytics to unveil learning strategies in a flipped classroom. Internet High. Educ. 33, 74–85. doi: 10.1016/j.iheduc.2017.02.001

PubMed Abstract | Crossref Full Text | Google Scholar

Kim, M. K., Kim, S. M., Khera, O., and Getman, J. (2014). The experience of three flipped classrooms in an urban university: an exploration of design principles. Internet High. Educ. 22, 37–50. doi: 10.1016/j.iheduc.2014.04.002

Crossref Full Text | Google Scholar

Kumar, D., Mathur, M., Sarkar, A., and Chauhan, M. (2024). E-classroom as a blended learning tool: a structural equation modelling analysis using modified technology acceptance model. Cureus 16:e56925. doi: 10.7759/cureus.56925

PubMed Abstract | Crossref Full Text | Google Scholar

Lee, J. (2020). A study on application of artificial intelligence in the field of education: focusing on blended learning. J. Digit. Converg. 18, 429–436.

Google Scholar

Liao, C.-H., and Wu, J.-Y. (2022). Deploying multimodal learning analytics models to explore the impact of digital distraction and peer learning on student performance. Comput. Educ. 190:104599. doi: 10.1016/j.compedu.2022.104599

Crossref Full Text | Google Scholar

Lim, C. P., Wang, T., and Graham, C. (2019). Driving, sustaining and scaling up blended learning practices in higher education institutions: a proposed framework. Innov. Educ. 1:1. doi: 10.1186/s42862-019-0002-0

Crossref Full Text | Google Scholar

Means, B., Toyama, Y., Murphy, R., and Baki, M. (2013). The effectiveness of online and blended learning: a meta-analysis of the empirical literature. Teach. Coll. Rec. 115, 1–47. doi: 10.1177/016146811311500307

Crossref Full Text | Google Scholar

Means, B., Toyama, Y., Murphy, R., Baki, M., and Jones, K. (2010). Evaluation of evidence-based practices in online learning: a meta-analysis and review of online learning studies. Lancaster: Center for Technology in Learning.

Google Scholar

Merilampi, S., Sirkka, A., Leino, M., Koivisto, A., and Finn, E. (2014). Cognitive mobile games for memory impaired older adults. J. Assist. Technol. 8, 207–223. doi: 10.1108/JAT-09-2013-0027

Crossref Full Text | Google Scholar

Mizza, D., and Rubio, F. (2020). Creating effective blended language learning courses: a research-based guide from planning to evaluation. Cambridge: Cambridge University Press.

Google Scholar

Morris, S. B. (2008). Estimating effect sizes from pretest-posttest-control group designs. Organ. Res. Methods. 11, 364–386. doi: 10.1177/1094428106291059

Crossref Full Text | Google Scholar

Müller, C., Stahl, M., Alder, M., and Müller, M. (2018). Learning effectiveness and students' perceptions in a flexible learning course. Eur. J. Open Dist. E-Learn. 21, 44–53. doi: 10.21256/zhaw-3189

Crossref Full Text | Google Scholar

Ofosu-Ampong, K. (2024). Artificial intelligence research: a review on dominant themes, methods, frameworks and future research directions. Telemat. Inform. Rep. 14:100127. doi: 10.1016/j.teler.2024.100127

Crossref Full Text | Google Scholar

Owston, R., York, D., and Murtha, S. (2013). Student perceptions and achievement in a university blended learning strategic initiative. Internet High. Educ. 18, 38–46. doi: 10.1016/j.iheduc.2012.10.002

Crossref Full Text | Google Scholar

Park, Y., and Doo, M. Y. (2024). Role of AI in blended learning: a systematic literature review. Int. Rev. Res. Open Dist. Learn. 25, 164–196. doi: 10.19173/irrodl.v25i1.7372

Crossref Full Text | Google Scholar

Phillips, A., Pane, J. F., Reumann-Moore, R., and Shenbanjo, O. (2020). Implementing an adaptive intelligent tutoring system as an instructional supplement. Educ. Technol. Res. Dev. 68, 1409–1437. doi: 10.1007/s11423-020-09745-w

Crossref Full Text | Google Scholar

Picciano, A. G. (2012). The evolution of big data and learning analytics in American higher education. Online Learn. 16, 9–20. doi: 10.24059/olj.v16i3.267

Crossref Full Text | Google Scholar

Rasheed, A. R., Kamsin, A., and Abdullah, N. A. (2020). Challenges in the online component of blended learning: a systematic review. Comput. Educ. 144:103701. doi: 10.1016/j.compedu.2019.103701

Crossref Full Text | Google Scholar

Rosenthal, R. (1979). The file drawer problem and tolerance for null results. Psychological Bulletin, 86, 638–641.

Google Scholar

Shi, L., Umer, A. M., and Shi, Y. (2023). Utilizing AI models to optimize blended teaching effectiveness in college-level English education. Cogent Educ. 10:2282804. doi: 10.1080/2331186X.2023.2282804

Crossref Full Text | Google Scholar

Simonsohn, U., Nelson, L. D., and Simmons, J. P. (2014). p-curve and effect size: Correcting for publication bias using only significant results. Perspect. Psychol. Sci. 9, 666–681.

Google Scholar

Slavin, R. E. (1986). Best-evidence synthesis: an alternative to metaanalytic and traditional reviews. Educ. Res. 15:511.

Google Scholar

Spanjers, I. A., Könings, K. D., Leppink, J., Verstegen, D. M., de Jong, N., Czabanowska, K., et al. (2015). The promised land of blended learning: quizzes as a moderator. Educ. Res. Rev. 15, 59–74. doi: 10.1016/j.edurev.2015.05.001

Crossref Full Text | Google Scholar

Tlili, A. (2024). Can artificial intelligence (AI) help in computer science education? A meta-analysis approach [¿Puede ayudar la inteligencia artificial (IA) en la educación en ciencias de la computación? Un enfoque metaanalítico]. Rev. Esp. Pedagog. 82, 469–490. doi: 10.22550/2174-0909.4172

Crossref Full Text | Google Scholar

Tlili, A., Garzón, J., Salha, S., Huang, R., Xu, L., Burgos, D., et al. (2023). Are open educational resources (OER) and practices (OEP) effective in improving learning achievement? A meta-analysis and research synthesis. Int. J. Educ. Technol. High. Educ. 20:54. doi: 10.1186/s41239-023-00424-3

Crossref Full Text | Google Scholar

Tlili, A., Salha, S., Wang, H., Huang, R., Rudolph, J., and Weidong, R. (2024). Does personalization really help in improving learning achievement? A meta-analysis. IEEE Int. Conf. Adv. Learn. Technol. (ICALT) 2024, 13–17. doi: 10.1109/ICALT61570.2024.00011

Crossref Full Text | Google Scholar

Tlili, A., Saqer, K., Salha, S., and Huang, R. (2025). Investigating the effect of artificial intelligence in education (AIEd) on learning achievement: A meta-analysis and research synthesis. Inf. Dev. doi: 10.1177/02666669241304407

Crossref Full Text | Google Scholar

Vaughan, N. (2014). Student engagement and blended learning: making the assessment connection. Educ. Sci. 4, 247–264. doi: 10.3390/educsci4040247

Crossref Full Text | Google Scholar

Vo, H. M., Zhu, C., and Diep, N. A. (2017). The effect of blended learning on student performance at course-level in higher education: a meta-analysis. Stud. Educ. Eval. 53, 17–28. doi: 10.1016/j.stueduc.2017.01.002

Crossref Full Text | Google Scholar

Wang, H., and Jan, N. (2022). A survey of multimedia-assisted English classroom teaching based on statistical analysis. J. Math. 2022, 1–11. doi: 10.1155/2022/4458478

Crossref Full Text | Google Scholar

Wang, H., Tlili, A., Huang, R., Cai, Z., Li, M., Cheng, Z., et al. (2023). Examining the applications of intelligent tutoring systems in real educational contexts: A systematic literature review from the social experiment perspective. Educ. Inf. Technol. 28, 9113–9148. doi: 10.1007/s10639-022-11555-x

Crossref Full Text | Google Scholar

Wanner, T., and Palmer, E. (2015). Personalising learning: exploring student and teacher perceptions about flexible learning and assessment in a flipped university course. Comput. Educ. 88, 354–369. doi: 10.1016/j.compedu.2015.06.006

Crossref Full Text | Google Scholar

Watson, J. (2008). Blending learning: the convergence of online and face-to-face education. Vienna: North American Council for Online Learning.

Google Scholar

Xu, L., Duan, P., Padua, S. A., and Li, C. (2022). The impact of self-regulated learning strategies on academic performance for online learning during COVID-19. Front. Psychol. 13:1047680. doi: 10.3389/fpsyg.2022.1047680

PubMed Abstract | Crossref Full Text | Google Scholar

Xu, Z., Zhao, Y., Zhang, B., Liew, J., and Kogut, A. (2022). A meta-analysis of the efficacy of self-regulated learning interventions on academic achievement in online and blended environments in K-12 and higher education. Behav. Inf. Technol. 42, 2911–2931. doi: 10.1080/0144929X.2022.2151935

Crossref Full Text | Google Scholar

Yan, Y., and Chen, H. (2021). Developments and emerging trends of blended learning: a document co-citation analysis (2003-2020). Int. J. Emerg. Technol. Learn. (iJET) 16, 149–164. doi: 10.3991/ijet.v16i24.25971

Crossref Full Text | Google Scholar

Yu, Z., Xu, W., and Sukjairungwattana, P. (2022). Meta-analyses of differences in blended and traditional learning outcomes and students' attitudes. Front. Psychol. 13:926947. doi: 10.3389/fpsyg.2022.926947

PubMed Abstract | Crossref Full Text | Google Scholar

Zhang, Q., Zhang, M., and Yang, C. (2022). Situation, challenges and suggestions of college teachers' blended teaching readiness. E-Edu. Res. 12, 46–53. doi: 10.13811/j.cnki.eer.2022.01.006

Crossref Full Text | Google Scholar

Appendix

Figure A1
Line graph depicting the p-curve of test results versus p-values. The blue line represents the observed p-curve, showing a steep drop from 80% at 0.01 to 10% at 0.02 and 0.03, declining to 0% at 0.04 and 0.05. The dashed green line indicates the null of 33% power, while the dotted red line shows the null of no effect. A legend provides details about power estimates and test results for right-skewness and flatness.

Figure A1. P-curve analysis.

Keywords: blended learning, hybrid learning, artificial intelligence, smart learning, learning achievement, learning performance, meta-analysis, quantitative evidence

Citation: Wu J, Tlili A, Salha S, Mizza D, Saqr M, López-Pernas S and Huang R (2025) Unlocking the potential of artificial intelligence in improving learning achievement in blended learning: a meta-analysis. Front. Psychol. 16:1691414. doi: 10.3389/fpsyg.2025.1691414

Received: 23 August 2025; Accepted: 08 October 2025;
Published: 31 October 2025.

Edited by:

Fahriye Altinay, Near East University, Cyprus

Reviewed by:

Erinc Ercag, University of Kyrenia, Cyprus
Deepika Kohli, Khalsa College of Education, India

Copyright © 2025 Wu, Tlili, Salha, Mizza, Saqr, López-Pernas and Huang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Jiajun Wu, bTIzMDkyMjAwMDE3QGNpdHl1LmVkdS5tbw==; Ahmed Tlili, YWhtZWQudGxpbGkyM0B5YWhvby5jb20=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.