Skip to main content

SYSTEMATIC REVIEW article

Front. Psychol., 19 September 2024
Sec. Cognitive Science

Which ERP components are effective in measuring cognitive load in multimedia learning? A meta-analysis based on relevant studies

Shuyu YuShuyu Yu1Lianghao TianLianghao Tian2Guohua Wang
Guohua Wang2*Shengxin NieShengxin Nie2
  • 1School of Educational Technology, Northwest Normal University, Lanzhou, China
  • 2School of Educational Technology, Faculty of Education, Henan Normal University, Xinxiang, China

The open and generative nature of multimedia learning environments tends to cause cognitive overload in learners, and cognitive load is difficult for researchers to observe objectively because of its implicit and complex nature. Event-related potentials (ERP), a method of studying potential changes associated with specific events or stimuli by recording the electroencephalogram (EEG), has become an important method of measuring cognitive load in cognitive psychology. Although many studies have relied on ERP output measurements to compare different levels of cognitive load in multimedia learning, the results of the effect of cognitive load on ERP have been inconsistent. In this study, we used a meta-analysis of evidence-based research to quantitatively analyze 17 experimental studies to quantitatively evaluate which ERP component (amplitude) is most sensitive to cognitive load. Forty five effect sizes from 26 studies involving 360 participants were calculated. (1) The results of the studies analyzed in subgroups indicated high level effect sizes for P300 and P200 (2) Analyses of moderating variables for signal acquisition did not find that different methods of signal acquisition had a significant effect on the measurement of cognitive load (3) Analyses of moderating variables for task design found that a task system with feedback was more convenient for the measurement of cognitive load, and that designing for 3 levels of cognitive load was more convenient for the measurement of cognitive load than for 2 levels of cognitive load. (4) Analyses of continuous moderating variables for subject characteristics did not find significant effects of age, gender, or sample size on the results.

1 Introduction

Multimedia learning has evolved with the proliferation of educational technology applications and increased opportunities to create multi-channel learning environments (Farkish et al., 2023; Liu et al., 2018; Mayer, 2006). The abundance of verbal and graphic information presented in teaching and learning through various technologies imposes a more complex and challenging cognitive load on learners. Educational researchers must employ advanced designs and techniques to assess cognitive load effectively (Çeken and Taşkın, 2022). Event-related potential (ERP) technology is a measurement technique developed based on EEG, which is able to capture the learners’ EEG activity within a specific time window during the learning process, thus providing a time-series interpretation of the dynamic changes in the cognitive load (Kramer, 1991). However, related studies face certain challenges. Firstly, researchers used different ERP components and measured various brain regions. Secondly, exploration of cognitive load measurement under specific factors and technical conditions has led to variability in findings. Therefore, this study conducted a quantitative review using meta-analysis to integrate existing literature findings, quantitatively assess the validity of ERP components as a measure, and analyze the factors that May affect the experimental results.

2 Related work

2.1 Cognitive load and ERP technology

Multimedia learning environments offer learners a wealth of information and diverse learning experiences, accompanied by complex cognitive load challenges (Cavanagh and Kiersch, 2023; Mutlu-Bayraktar et al., 2019). Learners often grapple with information overload across various perceptual channels and cognitive dimensions when engaging with multimedia content that incorporates multiple elements such as text, images, sound, and interaction. This complexity necessitates the learner’s cognitive system to not only process increased information but also integrate and interact across different media, thereby intensifying the difficulty of the learning task (Trypke et al., 2023). In response to this challenge, event-related potentials (ERPs) emerge as a particularly advantageous physiological method for directly measuring cognitive load. By recording EEG signals, ERPs can capture a learner’s neuroelectrical activity within specific time windows during the learning process, providing a time-series interpretation of dynamic changes in cognitive load (Ghani et al., 2020). The superiority of ERP in assessing cognitive load within multimedia learning environments is evident in several aspects. Firstly, ERP boasts high temporal resolution, enabling the tracking of rapid brain responses to various stimuli and revealing the temporal characteristics of learners’ processing of multimedia information. It accurately quantifies the distribution and temporal dynamics of cognitive load in learning tasks. Secondly, as a direct physiological measurement, ERP circumvents the limitations of subjective evaluation methods (Anmarkrud et al., 2019). Moreover, ERP proves sensitive to neural responses at varying levels of cognitive load, capturing changes in the brain at the millisecond level (Sun et al., 2022). This positions ERP as an ideal tool for investigating learners’ cognitive load in multimedia learning environments, contributing to a comprehensive understanding of the cognitive challenges in the learning process and providing objective physiological guidance for optimizing multimedia learning design.

2.2 ERP components

ERP, in contrast to frequency domain analysis of EEG, involves averaging signals across multiple EEG channels while minimizing or eliminating unwanted EEG activity within a specified time window. The ERP divides components according to their latency after stimulus onset, and is typically named according to their deflection direction and average expected latency (Kramer, 1991). However, the latency of ERP components is influenced by various factors, such as stimulus properties, task requirements, and individual differences. The length of latency does not necessarily directly reflect the cognitive significance of the component. Consequently, in addition to components identified within specific time windows (e.g., P3a, N2b), many ERP components are named based on their functions and features. Examples include RON (Readiness Potential), emphasizing its relevance to action preparation and execution (Berti and Schröger, 2003), MMN (Mismatch Negativity), highlighting its sensitivity to discrepancies between expected and actual stimuli (Kramer et al., 1995), and LPP (Late Positive Potential), denotes its positive potential characteristics and the post-stimulus phase in which it appears (Deeny et al., 2014). Although these metrics have been demonstrated to characterize different memory processing processes, their validity as measures of cognitive load is subject to debate. Therefore, it is crucial to quantify the effects of cognitive load on various ERP components and rationalize the system of ERP components as indicators of cognitive load as a whole.

Rolke et al. (2016) used EEG to measure the amplitude of ERP components such as the P300 while learners were performing a visual search task and showed that the amplitude of these components increased with increasing attentional engagement. Xu et al. (2020) analyzed the EEG of 10 participants during a task using a mental workload estimator developed based on EEG and ERPs, and concluded that the amplitudes of N100, P3a, and RON Decreased significantly with increasing MWL in the n-back condition. Aksoy et al. (2021) combined a virtual reality head-mounted display system with an electroencephalogram to measure ERP components such as P100 and P300, which were shown to be effective in describing the learner’s level of cognitive load when performing tasks using VR. In summary, ERPs are effective in measuring cognitive load in different multimedia learning environments. But at the same time, researchers have used many different ERP components, but which ones are most effective for measuring cognitive load in multimedia learning? This question needs to be addressed urgently.

2.3 Potential factors influencing the cognitive load of ERP measures

Recent studies employing ERPs for cognitive load assessment have highlighted various factors influencing the outcomes of measurements in multimedia learning environments. Given the diverse findings, further exploration is essential to identify the moderating factors affecting the validity of ERP measures for cognitive load. Beyond the primary focus on ERP components in this study, it was anticipated that numerous variables in the included studies could impact the final measurements. These variables fall into three categories: signal acquisition characteristics of the study, task design characteristics, and participants characteristics.

2.3.1 Signal acquisition

Based on our literature review, we analyzed three features of signal acquisition that May influence cognitive load measurements: brain region, electrode position and number of electrodes.

Deeny et al. (2014) noted that ERP components from different brain regions contribute differently to measures of cognitive load, suggesting that there May be potential moderating effects of signal acquisition in different brain regions. Identifying regional differences is therefore important for accurate assessment of cognitive load.

Many studies have exclusively measured ERP components using data from midline electrode sites like Fz, FCz, Cz, Pz, etc. These middle line positions, as suggested by Eschmann et al. (2018) and Meltzer et al. (2007, 2008), May offer more interpretative insights into cognitive processes compared to others. We therefore thought it necessary to explore whether the use of midline electrodes was more helpful than other electrodes in measuring cognitive load.

In addition, some studies have achieved unexpectedly positive experimental results by relying solely on ERP data from a single electrode or channel to interpret cognitive load (Sun et al., 2022). We therefore expected to compare the simple method of collecting data using a single electrode with the conventional method using multiple electrodes.

2.3.2 Task design

Various studies have employed diverse task designs to elicit cognitive load. We investigated the impact of three task design variables on ERP measures of cognitive load: number of tasks, number of cognitive load levels, interactive of task systems, form of learning resources.

The distinction between single-task and multi-task designs plays a crucial role. While multi-task designs mimic cognitive load in daily life, offering a comprehensive understanding of brain activity during the processing of multiple tasks, single-task designs enable a clearer analysis of neural activity for specific cognitive tasks (Brunken et al., 2003; Pashler, 1993; Schumacher et al., 2001). Both have their advantages, so we analyzed the number of tasks as a moderating variable to see whether cognitive load induced by multi-tasking is easier to measure than that induced by single-tasking in multimedia learning.

Many studies have designed controlled experiments with more than two groups to measure cognitive load (Daffner et al., 2011; Deeny et al., 2014; Folstein and Van Petten, 2011). We expect to find out whether multiple group experiments are more favorable for measurement. Therefore, in order to determine the necessity of a task with multiple load levels, we analyzed the number of load levels as a moderating variable.

Research has shown that increased learner interaction with the learning system can be effective in managing cognitive load (Darejeh et al., 2022; Paas et al., 2004). The presence or absence of interactivity in task systems is another critical factor. The absence of feedback May heighten cognitive load due to self-doubt or lead to cognitive idleness, as learners lack cues for the next step in the task, influencing experimental results (Clark, 1994). We therefore conducted a moderation analysis by considering whether the task system used by participants was interactive or non-interactive.

Imhof et al. (2011) found that sequential pictures and dynamic videos stimulate learners’ cognition differently and are likely to affect valid measures of cognitive load. Therefore, it is necessary to determine which of the two forms of learning materials is more appropriate for measuring cognitive load.

2.3.3 Participants characteristics

In addition, we analyzed the moderation of gender and age of the participants in the study, as Friedman (2003) and Güntekin and Başar (2007) have shown that gender and age have an effect on EEG. Finally, we also analyzed the effect of sample size on the results of cognitive load measurements, which could avoid small sample effects that could be detrimental to our study.

In this quantitative review of studies, we comprehensively review empirical studies from different databases on the use of ERPs to measure cognitive load in multimedia learning, investigating the ERP components used in these studies as well as other variables involved in the experimental treatment. The following research questions were developed accordingly:

RQ1 Which ERP components are more effective in measuring cognitive load?

RQ2.1 How do characteristics of signal acquisition affect measures of cognitive load?

RQ2.2 How do characteristics of task design affect measures of cognitive load?

RQ2.3 How do participant characteristics affect measures of cognitive load?

RQ3 Did publication bias influence our findings?

3 Materials and methods

3.1 Literature screening and inclusion

In order to ensure the quality and quantity of the original literature, authoritative English databases such as Web of Science, EBSCO, Science Direct, and Google Scholar were selected for the literature search. The combined logical search statement is TS = (workload or cognitive workload or working memory or mental workload) AND TS = (ERP or ERPs or event related potentials) AND TS = (learning or learner or student) Duplicate documents are removed after the search is completed.

After removing duplicates, studies were screened by title and abstract to exclude those that did not meet the inclusion criteria. When the abstract did not provide enough information, the study had to be screened in full text. Finally, articles with ineligible data were excluded according to the screening criteria, and studies that met all inclusion criteria were included in the meta-analysis.

3.2 Data collection

We extracted the following information from each study and coded it using Microsoft Excel: sample size, age, gender, ERP components, electrode position (Is it all on the midline), brain regions, number of tasks, number of levels of cognitive load, interaction, type of learning material.

3.3 Calculation of effect size

We conducted our analyses using the Comprehensive Meta-Analysis (CMA) software version 3.0, which is specifically designed for conducting meta-analyses. Following Rosenthal and DiMatteo (2001)’s recommendation, we selected the correlation coefficient (r) as the effect size. CMA software automatically transforms the effect size r into Fisher’s Z, bringing the distribution of r closer to a normal distribution and extending the range of values over the entire real number axis. Consequently, our study adopted the correlation coefficient, r, as the effect size for quantifying the impact of each ERP measure on cognitive load. According to Cohen (2016)’s criteria, the effect size r between 0.10 and 0.30 is defined as a small effect, between 0.30 and 0.50 as a medium effect, and above 0.50 as a large effect.

3.4 Heterogeneity test and elimination

Heterogeneity was quantified as the percentage of variation in effect size (i.e., the I2 statistic), with values of 75% and above indicating high heterogeneity, 50 to 75% indicating moderate heterogeneity, and 25–50% indicating low heterogeneity (Higgins and Green, 2008). We also assessed this by means of the chi-square test (Cochran’s Q statistic). Considering the diversity of the studies included, a high degree of heterogeneity between studies is expected, plus the sample sizes for each frequency of the study are not sufficiently large, it would be irresponsible to use random effects values to calculate the results for each indicator, with the risk of overestimating the size of the effect (Cooper et al., 2019). Therefore, if the heterogeneity I2 was >50% when performing the assessment of each indicator, we would first perform a sensitivity analysis, use the leave-one-out method to exclude studies with excessive heterogeneity, and use a fixed-effects model to derive the final results.

When conducting moderator variable analyses, we analyzed all included studies as a whole, as we wanted to be able to guide specific data collection and experimental design as a whole. Moderator variable analyses in this study examined the effect of a variable (integer or continuous) on the strength or direction of the relationship between cognitive load and measured outcomes (Rosenthal and DiMatteo, 2001).

For publication bias measures we used Egger regression and funnel plots, and if publication bias was shown to exist, we used trim and fill methods (Duval and Tweedie, 2000) to calculate bias-corrected estimates of the mean effect. The number of missing sample sizes was estimated to inform future quantitative studies. Our findings were considered to be more robust if there was no significant publication bias.

4 Results

After removing duplicates, 877 records were found in the database and reference list searches. The 130 remaining documents after abstract screening excluded measurements that did not include any ERP component or did not conduct controlled experiments by controlling for different levels of cognitive load. A detailed assessment of the full text of 59 studies identified additional studies that were not included because of the following criteria: insufficient data (23); no use of multimedia technology (11); not in English (3); other reasons (5). In the end, a total of 17 records met the inclusion criteria to be included, see Figure 1 for the process.

Figure 1
www.frontiersin.org

Figure 1. PRISMA study collection flow diagram.

4.1 Literature screening results

The choice of peak amplitude or mean amplitude in this review depends on which of the two of them better measures differences in levels of cognitive load in Specific studies. In addition to amplitude, four studies were included on latency, which has been widely shown to judge cognitive processing ability (Goodin, 1990; Ford et al., 1980). Sample characteristics, coded moderator variables and effect sizes are shown in Table 1. In addition, because these studies were conducted at the university level, we did not code the stage of the participants, instead coding for age.

Table 1
www.frontiersin.org

Table 1. Characteristics of the included literature.

After the effects are standardized, the combined overall ERP effect size r = 0.47 [0.41, 0.53] (I2 = 27.1%, Q = 60.3). A fixed effect size model was used because the heterogeneity was within acceptable limits (I2 < 50%). It shows that overall ERP components measuring cognitive load have moderate effect sizes.

4.2 ERP component

RQ1 Which ERP components are more effective in measuring cognitive load?

The study was analyzed by grouping the studies according to the different ERP components, and the effect size of each ERP component as a measure of cognitive load was calculated, as shown in Table 2. Indicators that lacked sufficient sample size to support them (k < 3) could not be judged to be valid, and a forest plot of valid indicators is shown in Figure 2.

Table 2
www.frontiersin.org

Table 2. ERP components effect size statistics.

Figure 2
www.frontiersin.org

Figure 2. Forest plot of ERP components that effectively measure cognitive load.

4.2.1 P300

Studies that used the P300 to measure cognitive load (k = 15) showed high levels of combined effect sizes, r = 0.53 [0.32, 0.63], p < 0.05, suggesting that P300 amplitudes in the high cognitive load task were significantly lower than P300 amplitudes in the low cognitive load task. Heterogeneity was low, Q = 20.41, p = 0.118, 𝜏2 = 0.28, I2 = 31.4%.

In subgroup calculations of electrode site effect sizes, r = 0.55 [0.42, 0.66] (k = 10) for the multiregional electrode group and r = 0.52 [0.21, 0.73] (k = 5) for the parietal electrode group, with a non-significant moderating effect (Q = 0.30, p = 0.843).

4.2.2 P200

The combined effect size for studies measuring P200 (k = 4) r = 0.55 [0.35, 0.70], p < 0.05, suggesting that P200 amplitudes were significantly lower in the high cognitive load task than in the low cognitive load task. Heterogeneity was not significant, Q = 0.7, p = 0.87, I2 = 0. Limited by the small number of studies on P200, there were no valid data from subgroup calculations of electrode site effect sizes or precise conclusions to be drawn from related studies.

4.2.3 P300 latency

The combined effect size for studies measuring P300 latencies (k = 3) r = 0.43 [0.16, 0.64], p < 0.05, suggesting that P300 latencies were significantly longer in the high cognitive load task than in the low cognitive load task. Heterogeneity was not significant (Q = 0.27, p = 0.87, I2 = 0).

4.2.4 N200

The combined effect size of studies (k = 5) using the N200 to measure cognitive load was not credible, r = 0.18 [0.02, 0.36], p = 0.076 > 0.05. Impact analyses by leave-one-out identified 1 study that was considered an outlier (Wei and Zhou, 2020). Withdrawing it resulted in an increase in mean effect size and a Decrease in heterogeneity, r = 0.38 [0.17, 0.55], Q = 0.73, p > 0.10, 𝜏2 = 0.03, I2 = 0%.

4.2.5 P100 and N100

Few studies have measured cognitive load using the P100 (k = 3) and N100 (k = 3), with a combined effect size of r = 0.28 [0.01, 0.51], p = 0.036 for the P100, and r = 0.39 [0.11, 0.62], p = 0.007 for the N100. Both showed a Decreasing trend in peak amplitude as cognitive load increased.

LPP, N300, P3a, P3b, RON and SPCN could not be analyzed separately due to insufficient sample size (k < 3), but their contributions to the overall ERP effect sizes were still statistically significant.

4.3 Moderation variables analysis

4.3.1 Signal acquisition

RQ2.1 How do characteristics of signal acquisition affect measures of cognitive load?

We analyzed the three characteristics of brain regions (Q = 3.264, p = 0.515), whether there were midline electrodes (Q = 0.26, p = 0.611) and the number of electrodes (Q = 0.08, p = 0.775) and found that none of these moderating variables were statistically significant (see Table 3).

Table 3
www.frontiersin.org

Table 3. Moderation analysis of signal acquisition as variable.

4.3.2 Task design

RQ2.2 How do characteristics of task design affect measures of cognitive load?

We analyzed the effect of the characteristics of the task design on the measurements, as shown in the table, the four characteristics of the task system with interactivity (Q = 4.840, p = 0.028), the number of tasks performed (Q = 0.536, p = 0.464), the number of levels of cognitive load (Q = 8.64, p = 0.004), and the type of learning material (Q = 2.009, p = 0.366). The effect sizes of task system with interactivity, and tertiary cognitive load level were found to have high effect sizes between groups. The effect size of the task system with interactivity (r = 0.58) was larger than that of the task system without interactivity (r = 0.43); the effect size of the level of tertiary cognitive load (r = 0.56) was significantly higher than that of the level of secondary cognitive load (r = 0.39), and both of them contributed the vast majority of the between-groups heterogeneity (Q = 0.833, p = 0.004) (see Table 4).

Table 4
www.frontiersin.org

Table 4. Moderation analysis of task design as variable.

4.3.3 Participant characteristics

RQ2.3 How do participant characteristics affect measures of cognitive load?

As the included ERP studies were all mixed sex and could not be grouped, Meta-regression analyses were attempted with female/all ratio as a covariate, and the results were not significant (p = 0.579, R2 = 0); nor were the results of meta-regression analyses of sample size (p = 0.313, R2 = 0.03) mean age (p = 0.929, R2 = 0) as a covariate, as shown in Table 5. In the analyses of sex ratio as a continuous moderator variable, three studies without a sex description were removed (Begum et al., 2014; Berti and Schröger, 2003; Daffner et al., 2011); in the analyses of age as a continuous moderator variable, the median of the age range was used for the studies that lacked a description of the mean age, which Daffner et al. (2011)‘s study was removed due to the large age span.

Table 5
www.frontiersin.org

Table 5. Meta-regression analysis of participants’ characteristics as covariates.

4.4 Publication bias

RQ3 Did publication bias influence our findings?

The funnel plot formed by the effect sizes of the ERP primary literature selected for this study was evenly distributed on the left and right with Z = 0.51 as the axis of symmetry, indicating that the selected primary literature publication bias was acceptable, as shown in Figure 3. Meanwhile, the test results of Egger’s regression method showed that t = 0.60, p = 0.55 > 0.05, indicating that the selected original literature publication bias is not significant, so the results of this study are robust.

Figure 3
www.frontiersin.org

Figure 3. Funnel plot for testing publication bias.

5 Discussion

In our study, we summarize articles that have investigated ERP components as well as event-related potential metrics under different cognitive loads. Firstly, our meta-analysis quantified the overall effect of cognitive load on the ERP components, followed by group analyses of the P100, P200, P300, P300Latency, N100, and N200 components, respectively. A number of moderating variables were found that affected the magnitude of the effect size. The remaining indicators that lacked sample size support and insignificant moderating variables are not discussed further.

5.1 ERP components

5.1.1 P300

P300 component is the most used ERP component and is widely used to assess cognitive load in single-task paradigms as well as multi-task paradigms (Kestens et al., 2023). The studies included in this meta-analysis that used the P300 amplitude to measure cognitive load all reported that it Decreased as cognitive load increased, and a high level of effect sizes (r = 0.54 > 0.50), coupled with the fact that it is well grounded in sufficient research (k = 15) to be judged as the preferred ERP indicator for measuring cognitive load.

5.1.2 P200

P200 component is a positive component of the ERP at the frontal-central scalp position (Zuckerman et al., 2023). Studies have looked at the amplitude of the P200 and found that it consistently Decreases with increasing cognitive workload (Deeny et al., 2014; Sun et al., 2022). It also has a high level of effect size in our results (r = 0.55 > 0.50), so it can also be used as a preferred measure of cognitive load.

5.1.3 Other ERP components

Almost all other ERP components had moderate effect sizes in the analyses (0.30 < r < 0.50), and P100 had low effect sizes (0.10 < r < 0.30). However, they can still be used as alternative indicators to measure cognitive load in multimedia learning environments, serving to complement the experimental data in several ways.

P300 peak latency measures the speed of information processing therefore the latency of the P300 component also shows a direct relationship with cognitive load (Sergeant et al., 1987). The three studies included in this review all found that as cognitive load increased, P300 peak latency also increased (Aksoy et al., 2021; Daffner et al., 2011; Folstein and Van Petten, 2011). To some extent, this suggests that P300 latency can be a good measure of cognitive load.

N200 component is the negative ERP component at the central parietal position of the scalp (Sur and Sinha, 2009). Goodin (1990) were the first to report an increase in the latency of the N200 with increasing cognitive workload. However, in four studies, Ke et al. (2016) found that N200 amplitude Decreased with increasing difficulty, suggesting that N200 is not a stable indicator of performance. Additionally, the research on its peak latency N200latency is very understudied (k < 3).

N100 component is a short latency ERP component that is produced primarily by the frontal cortex (Sur and Sinha, 2009). A similar Decrease in N100 amplitude was observed in all three included studies. The results of Kramer et al. (1995) also suggest that N100 amplitude Decreases with increasing cognitive workload.

P100 May be influenced by attention-driven, top-down modulation of visual processing (Gazzaley, 2011). In conditions of high cognitive load, P100 amplitude Decreased. Of the three studies we included, Pratt et al. (2011) increased cognitive load by performing a dual task, which was indeed more experimentally significant than the other two studies. To some extent, this suggests a more favorable triggering effect of the P100 in tasks affecting the allocation of attention.

5.2 Moderator variable

In the analysis of signal acquisition method characteristics as moderating variables, none of the three moderating variables were found to be statistically significant. This means that in the included experiments, different electrode numbers, electrode locations or electrode areas caused insignificant or offsetting changes in the measured ERP components. This suggests that we can modestly simplify our signal acquisition operations under the constraints of experimental equipment or time, and that we can measure substantial results by acquiring only the brain regions most relevant to the purpose of the experiment. This statement is not absolute, however, and although these three moderating variables did not show significance, this does not preclude them from playing an important role under specific experimental conditions or tasks. Therefore, further research May need to consider other factors or explore the effects of these variables in more detail.

In the analyses where task design features were used as moderating variables larger effect sizes were found for tertiary cognitive load levels than for secondary cognitive load levels. Experiments that set a level three cognitive load can generate more data and can verify the robustness and consistency of the experimental results by comparing multiple sets of data, thus increasing data complexity and reproducibility, and thus calculating a higher effect size. Analyses on whether the task system was interactive or not found higher effect sizes for task systems with interactivity. It has been shown that increasing learner interaction with the learning system can be effective in managing cognitive load (Darejeh et al., 2022; Paas et al., 2004). Therefore, the design or selection of the task system needs to provide timely feedback to the learners to prevent “cognitive idling” or attentional drift, which May threaten the accuracy of the experimental results.

In the analyses of subject characteristics as continuous moderating variables no significant effects of gender, age, and sample size on the measurement results were found. Preliminarily, it can be concluded that these subjects’ factors do not significantly affect the measurement results in the actual study of multimedia learning.

6 Limitations and future work

The lack of adequate sample sizes for some measures to support their effect sizes is likely due to the fact that some researchers only report measures that show statistically significant differences, which prevents the validity of some measures from being confirmed. The lack of sample size is compounded by the fact that there is a lot of specificity in the waveforms within the time window of the ERPs and the tendency for the measurements to be trivial. Therefore, it is necessary to design cognitive load elicitation tests to induce different degrees of cognitive load in subsequent studies, and use them as a basis for extracting and constructing ERP components, or even multimodal measures, in order to effectively characterize learners’ cognitive load.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Author contributions

SY: Conceptualization, Formal analysis, Investigation, Project administration, Supervision, Validation, Writing – original draft, Writing – review & editing. LT: Methodology, Software, Visualization, Writing – original draft. GW: Conceptualization, Data curation, Funding acquisition, Writing – review & editing. SN: Investigation, Resources, Software, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This research was funded by National Social Science Foundation “13th Five-Year Plan” 2020 general subject of education “‘Internet +’ to promote the innovative development of small-scale schools in western rural path and strategy research” (BCA200085).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Aksoy, M., Ufodiama, C. E., Bateson, A. D., Martin, S., and Asghar, A. U. R. (2021). A comparative experimental study of visual brain event-related potentials to a working memory task: virtual reality head-mounted display versus a desktop computer screen. Exp. Brain Res. 239 , 3007–3022. doi: 10.1007/s00221-021-06158-w

PubMed Abstract | Crossref Full Text | Google Scholar

Anmarkrud, O., Andresen, A., and Braten, I. (2019). Cognitive load and working memory in multimedia learning: conceptual and measurement issues. Educ. Psychol. 54 , 61–83. doi: 10.1080/00461520.2018.1554484

Crossref Full Text | Google Scholar

Begum, T., Reza, F., Ahmed, I., and Abdullah, J. M. (2014). Influence of education level on design-induced N170 and P300 components of event related potentials in the human brain. J. Integr. Neurosci. 13 , 71–88. doi: 10.1142/S0219635214500058

PubMed Abstract | Crossref Full Text | Google Scholar

Berti, S. (2008). Cognitive control after distraction: event-related brain potentials (ERPs) dissociate between different processes of attentional allocation. Psychophysiology 45 , 608–620. doi: 10.1111/j.1469-8986.2008.00660.x

PubMed Abstract | Crossref Full Text | Google Scholar

Berti, S., and Schröger, E. (2003). Working memory controls involuntary attention switching: evidence from an auditory distraction paradigm. Eur. J. Neurosci. 17 , 1119–1122. doi: 10.1046/j.1460-9568.2003.02527.x

Crossref Full Text | Google Scholar

Brunken, R., Plass, J. L., and Leutner, D. (2003). Direct measurement of cognitive load in multimedia learning. Educ. Psychol. 38 , 53–61. doi: 10.1207/S15326985EP3801_7

Crossref Full Text | Google Scholar

Cavanagh, T. M., and Kiersch, C. (2023). Using commonly-available technologies to create online multimedia lessons through the application of the cognitive theory of multimedia learning. Educ. Technol. Res. Dev. 71 , 1033–1053. doi: 10.1007/s11423-022-10181-1

PubMed Abstract | Crossref Full Text | Google Scholar

Çeken, B., and Taşkın, N. (2022). Multimedia learning principles in different learning environments: a systematic review. Smart Learn. Environ. 9 :19. doi: 10.1186/s40561-022-00200-2

Crossref Full Text | Google Scholar

Clark, R. E. (1994). Media will never influence learning. Educ. Technol. Res. Dev. 42 , 21–29.

Google Scholar

Cohen, J. (2016). A power primer. New York, NY, USA: American Psychological Association, 284. doi: 10.1037/14805-018

Crossref Full Text | Google Scholar

Combs, L., and Polich, J. (2006). P3a from auditory white noise stimuli. Clin. Neurophysiol. 117 , 1106–1112. doi: 10.1016/j.clinph.2006.01.023

PubMed Abstract | Crossref Full Text | Google Scholar

Cooper, H., Hedges, L. V., and Valentine, J. C. (2019). The handbook of research synthesis and Meta-analysis. Washington, DC, USA: Russell Sage Foundation. Available at: https://www.daneshnamehicsa.ir/userfiles/files/1/9-%20The%20Handbook%20of%20Research%20Synthesis%20and%20Meta-Analysis.pdf

Google Scholar

Daffner, K. R., Chong, H., Sun, X., Tarbi, E. C., Riis, J. L., McGinnis, S. M., et al. (2011). Mechanisms underlying age- and performance-related differences in working memory. J. Cogn. Neurosci. 23 , 1298–1314. doi: 10.1162/jocn.2010.21540

PubMed Abstract | Crossref Full Text | Google Scholar

Darejeh, A., Marcus, N., and Sweller, J. (2022). Increasing learner interactions with E-learning systems can either decrease or increase cognitive load depending on the nature of the interaction. Annee Psychol. 122, 405–437. doi: 10.3917/anpsy1.223.0405

Crossref Full Text | Google Scholar

Deeny, S., Chicoine, C., Hargrove, L., Parrish, T., and Jayaraman, A. (2014). A simple ERP method for quantitative analysis of cognitive workload in myoelectric prosthesis control and human-machine interaction. PLoS One 9 :e112091. doi: 10.1371/journal.pone.0112091

PubMed Abstract | Crossref Full Text | Google Scholar

Duval, S., and Tweedie, R. (2000). A nonparametric “trim and fill” method of accounting for publication Bias in Meta-analysis. J. Am. Stat. Assoc. 95 , 89–98. doi: 10.1080/01621459.2000.10473905

Crossref Full Text | Google Scholar

Eschmann, K. C. J., Bader, R., and Mecklinger, A. (2018). Topographical differences of frontal-midline theta activity reflect functional differences in cognitive control abilities. Brain Cogn. 123 , 57–64. doi: 10.1016/j.bandc.2018.02.002

PubMed Abstract | Crossref Full Text | Google Scholar

Farkish, A., Bosaghzadeh, A., Amiri, S. H., and Ebrahimpour, R. (2023). Evaluating the effects of educational multimedia design principles on cognitive load using EEG Signal analysis. Educ. Inf. Technol. 28 , 2827–2843. doi: 10.1007/s10639-022-11283-2

Crossref Full Text | Google Scholar

Folstein, J. R., and Van Petten, C. (2011). After the P3: late executive processes in stimulus categorization. Psychophysiology 48 , 825–841. doi: 10.1111/j.1469-8986.2010.01146.x

PubMed Abstract | Crossref Full Text | Google Scholar

Ford, J. M., Mohs, R. C., Pfefferbaum, A., and Kopell, B. S. (1980). “On the utility of P3 latency and RT for studying cognitive processes” in Progress in brain research, vol. 54, (Amsterdam, The Netherlands: Elsevier), 661–667.

Google Scholar

Friedman, D. (2003). Cognition and aging: a highly selective overview of event-related potential (ERP) data. J. Clin. Exp. Neuropsychol. 25 , 702–720. doi: 10.1076/jcen.25.5.702.14578

PubMed Abstract | Crossref Full Text | Google Scholar

Gazzaley, A. (2011). Influence of early attentional modulation on working memory. Neuropsychologia 49 , 1410–1424. doi: 10.1016/j.neuropsychologia.2010.12.022

PubMed Abstract | Crossref Full Text | Google Scholar

Ghani, U., Signal, N., Niazi, I. K., and Taylor, D. (2020). ERP based measures of cognitive workload: a review. Neurosci. Biobehav. Rev. 118 , 18–26. doi: 10.1016/j.neubiorev.2020.07.020

Crossref Full Text | Google Scholar

Goodin, D. S. (1990). Clinical utility of long latency ‘cognitive’ event-related potentials (P3): the pros. Electroencephalogr. Clin. Neurophysiol. 76 , 2–5. doi: 10.1016/0013-4694(90)90051-K

PubMed Abstract | Crossref Full Text | Google Scholar

Güntekin, B., and Başar, E. (2007). Brain oscillations are highly influenced by gender differences. Int. J. Psychophysiol. 65 , 294–299. doi: 10.1016/j.ijpsycho.2007.03.009

PubMed Abstract | Crossref Full Text | Google Scholar

Higgins, J. P., and Green, S. (2008). “Front matter” in Cochrane handbook for systematic reviews of interventions. 1st ed. (Hoboken, NJ, USA: Wiley).

Google Scholar

Imhof, B., Scheiter, K., and Gerjets, P. (2011). Learning about locomotion patterns from visualizations: effects of presentation format and realism. Comput. Educ. 57 , 1961–1970. doi: 10.1016/j.compedu.2011.05.004

Crossref Full Text | Google Scholar

Ke, Y., Wang, P., Chen, Y., Gu, B., Qi, H., Zhou, P., et al. (2016). Training and testing ERP-BCIs under different mental workload conditions. J. Neural Eng. 13 :016007. doi: 10.1088/1741-2560/13/1/016007

PubMed Abstract | Crossref Full Text | Google Scholar

Kestens, K., Van Yper, L., Degeest, S., and Keppler, H. (2023). The P300 auditory evoked potential: a physiological measure of the engagement of cognitive systems contributing to listening effort? Ear Hear. 44 , 1389–1403. doi: 10.1097/AUD.0000000000001381

PubMed Abstract | Crossref Full Text | Google Scholar

Kramer, A. F. (1991). “Physiological metrics of mental workload: a review of recent progress” in Multiple-Task Performance, 279–328.

Google Scholar

Kramer, A. F., Trejo, L. J., and Humphrey, D. (1995). Assessment of mental workload with task-irrelevant auditory probes. Biol. Psychol. 40 , 83–100. doi: 10.1016/0301-0511(95)05108-2

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, Y., Jang, B. G., and Roy-Campbell, Z. (2018). Optimum input mode in the modality and redundancy principles for university ESL students’ multimedia learning. Comput. Educ. 127 , 190–200. doi: 10.1016/j.compedu.2018.08.025

Crossref Full Text | Google Scholar

Mayer, R. E. (2006). “Coping with complexity in multimedia learning” in Handling complexity in learning environments. (Amsterdam, The Netherlands: Elsevier Ltd). Available at: https://books.google.co.jp/books?hl=zh-CN&lr=&id=buY4icX3LbAC&oi=fnd&pg=PA129&dq=Co**+with+complexity+in+multimedia+learning&ots=mnKtrHuAP_&sig=-HxrLgCnHcBqdCFAQ1MjivKaFmM&redir_esc=y#v=onepage&q=Co**%20with%20complexity%20in%20multimedia%20learning&f=false

Google Scholar

Meltzer, J. A., Negishi, M., Mayes, L. C., and Constable, R. T. (2007). Individual differences in EEG theta and alpha dynamics during working memory correlate with fMRI responses across subjects. Clin. Neurophysiol. 118 , 2419–2436. doi: 10.1016/j.clinph.2007.07.023

PubMed Abstract | Crossref Full Text | Google Scholar

Meltzer, J. A., Zaveri, H. P., Goncharova, I. I., Distasio, M. M., Papademetris, X., Spencer, S. S., et al. (2008). Effects of working memory load on oscillatory power in human intracranial EEG. Cereb. Cortex 18 , 1843–1855. doi: 10.1093/cercor/bhm213

PubMed Abstract | Crossref Full Text | Google Scholar

Mutlu-Bayraktar, D., Cosgun, V., and Altan, T. (2019). Cognitive load in multimedia learning environments: a systematic review. Comput. Educ. 141 :103618. doi: 10.1016/j.compedu.2019.103618

Crossref Full Text | Google Scholar

Paas, F., Renkl, A., and Sweller, J. (2004). Cognitive load theory: instructional implications of the interaction between information structures and cognitive architecture. Instr. Sci. 32 , 1–8. doi: 10.1023/B:TRUC.0000021806.17516.d0

Crossref Full Text | Google Scholar

Pashler, H. (1993). “Dual-task interference and elementary mental mechanisms” in Attention and performance 14: Synergies in experimental psychology, artificial intelligence, and cognitive neuroscience. (Cambridge, MA, US: The MIT Press), 245–264. Available at: https://psycnet.apa.org/record/1993-97600-010

Google Scholar

Pratt, N., Willoughby, A., and Swick, D. (2011). Effects of working memory load on visual selective attention: behavioral and electrophysiological evidence. Front. Hum. Neurosci. 5:57. doi: 10.3389/fnhum.2011.00057

PubMed Abstract | Crossref Full Text | Google Scholar

Rolke, B., Festl, F., and Seibold, V. C. (2016). Toward the influence of temporal attention on the selection of targets in a visual search task: an ERP study. Psychophysiology 53 , 1690–1701. doi: 10.1111/psyp.12734

PubMed Abstract | Crossref Full Text | Google Scholar

Rosenthal, R., and DiMatteo, M. R. (2001). Meta-analysis: recent developments in quantitative methods for literature reviews. Annu. Rev. Psychol. 52 , 59–82. doi: 10.1146/annurev.psych.52.1.59

Crossref Full Text | Google Scholar

Schumacher, E. H., Seymour, T. L., Glass, J. M., Fencsik, D. E., Lauber, E. J., Kieras, D. E., et al. (2001). Virtually perfect time sharing in dual-task performance: uncorking the central cognitive bottleneck. Psychol. Sci. 12 , 101–108. doi: 10.1111/1467-9280.00318

Crossref Full Text | Google Scholar

Sergeant, J., Geuze, R., and van Winsum, W. (1987). Event-related desynchronization and P300. Psychophysiology 24 , 272–277. doi: 10.1111/j.1469-8986.1987.tb00294.x

Crossref Full Text | Google Scholar

Shaw, E. P., Rietschel, J. C., Hendershot, B. D., Pruziner, A. L., Miller, M. W., Hatfield, B. D., et al. (2018). Measurement of attentional reserve and mental effort for cognitive workload assessment under various task demands during dual-task walking. Biol. Psychol. 134 , 39–51. doi: 10.1016/j.biopsycho.2018.01.009

Crossref Full Text | Google Scholar

Sun, Y., Ding, Y., Jiang, J., and Duffy, V. G. (2022). Measuring mental workload using ERPs based on FIR, ICA, and MARA. Comput. Syst. Sci. Eng. 41 , 781–794. doi: 10.32604/csse.2022.016387

Crossref Full Text | Google Scholar

Sur, S., and Sinha, V. K. (2009). Event-related potential: an overview. Ind. Psychiatry J. 18 , 70–73. doi: 10.4103/0972-6748.57865

PubMed Abstract | Crossref Full Text | Google Scholar

Suzuki, J., Nittono, H., and Hori, T. (2005). Level of interest in video clips modulates event-related potentials to auditory probes. Int. J. Psychophysiol. 55 , 35–43. doi: 10.1016/j.ijpsycho.2004.06.001

PubMed Abstract | Crossref Full Text | Google Scholar

Trypke, M., Stebner, F., and Wirth, J. (2023). Two types of redundancy in multimedia learning: a literature review. Front. Psychol. 14 :1148035. doi: 10.3389/fpsyg.2023.1148035

PubMed Abstract | Crossref Full Text | Google Scholar

Watter, S., Geffen, G. M., and Geffen, L. B. (2001). The n-back as a dual-task: P300 morphology under divided attention. Psychophysiology 38 , 998–1003. doi: 10.1111/1469-8986.3860998

PubMed Abstract | Crossref Full Text | Google Scholar

Wei, H., and Zhou, R. (2020). High working memory load impairs selective attention: EEG signatures. Psychophysiology 57 :e13643. doi: 10.1111/psyp.13643

PubMed Abstract | Crossref Full Text | Google Scholar

Xu, J., Ke, Y., Liu, S., Song, X., Xu, C., Zhou, G., et al. (2020). “Task-irrelevant auditory event-related potentials as mental workload indicators: a between-task comparison study” in 2020 42nd annual international conference of the IEEE engineering in Medicine & Biology Society (EMBC), vol. 2020, 3216–3219.

Google Scholar

Zuckerman, I., Laufer, I., and Mizrahi, D. (2023). Attachment style, emotional feedback, and neural processing: investigating the influence of attachment on the P200 and P400 components of event-related potentials. Front. Hum. Neurosci. 17 :1249978. doi: 10.3389/fnhum.2023.1249978

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: cognitive load, event-related potentials, multimedia learning, meta-analysis, electroencephalogram

Citation: Yu S, Tian L, Wang G and Nie S (2024) Which ERP components are effective in measuring cognitive load in multimedia learning? A meta-analysis based on relevant studies. Front. Psychol. 15:1401005. doi: 10.3389/fpsyg.2024.1401005

Received: 15 March 2024; Accepted: 06 September 2024;
Published: 19 September 2024.

Edited by:

Ion Juvina, Wright State University, United States

Reviewed by:

Bertille Somon, Office National d’Études et de Recherches Aérospatiales, France
Maria Paula Maziero, University of Texas Health Science Center at Houston, United States

Copyright © 2024 Yu, Tian, Wang and Nie. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Guohua Wang, d2doMTk4OTIwMDhAMTI2LmNvbQ==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.