Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Psychol., 19 December 2025

Sec. Quantitative Psychology and Measurement

Volume 16 - 2025 | https://doi.org/10.3389/fpsyg.2025.1664852

Automatic for the kids: assessment of cognitive efficiency in elementary school children’s auditory semantic processing

  • Institute of Educational Research, University of Wuppertal, Wuppertal, Germany

The present study examined time on task effects in semantic processes of listening comprehension on the word and sentence levels. The aim was to investigate whether response times in auditory semantic tasks show systematic relations with accuracy, and whether these tasks therefore provide valid information about the efficiency of semantic listening processes. A total of 480 German elementary school children in Grades 1–4 solved either a semantic categorization task (n = 205) or a sentence verification task (n = 275). Accuracy and response time data were analyzed using generalized linear mixed models. Random effects for persons and items were added to test whether the time on task effects were moderated by individual skill and item difficulty. Results showed that faster responses were associated with higher accuracy for both tasks. Moreover, this relationship was stronger for more skilled comprehenders and easier items. These findings suggest that response times add meaningful information about the efficiency of semantic listening processes and may complement accuracy-based measures in the assessment of listening component skills.

1 Introduction

Listening comprehension plays a central role in the language development of children and serves as a foundational skill for literacy acquisition (Kim and Pilcher, 2016). Although listening comprehension develops naturally and – at least for first language acquisition – without the need for formal instruction, listening is an active and multifaceted process in which listeners attend to an auditory source, understand vocabulary and grammar, interpret paraverbal information, and retain and integrate this within a larger context (Vandergrift, 1999). Therefore, listening is directly affected by language variables such as vocabulary, grammatical knowledge, and inference skills, and has been shown to draw on cognitive resources such as working memory and attentional control, which are necessary for maintaining and integrating incoming information into a coherent mental model (Florit et al., 2009; Kim, 2015a, 2016; Lervåg et al., 2018). According to Imhof’s (2010) model of listening, which is based on the Select-Organize-Integrate (SOI) model (Mayer, 1996), listeners first select relevant auditory and visual input for further processing. This incoming information must then be organized into meaningful units by recognizing words and accessing the mental lexicon, and by determining the grammatical structure and relationships between elements. Finally, the input is integrated with prior knowledge into a coherent model, which in turn helps to monitor the comprehension process and make necessary adjustments (Imhof, 2010). As auditory information is transient, working memory plays an important role in retaining the input long enough to organize it and build a coherent text representation. Especially when lower-level language processes, such as lexical access, are not yet efficient, listeners rely more heavily on contextual information to integrate meaning on the sentence or text level (Nation and Snowling, 1998; Stanovich et al., 1985). This increased reliance on context puts a higher burden on cognitive resources and leaves fewer resources for higher-order processes, such as drawing inferences or monitoring comprehension (Imhof, 2010; Perfetti, 1985, 1992).

The limitations of cognitive resources and the related consequences of inefficient language retrieval processes make it necessary that these processes become automatic for successful comprehension (Just and Carpenter, 1992; Just et al., 1996; Perfetti, 1985). We use the term automatic to represent rapid and low-effort access to mental representations under task-directed attention, in line with the verbal efficiency theory (Perfetti, 1985). We further use the term cognitive efficiency to describe the same phenomenon. Hui and Godfroid (2021) found that, for second language listeners, automaticity especially in lexical processes was important for propositional comprehension and in turn for general listening comprehension. This is analogous to processes in reading comprehension, in which the efficiency of word recognition and automatic activation of semantic networks support the efficiency in reading comprehension (Kim, 2015b; Perfetti, 1985, 1992). For listening, not only the transiency, but also the lack of control regarding the spoken input (i.e., the speed or amount of incoming information) can further lead to impaired comprehension when at least lexical-level comprehension processes are not automatic (Hui and Godfroid, 2021). Thus, whereas attentional resources play an important role especially for higher-level listening processes (e.g., Kim, 2015a; Shtyrov, 2010), automaticity in the processing of incoming oral language information and in the activation of semantic networks plays an important role in the efficiency and success of listening comprehension.

These considerations have implications for the assessment of listening skills. If automatic processes play a key role in listening comprehension, assessment procedures should not only measure how accurately listeners understand spoken language, but also how cognitively efficient they do so – that is, the extent to which listeners process spoken language quickly and with minimal cognitive effort. In the related cognitive domain of reading, a couple of processes, such as decoding or phonological recoding, but also sentence level processes such as semantic integration on the sentence level, are amenable to become automatic, and are automatic in experienced readers. Consequently, a number of reading tests that assess word- or sentence level processes have a speed component such that test scores increase not only with more correct responses, but also with shorter response times. This applies to paper-based tests at the sentence level that require judging as many sentences as possible within a given time (e.g., Salzburger Lese-Screening [Salzburg reading screening], SLS 2–9; Wimmer and Mayringer, 2014), to paper-based, word-level tests that require reading and understanding words quickly (e.g., Würzburger Leise Leseprobe [Würzburg silent reading probe], WLLP-R, Schneider et al., 2011), and to computer-based test batteries of reading component skills where item-level response times are integrated with accuracy to estimate proficiency (e.g., Prozessbezogene Diagnostik von Lesefähigkeiten im Grundschulalter [Process-oriented measurement of reading skills in primary school children], ProDi-L, Richter et al., 2017). In all these assessment procedures, the underlying assumption is that shorter response times are indicative of more highly automatized cognitive processes. As automatic cognitive processing is accurate and fast at the same time, given the right input (Shiffrin and Schneider, 1977), shorter response times are interpreted as indicative of higher proficiency levels.

The idea that short response times are akin to cognitive processes being highly routinized and thus fast and accurate at the same time is not without alternative and does not necessarily apply across the board. Specifically, in domains where tasks need a high degree of deliberate and controlled cognitive processing, such as problem solving (e.g., Nicolay et al., 2022), or hypertext reading (e.g., DeStefano and LeFevre, 2007), different interpretations of response times might apply. In these kinds of tasks, short response times might actually indicate disengagement from the task rather than automaticity or efficiency of cognitive processes. Where a task requires controlled cognitive processing to a large degree, and thus the investment of time, walking away from it by giving a quick response already might be a recipe for failure. Moreover, also within domains, time on task might have different associations with the odds of success, depending on both tasks and persons. In the domain of reading, for example, Walczyk (2000) proposed a model (“compensatory encoding model”) that claims that especially weak readers might compensate for their lack of skills by investment of additional time on task. From this perspective, when a task is cognitively challenging, response time and accuracy might be positively, rather than negatively related, and more so the less able a person is within the respective domain and the more difficult a task is.

A number of studies, both across and within domains, have provided evidence that indeed time on task effects are a function of tasks involving predominantly cognitive processes that are deliberate and consumptive of working memory, or largely automatic. First evidence for this idea was provided in a study by Goldhammer et al. (2014). These authors used field trial data from the OECD Programme for the International Assessment of Adult Literacies (PIAAC; see OECD, 2012) to investigate the relation between time on task and accuracy in two domains, problem solving in technology-rich environments and reading. They argued that problem solving is an activity that draws heavily on controlled cognitive processing, which follows almost from the definition of a “problem”: A problem exists if there is a discrepancy between a given and a goal state that, other than in a routine task, cannot be overcome by routine operations (Frensch and Funke, 1995). In such a situation, deliberate and controlled cognitive processing will be required to devise means suited to bridge the gap between the given and the goal state (Mayer and Wittrock, 2006).

In contrast to this, the reading tasks employed involved only the reading of single linear texts, thus involving cognitive operations from word decoding to the formation of a situation model, which most adult readers can carry out by means of automatic processing to a large degree (Schindler and Richter, 2018). In line with the authors’ hypotheses, in a generalized linear mixed model, the mean (fixed) effect of time on task was positive for problem solving, and negative for reading. Moreover, within domains, time on task effects were moderated by persons’ skill and items’ difficulty, as shown by the random effects’ covariance structure, that allowed random intercepts for persons and items to co-vary with random time on task slopes across persons and items. Positive time on task effects in problem solving increased with decreasing skill and increasing difficulty. Likewise, negative time on task effects in reading increased with increasing skill and decreasing difficulty. Thus, in problem solving, investment of time was crucial especially in hard tasks, and paid off especially for weaker problem solvers. In contrast, in reading, the association between speed and accuracy was strongest for easier tasks and skilled readers. This pattern of results was interpreted by the authors as being consistent with a dual-processing account (see Schneider and Shiffrin, 1977; Shiffrin and Schneider, 1977) of reading and problem solving, where reading tasks as those that were used in the PIAAC assessment are reliant on cognitive processes that are largely amenable to automatization, while problem solving requires controlled cognitive processing to a large degree. Further research extended and corroborated the idea that time on task is positively predictive of performance in complex cognitive tasks that require deliberate and controlled processing (Naumann and Goldhammer, 2017; Scherer et al., 2015; Vörös et al., 2021).

Vörös et al. (2021) replicated and extended the problem-solving part of Goldhammer et al.’s (2014) study and analyzed PIAAC problem solving data that came from the main assessment rather than the field trial. Drawing on this stronger data base, they found the same result that time on task effects were positive, and even more so in harder items. Moreover, they showed that task characteristics such as text complexity or technology load further increased the overall positive time on task effect, supporting to the idea that in problem solving, mobilization of cognitive resources requires taking time, and thereby positively contributes to task solution. Scherer et al. (2015) analyzed data from 2000 Finnish youths who completed an assessment of Complex Problem Solving. Using a structural equation modeling approach rather than a generalized mixed model as in Goldhammer et al. (2014), the authors also found positive associations between time on task and performance in solving complex problems. Moreover, both time on task and performance were positively predicted by reasoning ability, meaning that more intelligent students both performed better and spent more time to complete the assessment’s problem-solving tasks. Once again, this is in line with the idea that longer response times in complex cognitive tasks indicate not inefficiency, but engagement that adapts to the task’s difficulty. Naumann and Goldhammer (2017) extended the work by Goldhammer et al. (2014) through investigating a domain that can be viewed as comprising elements of both reading and problem solving, namely digital or hypertext reading, thus reading in an environment where it is left to the reader to select and integrate information from multiple sources according to both the sources’ credibility and usefulness. Such reading situations have explicitly been modeled as Reading as Problem Solving (RESOLV, Rouet et al., 2017) and have been shown to rely on problem solving skills (Naumann et al., 2025). From this perspective, the authors expected time on task effects that more closely resemble those observed in problem solving than those found for reading linear paragraphs, a prediction that was confirmed: Time on task effects were overall positive, and especially so for less skilled individuals and more difficult tasks. Moreover, they found that the features of digital text that require controlled processing, specifically navigation demands, moderated time on task effects. These effects were especially positive for items with high navigation demands.

In contrast to tasks that require controlled cognitive processing, reading tasks targeting lower-level reading processes such as phonological recoding (Naumann et al., 2016) or orthographic comparison (Naumann, 2018), tend to show negative, rather than positive, time on task effects. Notably, the magnitude of these negative time on task effects in both domains increased with increasing skill in the respective domain. In other words, especially in persons who are skilled in phonological recoding and in orthographic comparison, fast responses were also accurate. Furthermore, in the orthographic comparison domain, Naumann (2018) found that time on task effects were moderated by task difficulty, similar to the results reported by Goldhammer et al. (2014) for linear paragraphs: Negative time on task effects were strongest in easier tasks, likely because the item stimuli in these tasks were especially suited to be answered utilizing automatic cognitive processing. These results are in line with the view that word- and sentence-level reading component processes rely heavily on automatic processing, and that individual differences in word-level reading skill arise from differential automization and efficiency of these processes across readers (Karageorgos et al., 2020; Schindler and Richter, 2018).

Although the relationship between processing speed and accuracy has been extensively studied in reading, it remains an open question whether comparable modeling approaches can reliably capture efficiency in listening. Whereas reading and listening share many higher-order processes, especially the early, pre-comprehension processes are assumed to differ between the two which can lead to different demands on cognitive resources (Aryadoust, 2017; Imhof, 2010). In contrast to reading, listening unfolds in real time, relies on transient auditory input, and provides no opportunity for reinspection or self-paced control (Field, 2009). Listening is therefore assumed to rely more heavily on working memory and attentional resources compared to reading, where stable orthographic representations can support slower processing (Field, 2009; Florit et al., 2009; Jiang and Farquharson, 2018).

Due to this transient nature of spoken language and the resulting demands on working memory resources, listening component processes are crucial for learning and skill acquisition. From a time on task perspective, our overarching aim was to investigate whether the efficiency of listening component processes can be meaningfully assessed with tasks tapping into semantic processing, and using response times as a measure of cognitive efficiency. We thus examine whether response times systematically predict accuracy at the item and person levels, thereby testing the notion that integration of accuracy and response times could provide a more comprehensive account of efficiency in listening processes. Listening comprehension involves multiple levels of semantic processing that contribute to the construction of meaning. On the word level, listeners must efficiently access the meaning of words from their mental lexicon. On the sentence level, individual word meanings must be integrated into coherent propositions and monitored for plausibility within the sentence context. Both processes are important for successful comprehension and have been shown to rely on efficient, largely automatized processing (Imhof, 2010; Kintsch, 2005; Perfetti, 1985). In the present study, we therefore implemented tasks that measure auditory semantic processing on the word level (retrieval of word meanings) and on the sentence level (semantic integration). We investigated whether time on task effects in these tasks reflect cognitive efficiency rather than a measure of deliberate cognitive processing. To this end, we employed the parametrization introduced by Goldhammer et al. (2014) to accuracy and response time data from a semantic categorization task assessing the retrieval of word meanings, and a sentence verification task assessing semantic integration on the sentence level.

From the previously found patterns of time on task effects, and to test whether response times in these tasks show similar patterns to those associated with efficient, automatized, rather than to those of controlled or deliberate processing, we derived the following hypotheses:

Hypothesis 1: There is a negative time on task effect in the retrieval of word meanings, such that overall quicker retrieval of word meanings is associated with stronger performance.

Hypothesis 2: There is a negative association between person skill and person-specific time on task effects, such that while being negative overall, time on task effects are negative especially for persons with high levels of skill in the retrieval of word meanings.

Hypothesis 3: There is a negative association between item easiness and item-specific time on task-effects, such that while being negative overall, time on task effects are negative especially for easy items assessing the retrieval of word meanings.

Hypothesis 4: There is a negative time on task effect in the semantic integration at the sentence level so that overall quicker sentence processing is associated with stronger performance.

Hypothesis 5: There is a negative association between person skill and person-specific time on task effects, such that while being negative overall, time on task effects are negative especially for persons with high levels of skill of semantic integration at the sentence level.

Hypothesis 6: There is a negative association between item easiness and item-specific time on task-effects, such that while being negative overall, time on task effects are negative especially for easy items assessing the semantic integration at the sentence level.

2 Materials and methods

2.1 Subjects

A total of 480 elementary school children participated in the study. Of these, 205 children completed a semantic categorization task assessing the retrieval of word meanings, and 275 completed a sentence verification task assessing the semantic integration on the sentence level (see section measures below). Overall, 210 children were male, and 238 were female (missing gender information for 32 children). Table 1 gives the number of participating children broken down by task, gender, and grade level. From the children participating in the sentence verification task, 99 spoke German as their first language, and 50 spoke a language other than German as their first language (missing language background information for 97 children). For all children, a parent or guardian had given written consent for the children’s participation in the study. Also, school principals and teachers had given their consent for the study to be conducted in the respective schools and classrooms. For the children participating in the semantic categorization task, no information on language background was available due to schools’ data protection and privacy concerns.

Table 1
www.frontiersin.org

Table 1. Number of participants per task, gender, grade level and language background.

2.2 Measures

2.2.1 Retrieval of word meanings

The retrieval of word meanings was measured with a semantic categorization task from the ProDi-H (Prozessbezogene Diagnostik von Hörfähigkeiten im Grundschulalter [Process-oriented measurement of listening skills in primary school children]), a test battery for the measurement of listening component skills and the auditory equivalent of the ProDi-L (Richter et al., 2017). In the task, children first heard a category word (e.g., tools), followed by a target word (e.g., hammer) 200 ms later. Of the 35 presented items, 17 featured matching category-word pairs (e.g., food-chocolate) and 18 featured non-matching pairs (e.g., plants-bed). Classification tasks like these rely on participants to retrieve word meanings from their mental lexicon (Thompson-Schill and Gabrieli, 1999) and are strongly associated with semantic processing (Kutas and Iragui, 1998; Taikh et al., 2015). Common, everyday category names were chosen to ensure that a correct response was possible when the target word was known to the children. Item difficulty varied based on word frequency, familiarity, and the degree of relatedness between category and target words. Familiarity with the categories was rated by three independent teachers on a three-level scale (unadjusted intraclass correlation ICC = 0.84; cf. McGraw and Wong, 1996). The degree of relatedness for non-matching pairs as well as the degree of typicality for the respective category for matching pairs was quantified using cosine value from Latent Semantic Analysis (LSA). The cosine values were derived from a large German semantic space (dewak100; see Günther et al., 2015). The task was computer-based and both response accuracy as well as response time were recorded.

2.2.2 Semantic integration on the sentence level

Semantic integration processes were measured with a sentence verification task from the ProDi-H. Children were presented with short sentences and asked to determine whether they were true or false. This task required them to comprehend the sentence meaning and evaluate each statement based on their existing world knowledge (Richter et al., 2009). Task difficulty was influenced by two factors: (a) semantic complexity, defined by the number of propositions (which varied between 1 and 3), and (b) contextual predictability within the sentence (Frisson et al., 2005). A total of 48 sentences were presented – half of which were factually correct (Strawberries are red), and half of which were incorrect (e.g., Fire is gluey). Among the correct sentences, half were contextually predictable (e.g., Bananas are crooked and yellow) and the other half were less predictable (e.g., Flowers have roots, compared to the more expected flowers have stems). All test sentences used vocabulary that was familiar to elementary school children and maintained simple syntactic structures. Sentence difficulty for incorrect sentences was expected to be higher than for correct sentences. Furthermore, within the incorrect sentences, we assumed that a higher number of propositions would lead to more difficult items. When incorrect sentences contained more than one proposition, we assumed that those with only one false proposition would be more difficult to answer than sentences where all propositions are false. As with the semantic categorization task, both response accuracy and response time were recorded.

2.3 Procedure

Children were tested in their classrooms on tablet computers, with their teacher present. First, they were greeted by the experimenter, and explained that they could interrupt, or terminate their participation at any time, and if something went wrong or they had any question, they could ask for help at any time by raising their hand. The experimenter then started the script that administered the semantic categorization, or sentence verification task, respectively. Children gave their responses through the “f” and “j” keys on the keyboard. The “j” key for “yes” responses was marked with a green dot, the “f” key for “no” responses was marked with a red dot. Each task was framed within a cover story featuring an extraterrestrial “Reli,” who wanted to learn the Earthlings’ language but continued to make a lot of mistakes, and asked the children for help. “Reli” explained in detail the response keys, and that children best would keep their fingers on them during the test, but pressing only to answer an item. Children were instructed to work as fast and accurately as possible. Before the actual trials children completed two practice trials where feedback was provided. When the last child had completed the task, children were thanked for their participation and told that they or their parents or guardians could contact the researchers at any time in case they had further questions as to the purpose of the research.

2.4 Statistical modeling

To test for time on task effects, and their moderation by item difficulty and person skill, we used the parametrization introduced by Goldhammer et al. (2014), depicted in Equation 1. We expressed the log odds of a correct response by a fixed effect of time on task, a random person intercept, a random item intercept, and a random slope of time on task across both persons and items. Since children were clustered in classrooms, we included an additional random effect for classroom. Since children potentially varied considerably in their language proficiency due to different grade levels, we included grade level as additional three fixed effects (successive contrasts comparing grades 1–2, grades 2–3, and grades 3–4). Thus, the following (full) model was estimated:

ln ( p 1 p ) = β 0 + b 0 c + b 0 p + b 0 i + ( β 1 + b 1 i + b 1 p ) ( time on task ) + β 2 + β 3 + β 4     (1)

In this notation, p is the probability that an item is answered correctly. Greek letters (β) refer to fixed effects, and Latin letters (b) refer to random effects. β0 is a fixed intercept, β1 is the fixed effect of time on task, β2, β3, and β4 are the fixed effects for grade. b0c, b0p and b0i are the random classroom, person and item intercepts, respectively. The random person intercept can be interpreted as ability, the random item intercept as easiness. (Note that the parametrization is opposite to the usual parametrization in a Rasch model where the item parameter depicts difficulty). b1p and b1i are the random time on task slopes, respectively. In the full model, the covariance between b0p and b1p is estimated, as is the covariance between b0i and b1i.

To account for the skewness of the response time distributions, we used the natural logarithm. To facilitate model identification and estimation, in addition to taking the natural log we standardized the entire response time vector, including all items and persons. This meant that if there were differences in response time variance between persons or items, these were retained.

3 Results

Mean accuracy and latencies per grade level are given in Table 2. An a priori alpha level of 5% was set for all statistical tests. Where hypotheses were directed, tests were one-tailed. All model specifications and comparisons can be found in Table 3. The fixed effects are reported for the respective final models.

Table 2
www.frontiersin.org

Table 2. Mean accuracy and response time per grade level.

Table 3
www.frontiersin.org

Table 3. Model specification, comparisons, and fit statistics.

3.1 Retrieval of word meanings

3.1.1 Fixed effects

As expected, time on task had a significant negative fixed effect, b = −0.20 (SE = 0.05), z = −3.83, p < 0.001, meaning that on average faster responses were also more accurate. This result supports hypothesis 1. While in general accuracy was higher in each successive grade descriptively (see Table 2), the respective contrasts were not significant between grades 1 and 2, b = 0.22 (SE = 0.20), z = 1.12, p > 0.05, and grades 2 and 3, b = 0.21 (SE = 0.17), z = 1.23, p > 0.05, but only between grades 3 and 4, b = 0.53 (SE = 0.19), z = 2.82, p < 0.01.

3.1.2 Random effects

Persons. First, we examined if the time on task effect would be the same for all persons, or whether it would display variance across persons. To this end, we estimated a model that specified a random person intercept, but not a random person time on task slope (M0pwords). This model was compared to a model that estimated a random time on task slope but constrained the covariance between intercept and slope to be zero (M1pwords). This resulted in the model M1pwords specifying a random time on task slope across persons fitting the data significantly better than the model M0pwords, χ2(1) = 13.18, p < 0.001. Thus, the time on task slopes varied across persons. In a next step, we examined if this variance could in part be explained by persons’ skills. We estimated a model where the covariance between the random person intercept and the random time on task slope across persons was freed (M2words). This resulted in a correlation of random person intercept and random time on task slope across persons of −0.53, thus a negative correlation as expected. To test this correlation for significance, we compared this model to the model M1pwords, where the random time on task effect across persons was estimated, but could not co-vary with the random person intercept. This resulted in model M2words fitting the data significantly better than model M1pwords, χ2(1) = 7.46, p < 0.01. Combining the fixed time on task effect and the a posteriori estimates of the person-specific random time on task effects revealed that indeed most person-specific time on task effects were negative, and the more so the more skilled the persons were (see Figure 1). Thus, the more skilled persons were, the more fast responses were accurate at the same time. These results support hypothesis 2.

Figure 1
Scatter plot illustrating the negative correlation between the ability of word meaning retrieval (x-axis) and person ToT (time on task) effect (y-axis). Data points are scattered with a downward sloping trend line, indicating that as the ability to retrieve word meanings increases, the more negative the ToT effect becomes.

Figure 1. Person-specific time on task-effects in the semantic categorization task. Continuous line: Linear regression of person-specific time on task effects on ability. Dashed line: Time on task effect of zero. Dotted line: fixed time on task effect.

Items. In a first step, we examined if the fixed time on task would be the same for all items, or whether it would randomly vary across items. Thus, we first estimated a model M0iwords that specified only a random item intercept, but no random time on task slope across items, thus assuming that the time on task effect would be the same for all items. We then compared this model to a model M1iwords that specified a random time on task slope across items, but no covariance between the random item intercept and the random time on task slope across items. This resulted in the Model M1iwords fitting the data significantly better than the model M0iwords, χ2(1) = 6.96, p < 0.01, thus time on task slopes varied across items. In a next step we examined whether this variance could be in part explained by items’ difficulties. In model M2words, the covariance between the random item intercept and the random time on task slope across items was freed. This resulted in a correlation of −0.52 between the random item intercept and the random time on task slope across items, thus negative as expected. To test this correlation for significance, we compared the fit of model M2words to the fit of model M1iwords, which restricted the covariance between the random person intercept and the random time on task slope across items at zero. This resulted in model M2words fitting the data significantly better than model M1iwords, χ2(1) = 4.13, p < 0.05. Combining the fixed time on task effect and the a posteriori estimates of the item-specific random time on task effects meant that item-specific time on task effects were negative with one exception, but it was especially in easy items that fast responses were associated with correct responses (see Figure 2). In (relatively) difficult items, the item-specific time on task effects were negative as well, but to a smaller extent. These results support hypothesis 3.

Figure 2
Scatter plot showing a negative correlation between item easiness of word meaning retrieval on the x-axis and item ToT (time on task) effect on the y-axis. Data points are scattered, with a trend line sloping downward, indicating that as item easiness increases, the more negative the ToT effect becomes.

Figure 2. Item-specific time on task-effects in the semantic categorization task. Continuous line: Linear regression of item-specific time on task effects on item easiness. Dashed line: Time on task effect of zero. Dotted line: fixed time on task effect.

3.2 Semantic integration on the sentence level

3.2.1 Fixed effects

Similar to the findings for the retrieval of word meanings and in accordance with our expectations, there was a significant negative fixed effect of time on task, b = −0.44 (SE = 0.06), z = −7.57, p < 0.001. Thus, on average, faster responses were also more accurate across tasks and persons. This result supports hypothesis 4. Mean performance per grade is given in Table 2. Performance did not differ significantly between grades 1 and 2, b = −0.07 (SE = 0.23), z = −0.33, p > 0.05, and 2 and 3, b = 0.19 (SE = 0.17), z = 1.08, p > 0.05, while it differed between grades 3 and 4, b = 0.54 (SE = 0.22), z = 2.48, p < 0.05.

3.2.2 Random effects

Persons. First, we examined whether there was variance in time on task effects across persons. We employed the same procedure as with the retrieval of word meanings, comparing the fit of a model M1psentence that specified a random effect of time on task across persons, but no covariance between random slope and intercept, to the fit of a model M0psentence that constrained this variance component to be zero and only specified a random person intercept. This comparison indicated that model M1psentence fit the data significantly better, χ2(1) = 23.14, p < 0.001. This meant that persons differed in how steep the time on task slope was. In a next step, we examined if this variance would be explained by persons’ skills. We estimated a model M2sentence that, in addition to the random slope of time on task across persons, estimated the variance between the random time on task slope across persons and the random person intercept. This model indicated that random person intercept and time on task slope were negatively correlated at r = −0.38. As with the retrieval of word meanings, we tested this correlation for significance by comparing the fit of models M2sentence and M1psentence. This comparison indicated that M2sentence fit the data significantly better, χ2(1) = 7.65, p < 0.01. Combining the fixed time on task effect and a posteriori estimates of the random person specific deviations from the fixed time on task effect revealed that person-specific time on task effects were consistently negative with only 4 exceptions, and increasingly so with increasing skill (see Figure 3). This meant that especially for skilled sentence comprehenders, accurate responses were also fast. These results support hypothesis 5.

Figure 3
Scatter plot showing a negative correlation between the ability in sentence comprehension and person time on task (ToT) effect. Data points are scattered around a downward-sloping trend line, indicating that as the ability to comprehend sentences increases, the stronger the negative ToT effect becomes.

Figure 3. Person-specific time on task-effects in the sentence verification task. Continuous line: Linear regression of person-specific time on task effects on ability. Dashed line: Time on task effect of zero. Dotted line: fixed time on task effect.

Items. As in the previous analyses, we first examined whether time on task slopes varied across items. We compared a model M1isentence that specified a random time on task slope across items, but constrained the covariance between the random item slope and intercept to zero, to a model M0isentence that constrained the random time on task slope across items to be zero. This comparison indicated that model M1isentence fit the data significantly better than model M0isentence, χ2(1) = 23.95, p < 0.001. Model M2sentence specified the covariance between the random time on task slope across items and the random intercept as an additional parameter and revealed a negative correlation between the random slope and intercept of r = −0.42. Comparing this model’s fit to the fit of model M1isentence revealed that model M2sentence fit the data better than model M1isentence, χ2(1) = 4.64, p < 0.05. Combining the fixed time on task effect and the a posteriori estimates of the item-specific random time on task effects revealed that item-specific time on task effects were negative with one exception, but it was especially in easy items that fast responses were associated with correct responses (see Figure 4). In (relatively) difficult items, the item-specific time on task effects were negative as well, but to a smaller extent. These results replicated closely the pattern we found for the retrieval of word meanings, and support hypothesis 6.

Figure 4
Scatter plot illustrating the relationship between Item ToT (time on task) effect and item easiness of sentence comprehension. Data points show a negative correlation, with the line of best fit sloping downward from left to right. This indicates that as item easiness increases, the more negative the ToT effect becomes.

Figure 4. Item-specific time on task-effects in the sentence verification task. Continuous line: Linear regression of item-specific time on task effects on item easiness. Dashed line: Time on task effect of zero. Dotted line: fixed time on task effect.

4 Discussion

The present research investigated whether listening component tasks on the word (semantic categorization) and sentence levels (sentence verification) capture cognitive efficiency patterns that are characteristic of automatized processing. Specifically, we examined time on task effects as an indicator of whether response times systematically predict accuracy, reflecting the efficiency of semantic retrieval and integration processes during listening. In accordance with earlier research on automatized skills, we expected the tasks to show that quicker responses are associated with higher accuracy, and that this effect would be more pronounced for stronger comprehenders and easier items. The results confirmed the expected patterns, as quicker responses went along with higher accuracy for both the retrieval of word meanings and sentence-level semantic integration, thus indicating that the tasks captured meaningful variance in the efficiency of these processes.

As for the differences between individuals, person-level random intercepts and slopes were negatively correlated, implying that responding quickly was especially important for the accuracy of more skilled comprehenders. Again, this was the case for both word- and sentence level semantic processes. This result is consistent with the notion that the more routinized these skills become, the more they rely on quick and automatic processing. Although this was also the case for less accurate individuals in both semantic processes, their correlations were weaker in both tasks. For items, the negative correlation between the item-specific intercept and slope showed that quicker responses were more strongly correlated with higher accuracy when items had a higher mean accuracy, i.e., were easier to answer. Thus, especially for easy items, automatic processing was important for accurate responses. Again, this was the case for the semantic categorization task and the sentence verification task.

These results are in line with the idea that oral language comprehension on the word and sentence levels relies on automatized access to meaning rather than slow, deliberate processing. The observed time on task effects suggest that the two listening tasks successfully modeled this efficiency. Especially when comprehenders are more skilled or when the language input is easy to comprehend, it is important that language information is processed automatically with little cognitive effort. When word or sentence meanings were not retrieved correctly, i.e., when children did not understand the meaning of the words or sentences, they required more time to process the input before responding. This could result from a semantic search requiring a more deliberate cognitive effort and a higher strain on working memory, in line with the cognitive model of listening (Imhof, 2010) or the verbal efficiency theory (Perfetti, 1985). Nevertheless, this process took longer than the responses to correctly answered items in both tasks, even for children with lower accuracy, indicating that no matter the skill level of the children, access to the correct word or sentence meanings did not require as much cognitive effort. Therefore, in the present study with elementary school children, we could not detect different processing types as proposed for reading skills by the compensatory-encoding model (Walczyk, 2000), but listening skills at this age may have already been too advanced and the language input too easy to find these differences.

These results are nevertheless in line with studies on reading processes, where automatization on different levels has been shown to be an important predictor of successful comprehension (Roembke et al., 2019; Schindler and Richter, 2018). It can be argued that sentence- or text-level processes are not entirely automatic as these rely on memory to integrate prior knowledge or draw inferences, but the automaticity of lexical access and propositional encoding decrease the reliance on deliberate cognitive processing (Hui and Godfroid, 2021; Perfetti, 1985). Semantic integration in sentences relies on both of these processes, which in turn may mediate its efficiency.

Regarding differences between grade levels, a significant difference in accuracy was found between Grades 3 and 4, indicating a modest increase in listening efficiency toward the end of elementary school. However, no significant incremental gains were observed between the earlier grades. This contrasts with previous findings on similar processes in reading and listening, where larger developmental improvements in accuracy were found across the early school years (e.g., Dahdah et al., 2025; Richter et al., 2012; Verhoeven and Van Leeuwe, 2008). The smaller early gains in the present data may reflect cohort differences. Moreover, because the data are cross-sectional, developmental inferences should be made with caution.

In terms of the limitations of the present study, the assessments were limited to children in grades one through four. As listening comprehension – at least in one’s first language – begins to develop considerably earlier than elementary school (e.g., Fernald et al., 2006), semantic comprehension processes on the word and sentence levels may have already reached a point where they are predominantly automatized. It is reasonable to assume that semantic retrieval prior to elementary school may not yet occur automatically and require more cognitive effort, similar to reading processes (Walczyk, 2000; Walczyk et al., 2007). A look at these processes at earlier ages could thus give insight into when these processes switch from deliberate to automatic execution. Furthermore, the present analyses were based on two samples of children who each solved either the semantic categorization or the sentence verification task. Therefore, it was not possible to assess the correlations between response times and accuracy between both tasks. Similarly, it was not possible to directly compare the time on task effects between the word and sentence levels. Furthermore, we did not include additional measures such as intelligence or broader language abilities. Collecting these data was not feasible within the limited testing time available in the classroom setting. However, by including class and grade level into the analyses, we were able to at least in part account for variance related to these variables. Moreover, the items in the study were recorded in a studio environment and articulated clearly at a speech rate that was easy to follow for elementary school children. Whereas this was appropriate for examining whether our tasks capture efficiency under controlled conditions, it remains unclear how well the same efficiency indicators would perform in more realistic listening environments. Future studies could therefore incorporate more varied and challenging listening environments that include varying speech rates, less clear articulation, noise, or other signal degrading characteristics. This higher difficulty could put a higher burden on attentional resources (Murgia, 2024; Wild et al., 2012) but may also lead to a higher reliance on the automaticity of comprehension processes. Examining task performance under such conditions would help determine the generalizability of efficiency-based listening assessments.

In sum, the results of this study show that for auditory semantic processes on the word and sentence levels, shorter response times were systematically associated with higher accuracy, particularly among skilled comprehenders and for easier items. This pattern supports the assumption that our tasks capture cognitive efficiency in line with automatized processing, rather than reflecting deliberate or effortful processing.

The present results also have important implications for the assessment of listening component skills. They suggest that response latencies can provide valid information about the efficiency of semantic listening processes and can therefore complement accuracy-based measures in the evaluation of language proficiency. In other words, interpreting shorter response times as reflecting cognitive efficiency appears reasonable for these types of listening tasks. Future research should extend this approach to more complex listening skill components, such as coherence building or inference generation, and to younger children for which these processes are not yet automatized, in order to determine the advantages and limitations of efficiency-based assessments in listening comprehension.

Data availability statement

The datasets presented in this study and the analysis script needed to reproduce the reported analyses are openly available in an OSF repository at: https://osf.io/gcm3n.

Ethics statement

Ethical approval was not required for the study involving human samples in accordance with the local legislation and institutional requirements because the minimal-risk data collection required only institutional permission from the schools and parental authorization. Written informed consent for participation in this study was provided by the participants’ legal guardians/next of kin.

Author contributions

PD: Writing – original draft, Writing – review & editing, Conceptualization, Formal analysis, Methodology, Project administration, Data curation, Visualization. IO: Writing – review & editing, Project administration, Data curation. JN: Writing – original draft, Writing – review & editing, Conceptualization, Formal analysis, Methodology, Project administration, Data curation, Visualization.

Funding

The author(s) declared that no financial support was not received for this work and/or its publication.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that Generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Aryadoust, V. (2017). An integrated cognitive theory of comprehension. Int. J. Listen. 33, 71–100. doi: 10.1080/10904018.2017.1397519

Crossref Full Text | Google Scholar

Dahdah, P., Naumann, J., Haffmans, M.-B., Schindler, J., and Richter, T. (2025). Development of the retrieval of word meanings and its impact on reading comprehension throughout elementary school: a longitudinal study. J. Educ. Psychol. 117, 1285–1312. doi: 10.1037/edu0000962

Crossref Full Text | Google Scholar

DeStefano, D., and LeFevre, J.-A. (2007). Cognitive load in hypertext reading: a review. Comput. Hum. Behav. 23, 1616–1641. doi: 10.1016/j.chb.2005.08.012

Crossref Full Text | Google Scholar

Fernald, A., Perfors, A., and Marchman, V. A. (2006). Picking up speed in understanding: speech processing efficiency and vocabulary growth across the 2nd year. Dev. Psychol. 42, 98–116. doi: 10.1037/0012-1649.42.1.98,

PubMed Abstract | Crossref Full Text | Google Scholar

Field, J. (2009). Listening in the language classroom. Cambridge: Cambridge University Press.

Google Scholar

Florit, E., Roch, M., Altoè, G., and Levorato, M. C. (2009). Listening comprehension in preschoolers: the role of memory. Br. J. Dev. Psychol. 27, 935–951. doi: 10.1348/026151008X397189,

PubMed Abstract | Crossref Full Text | Google Scholar

Frensch, P. A., and Funke, J. (1995). Complex problem solving: the European perspective. Mahwah: Erlbaum.

Google Scholar

Frisson, S., Rayner, K., and Pickering, M. J. (2005). Effects of contextual predictability and transitional probability on eye movements during reading. J. Exp. Psychol. Learn. Mem. Cogn. 31, 862–877. doi: 10.1037/0278-7393.31.5.862

Crossref Full Text | Google Scholar

Goldhammer, F., Naumann, J., Stelter, A., Tóth, K., Rölke, H., and Klieme, E. (2014). The time on task effect in reading and problem solving is moderated by task difficulty and skill: insights from a computer-based large-scale assessment. J. Educ. Psychol. 106, 608–626. doi: 10.1037/a0034716

Crossref Full Text | Google Scholar

Günther, F., Dudschig, C., and Kaup, B. (2015). LSAfun - an R package for computations based on latent semantic analysis. Behav. Res. Methods 47, 930–944. doi: 10.3758/s13428-014-0529-0,

PubMed Abstract | Crossref Full Text | Google Scholar

Hui, B., and Godfroid, A. (2021). Testing the role of processing speed and automaticity in second language listening. Appl. Psycholinguist. 42, 1089–1115. doi: 10.1017/S0142716420000193

Crossref Full Text | Google Scholar

Imhof, M. (2010). “What is going on in the mind of a listener? The cognitive psychology of listening” in Listening and human communication in the 21st century. ed. A. D. Wolvin. 1st ed (Hoboken: Wiley), 97–126.

Google Scholar

Jiang, H., and Farquharson, K. (2018). Are working memory and behavioral attention equally important for both reading and listening comprehension? A developmental comparison. Read. Writ. 31, 1449–1477. doi: 10.1007/s11145-018-9840-y,

PubMed Abstract | Crossref Full Text | Google Scholar

Just, M. A., and Carpenter, P. A. (1992). A capacity theory of comprehension: individual differences in working memory. Psychol. Rev. 99, 122–149. doi: 10.1037/0033-295X.99.1.122,

PubMed Abstract | Crossref Full Text | Google Scholar

Just, M. A., Carpenter, P. A., and Keller, T. A. (1996). The capacity theory of comprehension: new frontiers of evidence and arguments. Psychol. Rev. 103, 773–780. doi: 10.1037/0033-295X.103.4.773,

PubMed Abstract | Crossref Full Text | Google Scholar

Karageorgos, P., Richter, T., Haffmans, M.-B., Schindler, J., and Naumann, J. (2020). The role of word-recognition accuracy in the development of word-recognition speed and reading comprehension in primary school: a longitudinal examination. Cogn. Dev. 56:100949. doi: 10.1016/j.cogdev.2020.100949

Crossref Full Text | Google Scholar

Kim, Y. (2015a). Language and cognitive predictors of text comprehension: evidence from multivariate analysis. Child Dev. 86, 128–144. doi: 10.1111/cdev.12293,

PubMed Abstract | Crossref Full Text | Google Scholar

Kim, Y.-S. G. (2015b). Developmental, component-based model of reading fluency: an investigation of predictors of word-reading fluency, text-reading fluency, and reading comprehension. Read. Res. Q. 50, 459–481. doi: 10.1002/rrq.107

Crossref Full Text | Google Scholar

Kim, Y.-S. G. (2016). Direct and mediated effects of language and cognitive skills on comprehension of oral narrative texts (listening comprehension) for children. J. Exp. Child Psychol. 141, 101–120. doi: 10.1016/j.jecp.2015.08.003,

PubMed Abstract | Crossref Full Text | Google Scholar

Kim, Y.-S. G., and Pilcher, H. (2016). “What is listening comprehension and what does it take to improve listening comprehension?” in Interventions in learning disabilities. eds. R. Schiff and R. M. Joshi, vol. 13 (Berlin: Springer International Publishing), 159–173.

Google Scholar

Kintsch, W. (2005). An overview of top-down and bottom-up effects in comprehension: the CI perspective. Discourse Process. 39, 125–128. doi: 10.1080/0163853X.2005.9651676

Crossref Full Text | Google Scholar

Kutas, M., and Iragui, V. (1998). The N400 in a semantic categorization task across 6 decades. Electroencephalogr. Clin. Neurophysiol./Evoked Potentials Section 108, 456–471. doi: 10.1016/S0168-5597(98)00023-9,

PubMed Abstract | Crossref Full Text | Google Scholar

Lervåg, A., Hulme, C., and Melby-Lervåg, M. (2018). Unpicking the developmental relationship between oral language skills and reading comprehension: it’s simple, but complex. Child Dev. 89, 1821–1838. doi: 10.1111/cdev.12861

Crossref Full Text | Google Scholar

Mayer, R. E. (1996). Learning strategies for making sense out of expository text: the SOI model for guiding three cognitive processes in knowledge construction. Educ. Psychol. Rev. 8, 357–371. doi: 10.1007/BF01463939

Crossref Full Text | Google Scholar

Mayer, R. E., and Wittrock, M. C. (2006). “Problem Solving” in Handbook of educational psychology. eds. P. A. Alexander and P. H. Winne. 2nd ed (Mahwah: Erlbaum), 287–304.

Google Scholar

McGraw, K. O., and Wong, S. P. (1996). Forming inferences about some intraclass correlation coefficients. Psychol. Methods 1, 30–46. doi: 10.1037/1082-989X.1.1.30

Crossref Full Text | Google Scholar

Murgia, S. (2024). “Chapter five—listening challenges in children: comprehension and effort in noisy and voice-degraded conditions” in Psychology of Learning and Motivation, ed. K. D. Federmeier, vol. 81 (Cambridge: Academic Press), 157–217.

Google Scholar

Nation, K., and Snowling, M. J. (1998). Semantic processing and the development of word-recognition skills: evidence from children with reading comprehension difficulties. J. Mem. Lang. 39, 85–101. doi: 10.1006/jmla.1998.2564

Crossref Full Text | Google Scholar

Naumann, J. (2018). Geschwindigkeit und Akkuratheit beim Lesen: Kommunalitäten über Leseaufgaben und Lebensalter [speed and accuracy in reading: communalities across tasks and age] [conference presentation]. Frankfurt am Main, Germany: DGPs Biannual Meeting.

Google Scholar

Naumann, J., Brandhuber, L., Goldhammer, F., Eichmann, B., and Greiff, S. (2025). For skilled comprehenders digital reading is a routine task, for unskilled comprehenders it is a problem. Learn. Individ. Differ. 122:102745. doi: 10.1016/j.lindif.2025.102745

Crossref Full Text | Google Scholar

Naumann, J., and Goldhammer, F. (2017). Time-on-task effects in digital reading are non-linear and moderated by persons' skills and tasks' demands. Learn. Individ. Differ. 53, 1–16. doi: 10.1016/j.lindif.2016.10.002

Crossref Full Text | Google Scholar

Naumann, J., Richter, T., Neeb, Y., Knoepke, J., and Isberner, M.-B. (2016), The time on task effect in high vs. low level reading tasks [conference presentation]. SSSR annual meeting, Porto, Portugal.

Google Scholar

Nicolay, B., Krieger, F., Stadler, M., Vainikainen, M.-P., Lindner, M. A., Hansen, A., et al. (2022). Examining the development of metacognitive strategy knowledge and its link to strategy application in complex problem solving – a longitudinal analysis. Metacogn. Learn. 17, 837–854. doi: 10.1007/s11409-022-09324-9

Crossref Full Text | Google Scholar

OECD (2012). Literacy, numeracy and problem solving in technology-rich environments: framework for the OECD survey of adult skills. Paris: OECD.

Google Scholar

Perfetti, C. A. (1985). Reading ability. Oxford: Oxford University Press.

Google Scholar

Perfetti, C. A. (1992). “The representation problem in reading acquisition” in Reading acquisition. eds. P. B. Gough, L. C. Ehri, and R. Treiman (Mahwah: Lawrence Erlbaum Associates, Inc), 145–174.

Google Scholar

Richter, T., Isberner, M. B., Naumann, J., and Kutzner, Y. (2012). Prozessbezogene Diagnostik von Lesefähigkeiten bei Grundschulkindern [process-based measurement of reading skills in primary school children]. Z. Pädagog. Psychol. 26, 313–331. doi: 10.1024/1010-0652/a000079

Crossref Full Text | Google Scholar

Richter, T., Naumann, J., Isberner, M.-B., Neeb, Y., and Knoepke, J. (2017). ProDi-L: Prozessbezogene Diagnostik von Lesefähigkeiten im Grundschulalter [ProDi-L: Process-oriented measurement of reading skills in primary school children]. Göttingen: Hogrefe.

Google Scholar

Richter, T., Schroeder, S., and Wöhrmann, B. (2009). You don't have to believe everything you read: background knowledge permits fast and efficient validation of information. J. Pers. Soc. Psychol. 96, 538–558. doi: 10.1037/a0014038,

PubMed Abstract | Crossref Full Text | Google Scholar

Roembke, T. C., Hazeltine, E., Reed, D. K., and McMurray, B. (2019). Automaticity of word recognition is a unique predictor of reading fluency in middle-school students. J. Educ. Psychol. 111, 314–330. doi: 10.1037/edu0000279

Crossref Full Text | Google Scholar

Rouet, J.-F., Britt, M. A., and Durik, A. M. (2017). RESOLV: readers' representation of reading contexts and tasks. Educ. Psychol. 52, 200–215. doi: 10.1080/00461520.2017.1329015

Crossref Full Text | Google Scholar

Scherer, R., Greiff, S., and Hautamäki, J. (2015). Exploring the relation between time on task and ability in complex problem solving. Intelligence 48, 37–50. doi: 10.1016/j.intell.2014.10.003

Crossref Full Text | Google Scholar

Schindler, J., and Richter, T. (2018). “24. Reading comprehension: individual differences, disorders, and underlying cognitive processes” in Handbook of communication disorders: theoretical, empirical, and applied linguistic perspectives. eds. E. Dattner and D. Ravid (Berlin: De Gruyter Mouton), 503–524.

Google Scholar

Schneider, W., Blanke, I., Faust, V., and Küspert, P. (2011). WLLP-R Würzburger Leise Leseprobe – Revision [Würzburg silent reading probe -revision]. Göttingen: Hogrefe.

Google Scholar

Schneider, W., and Shiffrin, R. M. (1977). Controlled and automatic human information processing: I. detection, search, and attention. Psychol. Rev. 84, 1–66. doi: 10.1037/0033-295X.84.1.1

Crossref Full Text | Google Scholar

Shiffrin, R. M., and Schneider, W. (1977). Controlled and automatic human information processing: part II. Perceptual learning, automatic attending and a general theory. Psychol. Rev. 84, 127–190. doi: 10.1037/0033-295X.84.2.127

Crossref Full Text | Google Scholar

Shtyrov, Y. (2010). Automaticity and attentional control in spoken language processing: neurophysiological evidence. Ment. Lexicon 5, 255–276. doi: 10.1075/ml.5.2.06sht

Crossref Full Text | Google Scholar

Stanovich, K. E., Nathan, R. G., West, R. F., and Vala-Rossi, M. (1985). Children's word recognition in context: spreading activation, expectancy, and modularity. Child Dev. 56, 1418–1428. doi: 10.2307/1130461

Crossref Full Text | Google Scholar

Taikh, A., Hargreaves, I. S., Yap, M. J., and Pexman, P. M. (2015). Semantic classification of pictures and words. Q. J. Exp. Psychol. 68, 1502–1518. doi: 10.1080/17470218.2014.975728,

PubMed Abstract | Crossref Full Text | Google Scholar

Thompson-Schill, S. L., and Gabrieli, J. D. E. (1999). Priming of visual and functional knowledge on a semantic classification task. J. Exp. Psychol. Learn. Mem. Cogn. 25, 41–53. doi: 10.1037/0278-7393.25.1.41,

PubMed Abstract | Crossref Full Text | Google Scholar

Vandergrift, L. (1999). Facilitating second language listening comprehension: acquiring successful strategies. ELT J. 53, 168–176. doi: 10.1093/elt/53.3.168

Crossref Full Text | Google Scholar

Verhoeven, L., and Van Leeuwe, J. (2008). Prediction of the development of reading comprehension: a longitudinal study. Appl. Cogn. Psychol. 22, 407–423. doi: 10.1002/acp.1414

Crossref Full Text | Google Scholar

Vörös, Z., Kehl, D., and Rouet, J.-F. (2021). Task characteristics as source of difficulty and moderators of the effect of time-on-task in digital problem-solving. J. Educ. Comput. Res. 58, 1494–1514. doi: 10.1177/0735633120945930

Crossref Full Text | Google Scholar

Walczyk, J. J. (2000). The interplay between automatic and control processes in reading. Read. Res. Q. 35, 554–566. doi: 10.1598/RRQ.35.4.7

Crossref Full Text | Google Scholar

Walczyk, J. J., Wei, M., Grifith-Ross, D. A., Goubert, S. E., Cooper, A. L., and Zha, P. (2007). Development of the interplay between automatic processes and cognitive resources in reading. J. Educ. Psychol. 99, 867–887. doi: 10.1037/0022-0663.99.4.867

Crossref Full Text | Google Scholar

Wild, C. J., Yusuf, A., Wilson, D. E., Peelle, J. E., Davis, M. H., and Johnsrude, I. S. (2012). Effortful listening: the processing of degraded speech depends critically on attention. J. Neurosci. 32, 14010–14021. doi: 10.1523/JNEUROSCI.1528-12.2012,

PubMed Abstract | Crossref Full Text | Google Scholar

Wimmer, H., and Mayringer, H. (2014). Salzburger lese-screening für die Schulstufen 2–9 [Salzburg reading screening for the grades 2–9]. Göttingen: Hogrefe.

Google Scholar

Keywords: time on task, listening comprehension, cognitive efficiency, semantic processing, automaticity

Citation: Dahdah P, Osipov I and Naumann J (2025) Automatic for the kids: assessment of cognitive efficiency in elementary school children’s auditory semantic processing. Front. Psychol. 16:1664852. doi: 10.3389/fpsyg.2025.1664852

Received: 12 July 2025; Revised: 17 November 2025; Accepted: 27 November 2025;
Published: 19 December 2025.

Edited by:

João Marôco, Lusofona University, Portugal

Reviewed by:

Yan Wu, Northeast Normal University, China
Kathryn Toffolo, University of Rochester, United States

Copyright © 2025 Dahdah, Osipov and Naumann. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Patrick Dahdah, ZGFoZGFoQHVuaS13dXBwZXJ0YWwuZGU=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.