SYSTEMATIC REVIEW article

Front. Psychol., 20 June 2025

Sec. Psychology of Language

Volume 16 - 2025 | https://doi.org/10.3389/fpsyg.2025.1605898

This article is part of the Research TopicReviews in Psychology of Language - Volume IIView all articles

A cross-age odyssey of cognitive reading attributes: a scoping review

  • 1Faculty of Education, Universiti Kebangsaan Malaysia, Bangi, Selangor, Malaysia
  • 2Research Centre of Education Leadership and Policy, Faculty of Education, Universiti Kebangsaan Malaysia, Bangi, Selangor, Malaysia
  • 3Centre of Teaching and Learning Innovation, Faculty of Education, Universiti Kebangsaan Malaysia, Bangi, Selangor, Malaysia

Reading is a fundamental cognitive-linguistic process that involves the dynamic interaction of multiple interrelated cognitive and perceptual mechanisms. Existing reading models are often limited in fully capturing the intricate relationships between reading attributes across different age groups. This review aims to compare the cognitive attributes utilized in reading assessments for young and adult readers. Using a scoping review methodology, the study analyzed 47 empirical studies selected through a systematic search of reputable academic databases, Scopus, Web of Science, and Google Scholar, from an initial pool of 331 publications. The selection followed a screening process based on three inclusion and exclusion criteria: types of publications, language, and skills assessed. A distinctive pattern emerges in the assessment of cognitive reading skills across age groups, with adult readers evaluated on a wider array of attributes that encompass both fundamental skills and higher order cognitive abilities. In contrast, young readers' assessments tend to center on a narrower spectrum of subskills, primarily emphasizing literal and interpretive comprehension. This developmental pattern calls for the refinement of existing assessment models to better capture the progressive nature of cognitive reading development. Tailoring assessment tools and instructional strategies to align with learners' cognitive demands is imperative for internal stakeholders, while external stakeholders are urged to develop age-appropriate assessments. Future research could address the study's limitations by exploring advanced technologies, such as eye tracking, conducting rigorous reviews, performing cross-linguistic comparative studies, and evaluating diverse assessment methods to enhance the accuracy, effectiveness, and generalizability of reading assessments across various contexts.

1 Introduction

Reading is a fundamental cognitive-linguistic process that involves the dynamic interaction of multiple interrelated cognitive and perceptual mechanisms. As a core aspect of language comprehension, reading requires the coordination of lexical access, working memory, syntactic parsing, and semantic integration, enabling individuals to construct meaning from written texts (Hosseini Alast and Baleghizadeh, 2024). Unlike other polytomous language skills such as speaking and writing, which involve productive language use and require rater involvement for assessment (Mohd Noh and Mohd Matore, 2022, 2020), reading is a visually mediated process that necessitates efficient orthographic, phonological, and semantic activation to support fluent comprehension (Teng and Mizumoto, 2024). The psychological mechanisms underlying reading are deeply interconnected with broader cognitive functions, including executive control, attentional regulation, and memory retrieval, which contribute to individual differences in reading proficiency. Due to its complexity and the diverse theoretical perspectives that frame it, including developmental, cognitive, pedagogical, and computational frameworks, reading requires a comprehensive assessment approach capable of capturing its full range of cognitive attributes and skills. Within the psychology of language, multiple theoretical frameworks have been proposed to explain how cognitive and linguistic processes converge during reading. Classic reading models, such as psycholinguistic reading model (Goodman, 1967) and information processing model (Gough, 1972) provide insights into how readers derive meaning from texts. More recent perspectives such as the interactive-compensatory model (Stanovich, 1980). More recent models, like the cognitive processing model (Khalifa and Weir, 2009) and the direct and indirect effects model of reading (Kim, 2020), emphasize the hierarchical and interactive nature of reading, where multiple cognitive and contextual skills directly and indirectly contribute to comprehension.

While these models offer valuable insights into reading mechanisms, they often overlook developmental variations in cognitive reading processes. The way young and adult readers engage with text differs not only in terms of linguistic complexity but also in the cognitive operations employed during comprehension. Research has shown that adult readers rely more on higher-order cognitive functions, such as inference-making and semantic integration, while young readers focus more on basic decoding and literal comprehension (Sieo et al., 2022). However, existing reading models do not always explicitly address the gradual shifts in cognitive reading attributes throughout the lifespan. To address this gap, this review examines how cognitive reading attributes develop across different age groups by synthesizing findings from empirical studies. Another key limitation in traditional reading models is their reliance on broad theoretical constructs rather than fine-grained, diagnostic insights into individual cognitive mechanisms. Cognitive Diagnostic Assessment (CDA) emerges as an innovative framework capable of providing such granularity, offering a psychometric approach that systematically examines specific cognitive reading attributes. Distinct from conventional reading assessments, which yield general single proficiency scores, CDA enables a detailed diagnosis of strengths and weaknesses in core reading processes. This methodological advancement aligns with contemporary perspectives in reading cognition, developmental psycholinguistics, and the psychology of language by emphasizing a learner-specific approach to reading assessment. Despite its potential, research integrating CDA into developmental reading studies remains underexplored. While existing CDA studies have successfully retrofitted high-stakes assessments to diagnose reading attributes (Safari and Ahmadi, 2023), fewer studies have systematically applied CDA to trace the evolution of cognitive reading attributes across different age groups. Thus, the current review aims to identify and compare the reading attributes used in previous research to measure the reading skills of individuals from different age groups.

2 Literature review

2.1 Reading models

The diversity of reading models reflects various theoretical perspectives on how individuals engage with and comprehend texts. Table 1 presents the different models of reading (Goodman, 1967). Early conceptualization of reading as a psycholinguistic guessing game suggests that readers rely on contextual cues, linguistic structure, and prior knowledge to make predictions during reading, emphasizing the role of higher-level cognitive processes in guiding comprehension. This model exemplifies a top-down approach to reading, where meaning is primarily constructed through the reader's background knowledge and cognitive expectations rather than solely relying on the text itself. In contrast, Gough's (1972) and LaBerge and Samuels's (1974) stage-based models also known as the information processing model describe reading as a sequential process where readers decode text progressively, from letters to words to complete comprehension, highlighting the role of automaticity in word recognition and information processing. These models represent a bottom-up approach to reading, where comprehension is built from the accurate and systematic decoding of smaller units of text, such as letters and words before higher-level meaning can be constructed.

Table 1
www.frontiersin.org

Table 1. Types of reading models and their characteristics.

Apart from these stage-based reading models, component-based models, such as the sociocultural reading model, describe how a reader's background knowledge, conceptual abilities, and process strategies work together to create comprehension (Coady, 1979). Similarly, the Simple View of Reading conceptualizes reading comprehension as the product of two core components: decoding and linguistic comprehension, emphasizing that both are necessary for proficient reading (Hoover and Gough, 1990). This model highlights that weaknesses in either decoding or comprehension can hinder overall reading ability, underscoring the importance of balanced instruction targeting both skills. Meanwhile, the interactive reading model integrates both top-down and bottom-up processes, emphasizing a compensatory mechanism where strengths in one cognitive process balance weaknesses in another (Stanovich, 1980; Rumelhart, 1977). Rumelhart's (1977) interactive model synthesizes these perspectives by proposing that reading involves a dynamic interaction of both text-driven (perceptual) and knowledge-driven (cognitive) processes occurring simultaneously. Whereas, Stanovich's (1980) interactive-compensatory model extends the idea by suggesting that strengths in one cognitive process, such as contextual knowledge, can offset weaknesses in another, such as phonological decoding.

Expanding on these foundational models, the cognitive processing model of reading (Khalifa and Weir, 2009) integrates cognitive psychology perspectives to describe reading as a dynamic process involving both bottomup and top-down processing. It emphasizes the importance of goal setting during reading and distinguishes between local comprehension (understanding specific details) and global comprehension (grasping overall meaning), illustrating the complex interplay of cognitive skills required for effective reading. Similarly, the integrated reading model, also referred to as the Direct and Indirect Effects Model of Reading (DIER; Kim, 2020) proposes reading comprehension as an outcome of hierarchical, dynamic, and interactive contributions from multiple language, cognitive, and contextual skills. This model explains how some skills, such as word recognition, directly impact comprehension, while others, such as working memory and inferencing, provide indirect support by facilitating the integration of information. These models collectively aim to explain the complex, multi-stage nature of reading, emphasizing the interaction of perceptual, cognitive, and contextual mechanisms rather than rigid classifications, while acknowledging the dynamic and adaptive nature of reading comprehension.

2.2 Reading taxonomies and cognitive structures

Reading taxonomies provide systematic classification framework that organizes and categorizes the various cognitive skills, processes, and components involved in reading comprehension. These reading taxonomies underscore the varied approaches to categorizing and assessing the intricate skills involved in reading, distinguishing them from reading models, which focus on the theoretical frameworks of the reading process itself. Each taxonomy emphasizes different aspects of reading comprehension and cognitive processes. Reading taxonomy by Davis (1968) focuses on sub-skills like recalling word meanings, making inferences, and understanding the author's technique, which was empirically tested on students. Munby's (1978) taxonomy, often used in curriculum development, outlines 19 sub-skills, including identifying main ideas and interpreting text through various cohesion devices, though it has been critiqued for lacking empirical validation. Heaton's (1991) taxonomy provides a detailed list of 14 sub-skills, such as deducing word meaning and understanding conceptual meaning, aimed at assessing specific reading skills. Meanwhile, Hughes and Hughes's (2020) divide reading skills into expedient and careful reading, further categorizing them into sub-skills like identifying discourse markers and interpreting complex sentences. Whereas, Anderson and Krathwohl (2001) updated version of Bloom's Taxonomy introduces a cognitive dimension to reading by classifying skills into six levels: remembering, understanding, applying, analyzing, evaluating, and creating. This hierarchical model is echoed in Luebke and Lorié (2013) divides reading skills into four sub-skills: identification, understanding and analyzing, making inferences, and application. Each of these taxonomies provides a different lens through which reading comprehension can be assessed, from the recall of explicit information to higher-order cognitive skills such as critical thinking and application. Collectively, these taxonomies highlight the complexity of reading comprehension and offer valuable tools for educators and researchers to assess and enhance reading skills in a structured and systematic manner.

2.3 The exploration of reading attributes through cognitive diagnostic assessment

An attribute refers to a specific cognitive skill or ability that contributes to task performance, particularly in educational contexts (Wang et al., 2021). Cognitive attributes are the foundational components underlying a learner's ability to process, interpret, and respond to tasks, making them essential units for understanding academic performance (Zhang et al., 2024). Meanwhile, cognitive reading attributes are the cognitive skills involved in the process of reading and comprehending text. These attributes encompass a range of cognitive processes, including word recognition, decoding, inferencing, summarizing, comprehension monitoring, and understanding text structures (Li et al., 2021). Together, they influence a reader's ability to extract, interpret, and construct meaning from written material. Understanding these attributes is crucial for assessing reading proficiency and guiding instructional strategies tailored to individual learner needs. The underlying sub-skills of reading have been explored in previous research studies using various approaches, including factor analysis (Nightingale et al., 2023; Lin, 2024), structural equation modeling (SEM) (Lee and Lee, 2023; Peng et al., 2021), item response theory (IRT) (Geramipour et al., 2021; Polat et al., 2022), and cognitive diagnostic modeling (CDM) (Askari and Karami, 2024; Chen et al., 2023), each providing unique insights into the cognitive processes and relationships among reading components.

Cognitive reading attributes have been investigated within CDA principles by examining how specific cognitive skills contribute to reading comprehension and how they can be accurately measured (Mei and Chen, 2022). CDA represents a modern advancement in educational measurement by integrating psychological and measurement theories that offers insights into students' cognitive processes and addresses the limitations of conventional assessment methods (Maas et al., 2024). CDA provides a holistic and contextual understanding of individual cognitive processes by diagnosing specific cognitive reading attributes, drawing on insights from psychological theories and advanced measurement techniques (Ketabi et al., 2021). Other approaches often provide a narrow view of students' cognitive capabilities, falling short to diagnostically capture the intricacies of reading comprehension and the diverse cognitive reading attributes involved (von Davier and Lee, 2019).

The evolution of CDA is deeply rooted in the historical development of educational psychology and measurement theories. Early educational psychology prioritized intelligence measurement through specific tests, focusing on intellectual and academic skills (van der Linden, 2016). However, as understanding of human cognition expanded, theories such as cognitive development theory (Piaget, 1950) and information processing theory (Miller, 1956) highlighted the importance of cognitive processes like information handling, storage, and retrieval—key elements in reading comprehension. Concurrent advancements in measurement theories supported a more holistic and contextual approach to assessing these cognitive processes (Miller and Lovler, 2020). This integration ultimately led to the development of CDA models, which enable the exploration of cognitive reading attributes across various contexts, including language skills (Sessoms and Henson, 2018).

One common method used within CDA studies for investigating reading attributes is the retrofitting of existing high-stakes assessments for diagnostic purposes. Retrofitting involves reanalyzing existing reading tests to ensure they provide relevant and useful diagnostic information (Sessoms and Henson, 2018). This approach has been applied to large-scale assessments such as the Program for International Student Assessment (PISA; Chen and Chen, 2016) and the Test of English as a Foreign Language (TOEFL; Safari and Ahmadi, 2023), among others. Retrofitting allows educators to gain diagnostic insights from assessments not originally designed for this purpose, although challenges remain, such as ensuring a sufficient number of items to measure specific cognitive reading attributes accurately (Gierl and Cui, 2008). Despite the efficiencies of retrofitting, some researchers advocate for the development of entirely new CDA items and assessments tailored specifically to measure reading attributes. Although more resource intensive, this approach allows for a more detailed and personalized analysis of learners' cognitive strengths and weaknesses in reading. For example, a study in Turkey developed a diagnostic reading test for university students (Toprak and Cakir, 2021). Developing new CDA items ensures that the assessments are specifically designed to measure key cognitive reading attributes, such as decoding, inferencing, and comprehension, providing more precise diagnostic information and supporting targeted instructional interventions (Ketabi et al., 2021).

3 Methods

3.1 Design

The current study employed a scoping review method as it is appropriate for gaining an in-depth understanding of a broad issue and charting existing research to better inform future studies (Munn et al., 2022). The key difference between a scoping review and a systematic literature review is that while a systematic review focuses on synthesizing specific evidence to answer a well-defined research question (Sabtu and Mohd Matore, 2024), a scoping review aims to map the existing literature on a broader topic, identifying gaps and trends without necessarily assessing the quality of individual studies (Nadmilail et al., 2023). It is particularly useful for analyzing emerging evidence when it is uncertain if more specific questions can be effectively addressed by a precise systematic review (Pollock et al., 2021). Scoping reviews are ideal for evaluating and understanding the extent of knowledge in a developing field, as well as for identifying, mapping, reporting, or discussing the characteristics and concepts within that area (Peters et al., 2022). The scoping review was conducted based on the guidelines recommended by Levac et al. (2010) and Arksey and O'Malley (2005) which involves five stages: (i) identifying the research objectives, (ii) identifying relevant studies, (iii) selecting the studies, (iv) charting the data, and (v) collating, summarizing, and reporting the results.

3.2 Review protocol

The review was guided by the Preferred Reporting Items for Systematic reviews and Meta-Analyses extension for Scoping Reviews (PRISMA-ScR). The review process included the identification of relevant studies, their selection based on predefined criteria, and the abstraction and synthesis of key data. All screening and selection processes were conducted independently by two reviewers, following the PRISMA-ScR guidelines to minimize bias and ensure the reliability of the findings.

This rigorous process ensures the inclusion of high-quality evidence to address the research questions posed in this study.

3.3 Research question formulation

The research questions were formulated based on the Population, Interest and Context (PICo) framework (Lockwood et al., 2015). The PICo framework aids researchers in crafting appropriate research questions for reviews by focusing on the core elements (Mohamed Shaffril et al., 2021). The current review has applied these components by examining, (i) Population: readers across different age groups, (ii) Interest: cognitive reading attributes, and (iii) Context: cognitive diagnostic assessment studies. As a result, the review is guided by three research questions:

1. What are the publication trends in cognitive diagnostic assessment studies on reading, specifically in relation to test types, sample demographics and age distributions?

2. How do cognitive diagnostic assessment studies conceptualize and operationalize key reading attributes for young and adult readers?

3. How do the similarities and differences in cognitive reading attributes assessed in young and adult readers reflect their cognitive and linguistic development?

3.4 Database

The review process began with the development of a systematic search strategy. This involved searching through three reputable academic databases, Scopus, Web of Science and Google Scholar. These databases have been used by previous researchers in reviewing the latest publication trends in various fields, providing wide coverage of peer-reviewed journals, conference papers, and other scholarly works (Sakaria et al., 2023). These databases were selected due to their extensive reach and credibility, ensuring that the review captures a broad spectrum of relevant and up-to-date research findings (Sabtu and Mohd Matore, 2024). Scopus, an esteemed database, aggregates content from over 30,000 journals representing a wide range of subject fields, including education, measurement, and language, enabling researchers to identify trends, gaps, and advancements in the field of study (Masdoki et al., 2021). This database, encompassing contributions from more than 11,000 publishers, undergoes a rigorous peer-review process. On the other hand, the Web of Science (WoS) is a platform providing access to a robust database across numerous academic disciplines (Singh et al., 2021). Originally established by the Institute for Scientific Information (ISI) and currently monitored by Clarivate Analytics, WoS covers over 30,000 journals spanning more than 250 disciplines (Birkle et al., 2020). Meanwhile, Google Scholar provides extensive access to a wide range of academic literature, including peer-reviewed articles, theses, books, and conference papers across various disciplines (Martín-Martín et al., 2021). Additionally, Google Scholar's extensive indexing of both traditional and gray literature ensures a thorough and inclusive search of relevant studies.

3.5 Systematic search strategies

This stage involves three steps; identification, screening, and eligibility, as illustrated in Figure 1.

Figure 1
www.frontiersin.org

Figure 1. PRISMA flowchart of the study.

3.5.1 Identification

The identification process begins with selecting appropriate keywords. To maximize the retrieval of relevant articles for the current review, the keywords are categorized into two groups: reading assessment and cognitive diagnostic assessment. Related terms for each keyword were identified and shown in Table 2. By combining the keywords and related terms, the search process was conducted on the three databases with respective search strings as displayed in Table 3. The search on Scopus and Web of Science used a structured string incorporating Boolean operators, proximity operators, and wildcard symbols * to ensure relevant results. Boolean operators (OR) broadened the search by including synonyms like “reading assess*” OR “reading test*” OR “reading evaluat*”. Proximity operators (W/) refined the results by ensuring terms like “assess* W/reading” and “test* W/reading” appeared close together, maintaining relevance. Wildcards (*) captured word variations, such as “assess*” for “assess”, “assessment”, or “assessed”. Keywords like “cognitive diagnos*” and “classification AND diagnos*” ensured a focus on cognitive and diagnostic aspects. This strategy effectively targeted studies on reading assessment and cognitive diagnostics. Whereas, the search of the Google Scholar database was conducted using the Publish or Perish version 8.17 for Windows software, a tool designed to retrieve and analyze academic citations. This software allowed for efficient querying of Google Scholar using the same structured keywords and phrases as the Scopus and Web of Science searches. The search yielded a total of 331 publications, with 73 articles retrieved from Scopus, 68 from Web of Science, and 190 from Google Scholar.

Table 2
www.frontiersin.org

Table 2. Keywords and related terms.

Table 3
www.frontiersin.org

Table 3. Search string.

3.5.2 Screening

Following the identification stage, 78 duplicate records were removed, leaving 253 records for initial screening. During the screening stage, the records were assessed for relevance based on predefined inclusion and exclusion criteria as presented in Table 4. To ensure transparency and reduce the risk of selection bias, two independent reviewers conducted the title and abstract screening. Discrepancies between the reviewers were resolved through discussion until consensus was reached. If disagreements persisted, a third reviewer was consulted. The first criterion was the types of publications. Only publications on empirical studies were selected but not empirical studies using simulation data and other types of studies, including conceptual and review studies. The second criterion applied was language. Only publications in the English language were accepted. As a result, 21 records were excluded, with 12 records removed due to inappropriate paper types such as conceptual papers and book chapters, while nine records were excluded due to language limitations such as Chinese language and Persian language. This reduced the number of reports to 232, which were then sought for retrieval. However, three reports could not be retrieved, leaving a total of 229 reports for further evaluation.

Table 4
www.frontiersin.org

Table 4. Screening criteria.

3.5.3 Eligibility

In the eligibility stage, the remaining reports were assessed in detail to determine their suitability for the review. Again, two reviewers independently assessed the full-text articles using the predefined inclusion and exclusion criteria. This dual-reviewer approach enhanced the methodological rigor of the review process. A total of 181 reports were excluded for various reasons: 35 were review papers, 33 focused on different skills, 83 addressed subjects outside the scope of the review such as Mathematics, diagnostic assessment and computerized adaptive testing (CAT), 19 were statistical reviews, and 11 followed different study types that did not meet the inclusion criteria, such as action research. Following this rigorous screening and eligibility process, 47 studies were deemed relevant and included in the final review. This multi-step process ensured the inclusion of only the most pertinent and high-quality studies, allowing for a robust and reliable scoping review of cognitive reading attributes across different age groups. The next step involved graphically presenting the selected articles. Summaries were created for each article, detailing the author, year, title of the study, methodologies, population, samples and test names. The final phase of this review framework included compiling, summarizing, and reporting the results. All data were stored and processed using Mendeley Desktop Version 1.19.8. A synthesis of the literature was then created by summarizing the key points and presenting them in text, tables, and figures. All 47 publications were thoroughly retrieved and reviewed in their entirety to ensure they met the inclusion criteria.

3.6 Systematic coding

Each publication was coded based on six variables: study design, test names, samples, sample types, sample ages, and cognitive attributes. Table 5 outlines the definitions and possible codes for each variable. The study design refers to the scientific framework of the study, categorized as either a retrofitting study or a true study. Test names indicate the reading assessments used, including wellknown standardized tests such as the Progress in International Reading Literacy Study (PIRLS) and the Program for International Student Assessment (PISA), as well as newly developed reading tests. The samples represent the test-takers, which include primary school students, high school students, undergraduate students, and postgraduate students. Sample types further classify test-takers based on their reading development stage, distinguishing between young and adult readers. Meanwhile, sample age specifies whether the exact age of the participants is mentioned or not. Lastly, cognitive attributes refer to the specific reading skills assessed in each study, such as finding explicit information (EXP), making inferences (INF), generalizing or synthesizing main ideas (GEN), interpreting texts (INT), lexical or vocabulary knowledge (LEX), and evaluating or analyzing texts (EVA). By coding publications according to these variables, the review systematically organizes reading research, allowing for a structured analysis of trends, patterns, and gaps in the field. This approach enhances the understanding of how reading skills have been assessed across various studies and contexts. The preparation of data for presentation and synthesis, including handling missing summary statistics, data conversions, and visual display of individual study results, was conducted using Microsoft Excel 2021.

Table 5
www.frontiersin.org

Table 5. Article coding for the present review.

4 Results

4.1 RQ1: What are the publication trends in cognitive diagnostic assessment studies on reading, specifically in relation to test types, sample demographics and age distributions?

The analysis identified 11 articles focusing on young readers, with each study utilized different types of reading tests. Table 6 shows the reading tests and sample profiles used in the studies. Only one study developed new reading assessment items (Nallasamy and Khairani, 2022), while the other 10 studies retrofitted existing non-diagnostic reading assessment items to extract the diagnostic reports such as from a large-scale K−12 ELP assessment measuring the social and academic language (Yumsek, 2023), Progress in International Reading Literacy Study (PIRLS; Toprak-yildiz, 2021), Program for International Student Assessment (PISA) (Chen and Chen, 2016), Grade 6 provincial reading achievement assessment (Jang et al., 2013, 2015), Grade 5 and 6 reading comprehension test (Liu and Bian, 2021), and the National Matriculation English Test (NMET) (Fan and Yan, 2020). In regards to the types of the samples, seven of these studies focused on primary school students, while the remaining four targeted secondary school students, as shown in Figure 2. Among the primary school samples, four studies involved students aged ten (Nallasamy and Khairani, 2022; Thi and Loye, 2019), and three studies involved students aged 12 (Jang et al., 2013, 2015). Regarding the secondary school samples, two studies focused on students aged 15 (Chen and Chen, 2016, 2015), while one study each involved students aged 17 (Fan and Yan, 2020) and 18 (Yumsek, 2023), respectively.

Table 6
www.frontiersin.org

Table 6. Reading tests and sample profiles in CDA studies involving young readers.

Figure 2
www.frontiersin.org

Figure 2. Publication counts based on sample types and ages.

A total of 36 articles sampling adult readers were further analyzed after the eligibility procedure. The reviewed studies employed a variety of reading tests and sample groups. Table 7 depicts the names of the reading tests, the types and the age of the samples. Findings reveal that only five studies developed new reading assessment items (Askari and Karami, 2024; Toprak and Cakir, 2021; Alavi and Ranjbaran, 2018; Liu and Read, 2023; Toprak-Yildiz, 2022). In contrast, the remaining 31 studies employed a retrofitting approach by adapting existing tests to analyze cognitive attributes of reading skills and to produce diagnostic profiling reports using different types of tests such as high-stakes language proficiency test (Boori et al., 2023), test for English Majors (Chen et al., 2023), International English Language Testing System (IELTS) (Jang et al., 2023) and high-stakes university entrance examination (Tonekaboni et al., 2021). Meanwhile, the study samples represent five distinct groups: undergraduate students, postgraduate students, undergraduate student candidates, postgraduate student candidates, and specific international examination test takers. While the majority of the studies did not report the ages of the samples, 12 studies did provide age ranges. None of these studies focused on a single age group; instead, the samples encompassed various age ranges. The youngest participants were 18 years old (Toprak-Yildiz, 2022) and the oldest was 59 years old (Jang et al., 2023).

Table 7
www.frontiersin.org

Table 7. Reading tests and sample profiles in CDA studies involving adult readers.

Figure 3 displays the publication counts categorized by test types and sample types. Among the five studies that developed new assessment items, all utilized undergraduate students (Toprak and Cakir, 2021; Alavi and Ranjbaran, 2018), with one study also combining undergraduate and postgraduate students (Askari and Karami, 2024). This kind of study, also known as true CDA studies, undergoes a rigorous process of designing cognitive models, constructing Qmatrix, developing and validating new items and finally profiling test-takers mastery classes. For retrofitting studies that adapted existing tests, 15 studies involved undergraduate students (Chen et al., 2023), seven involved postgraduate students (Boori et al., 2024), two focused on undergraduate student candidates (Hemati and Baghaei, 2020) and, respectively, five studies targeted postgraduate student candidates (Javidanmehr et al., 2019) and specific examination test takers (Jang et al., 2023). Retrofitting is ubiquitous in CDA studies due to the need to adapt existing assessments to align with cognitive diagnostic models, allowing researchers to extract detailed information about the test-takers strengths and weaknesses without developing entirely new tests (Mohd Noh and Mohd Matore, 2024).

Figure 3
www.frontiersin.org

Figure 3. Publication counts based on test types and samples.

4.2 RQ2: How do cognitive diagnostic assessment studies conceptualize and operationalize key reading attributes for young and adult readers?

The analysis of CDA studies reveals distinct approaches to defining and measuring reading attributes for both young and adult readers. Across the reviewed studies, reading attributes are conceptualized based on cognitive processing theories and language comprehension models. Based on the coding process, 13 codes of cognitive reading attributes emerged with at least one occurrence. Table 8 highlights the essential reading attributes employed in CDA studies to evaluate the reading skills of young readers. Topping the list are finding explicit information (EXP) and making inferences (INF). These are followed by generalizing or synthesizing main ideas (GEN), making interpretation of texts (INT), lexical or vocabulary knowledge (LEX), evaluating or analyzing the text (EVA), grammatical knowledge (GRAM), syntactic knowledge (SYN) and summarizing (SUM). Finally, others represent four other attributes with <3 occurrences, including sequencing, understanding charts, expressing in written forms and recognizing authors' styles.

Table 8
www.frontiersin.org

Table 8. Key reading attributes of young readers in CDA studies.

Table 9 presents key reading attributes of adult readers assessed in CDA studies. Regarding the number of attributes used to assess reading skills among adult readers, they ranged between 3 (Jang et al., 2023) and 10 (Kim, 2015). The coding process has produced 21 codes with at least one occurrence of attributes used in the reviewed studies to measure the reading skills of adult readers. The key attributes are led by lexical knowledge (LEX), making inferences (INF) and finding explicit information (EXP). These are followed by syntactic knowledge (SYN), summarizing ideas (SUM), grammar knowledge (GRAM), generalizing or synthesizing (GEN), identifying main ideas (MI), making interpretation (INT), connecting ideas (CON), recognizing authors' styles (AUT), scanning or skimming (SCAN), understanding paragraphs (PARA), evaluating (EVA), building a mental model (BUI), identifying supporting details (DET). Others refer to other attributes with <3 occurrences including negation, applying background knowledge, predicting, pragmatic knowledge and holding memory.

Table 9
www.frontiersin.org

Table 9. Key reading attributes of adult readers assessed in CDA studies.

4.3 RQ3: How do the similarities and differences in cognitive reading attributes assessed in young and adult readers reflect their cognitive and linguistic development?

A comparison of the cognitive reading attributes assessed in young and adult readers highlights key similarities and differences, which reflect developmental changes in cognitive and linguistic processing. Table 10 presents the distribution of attributes across the two groups, showing that while both young and adult readers rely on fundamental comprehension skills such as finding explicit information and making inferences, the complexity and scope of assessed attributes increase with age.

Table 10
www.frontiersin.org

Table 10. Frequency and percentage of attributes used in assessing reading skills of adult and young readers.

For young readers, the findings indicate that “finding explicit information” and “making inferences” are the most emphasized skills, with both being assessed 100% of the time. These skills are fundamental in evaluating reading comprehension, as they help determine a reader's ability to understand and extract meaning from texts. Moderately assessed skills include “synthesizing” and “making interpretations” (55%), which suggest an effort to measure higher-order thinking, along with “lexical knowledge” and “evaluating content” (45%), highlighting the role of vocabulary and critical assessment in reading. Meanwhile, “grammar knowledge” (36%) and “summarizing ideas” (27%) are given less emphasis, suggesting that structural language knowledge and text summarization are secondary to comprehension skills. Less frequently assessed attributes include “understanding graphs”, “expressing in written forms”, and “syntactic knowledge” (18%), which may reflect their lesser role in early reading development. Finally, “sequencing” and “recognizing authors' styles” are the least prioritized skills, assessed only 9% of the time. The findings suggest a strong focus on comprehension and meaning-making, with less emphasis on linguistic structure and stylistic analysis in assessing young readers' abilities.

As for adult readers, the findings reveal that “lexical knowledge” (86%), “making inferences” (81%), and “finding explicit information” (76%) are the most frequently assessed attributes in evaluating adult readers' skills. This suggests a strong emphasis on vocabulary knowledge, the ability to draw conclusions, and identifying direct information from texts. “Syntactic knowledge” (46%) and “summarizing ideas” (35%) are also significant, reflecting the importance of understanding sentence structures and condensing key points. “Grammar knowledge” (27%), along with “identifying the main idea” and “synthesizing” (24% each), highlights a moderate focus on language rules and higherorder thinking. “Making interpretation” (22%) indicates a role in deeper text analysis. Several other attributes, including “understanding paragraphs”, “scanning or skimming”, “connecting ideas”, and “recognizing authors' styles” (14% each), receive less emphasis, suggesting they are assessed selectively. “Identifying supporting details” and “evaluating content” (11% each) are considered minor components, while “building a mental model” (8%) is assessed infrequently. The least prioritized attributes include “negation” and “applying background knowledge” (5% each), as well as “predicting sequences of events”, “pragmatic knowledge”, and “holding memory” (3% each) receiving minimal attention. The findings highlight a strong focus on comprehension, vocabulary, and sentence structure in adult reading assessments, while aspects related to memory, background knowledge, and prediction are less emphasized.

Moreover, studies involving adult readers have assessed a greater number of attributes, totaling 21, compared to 13 attributes identified in studies on young readers, as illustrated in Figure 4. Notably, certain attributes were exclusively examined in CDA studies on adult readers, including understanding paragraphs, connecting ideas, applying background knowledge, and pragmatic knowledge. Conversely, some attributes were uniquely assessed among young readers, such as understanding graphs and sequencing.

Figure 4
www.frontiersin.org

Figure 4. Comparison of attributes employed to assess reading skills of young and adult readers.

5 Discussion

5.1 Developmental progression of cognitive reading skills

The findings of this review indicate that reading frameworks applied across age groups are shaped by varying conceptualizations of cognitive reading processes. In young readers, the dominant attributes, finding explicit information and making inferences, suggest a framework that primarily focuses on basic comprehension and text-based understanding. These attributes align with early reading development models, which emphasize the acquisition of literal comprehension skills as the foundation for reading proficiency (Kim and Suk, 2020). The analysis reveals that young readers are predominantly assessed on literal and interpretive levels of comprehension (Nurjanah and Putri, 2020), reflecting a framework that prioritizes understanding surface-level information and basic meaning derivation. This observation also aligns with findings from neurocognitive studies, where children rely more on literal processing, supported by stronger inter-regional connectivity (Liu et al., 2018). However, the limited inclusion of syntactic knowledge, along with the absence of critical and creative comprehension skills such as evaluating content or connecting ideas, suggests that early reading frameworks place less emphasis on complex higher-order cognitive processes, which are essential for developing more sophisticated reading abilities.

Conversely, the more diverse set of attributes in adult readers' assessments highlights a broader cognitive spectrum, where vocabulary mastery, sentence structure, and higher-order skills such as synthesizing ideas and evaluating content are crucial components of reading competence. This shift signals a more holistic reading framework in adult readers, consistent with interactive and strategic reading models that require integrating both text-based and knowledge-based processes (Khalifa and Weir, 2009). The broader cognitive spectrum observed in adult readers is further supported by neurocognitive evidence, which indicates that adults not only rely on semantic processing networks (Liu et al., 2024). As reading development progresses, semantic, pragmatic, and syntactic networks play a crucial role in predicting reading ability at earlier stages of life, while more complex integrative networks emerge as stronger predictors of reading proficiency later in life (Horowitz-Kraus et al., 2024). This developmental trajectory suggests that the future-reading network, which involves the integration of multiple cognitive processes, becomes more prominent as readers mature, allowing adult readers to engage with narratives in a more sophisticated manner.

5.2 Diagnostic precision in measuring higher-order reading skills

The diagnostic precision of cognitive diagnostic assessment models in capturing higher-order reading skills such as synthesizing, evaluating, and interpreting texts remains a contentious issue across different age groups. While these skills are more frequently assessed in adult readers, their limited use in young readers' assessments could stem from both genuine developmental constraints and methodological limitations. Higher-order cognitive processes often require abstract reasoning, metacognitive awareness, and the integration of prior knowledge, abilities that are typically more advanced in older readers. This aligns with the hierarchical structure of reading development, where early stages focus on literal comprehension and basic meaning retrieval, while more advanced stages involve critical appraisal and inferential thinking (Chen et al., 2023). Nevertheless, the limited inclusion of higherorder skills in young readers' assessments may also be attributed to the inherent complexity of these skills and the difficulty of designing reliable, scalable tasks that isolate and measure them independently (Molokopeeva and Simard, 2024). Unlike lower-level processes, which can be assessed using controlled measures with clear indicators of specific deficits, higher-order reading involves interconnected cognitive processes that operate simultaneously (Alderson et al., 2015).

Moreover, automaticity in lower-level skills such as retrieving explicit information and lexical knowledge is essential for freeing up cognitive resources needed for more complex comprehension tasks (Suzuki and Elgort, 2023). Therefore, CDA models may prioritize lower-level attributes to establish a solid foundation before progressing to higher-order skills. However, this approach risks oversimplifying the reading process and underestimating young readers' potential to engage in advanced cognitive tasks when provided with appropriate scaffolding and instructional support. Developing ageappropriate assessments that integrate both fundamental and higher-order skills is essential to providing a more detailed diagnostic profile of young readers' cognitive abilities. Additionally, CDA models often rely on retrofitting existing assessments, which are not originally designed to probe such complex cognitive processes (Ketabi et al., 2021). Consequently, the underrepresentation of higher-order skills might not solely reflect developmental readiness but rather the inadequacy of assessment tools to elicit and diagnose these abilities. Addressing this limitation requires the development of customized assessment instruments that integrate distinct item formats, interactive tasks, and multimodal stimuli to better capture the gradual emergence of higher-order thinking in young readers (Hemati and Baghaei, 2020).

6 Conclusion

This study has addressed all the three research questions by critically analyzing the cognitive attributes used in reading assessments across different age groups and providing valuable insights into how reading skills are conceptualized and measured from a cognitive perspective. By comparing the attributes used in assessments of young and adult readers, the study highlights the distinct cognitive demands that readers encounter at different stages of development. The findings underscore the importance of age-specific approaches to reading instruction and assessment, which are essential for fostering reading proficiency across the lifespan. Ultimately, this research contributes to the broader discourse on educational assessment by advocating for more differentiated and contextually relevant assessment practices. By aligning reading assessments with the cognitive capabilities of learners, educators, policymakers, and assessment developers can better support the development of literacy skills to ensure that all individuals are equipped with the reading abilities necessary for their academic and professional success.

6.1 Theoretical and practical implications

This study contributes to the theoretical understanding of reading skill development across different age groups by highlighting the distinct cognitive attributes prioritized in assessing young and adult readers. The findings support and extend existing CDA frameworks by offering a comparative analysis of how reading skills are conceptualized and measured at various stages of life. By distinguishing between foundational skills emphasized in young learners and more complex, abstract skills in adults, the study provides a comprehensive perspective on the cognitive processes underlying reading proficiency. This reinforces the need for age-specific theoretical models that account for the evolving nature of reading comprehension and the adaptive strategies employed by readers as they mature.

The review also suggested that internal and external stakeholders should adjust their approaches based on the findings. Concerning internal stakeholders, especially educators, who play a key role in decision-making that directly impacts the institution's operations and pedagogical direction (Chan and Chou, 2020), the findings have direct implications for the design and implementation of reading instructional strategies and assessments. The identification of key reading attributes for different age groups enables educators to tailor their teaching methods and assessment tools to better align with the cognitive abilities of their students. For young learners, the emphasis on attributes like making inferences and finding explicit information suggests that teaching strategies should focus on enhancing these critical interpretive skills. In contrast, for adult learners, greater attention should be paid to vocabulary development, syntactic knowledge, and summarization techniques. Although readers of all ages must develop a range of reading skills, young learners acquire language implicitly through exposure, building interpretive abilities like analysis and inference that are essential for academic success. In contrast, adult learners benefit from explicit instruction in vocabulary, syntax, and summarization, which are vital for comprehending more complex texts and engaging in professional or academic reading. Thus, these distinctions highlight the need for a differentiated instructional approach that accommodates the developmental and cognitive differences between young and adult learners.

Regarding external stakeholders, such as policymakers, curriculum developers, educational authorities, and assessment developers, who influence decisions, shape policies, and establish frameworks impacting the institution's operations within the broader educational ecosystem (Agnew, 2021), the study highlights the need for reading assessments that accurately reflect the cognitive demands placed on readers of different ages. The findings advocate for a more differentiated approach to assessment design, one that recognizes the varying complexities of reading tasks faced by young and adult readers. This has implications for standardized testing and literacy programs, suggesting that a one-size-fits-all model is inadequate. External stakeholders are encouraged to consider these insights when formulating educational policies, designing literacy interventions, and developing new assessment tools that are both equitable and effective across diverse populations.

6.2 Limitations and recommendations

Despite offering valuable insights, this study acknowledges several limitations that highlight opportunities for further research. The study focuses on reading assessment research through the cognitive diagnostic approach, which provides a fine-grained understanding of reading attributes but may not fully capture perspectives from other emerging assessment frameworks. The use of the scoping review methodology in identifying, screening, and analyzing publications facilitates a comprehensive mapping of the literature, yet it does not offer the quantitative synthesis that could be achieved through meta-analyses. Additionally, the study is confined to English language resources and reading assessments in the context of English as a first, second, or foreign language, which may restrict the generalizability of the findings across diverse linguistic and cultural settings. The selection of studies from three reputable academic databases with a focus on empirical studies ensures the inclusion of high-quality publications but may exclude valuable insights from other sources such as conference proceedings, book chapters, and gray literature.

To address these limitations, future research could explore several key areas. Future reviews on advanced technologies in reading assessment, such as eye tracking and electroencephalography (EEG), are needed to provide objective, real-time insights into readers' cognitive processes and offer a more holistic understanding of reading behaviors. Additionally, future studies could undertake more rigorous review methodologies, such as meta-analyses and bibliometric studies, to provide a more extensive understanding of prevailing trends and gaps in the literature on cognitive reading attributes. Cross-linguistic comparative research would also be valuable in examining the consistency of cognitive reading attributes across different languages and cultural contexts, thereby enhancing the generalizability of the findings. Furthermore, comparative investigations between conventional and technology-assisted assessments could yield critical insights into the effectiveness, adaptability, and efficiency of different assessment approaches. Finally, evaluating the strengths and limitations of various reading assessment methods would help identify the most reliable and valid techniques for measuring reading proficiency across diverse age groups and educational settings. Addressing these avenues would significantly enrich the understanding of cognitive reading attributes and contribute to the advancement of reading assessment practices in various linguistic and educational contexts.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

MMN: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Resources, Software, Visualization, Writing – original draft, Writing – review & editing. MMM: Funding acquisition, Resources, Supervision, Validation, Visualization, Writing – review & editing, Conceptualization. NS: Supervision, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. The study was funded by Ganjaran Penerbitan GP-KO21854 and TAP-KO21854 Universiti Kebangsaan Malaysia.

Acknowledgments

The authors would like to extend gratitude to Kumpuluan Penyelidikan Universiti (KPU) Penilaian Pendidikan, Faculty of Education, UKM Bangi for continuous support.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Gen AI was used in the creation of this manuscript.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Agnew, J. (2021). Internal and external stakeholder impact on curriculum design. Acad. Lett. 1–4. doi: 10.20935/AL2028

Crossref Full Text | Google Scholar

Alavi, M., and Ranjbaran, F. (2018). Constructing and validating a Q-Matrix for cognitive diagnostic analysis of a reading comprehension test battery. J. Engl. Lang. Teach. Learn. 21, 1–15. Available online at: https://elt.tabrizu.ac.ir/article_17207_d9a9150ba8996ecfe90af3e07b905503.pdf

Google Scholar

Alderson, J. C., Brunfaut, T., and Harding, L. (2015). Towards a theory of diagnosis in second and foreign language assessment: insights from professional practice across diverse fields. Appl. Linguist. 36, 236–260. doi: 10.1093/applin/amt046

Crossref Full Text | Google Scholar

Anderson, L. W., and Krathwohl, D. (2001). A Taxonomy for Learning, Teaching and Assessing: A Revision of Bloom's Taxonomy of Educational Objectives. New York, NY: Longman.

Google Scholar

Arksey, H., and O'Malley, L. (2005). Scoping studies: towards a methodological framework. Int. J. Soc. Res. Methodol. Theory Pract. 8, 19–32. doi: 10.1080/1364557032000119616

Crossref Full Text | Google Scholar

Askari, M., and Karami, H. (2024). On the relationship between sensory learning styles and reading subskill profiles: an application of Fusion Model. Lang. Relat. Res. 15, 245–274. doi: 10.2139/ssrn.4873700

Crossref Full Text | Google Scholar

Baghaei, P., and Ravand, H. (2015). A cognitive processing model of reading comprehension in English as a foreign language using the linear logistic test model. Learn. Individ. Differ. 43, 100–105. doi: 10.1016/j.lindif.2015.09.001

Crossref Full Text | Google Scholar

Birkle, C., Pendlebury, D. A., Schnell, J., and Adams, J. (2020). Web of science as a data source for research on scientific and scholarly activity. Quant. Sci. Stud. 1, 363–76. doi: 10.1162/qss_a_00018

Crossref Full Text | Google Scholar

Boori, A. A., Ghazanfari, M., Ghonsooly, B., and Baghaei, P. (2023). The construction and validation of a Qmatrix for cognitive diagnostic analysis: the case of the reading comprehension section of the IAUEPT. Int. J. Lang. Test. 13, 31–53. doi: 10.22034/IJLT.2023.383112.1227

PubMed Abstract | Crossref Full Text | Google Scholar

Boori, A. A., Ghazanfari, M., Ghonsooly, B., and Baghaei, P. (2024). A cognitive diagnostic modeling analysis of the reading comprehension section of an Iranian high-stakes language proficiency test. Int. J. Lang. Test. 14, 17–33. doi: 10.22034/ijlt.2023.399561.1256

PubMed Abstract | Crossref Full Text | Google Scholar

Cai, Y., and Chen, H. (2022). The fluctuating effect of thinking on language performance: new evidence for the Island Ridge Curve. Lang. Assess Q. 19, 465–479. doi: 10.1080/15434303.2022.2080553

Crossref Full Text | Google Scholar

Chan, S. J., and Chou, C. (2020). Who influences higher education decision-making in Taiwan? An analysis of internal stakeholders. Stud. High Educ. 45, 2101–2109. doi: 10.1080/03075079.2020.1823646

Crossref Full Text | Google Scholar

Chen, H., Cai, Y., and de la Torre, J. (2023). Investigating second language (L2) reading subskill associations: a cognitive diagnosis approach. Lang. Assess Q. 20, 166–89. doi: 10.1080/15434303.2022.2140050

Crossref Full Text | Google Scholar

Chen, H., and Chen, J. (2015). Exploring reading comprehension skill relationships through the G-DINA model. Educ. Psychol. 36, 1049–1064. doi: 10.1080/01443410.2015.1076764

Crossref Full Text | Google Scholar

Chen, H., and Chen, J. (2016). Retrofitting non-cognitive-diagnostic reading assessment under the Generalized DINA Model framework. Lang. Assess Q. 13, 218–230. doi: 10.1080/15434303.2016.1210610

Crossref Full Text | Google Scholar

Coady, J. (1979). “A psycholinguistic model of the ESL reader,” in Reading in A Second Language, eds R. Mackay, B. Barkman, and R. R. Jordan (Rowley, MA: Newbury House).

Google Scholar

Davis, F. B. (1968). Research in comprehension in reading. Read. Res. Q. 3:499. doi: 10.2307/747153

Crossref Full Text | Google Scholar

Du, W., and Ma, X. (2021). Probing what's behind the test score: application of multi-CDM to diagnose EFL learners' reading performance. Read Writ. 34, 1441–1466. doi: 10.1007/s11145-021-10124-x

Crossref Full Text | Google Scholar

Fan, T., and Yan, X. (2020). Diagnosing English reading ability in Chinese senior high schools. Stud. Educ. Eval. 67, 1–12. doi: 10.1016/j.stueduc.2020.100931

Crossref Full Text | Google Scholar

George, A. C., and Robitzsch, A. (2021). Validating theoretical assumptions about reading with cognitive diagnosis models. Int. J. Test. 21, 105–29. doi: 10.1080/15305058.2021.1931238

PubMed Abstract | Crossref Full Text | Google Scholar

Geramipour, M., Talebzadeh, H., and Mahdi, S. (2021). The optimal cognitive diagnostic model (CDM) for the grammar section of MA entrance examination of state universities for EFL candidates. Lang. Relat. Res. 12, 187–218. doi: 10.29252/LRR.12.1.6

Crossref Full Text | Google Scholar

Gierl, M. J., and Cui, Y. (2008). Defining characteristics of diagnostic classification models and the problem of retrofitting in cognitive diagnostic assessment. Meas. Interdiscip. Res. Perspect. 6, 263–268. doi: 10.1080/15366360802497762

Crossref Full Text | Google Scholar

Goodman, K. S. (1967). Reading: a psycholinguistic guessing game. J. Read Spec. 6, 126–135. doi: 10.1080/19388076709556976

Crossref Full Text | Google Scholar

Gough, P. B. (1972). “One second of reading,” in Language by Ear and by Eye, eds J. F. Kavanagh, and I. G. Mattingly (Cambridge: The MIT Press), 331–358.

Google Scholar

He, L., Xiong, L., and Min, S. (2022). Diagnosing listening and reading skills in the Chinese EFL context: performance stability and variability across modalities and performance levels. System. 106, 1–19. doi: 10.1016/j.system.2022.102787

Crossref Full Text | Google Scholar

Heaton, J. (1991). Writing English Language Test. New York, NY: Longman.

Google Scholar

Hemati, S. J., and Baghaei, P. (2020). A cognitive diagnostic modeling analysis of the English reading comprehension section of the Iranian National University Entrance Examination. Int. J. Lang. Test. 10, 11–32. Available online at: https://www.ijlt.ir/article_114278_022ab6b64d4bc8acba9c561846828270.pdf

Google Scholar

Hoover, W. A., and Gough, P. B. (1990). The simple view of reading. Read Writ. Interdiscip. J. 2, 127–160. doi: 10.1007/BF00401799

Crossref Full Text | Google Scholar

Horowitz-Kraus, T., Meri, R., Holland, S. K., Farah, R., Rohana, T., Haj, N., et al. (2024). Language first, cognition later: different trajectories of subcomponents of the future-reading network in processing narratives from kindergarten to adolescence. Brain Connect. 14, 60–69. doi: 10.1089/brain.2023.0012

PubMed Abstract | Crossref Full Text | Google Scholar

Hosseini Alast, S. F., and Baleghizadeh, S. (2024). The interplay of glossing with text difficulty and comprehension levels. Lang. Teach. Res. 28, 1201–1230. doi: 10.1177/13621688211020423

Crossref Full Text | Google Scholar

Hughes, A., and Hughes, J. (2020). Testing for Language Teachers. 3rd Edn. Cambridge: Cambridge University Press.

Google Scholar

Jang, E. E. (2009a). Cognitive diagnostic assessment of L2 reading comprehension ability: Validity arguments for Fusion Model application to LanguEdge assessment. Lang. Test. 26, 31–73. doi: 10.1177/0265532208097336

Crossref Full Text | Google Scholar

Jang, E. E. (2009b). Demystifying a Q-Matrix for making diagnostic inferences about L2 reading skills. Lang. Assess Q. 6, 210–238. doi: 10.1080/15434300903071817

Crossref Full Text | Google Scholar

Jang, E. E., Barron, C., Kim, H., and Russell, B. (2023). Mixed methods investigation into test score users' perspectives about IELTS reading skill profiles. Lang. Teach. Res. Q. 37, 231–247. doi: 10.32038/ltrq.2023.37.13

Crossref Full Text | Google Scholar

Jang, E. E., Dunlop, M., Park, G., and Van Der Boom, E. H. (2015). How do young students with different profiles of reading skill mastery, perceived ability, and goal orientation respond to holistic diagnostic feedback? Lang. Test. 32, 1–25. doi: 10.1177/0265532215570924

Crossref Full Text | Google Scholar

Jang, E. E., Dunlop, M., Wagner, M., Kim, Y., and Gu, Z. (2013). Elementary school ELLs' reading skill profiles using Cognitive Diagnosis Modeling: roles of length of residence and home language environment. Lang. Learn. 63, 400–436. doi: 10.1111/lang.12016

Crossref Full Text | Google Scholar

Javidanmehr, Z., Reza, M., and Sarab, A. (2019). Retrofitting non-diagnostic reading comprehension assessment: application of the G-DINA Model to a high-stakes reading comprehension test. Lang. Assess. Q. 16, 1–18. doi: 10.1080/15434303.2019.1654479

Crossref Full Text | Google Scholar

Ketabi, S., Alavi, S. M., and Ravand, H. (2021). Diagnostic test construction: insights from cognitive diagnostic modeling. Int. J. Lang. Test. 11, 22–35. Available online at: https://www.ijlt.ir/article_128357_68e74c779145d7be54c24f610a549879.pdf

Google Scholar

Khalifa, H., and Weir, C. J. (2009). Examining Reading: Research and Practice in Assessing Second Language Reading. Cambridge: Cambridge University Press. Available online at: http://hdl.handle.net/10547/224632 (accessed February 10, 2025).

PubMed Abstract | Google Scholar

Kim, A. Y. (2015). Exploring ways to provide diagnostic feedback with an ESL placement test: cognitive diagnostic assessment of L2 reading ability. Lang. Test. 32, 227–258. doi: 10.1177/0265532214558457

Crossref Full Text | Google Scholar

Kim, Y., and Suk, G. (2020). Hierarchical and dynamic relations of language and cognitive skills to reading comprehension: testing the Direct and Indirect Effects Model of Reading (DIER). J. Educ. Psychol. 94:659. doi: 10.1037/edu0000407

Crossref Full Text | Google Scholar

Kim, Y. S. G. (2020). Toward integrative reading science: the direct and indirect effects model of reading. J. Learn. Disabil. 53, 469–491. doi: 10.1177/0022219420908239

PubMed Abstract | Crossref Full Text | Google Scholar

LaBerge, D., and Samuels, S. J. (1974). Toward a theory of automatic information processing in reading. Cogn. Psychol. 6, 293–323. doi: 10.1016/0010-0285(74)90015-2

Crossref Full Text | Google Scholar

Lee, H., and Lee, J. H. (2023). Extending the simple view of reading in second and foreign language learning: a meta-analytic structural equation modeling approach. Rev. Educ. Res. 94, 467–500. doi: 10.3102/00346543231186605

PubMed Abstract | Crossref Full Text | Google Scholar

Lee, Y. W., and Sawaki, Y. (2009). Application of three cognitive diagnosis models to ESL reading and listening assessments. Lang. Assess Q. 6, 239–263. doi: 10.1080/15434300903079562

Crossref Full Text | Google Scholar

Levac, D., Colquhoun, H., and O'Brien, K. K. (2010). Scoping studies: advancing the methodology. Represent. Interv. 5, 1–18. doi: 10.1186/1748-5908-5-69

PubMed Abstract | Crossref Full Text | Google Scholar

Li, H., Hunter, C. V., and Lei, P. W. (2015). The selection of cognitive diagnostic models for a reading comprehension test. Lang. Test. 33, 391–409. doi: 10.1177/0265532215590848

Crossref Full Text | Google Scholar

Li, H., and Suen, H. K. (2013). Constructing and validating a Q-matrix for cognitive diagnostic analyses of a reading test. Educ. Assess. 18, 1–25. doi: 10.1080/10627197.2013.761522

Crossref Full Text | Google Scholar

Li, S., and Wang, J. (2017). “Diagnostic assessment of L2 Chinese learners' reading comprehension ability,” in Chinese as A Second Language Assessment, eds D. Zhang, and C. Lin (Singapore: Springer), 183–202.

Google Scholar

Li, Y., Zhen, M., and Liu, J. (2021). Validating a reading assessment within the cognitive diagnostic assessment framework: Q-matrix construction and model comparisons for different primary grades. Front. Psychol. 12:e786612. doi: 10.3389/fpsyg.2021.786612

PubMed Abstract | Crossref Full Text | Google Scholar

Lin, J. (2024). Developing a reading proficiency scale for Chinese as a second language: a confirmatory factor analysis approach. Lang. Aware. 33, 597–624. doi: 10.1080/09658416.2023.2291521

Crossref Full Text | Google Scholar

Liu, H., and Bian, Y. (2021). Model selection for cogitative diagnostic analysis of the reading comprehension test. Front. Psychol. 12:e644764. doi: 10.3389/fpsyg.2021.644764

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, X., Gao, Y., Di, Q., Hu, J., Lu, C., Nan, Y., et al. (2018). Differences between child and adult large-scale functional brain networks for reading tasks. Hum. Brain Mapp. 39, 662–679. doi: 10.1002/hbm.23871

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, X., and Read, J. (2021). Investigating the skills involved in reading test tasks through expert judgement and verbal protocol analysis: convergence and divergence between the two methods. Lang. Assess Q. 18, 357–381. doi: 10.1080/15434303.2021.1881964

Crossref Full Text | Google Scholar

Liu, X., and Read, J. (2023). Designing a new diagnostic reading assessment for a local post-admission assessment program: a needs-driven approach. Educ. Linguist. 61, 83–102. doi: 10.1007/978-3-031-33541-9_5

Crossref Full Text | Google Scholar

Liu, X., Zhang, L., Yu, S., Bai, Z., Qi, T., Mao, H., et al. (2024). The effects of age and reading experience on the lifespan neurodevelopment for reading comprehension. J. Cogn. Neurosci. 36, 239–260. doi: 10.1162/jocn_a_02086

PubMed Abstract | Crossref Full Text | Google Scholar

Lockwood, C., Munn, Z., and Porritt, K. (2015). Qualitative research synthesis: methodological guidance for systematic reviewers utilizing meta-aggregation. Int. J. Evid. Based Healthc. 13, 179–187. doi: 10.1097/XEB.0000000000000062

PubMed Abstract | Crossref Full Text | Google Scholar

Luebke, S., and Lorié, J. (2013). Use of Bloom's Taxonomy in developing reading comprehension specifications. J. Appl. Test Technol. 14, 1–27. Available online at: https://jattjournal.net/index.php/atp/article/view/45250/36643

Google Scholar

Maas, L., Madison, M. J., and Brinkhuis, M. J. S. (2024). Properties and performance of the one-parameter loglinear cognitive diagnosis model. Front. Educ. 9:e1287279. doi: 10.3389/feduc.2024.1287279

Crossref Full Text | Google Scholar

Martín-Martín, A., Thelwall, M., Orduna-Malea, E., and Delgado López-Cózar, E. (2021). Google Scholar, Microsoft Academic, Scopus, Dimensions, Web of Science, and OpenCitations' COCI: a multidisciplinary comparison of coverage via citations. Scientometrics. 126, 871–906. doi: 10.1007/s11192-020-03690-4

PubMed Abstract | Crossref Full Text | Google Scholar

Masdoki, M., Din, R., and Effendi Ewan Mohd Matore, M. (2021). Teaching 4.0 competency in higher learning institutions: a systematic mapping review. Int. J. Learn. Teach. Educ. Res. 20, 217–31. doi: 10.26803/ijlter.20.10.12

Crossref Full Text | Google Scholar

Mei, H., and Chen, H. (2022). Cognitive diagnosis in language assessment: a thematic review. RELC J. 55, 1–9. doi: 10.1177/00336882211072898

Crossref Full Text | Google Scholar

Miller, G. A. (1956). The magical number seven, plus or minus two: some limits on our capacity for processing information. Psychol. Rev. 63, 81–97. doi: 10.1037/h0043158

Crossref Full Text | Google Scholar

Miller, L. A., and Lovler, R. L. (2020). Foundations of Psychological Testing: A Practical Approach. 6th Edn. Los Angeles, CA: SAGE.

Google Scholar

Mirzaei, A., Vincheh, M. H., and Hashemian, M. (2020). Retrofitting the IELTS reading section with a general cognitive diagnostic model in an Iranian EAP context. Stud. Educ. Eval. 64, 1–10. doi: 10.1016/j.stueduc.2019.100817

Crossref Full Text | Google Scholar

Mohamed Shaffril, H. A., Samsuddin, S. F., and Abu Samah, A. (2021). The ABC of systematic literature review: the basic methodological guidance for beginners. Qual. Quant. 55, 1319–1346. doi: 10.1007/s11135-020-01059-6

Crossref Full Text | Google Scholar

Mohammed, A., Kareem, A., Dawood, S., Alghazali, T., Khlaif, Q., Sabti, A. A., et al. (2023). A cognitive diagnostic assessment study of the Reading Comprehension Section of the Preliminary English Test (PET). Int. J. Lang. Test. 13, 1–20. doi: 10.22034/ijlt.2022.362849.1195

PubMed Abstract | Crossref Full Text | Google Scholar

Mohd Noh, M. F., and Mohd Matore, M. E. E. (2020). Rating performance among raters of different experience through multi-facet Rasch measurement (MFRM) model. J. Meas. Eval. Educ. Psychol. 11, 147–162. doi: 10.21031/epod.662964

Crossref Full Text | Google Scholar

Mohd Noh, M. F., and Mohd Matore, M. E. E. (2022). Rater severity differences in English language speaking assessment based on rating experience, training experience and teaching experience through many-faceted Rasch measurement analysis. Front. Psychol. 13:e941084. doi: 10.3389/fpsyg.2022.941084

PubMed Abstract | Crossref Full Text | Google Scholar

Mohd Noh, M. F., and Mohd Matore, M. E. E. (2024). Conundrum and considerations in Cognitive Diagnostic Assessment for language proficiency evaluation. Inf. Manag. Bus. Rev. 16, 63–72. doi: 10.22610/imbr.v16i2(I).3690

Crossref Full Text | Google Scholar

Molokopeeva, T., and Simard, D. (2024). Interaction between levels of text representation and working memory during L2 reading comprehension: what about it? Int. J. Appl. Linguist. 34, 568–585. doi: 10.1111/ijal.12516

Crossref Full Text | Google Scholar

Munby, J. (1978). Communicative Syllabus Design. Cambridge: Cambridge University Press.

Google Scholar

Munn, Z., Pollock, D., Khalil, H., Alexander, L., McLnerney, P., Godfrey, C. M., et al. (2022). What are scoping reviews? Providing a formal definition of scoping reviews as a type of evidence synthesis. JBI Evid. Synth. 20, 950–952. doi: 10.11124/JBIES-21-00483

PubMed Abstract | Crossref Full Text | Google Scholar

Nadmilail, A. I., Mohd Matore, M. E. E., Maat, S. M., and Sheridan, L. (2023). Broad vs. narrow traits: A scoping review of measuring personality traits in teacher selection using the situational judgment test. Front. Psychol. 14:e1217321. doi: 10.3389/fpsyg.2023.1217321

PubMed Abstract | Crossref Full Text | Google Scholar

Nallasamy, R., and Khairani, A. Z. (2022). Validating reading comprehension assessment under the GDINA Model. Malaysian J. Soc. Sci. Humanit. 7, 1–12. doi: 10.47405/mjssh.v7i11.1877

PubMed Abstract | Crossref Full Text | Google Scholar

Nightingale, E., Branum-Martin, L., and Greenberg, D. (2023). Modeling the relations between reading skills and assessment methods in adults who struggle with foundational reading skills. Read Writ. 36, 2175–2198. doi: 10.1007/s11145-022-10384-1

Crossref Full Text | Google Scholar

Nurjanah, R. L., and Putri, S. R. (2020). The effect of literal comprehension on the higher levels of comprehension in reading skill: a longitudinal case study. Engl. Lang. Lit. Int. Conf. 5, 471–476. Available online at: https://jurnal.unimus.ac.id/index.php/ELLIC/article/view/9692/6267

Google Scholar

Peng, P., Lee, K., Luo, J., Li, S., Joshi, R. M., Tao, S., et al. (2021). Simple view of reading in Chinese: a one-stage meta-analytic structural equation modeling. Rev. Educ. Res. 91, 3–33. doi: 10.3102/0034654320964198

PubMed Abstract | Crossref Full Text | Google Scholar

Peters, M. D. J., Godfrey, C., McInerney, P., Khalil, H., Larsen, P., Marnie, C., et al. (2022). Best practice guidance and reporting items for the development of scoping review protocols. JBI Evid. Synth. 20, 953–968. doi: 10.11124/JBIES-21-00242

PubMed Abstract | Crossref Full Text | Google Scholar

Piaget, J. (1950). The Psychology of Intelligence. New York, NY: Harcourt Brace.

Google Scholar

Polat, M., Toraman, Ç., and Turhan, N. S. (2022). Reliability analysis of PISA 2018 reading literacy student qestionnaire based on item response theory (IRT): Turkey sample. Int. J. Curric. Instr. 14, 1004–1028. Available online at: https://ijci.net/index.php/IJCI/article/view/906/469

Google Scholar

Pollock, D., Davies, E. L., Peters, M. D. J., Tricco, A. C., Alexander, L., McInerney, P., et al. (2021). Undertaking a scoping review: a practical guide for nursing and midwifery students, clinicians, researchers, and academics. J. Adv. Nurs. 77, 2102–2113. doi: 10.1111/jan.14743

PubMed Abstract | Crossref Full Text | Google Scholar

Ranjbaran, F., and Alavi, S. M. (2017). Developing a reading comprehension test for cognitive diagnostic assessment: a RUM analysis. Stud. Educ. Eval. 55, 167–179. doi: 10.1016/j.stueduc.2017.10.007

Crossref Full Text | Google Scholar

Ravand, H. (2016). Application of a Cognitive Diagnostic Model to a high-stakes reading comprehension test. J. Psychoeduc. Assess. 34, 782–799. doi: 10.1177/0734282915623053

Crossref Full Text | Google Scholar

Ravand, H. (2019). “Application of a hierarchical diagnostic classification model in assessing reading comprehension,” in Quantitative Data Analysis for Language Assessment Volume II, eds V. Aryadoust, and M. Raquel (London; New York, NY: Routledge), 79–98.

Google Scholar

Ravand, H., and Robitzsch, A. (2018). Cognitive diagnostic model of best choice: a study of reading comprehension. Educ. Psychol. 38, 1255–1277. doi: 10.1080/01443410.2018.1489524

Crossref Full Text | Google Scholar

Rumelhart, D. E. (1977). “Toward and interactive model of reading,” in Attention and Performance ed. S. Dornic (Hillsdale, NJ: Erlbaum).

Google Scholar

Sabtu, S. H., and Mohd Matore, M. E. E. (2024). Systematic literature review on the implementation of the Six Sigma approach in education. Int. J. Eval. Res. Educ. 13, 262–270. doi: 10.11591/ijere.v13i1.26196

Crossref Full Text | Google Scholar

Safari, F., and Ahmadi, A. (2023). Developing and evaluating an empirically-based diagnostic checklist for assessing second language integrated writing. J. Second Lang. Writ. 60, 1–15. doi: 10.1016/j.jslw.2023.101007

Crossref Full Text | Google Scholar

Sakaria, D., Maat, S. M., and Mohd Matore, M. E. E. (2023). Factors influencing mathematics teachers' pedagogical content knowledge (PCK): a systematic review. Pegem. Egit. Ogr. Derg. 13, 1–14. doi: 10.47750/pegegog.13.02.01

Crossref Full Text | Google Scholar

Sawaki, Y., Kim, H., and Gentile, C. (2009). Q-Matrix construction: Defining the link between constructs and test items in large-scale reading and listening comprehension assessments. Lang. Assess Q. 6, 190–209. doi: 10.1080/15434300902801917

Crossref Full Text | Google Scholar

Sessoms, J., and Henson, R. A. (2018). Applications of diagnostic classification models: a literature review and critical commentary. Meas. Interdiscip. Res. Perspect. 16, 1–17. doi: 10.1080/15366367.2018.1435104

PubMed Abstract | Crossref Full Text | Google Scholar

Shahmirzadi, N., and Marashi, H. (2023). Cognitive diagnostic assessment of reading comprehension for high- stakes tests: Using GDINA model. Lang. Test Focus Int. J. 8, 1–16. doi: 10.32038/ltf.2023.08.01

Crossref Full Text | Google Scholar

Sieo, W. M., Amini, M., Lee, K. F., Lin, S. E., Tee, J. N., Siau, C. S., et al. (2022). Reading strategies in teaching and learning English as a foreign language: a mixed-method study. Pertanika J. Soc. Sci/ Humanit. 30, 475–499. doi: 10.47836/pjssh.30.2.04

Crossref Full Text | Google Scholar

Singh, V. K., Singh, P., Karmakar, M., Leta, J., and Mayr, P. (2021). The journal coverage of Web of Science, Scopus and Dimensions: a comparative analysis. Scientometrics. 126, 5113–5142. doi: 10.1007/s11192-021-03948-5

PubMed Abstract | Crossref Full Text | Google Scholar

Sook Yi, Y. (2016). Probing the relative importance of different attributes in L2 reading and listening comprehension items: an application of cognitive diagnostic models. Lang. Test. 34, 1–9. doi: 10.1177/0265532216646141

Crossref Full Text | Google Scholar

Stanovich, K. E. (1980). Toward an interactive-compensatory model of individual differences in the development of reading fluency. Read. Res. Q. 16, 32–71. doi: 10.2307/747348

Crossref Full Text | Google Scholar

Sun, H., and Hwang, Y. (2023). Enhancing EFL learners' English reading performance through the diagnostic feedback of CDA. Korean J. Engl. Lang. Linguist. 23, 535–553. doi: 10.15738/kjell.23..202307.535

Crossref Full Text | Google Scholar

Suzuki, Y., and Elgort, I. (2023). “Measuring automaticity in a second language: a methodological synthesis of experimental tasks over three decades (1990-2021),” in Practice and Automatization in Second Language Research: Perspectives from Skill Acquisition Theory and Cognitive Psychology, ed. Y. Suzuki (New York, NY: Routledge), 206–34.

Google Scholar

Tabatabaee-Yazdi, M. (2020). Hierarchical diagnostic classification modeling of reading comprehension. SAGE Open. 10, 1–13. doi: 10.1177/2158244020931068

PubMed Abstract | Crossref Full Text | Google Scholar

Teng, M. F., and Mizumoto, A. (2024). Validation of metacognitive knowledge in vocabulary learning and its predictive effects on incidental vocabulary learning from reading. Int. Rev. Appl. Linguist Lang. Teach. 62, 1–32. doi: 10.1515/iral-2023-0294

Crossref Full Text | Google Scholar

Thi, D. T. D., and Loye, N. (2019). Cognitive diagnostic analyses of the Progress in International Reading Literacy Study (PIRLS) 2011 results. Mes Éval. Éduc. 42, 127–166. doi: 10.7202/1084131ar

Crossref Full Text | Google Scholar

Tonekaboni, F. R., Ravand, H., and Rezvani, R. (2021). The construction and validation of a Q-matrix for a high-stakes reading comprehension test: a G-DINA study. Int. J. Lang. Test. 11, 58–87. Available online at: https://www.ijlt.ir/article_128361.html

Google Scholar

Toprak, T. E., and Cakir, A. (2021). Examining the L2 reading comprehension ability of adult ELLs: developing a diagnostic test within the cognitive diagnostic assessment framework. Lang. Test. 38, 106–131. doi: 10.1177/0265532220941470

Crossref Full Text | Google Scholar

Toprak-yildiz, T. E. (2021). An international comparison using cognitive diagnostic assessment: fourth graders' diagnostic profile of reading skills on PIRLS 2016. Stud. Educ. Eval. 70, 1–10. doi: 10.1016/j.stueduc.2021.101057

Crossref Full Text | Google Scholar

Toprak-Yildiz, T. E. (2022). Developing a cognitive model of L2 reading comprehension for evaluative purposes in an EMI context. Stud. Educ. Eval. 75, 1–11. doi: 10.1016/j.stueduc.2022.101200

Crossref Full Text | Google Scholar

van der Linden, W. J. (2016). Handbook of Item Response Theory Volume 1: Model. London; New York, NY: CRC Press.

Google Scholar

von Davier, M., and Lee, Y. S. (2019). Handbook of Diagnostic Classification Models. Cham: Springer.

Google Scholar

Wang, D., Cai, Y., and Tu, D. (2021). Q-matrix estimation methods for Cognitive Diagnosis Models: Based on partial known Q-Matrix. Multivariate Behav. Res. 56, 514–26. doi: 10.1080/00273171.2020.1746901

PubMed Abstract | Crossref Full Text | Google Scholar

Wang, M. (2023). Bridging assessment and learning: a cognitive diagnostic analysis of a large-scale Spanish proficiency test. Porta Lingarium. 40, 9–24. doi: 10.30827/portalin.vi40.15930

Crossref Full Text | Google Scholar

Yumsek, M. (2023). Educational L2 constructs and diagnostic measurement. Lang. Test Asia. 13, 1–23. doi: 10.1186/s40468-022-00214-0

Crossref Full Text | Google Scholar

Zhang, H., Wu, X., and Ju, M. (2024). Developing a cognitive model of solid geometry based on Interpretive Structural Modeling method. Heliyon. 10, 1–12. doi: 10.1016/j.heliyon.2024.e27063

PubMed Abstract | Crossref Full Text | Google Scholar

Zhao, H., Wang, W., and Huang, Y. (2020). A cognitive analysis of an English reading test through the GDINA model A cognitive analysis of an English reading test through the G-DINA model. J. Phys. Conf. Series. 1629, 1–7. doi: 10.1088/1742-6596/1629/1/012037

Crossref Full Text | Google Scholar

Keywords: reading skills, cognitive reading attributes, language assessment, scoping review, cognitive diagnostic

Citation: Mohd Noh MF, Mohd Matore MEE and Sulaiman NA (2025) A cross-age odyssey of cognitive reading attributes: a scoping review. Front. Psychol. 16:1605898. doi: 10.3389/fpsyg.2025.1605898

Received: 04 April 2025; Accepted: 28 May 2025;
Published: 20 June 2025.

Edited by:

Antonio Bova, Catholic University of the Sacred Heart, Italy

Reviewed by:

Ugur Özbilen, Akdeniz University, Türkiye
Prajit Dhar, University of Potsdam, Germany

Copyright © 2025 Mohd Noh, Mohd Matore and Sulaiman. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Mohd Effendi Ewan Mohd Matore, ZWZmZW5kaUB1a20uZWR1Lm15

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.