Impact Factor 2.323

The 1st most cited journal in Multidisciplinary Psychology

Original Research ARTICLE

Front. Psychol., 10 May 2017 |

Learning Potential in Narrative Writing: Measuring the Psychometric Properties of an Assessment Tool

Léia G. Gurgel1, Mônica M. C. de Oliveira1, Maria C. R. A. Joly2 and Caroline T. Reppold1*
  • 1Research Laboratory in Psychological Assessment, Federal University of Health Sciences of Porto Alegre, Porto Alegre, Brazil
  • 2Postgraduate Program in Human Development and Health Processes, University of Brasília, Brasília, Brazil

Objective: The Computerized and Dynamic Writing Test (TIDE) is designed to examine the learning potential of adolescents in narrative writing. This was a validation study of the TIDE based on its internal structure. Learning potential is responsible for cognitive modifiabilty according to the Theory of Cognitive Structural Modifiability (CSM) developed by Feüerstein.

Method: Included 304 participants between 10 and 17 years of age from schools in the South of Brazil. The data collection involved student groups that were divided according to age and school grade. Each participant reponded to the TIDE for an average of 50 min in the school's computer lab. The participants' selection criteria were: being regularly enrolled in the fifth to eighth grade and providing an informed consent form signed by a responsible caregiver.

The exclusion criteria included: neurological problems, having been held back in school for two or more years, not cooperating, not completing the test for any reason and physical conditions impeding the assessment.

Results: The Kendall test indicated agreement between two evaluators, who corrected the participants' first and second texts that resulted from applying the TIDE. The TIDE is divided into three modules. Factor analysis was applied to the first module (pre-test), which revealed a division in two factors, and to the second module (instructional module), which was divided in three factors. The reliability of the TIDE items was verified using Cronbach's Alpha with coefficients >0.7. The analysis of the third module (post-test) was based on McNemar's Test and showed statistically significant results that demonstrated an evolution in the participants' learning potential.

Conclusion: The TIDE proved to be valid and is considered a relevant tool for speech, language, hearing, psychological and educational assessment. The original nature of the tool presented here is highlighted, based on the dynamic assessment method, offering data on a narrative writing learning method as well as its possible adaptation to other contexts and languages. In addition, the computer-based nature of the tool is emphasized, enabling its more precise application and analysis of participant performance, in addition to its lower cost, reduced application bias and ability to test more than one person simultaneously.


Interest in dynamic and assisted assessment arose around 1960, based on the development of interactive evaluation procedures. Dynamic assessment is a cognitive assessment modality developed under the influence of Vygotsky's Sociocultural Theory and the work of Reuven Feuerstein in the 1970s (Sternberg and Grigorenko's, 2002; Camilleri and Law, 2013). This approach promotes assessment combined with mediation with the aim of analyzing participants' learning potential in a specific achievement task. The theory implies that assessment and teaching should be integrated to broadly understand the abilities of the participants being assessed, exploring their potential and observing changes in their performance during the task (Alavi and Taghizadeh, 2014). This assessment modality presupposes the plasticity of human learning, allowing for variations in how subjects think, as well as alternative diagnostic methods for participants who do not perform well on conventional norm-referenced tests. Alternative explanations are sought for weak learning and problem-solving performances, as well as greater understanding of the participants' difficulties. Therefore, the assessment is based on the instructions given and interactions that take place in the course of the assessment in order to gain a deeper understanding of the participants' task response process (Vygotsky, 1998a,b; Enumo, 2005; Fazlollahi et al., 2015). This type of assessment intervention is also called a “mediated learning experience” (Pena et al., 2014). Learning potential is responsible for cognitive modifiabilty according to the Theory of Cognitive Structural Modifiability (CSM) developed by Feüerstein (Feuerstein et al., 1979). It is quantified by instruments that can predict changes in initial ability patterns compared to those observed after the trainng and refers to the difference in performance between aided and unaided learning. For the use of these instruments to be appropriate, it is necessary that they have psychometric studies to search for internal structure validity (Davidson et al., 2016).

When comparing the static and dynamic assessment modalities, it is observed that the former focuses on the results the participants obtain and the profile these results produce in a standardized and structured manner. Static assessment modalities are more objective than dynamic assessment modalities and better assess participants with average performance levels. On the other hand, dynamic assessments are able to assess changes based on mediations, interaction and feedback, interpreting the results by focusing on cognitive modifiability (Linhares et al., 2006). Such assessments consider the effects of the participant's culture, linguistic differences and performance on the test itself. It is a valid modality for language tests and for increasing scores in statistical tests (Jacobs, 2001). In view of these attributes, dynamic procedures have been used more frequently since the 1970s. At the same time, there has been increasing criticism of static tests, drawing a contrast between the two assessment methods (Sternberg and Grigorenko's, 2002; Camilleri and Law, 2013).

More specifically, dynamic tools are considered important resources for assessing the writing performance of children and adolescents (Dias and Enumo, 2006). Writing assessment has been described and accomplished in various manners, including formal methods, such as standardized tests and informal methods, such as qualitative observation techniques. This choice tends to be guided by the evaluator according to the participants' needs and the context in which the test takes place (Wolfe et al., 2016). Dynamic writing assessment techniques are in line with the proposals of Wardle and Roozen (2012) and Yancey (1999). According to these authors, the future of writing assessment tends to be more expansive in order to fully capture participants' learning of the writing process, including understanding why participants succeed or fail in tasks related to the construct. Nevertheless, most studies still focus on the use of static resources and tests whose scores are based on school curriculum (Wilson et al., 2016). In this context, this study aims to seek evidence of validity of the Computerized and Dynamic Writing Test (TIDE) based on its internal structure and to describe the response process in the Brazilian context. This test is based on the dynamic assessment method, which aims to assess adolescents' learning potential in narrative text writing. The search for evidence of validity based on the internal structure is the phase in which the correlations between the items and subtests are established (using factorial analyses, for example) including evidence of reliability (American Educational Research Association (AERA), American Psychological Association (APA), National Council on Measurement in Education (NCME), 2014).

In general, quantitative and qualitative forms are included in writing assessments, whether they are based on the school curriculum or not, considering the dimensionality of the written production with regard to the students' quality and productivity (Kim et al., 2014). In this sense, there is a need to develop and seek evidence of the validity of tools that identify the characteristics of the students' writing, such as the development of sentence construction, grammar and punctuation, as well as structural aspects, such as cohesion and logical sequence (Wilson et al., 2016). The number of writing assessment instruments in the literature is limited, especially for adolescents. Existing tests do not assess the construct comprehensively (including at the word, sentence and paragraph levels) with micro- and macro-structural grammatical and orthographic aspects. In addition, the tests present little or no evidence of validity (Richards, 2015). Therefore, studies seeking evidence of validity (as described in this article) are important to changing this reality and producing techniques and quality assessments that allow for appropriate decision making about assessments and intervention programs (Plake and Wise, 2014).

It is important to use instruments that assess the writing process and pay attention to the strategies the subjects use throughout the production (Flower and Hayes, 1981; Richards, 2015). To meet this need, TIDE (Joly and Schiavoni, 2013) was organized according to the written structure production described by Hayes and Flower (1980), considering the processes of planning, translating and reviewing. In a preliminary psychometric study seeking evidence of the internal validity of this tool, statistically significant differences were found between the number of words and school performance in Portuguese, considering the text at the beginning and the end of the task. The authors highlighted the need for further studies using the tool under analysis (Joly et al., 2015). TIDE stands out when compared to other static or dynamic writing assessment instruments in the literature, as it allow specific recommendations to be made to the student/respondent about possible measures to improve his/her performance, instead of merely describing that performance (Sternberg and Grigorenko's, 2002). It also evaluates adolescents' narrative writing skills and analyzes the subject's learning potential.

Materials and Methods


The number of study participants was calculated to comply with the “items/subject index” criterion used for sample calculations when factor analyses are required (Hair et al., 2006). The sample was comprised of 304 participants, who were chosen at the public and private schools that participated in the research. The schools were invited to participate on a convenience basis and were located in the Central, Eastern and Northern regions of Porto Alegre, RS, Brazil. The study participants were fifth to eighth graders, the largest proportion of whom was in the seventh grade (37.1%). The participants' age ranged from 10 to 17 years and 43.2% were 12 or 13 years old. In terms of gender, 64.5% of the sample was female and only 8.2% came from private schools. In total, 34.6% of the responsible guardians/caretakers considered the adolescents' performance to be good while 33.9% reported the presence of problems related to attention.

The Brazilian educational system consists of early childhood education, primary education and secondary education. According to the school data, almost half of the participants started school at 6 years of age (48.8%), 82.1% attended childhood education and 9.1% had reading or writing problems, according to their parents and/or caregivers (who completed questionnaires with information on the participants). The participants' selection criteria were: being regularly enrolled in the fifth to eighth grade and providing an informed consent form signed by their responsible caregivers. The exclusion criteria included: neurological problems, having been held back in school for two or more years, not cooperating, not completing the test for any reason and physical conditions impeding the assessment.


The goal of the TIDE (Joly and Schiavoni, 2013) is to assess the learning potential of primary education students in the production of written narrative texts. It was constructed based on the dynamic assessment principles of learning potential, in accordance with Sternberg and Grigorenko's (2002) theoretical premises and Hayes and Flower (1980) theoretical structure of written production. Its internal structure consists of a pre-test, instruction module and post-test. In the pre-test, the participant provides information, such as name, grade and age. The test then starts with instructions on the proposed task: the writing of a narrative text on the topic: “Hero for a day.” The participant is then asked to read his/her text and make any corrections that he/she deems necessary. No hints are provided to aid the respondent's performance.

The instruction module, in turn, supposes that strategies are offered to facilitate the learning of text writing, and is based on Hayes and Flower (1980) processes. This written composition model does not describe successive stages but rather the processes involved. The first of these processes is planning, which serves to extract information on the nature of the task and on long-term memory and use it to establish a plan that guides the production of a text, including three subgroups: idea production, organization and target setting. The second process is translation, which is intended to transform the planning of the message into written sentences, which should comply with the standards of written language. The third process is reviewing, which involves evaluating the text produced and improving its quality. This process contains two sub-processes: reading and editing.

The individual receives advice on how to write a narrative text and the elements to be included, such as characters, scene, situation and a problem. He/she is asked to verify whether his/her text contains all of these elements. Questions are then asked to lead the participant to reflect on his/her written construction. He/she is expected to answer the questions and revise the text if any of the elements are missing, according to his/her own judgment. The first questions focus on the characters in the story (e.g., “Who are the main characters?,” “What do they look like?”). Questions are then asked about the scene, situation, problem faced, response to the problem, action, solution and reaction of the characters in the story. Finally, the participant is asked to revise his/her text based on final questions. Afterwards, he/she is instructed to review whether the sentences are complete, whether proper names start with capital letters, whether the words are spelled correctly and whether punctuation marks are used appropriately. The result is analyzed on a three-point Likert scale (participant did not add the answer to the question to the text, participant partially added the answer to the question to the text or participant fully added the answer to the question to the text). The instruction module has 19 items.

Finally, in the post-test, the participant analyzes his/her textual output as a whole. He/she has the opportunity to reread and edit the text, if necessary, without interference or hints from the evaluator. Because it is a computerized test, the participant can access and change any item in the module that he/she deems necessary and may do so as many times as needed. The changes made while editing the text appear on the screen in a different color. In the instruction module, each element of the narrative text has a different color so that it becomes clear to the student how much he/she has altered or added to each of these elements of the narrative.

The evaluators correct the results. Automated correction is not used. The evaluators compare the initial and reformulated texts to verify the participant's performance after the instructions were provided. In the present study, two evaluators assessed the texts: a speech therapist and a psychologist, both of whom are extensively trained post-graduate students The evaluation consists comparing the participants' initial and final texts. Four specific aspects are observed in each participant's texts (initial and final/post-intervention): (1) changes made throughout the text in relation to the context of the theme and the characters; (2) changes made in relation to the development of the text (problem situation/proposed solution/action to solve the problem/character's reaction); (3) outcome of the theme and (4) overall content. The texts are assessed on a three-point Likert scale (did not comply with the criterion, partially complied with the criterion or fully complied with the criterion). The difference between the punctuation of the initial and final texts is assessed for each participant. In this last phase, it is possible to measure the participant's proximal development zone—i.e., the difference between each participant's potential and actual performance (Vygotsky, 1988).

Data Collection Procedures

The parents or responsible caregivers of the students from the participating schools were contacted to be informed about the research and to sign the Informed Consent Form. Only those students whose parents or responsible caregivers gave their informed consent participated in the research. The data collection took place in a private room at each school, in silence, without external interference, during regular class times, in groups of up to three participants and was monitored by three researchers. The groups were organized according age and school grade. The students received the instructions and took the test individually. The application of the complete TIDE (including the pretest, instruction model and post-test) took an average of 60 min.

Notebook computers with basic hardware capacity were used to support the TIDE software. After the test applications, the software automatically stored the tests in a database. At the end of each application session, a backup of the data was made on a removable disk. It is highlighted that this research did not pose risks to the participants and that individual privacy was preserved. Ethical aspects of research involving human beings were complied with. Approval was obtained from the Research Ethics Committee at the Federal University of Health Sciences of Porto Alegre, under protocol 502.515.

Data Analysis Procedure

The test answers were corrected and scored by two independent evaluators who were fully trained to apply and interpret the data according to the TIDE's preset criteria, as described in the tool. Any disagreements about the corrections were resolved by consensus and joint analysis. Exploratory factor analyses with varimax rotation were used, which is an orthogonal method that minimizes the number of variables in each cluster, resulting in a clearer separation between the factors. Factor analysis was used in the pre-test and instruction modules, since this type of analysis studies the relationship between variables and provides evidence on the quality of the instrument's structure. It is used to identify the latent dimensions and underlying variables, contributing to the theoretical interpretation of the test (Hair et al., 2006). Extraction via Principal Components with the varimax rotation method was adopted, given the structure of the items being analyzed. In addition, Cronbach's Alpha was calculated to analyze the pre-test and the instruction modules to verify the tool's reliability.

McNemar's Test was used to analyze the pre-test in order to determine the participants' evolution by comparing the analyses of each participant's final and initial texts, as this test is used to analyze related samples. Kendall's Agreement Test was used to analyze the agreement between the evaluators of the texts. The search for evidence of validity based on the external criterion of age was performed using the Kruskall Wallis test for independent samples and the chi-square test or Mann-Whitney test was used for school grade and gender. Statistical analysis of the collected data was developed in the Statistical Package for Social Science (SPSS), version 19.0.


The Kendall Test was applied initially to verify the agreement between the two evaluators' assessments, showing appropriate results that indicate the similarity between the researchers' analyses of the participants' texts. No correlations were found between the evaluators' scores of the participants' texts. The results were as follows: question 1: tau = 0.70 (p < 0.01); question 2: tau = 0.74 (p < 0.01); question 3: tau = 0.69 (p < 0.01); question 4: tau = 0.92 (p < 0.01). Based on this information, the items and correction protocol were considered to be sufficiently clear, permitting the uniformity of the process. For the purpose of analysis, we used the scores of Evaluator #2.

Factor analysis was subsequently applied to obtain evidence of validity based on the internal structure of the proposed tool. Therefore, the following steps were followed separately for the pre-test and instruction modules: (1) analysis of the conditions for the factor analysis (using the Kaiser-Meyer-Olkin (KMO) tests and Bartlett's sphericity test) and (2) definition of the number of factors in each test module (exploratory factor analyses). For the pre-test module, the (KMO) tests and Bartlett's sphericity test were applied to indicate the appropriateness of the data for the factor analysis. The former indicates the proportion of data variance, with values closer to one indicating that the sample is more appropriate for the factor analysis. In this study, the result was 0.76, representing an average result. The latter indicates whether the matrix demonstrates correlation among the data and the result was positive (p < 0.001).

In Step 2 of the pretest module, the presence of two factors was observed. Factor 1 corresponded to 41.62% of the explained variance between the items and Factor 2 corresponds to 15.90%. Of all of the items, four were related to Factor 1 (situation, response, action and solution), and three were related to Factor 2 (characters, scene and reaction). With respect to the communalities (the total amount of variance an original variable shares with the others), the minimum acceptable coefficient was 0.30 (Hair et al., 2009). All items in the test had higher coefficients. The item “solution” in Factor 1 and the item “scene” in Factor 2 presented the highest communality coefficients and were shown to be the most representative of each factor. Table 1 shows the coefficients of each component and each item.


Table 1. Factorial solution with factor loadings, communalities and Cronbach's Alpha for TIDE's pre-test module items (N = 304).

The (KMO) tests and Bartlett's sphericity test were also applied for the instruction module. The result was 0.92 for the former and the significance was p < 0.001 for the latter. Representing a very good result. In the instruction module, the presence of three factors was observed. In terms of variance, Factors 1, 2, and 3 presented 40.16, 8.28, and 6.73%, respectively. All items presented coefficients greater than the minimum for the communalities, ranging from 0.35 to 0.76. The items with the highest coefficients were “How is this problem solved?” in Factor 1, “How are they physically?” in Factor 2 and “What situations do they experience in the story?” in Factor 3. Table 2 displays the coefficient of each component and item.


Table 2. Factorial solution with factor loadings, communalities and Cronbach's Alpha for TIDE's instruction module items (N = 304).

Item reliability in the pre-test and the instruction module was verified using Cronbach's Alpha. It was observed that the TIDE coefficients were high, corresponding to 0.76 for the pretest and 0.91 for the instruction module. Internal consistency was demonstrated for the items in each of the modules tested.

McNemar's test was applied to analyze the post-test module and was used for dependent, paired samples (Hair et al., 2006). As the test requires, the answers were dichotomized and grouped under “did not comply with the criterion” and “complied with the criterion” (including “fully complied with the criterion” and “partially complied with the criterion”), thereby considering whether or not there were changes in the participants' behavior during the course of the test. Statistically significant positive results were observed, demonstrating the rejection of the hypothesis that the response percentages were the same in the initial and final texts. In other words, a relevant change was observed when considering the initial condition of the subjects' texts. Table 3 shows the results of the post-test obtained through the statistical analysis of the answers coded by McNemar's Test.


Table 3. Statistical analysis results of McNemar's Test for four questions in the TIDE's post-test (N = 304).

The qualitative analysis of the participants' performance revealed that, for the first aspect (changes throughout the text in the contextualization of the theme and the characters), 35.5% of the participants partially complied with the criterion in the initial text and fully in the final text, showing considerable evolution in the text production. For the second aspect (related to the development of the text, the problem situation, proposed solution, action to solve the problem and characters' reaction), a significant increase was observed in the proportion of participants who partially complied with the criterion at the beginning and fully complied at the end (21.1%). As for the third aspect (related to the outcome of the theme), the percentage of participants who still complied partially with the criterion stood out (27%). In the fourth and final aspect (regarding the content in general), a large proportion of participants continued not to comply with the criterion (76.2%). Based on the analysis of the post-test, the potential of the TIDE to diagnose the participants' difficulties was shown, as well as whether the intervention strategies resulted in better performance. These results are detailed in Table 4, which shows the descriptive statistics of the post-test results.


Table 4. Descriptive statistics for assessment of the initial and final texts in the TIDE's post-test (N = 304).

In regard to the participant's age (organized in quartiles of < 12, 12, 13, and 14 years or older), it was observed that there was no significant association in the pre-test module (x2 = 6.75, linear by linear association = 1.17, and p > 0.005), demonstrating that there was no association between age and performace on the test. Thus, the test is configured as an instrument with a wide range of possible applications, as it does not undergo changes in the responses according to the ages of the individuals submitted to this evaluation. Similarly in the instruction module, significant changes in performance were not observed using the Kruskal Wallis test (p > 0.005), as the estimated score medians were similar among the age groups analyzed.

The participants' school grade, which was divided into four categories (fifth, sixth, seventh and eighth grades) also did not show any association with the subjects' performance in the pre-test module (x2 = 0.20, p = 0.16 and linear by linear association = 2.43). As in the instruction module, the subjects' school grade was observed to have no influence on the responses given to each item (p > 0.005).

Finally, the gender variable was also considered as a possible influence on the results of the present study. However, this was also not significantly related to the pre-test module or the instructional module. The pre-test module score was not associated with gender (x2 = 5.74, p > 0.005, and linear by linear association = 5.58) and the median instructional module score was similar for boys and girls, rendering the difference statistically non-significant.


Health and education researchers and professionals frequently point to the need for evaluation methods that can be adapted and used in linguistically, culturally, socially and educationally different populations. Dynamic assessment is a suitable method for this purpose. Therefore, the development and use of tests, such as the TIDE that examine participants' learning potential are considered essential (Camilleri and Law, 2013). To demonstrate the suitability of the tool, two additional aspects were addressed in this study, thus broadening its objectives: (1) Does the TIDE present appropriate psychometric properties that permit its use and are related to evidence of reliability and validity based on its internal structure? (2) Was the capacity of the TIDE to diagnose difficulties demonstrated?

The first inquiry that guided this study was intended to demonstrate the reliability and psychometric evidence of the tool and permit its use. In this sense, the initial goal was to analyze reliability by calculating the Cronbach's Alpha coefficients of the pre-test and instruction modules. The results indicated that the TIDE's Cronbach's Alpha coefficients were high, as they fell within the ideal range (> 0.70) established by Streiner and Norman (2003). The instruction module presented the highest coefficient and the tool was found to be precise overall.

Regarding the search for evidence of validity based on the internal structure, exploratory factorial analysis (EFA) was also applied to the pre-test and instruction modules. The EFA results revealed that the pre-test module was divided into two factors that support the structure of a narrative (Pinto et al., 2015). Factor 1 grouped items related to the main conflict present in the story (including situation, response, action and solution) and also characterized central issues guiding the text's development process, such as the production of ideas and the establishment of targets, in accordance with the model developed by Flower and Hayes (1981). Factor 2 included items related to the context in which the story takes place, including the characters, scene and the characters' reactions, which is also in line with the organization of ideas (Flower and Hayes, 1981). It is important to highlight that this factor presented the highest explained variance, demonstrating the relevant portion of common variance the factor extracted from the data set.

The factor loadings of the items ranged from 0.55 to 0.86, which are higher than the minimum loadings established in the literature (0.30), indicating that the TIDE items are highly relevant to assessing text production in the narrative model (Hair et al., 2009). In view of these results, it was also demonstrated that the orientation procedure in the first text production phase (and selected in the literature) was also valid (Flower and Hayes, 1981). Other studies have included this method as the basis for written production, including those by Matuchniak et al. (2014), Zheng et al. (2015), and Bruning et al. (2013).

It is highlighted that the writing process also implies metacognitive functioning, as proposed by van der Stel and Veenman (2014), which acts in an articulated manner with the text production structure, resulting in a narrative that meets the expected regulatory standards. This is the case especially due to the participant's ability to reflect and rethink his/her writing process. In this sense, the orientations and instructions provided during the TIDE promote the participant's reflection on his/her own production. Participants can develop skills related to analysis, control, planning, target setting and task revision, which take place at the beginning and the end. This process is reflective, as observed in the parameters considered in the TIDE's pre-test and instruction modules. It is also related to cognitive models, as the narrative task is sometimes considered a core cognitive action, constituting human thought. At other times, it is considered a cognitive processing model, although it is more paradigmatic and is based on more social phenomena (Genereux and McKeough, 2007).

The planning, translation and review steps of the writing model developed by Hayes and Flower's (1980) are based on the cognitive processes of text composition. Thus, the more experienced the writing participant is, the better he/she will produce objectives, storage and automation of the repertoire, in addition to reflecting on his/her writing and interpretation, including memory and motivation (Erhard et al., 2014). The internal structure of the TIDE is therefore appropriate for the assessment in question, as observed in the study conducted by Joly et al. (2015).

The EFA of the instruction module identified three factors, confirming the model proposed by Flower and Hayes (1981) for the text writing process. Thus, the TIDE can be considered to be didactically divided into three phases: planning, translation and review. The first phase is divided into the production of ideas and organization and setting of targets, while the third is divided into assessment and revision. When comparing the model by Flower and Hayes (1981) with the division into factors observed in the TIDE, Factor 1 included the items aimed at the production of ideas and setting of targets (belonging to planning) and their translation. Factor 2 included items focused on the organization of the ideas and the translation. Factor 3 included review items, which add complementary information to the text's main information (included in the items belonging to Factors 1 and 2 of the test).

For example, the model developed by Flower and Hayes (1981) does not presuppose a definitive predefined time model. Some participants take more time in the planning phase, while others put more energy into the revision phase, demonstrating variation among the participants that should also be considered and the need to make the analysis of the premise more flexible. In addition, considering the variation of the cognitive processes the participants use, the writing process should not be considered to be a strict sequence of phases, but rather a “set of optional cases.” For example, the idea production phase may need to be revised. Therefore, the writing process is flexible, despite its pre-established ideal hierarchy and despite being a goal-oriented process (Hayes and Flower, 1980). These aspects were verified in the instruction module's EFA and therefore support the classification of the TIDE's instruction module.

The tools constructed based on dynamic assessment (as shown thus far) are in line with Yancey (1999) and Wardle and Roozen (2012) as a process that values participants' learning. In this sense, the present article provides important contributions, presenting the development of a tool that enables the potential learning of writing to be analyzed. This fact can be particularly observed in the TIDE's final module (the post-test), which enabled a positive and satisfactory response to the second inquiry in this study, which was related to the potential learning and evolution in the adolescents' performance when writing the proposed narrative text.

The TIDE's post-test considers three main points: cohesion, coherence and structure. These three points are included in the four post-test questions, which provide the results of the participants' evolution, comparing the initial and final production. In addition, other relevant text production issues are considered in the post-test, such as the presentation of a beginning that permits expansion on the theme, characters and their descriptions, a scene, the resolution of the problems and an ending, thereby comprising the development of a story with a beginning, middle and end and a coherent and cohesive logical order (Genereux and McKeough, 2007; Pinto et al., 2015).

It was observed that age, school grade and sex were not factors that influenced the subjects' responses, as was the case in the pre-test module, also demonstrating the homogeneity of the sample in regard to the responses given to this module's items, as well as the instrument's stability. Further studies should be conducted to clarify and deepen the understanding of these issues, especially by using other school samples. This is due to the fact that, according to dos Santos and Befi-Lopes (2016), there is an increasing improvement in writing ability as students advance in school grades. Initially, students' writing reflects their speech, using simple syntax and orally-based vocabulary. From the fifth grade onwards, students begin to organize their sentences hierarchically, using more formality and graphic mastery and developing the linguistic and cognitive skills of written language.

The results for the four questions mentioned earlier showed statistically significant coefficients, demonstrating the appropriateness of the answers provided. Thus, it can be assumed that, among the four criteria the evaluators considered in the initial and final texts of each participant, it appears that the instructions the participants received during the course of the instruction module did not only influence the overall content. This fact can be observed in the large number of participants who did not comply with the criterion initially and continued not to comply in their final text. The remaining questions are easier to change in the text and are more open to improvements, while the content (which represents the core idea in the story) requires more effort and modification for its qualitative evolution.

In response this study's second research question (based on the analysis of the post-test), the TIDE's potential to diagnose the participants' difficulties and learning potential was demonstrated. The participants' learning potential was based on the premises of hints and instructions. It has been confirmed that the dynamic assessment method, as originated by Vygotsky (1987), enables the verification of emerging skills, developmental possibilities and the quality of mediated teaching. It also allows for the designation of the participants' proximal development zone through a more qualitative analysis, providing information and describing the participants' trajectory of written production (Poehner and Lantolf, 2013).

Regarding the final analysis of the test the TIDE proposes, no established standard exists to share the diagnosis observed in the dynamic testing. Nevertheless, Poehner et al. (2015) have commented that three manners are described in the literature: the creation of student categories, the proposal of scores and the description of student profiles. In this sense, the TIDE is scored to indicate a participant's performance based on a comparison of his/her initial and final texts, which supports the determination of his/her performance and learning potential. The TIDE also enables the respondent's profile to be analyzed through a qualitative analysis of his/her productions.

Finally, the computerized nature of the dynamic assessment of writing is discussed in order to promote reflection on the certainty and validity of this type of resource to assess adolescents' learning potential. Regarding this theme, Poehner and Lantolf (2013) showed that the first attempt to use computerized dynamic assessment was through the “Lerntest” tool, which was a learning test developed by Guthke and Beckmann (2000). Other tools were later developed, such as the “C-DA,” conceived by Tzuriel and Shamir (2002), which assesses cognitive aspects of mathematics learning, providing information on the student's learning. According to the authors, both tools are appropriate for computerized assessments. In this sense, the computerized assessment environment and its possible interaction are crucial points for evaluation and mediated learning activities. Also based on Vygotsky's theory of the proximal development zone, the potential for dynamic assessment based on external forms of mediation is highlighted, which can include dialogical interaction with an advisor and the use of resources, such as graphs, diagrams and models (Alderson, 2007; Poehner et al., 2015), including computers in this case.

There is a difference between the dynamic testing method developed by (Poehner et al., 2015) and the TIDE The former method interacts with participants to demonstrate when the answer is wrong through the sentence “That is not the correct answer.” On the other hand, due to its more comprehensive nature and because it proposes a freer task, the TIDE questions the participants about the text's components, teaching them the core aspects of a narrative text and the appropriate writing process in general, without any feedback on possible right or wrong answers. The main idea of the TIDE, as described by Tzuriel (2000), is to attempt to transform a students' impulsive writing style into a reflective model, granting them the opportunity to reflect on their writing process in a more conscious manner.

It is important to note that the information provided in this test is not specific to each participant and does not depend on his/her performance. In the TIDE, assessment centers on the inquiries the participants are confronted with and on the transfer of the questions asked during the instrumental module to the actual text. A participant's negative response to questions will point to topics that need to be changed, thus leading him/her to improve his/her text afterwards. Likewise, his/her positive response to questions will point to appropriate aspects of the text produced, indicating that there is no need to modify the material.

The purpose of the TIDE is to assess the potential to learn writing, using strategies that teachers would use to help students improve their narrative writing, such as a structured outline, asking them questions and encouraging editing and revising. The TIDE's purpose is to standardize this intervention in order to offer an instrument to the academic community that is capable of evaluating such issues, using technological resources.

Other authors also use digital platforms to stimulate students' narrative writing, as in the case of Malaysia (Annamalai et al., 2013), transforming the platform into space to learn narrative writing, demonstrating the participants' evolution in comparison with traditional methods. Therefore, there is a clear need for a preliminary analysis of aspects related to the participants that will enable the application of computerized writing tests, such as lack of experience, familiarity and knowledge about using this type of tool (Jeong, 2014).

Writing differences were also observed among the participants according to sex and age, with women performing better. A dichotomy was found with regard to age in that the older participants had more mature writing skills but less motivation, which appeared to decline with age (Troia et al., 2013). Therefore, new studies using the TIDE are being prepared, taking into account and analyzing data on participants and related constructs and looking for evidence of validity based on external criteria.


As observed in the literature, there is still a lack of investment in the assessment of students' written expression. Nevertheless, as a result of the more widespread use of computers, mobile phones and text messages, ideas are increasingly being transmitted in written form. Therefore, precise assessment methods are important for the analysis of and early intervention in persistent writing problems, including in the composition and construction of written texts (Berninger et al., 2015). In this sense, the present article offers relevant contributions to the specialized literature, as it presents data on the process of searching for evidence of the validity of the internal structure of an innovative test to assess adolescent narrative writing. Thus, the lack of proposed tools that specifically focus on adolescents is minimized.

In conclusion, (TIDE) is an appropriate and precise tool for assessing adolescent narrative writing that focuses on this population's learning potential (particularly in Brazil), which is the focus of this study in searching for evidence of validity. The original nature of the tool presented here is highlighted, based on the dynamic assessment method, offering data on a learning method for narrative writing as well as its possible adaptation to other contexts and languages. In addition, the computer-based nature of the tool is emphasized, enabling more precise application and analysis of the participants' performance, thus lowering costs, reducing application bias and enabling the testing of more than one person simultaneously (Bauer et al., 2012). On the other hand, the tool has a limitation in that it lacks flexibility in the instructions provided throughout the assessment process, as the computerized tool is pre-programmed. The instructions do not change during the course of the test, although the authors point out the possibility of using other instructional resources for participants who are more experienced with the task (Alderson, 2007; Poehner et al., 2015).

Finally, this study is a pioneer in its field. In addition to presenting an innovative tool and evidence of the validity of its internal structural, it aims to stimulate further research on this topic. In line with the suggestions by Muniz et al. (2015), more studies are needed to investigate the relationship between the assessment of learning potential and related constructs, such as participants' educational performance and school grade.

Ethics Statement

This study was carried out in accordance with the recommendations of Research Ethics Committee at Universidade Federal de Ciências da Saúde de Porto Alegre with written informed consent from all subjects. All subjects gave written informed consent in accordance with the Declaration of Helsinki. The protocol was approved by the Research Ethics Committee at Universidade Federal de Ciências da Saúde de Porto Alegre, under protocol 502.515.

Author Contributions

LGG and CTR - contributed to the conception and design of the work, as well as the acquisition, analysis, and interpretation of data for the work. MMCO and MCRAJ - drafted the work and revised it critically for important intellectual content.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.


American Educational Research Association (AERA), American Psychological Association (APA), National Council on Measurement in Education (NCME) (2014). Standards for Educational and Psychological Testing. American Educational Research Association, Washington, DC.

Alavi, S. M., and Taghizadeh, M. (2014). Dynamic assessment of writing: the impact of implicit/explicit mediations on L2 learners' internalization of writing skills and strategies. Educ. Assess. 19, 1–16. doi: 10.1080/10627197.2014.869446

CrossRef Full Text | Google Scholar

Alderson, J. C. (2007). “The challenge of (diagnostic) testing: do we know what we are measuring?,” in Language Testing Reconsidered, eds J. Fox, M. Wesche, D. Bayliss, L. Cheng, C. Turner, and C. Doe (Ottawa, ON: University of Ottawa Press), 21–39.

Google Scholar

Annamalai, N., Eng, T. K., and Abdullah, A. (2013). Exploring the effects of an online writing platform on students' performances in narrative writing. Malays. J. Distance Educ. 15, 1–18.

Google Scholar

Bauer, R. M., Iverson, G. L., Cernich, A. N., Binder, L. M., Ruff, R. M., and Naugle, R. I. (2012). Computerized neuropsychological assessment devices: joint position paper of the american academy of clinical neuropsychology and the national academy of neuropsychology. Clin. Neuropsychol. 26, 177–196. doi: 10.1080/13854046.2012.663001

PubMed Abstract | CrossRef Full Text | Google Scholar

Berninger, V. W., Nagy, W., Tanimoto, S., Thompson, R., and Abbott, R. D. (2015). Computer instruction in handwriting, spelling, and composing for students with specific learning disabilities in grades. Comput. Educ. 81, 154–168. doi: 10.1016/j.compedu.2014.10.005

PubMed Abstract | CrossRef Full Text | Google Scholar

Bruning, R., Dempsey, M., Kauffman, D. F., McKim, C., and Zumbrunn, S. (2013). Examining dimensions of self-efficacy for writing. J. Educ. Psychol. 105, 25. doi: 10.1037/a0029692

CrossRef Full Text | Google Scholar

Camilleri, B., and Law, J. (2013). Dynamic assessment of word learning skills of preschool children with primary language impairment. Int. J. Speech Lang. Pathol. 16, 507–516. doi: 10.3109/17549507.2013.847497

CrossRef Full Text

Davidson, C. A., Johannesen, J. K., and Fiszdon, J. M. (2016). Role of learning potential in cognitive remediation: construct and predictive validity. Schizophr. Res. 171, 117–124. doi: 10.1016/j.schres.2016.01.044

PubMed Abstract | CrossRef Full Text | Google Scholar

Dias, T. L., and Enumo, S. R. F. (2006). Criatividade e dificuldade de aprendizagem: avaliação com procedimentos tradicional e assistido. Psicol. Teoria e Pesqui. 22, 069–078. doi: 10.1590/S0102-37722006000100009

CrossRef Full Text | Google Scholar

Enumo, S. R. F. (2005). Avaliação assistida para crianças com necessidades educacionais especiais: um recurso auxiliar na inclusão escolar. Rev. Bras. de Educ. Espec. 11, 335–354. doi: 10.1590/S1413-65382005000300003

CrossRef Full Text | Google Scholar

Erhard, K., Kessler, F., Neumann, N., Ortheil, H. J., and Lotze, M. (2014). Professional training in creative writing is associated with enhanced fronto-striatal activity in a literary text continuation task. Neuroimage 100, 15–23. doi: 10.1016/j.neuroimage.2014.05.076

PubMed Abstract | CrossRef Full Text | Google Scholar

Fazlollahi, E., Marefat, F., and Vaezi, S. (2015). The application of dynamic assessment: is it worth the effort?. J. Lang. Teach. Res. 6, 985–992. doi: 10.17507/jltr.0605.10

CrossRef Full Text | Google Scholar

Feuerstein, R., Rand, Y., and Hoffman, M. (1979). The Dynamic Assessment of Retarded Performers: The Learning Potential Assessment Device (LPAD). Baltimore, MD: University Park Press.

Google Scholar

Flower, L., and Hayes, J. R. (1981). A cognitive process theory of writing. Coll. Comp. Commun. 32, 365–387. doi: 10.2307/356600

CrossRef Full Text | Google Scholar

Genereux, R., and McKeough, A. (2007). Developing narrative interpretation: structural and content analyses. Br. J. Educ. Psychology 7(Pt 4), 849–872. doi: 10.1348/000709907X179272

CrossRef Full Text | Google Scholar

Guthke, J., and Beckmann, J. F. (2000). “The learning test concept and its applications in practice,” in Dynamic Assessment: Prevailing Models and Applications, eds C. S. Lidz and J. G. Elliott (Amsterdam: Elsevier), 17–69.

Hair, J., Anderson, R., Tatham, R., and Black, W. (2006). Multivariate Data Analysis, 6th Edn. Upper Saddle River, NJ: Pearson Education, Inc.

Hair, J. F., Black, W. C., Babin, B. J., Anderson, R. E., and Tatham, R. L. (2009). Análise Multivariada de dados. Porto Alegre: Bookman.

Hayes, J. R., and Flower, L. S. (1980). “Identifying the organization of writing processes,” in Cognitive Processes in Writing, eds L. W. Gregg and E. R. Steinberg (Hillsdale, NJ: Lawrence Erlbaum Associates), 3–30.

Google Scholar

Jacobs, E. L. (2001). The effects of adding dynamic assessment components to a computerized preschool language screening test. Commun. Disord. Q. 22, 217–226. doi: 10.1177/152574010102200407

CrossRef Full Text | Google Scholar

Jeong, H. (2014). A comparativestudyofscoresoncomputer-based tests and paper-based tests. Behav. Inf. Technol. 33, 410–422. doi: 10.1080/0144929X.2012.710647

CrossRef Full Text | Google Scholar

Joly, M. C. R. A., and Schiavoni, A. (2013). Teste Informatizado e Dinamico de Escrita. Projeto de pesquisa. Universidade de Brasília: Programa de Pós-graduação em Processos de Desenvolvimento e Aprendizagem, Brazil.

Joly, M. C. R. A., Schiavoni, A., Agostinho, A., and Dias, A. S. (2015). Avaliação dinâmica e produção textual: evidências de validade para o TIDE no Fundamental. Cadernos Pós Graduação Distúrb. Desenvol. 15, 50–62.

Kim, Y., Otaiba, S. A., Folsom, J. S., Greulich, L., and Puranik, S. (2014). Evaluating the dimensionality of first-grade written composition. J. Speech Lang. Hear. Res. 57, 199–211. doi: 10.1044/1092-4388(2013/12-0152)

PubMed Abstract | CrossRef Full Text | Google Scholar

Linhares, M. B. M., Escolano, A. C. M., and Enumo, S. R. F. (2006). “Avaliação cognitiva assistida: fundamentos teórico-conceituais e contribuições,” in Avaliação Assistida: Fundamentos, Procedimentos e Aplicabilidade, eds M. B. M. Linhares, A. C. M. Escolanoand, and S. R. F. Enumo (São Paulo: Casa do Psicólogo), 16–32.

Matuchniak, T., Olson, C. B., and Scarcella, R. (2014). Examining the text-based, on-demand, analytical writing of mainstreamed Latino English learners in a randomized field trial of the pathway project intervention. Read. Writ. 27, 973–994. doi: 10.1007/s11145-013-9490-z

CrossRef Full Text | Google Scholar

Muniz, M., Seabra, A. G., and Primi, R. (2015). A relação entre potencial de aprendizagem e desempenho acadêmico: predição pelo teste dinβmico informatizado de raciocínio indutivo para crianças. Estud. de Psicol. 32, 343–356. doi: 10.1590/0103-166X2015000300001

CrossRef Full Text | Google Scholar

Pena, E. D., Gillam, R. B., and Bedore, L. M. (2014). Dynamic assessment of narrative ability in English accurately identifies language impairment in English language learners. J. Speech Lang. Hear. Res. 57, 2208. doi: 10.1044/2014_JSLHR-L-13-0151

PubMed Abstract | CrossRef Full Text | Google Scholar

Pinto, G., Tarchi, C., and Bigozzi, L. (2015). The relationship between oral and written narratives: a three year longitudinal study of narrative cohesion, coherence, and structure. Br. J. Educ. Psychol. 85, 551–569. doi: 10.1111/bjep.12091

PubMed Abstract | CrossRef Full Text | Google Scholar

Plake, B. S., and Wise, L. L. (2014). What is the role and importance of the revised AERA, APA, NCME Standards for educational and psychological testing?. Educ. Meas. Issues Pract. 33, 4–12. doi: 10.1111/emip.12045

CrossRef Full Text | Google Scholar

Poehner, M. E., and Lantolf, J. P. (2013). Bringing the ZPD into the equation: capturing L2 development during Computerized Dynamic Assessment (C-DA). Lang. Teach. Res. 17, 323–342. doi: 10.1177/1362168813482935

CrossRef Full Text | Google Scholar

Poehner, M. E., Zhang, J., and Lu, X. (2015). Computerized dynamic assessment (C-DA): diagnosing L2 development according to learner responsiveness to mediation. Lang. Test. 32, 337–357. doi: 10.1177/0265532214560390

CrossRef Full Text | Google Scholar

Richards, S. A. (2015). Characteristics, assessment, and treatment of writing difficulties in college students with language disorders and/or learning disabilities. Top. Lang. Disord. 35, 329–344. doi: 10.1097/TLD.0000000000000069

CrossRef Full Text | Google Scholar

Santos, M. T. M., and Befi-Lopes, D. M. (2016). Análise da produção de narrativa escrita de escolares do ensino fundamental. Distúrb. Comun. 28, 231–243.

Google Scholar

Sternberg, R. J., and Grigorenko, E. L. (2002). Dynamic Testing: The Nature and Measurement of Learning Potential. New York, NY: Cambridge University Press.

Google Scholar

Streiner, D. L., and Norman, G. R. (2003). Health Measurement Scales: A Practical Guide to their Development and Use, 3rd Edn. Oxford: Oxford University Press.

Troia, G. A., Harbaugh, A. G., Shankland, R. K., Wolbers, K. A., and Lawrence, A. M. (2013). Relationships between writing motivation, writing activity, and writing performance: effects of grade, sex, and ability. Read. Writ. 26, 17–44. doi: 10.1007/s11145-012-9379-2

CrossRef Full Text | Google Scholar

Tzuriel, D. (2000). Dynamic assessment of young children: educational and intervention perspectives. Educ. Psychol. Rev. 12, 385–435. doi: 10.1023/A:1009032414088

CrossRef Full Text | Google Scholar

Tzuriel, D., and Shamir, A. (2002). The effects of mediation in computer assisted dynamic assessment. J. Comp. Assist. Learn. 18, 21–32. doi: 10.1046/j.0266-4909.2001.00204.x

CrossRef Full Text | Google Scholar

van der Stel, M., and Veenman, M. V. (2014). Metacognitive skills and intellectual ability of young adolescents: a longitudinal study from a developmental perspective. Eur. J. Psychol. Educ. 29, 117–137. doi: 10.1007/s10212-013-0190-5

CrossRef Full Text | Google Scholar

Vygotsky, L. S. (1987). “Thinking and speech,” in Problems of General Psychology, Vol. 1. The collected works of L. S. Vygotsky, eds R. W. Rieber and A. S. Carton (New York, NY: Plenum), 39–285.

Vygotsky, L. S. (1988). A Formação Social da Mente - o Desenvolvimento dos Processos Psicológicos Superiores. São Paulo: Martins Fontes.

Vygotsky, L. S. (1998a). “The problem of age,” in Child Psychology, The collected works of L. S. Vygotsky, eds R. W. Rieber and A. S. Carton (New York, NY: Plenum), 187–206.

Vygotsky, L. S. (1998b). O Desenvolvimento Psicológico Na Infância. São Paulo: Martins Fontes.

Wardle, E., and Roozen, K. (2012). Addressing the complexity of writing development: toward na ecological model of assessment. Assess. Writing 17, 106–119. doi: 10.1016/j.asw.2012.01.001

CrossRef Full Text | Google Scholar

Wilson, J., Olinghouse, N. G., McCoach, B., Santangelo, T., and Andrada, G. N. (2016). Comparing the accuracy of different scoring methods for identifying sixth graders at risk of failing a state writing assessment. Assess. Writing 27, 11–23. doi: 10.1016/j.asw.2015.06.003

CrossRef Full Text | Google Scholar

Wolfe, E. W., Song, T., and Jiao, H. (2016). Features of difficult-to-score essays. Assess. Writing 27, 1–10. doi: 10.1016/j.asw.2015.06.002

CrossRef Full Text | Google Scholar

Yancey, K. B. (1999). Looking back as we look forward: historicizing writing assessment. Coll. Comp. Commun. 50, 483–503. doi: 10.2307/358862

CrossRef Full Text | Google Scholar

Zheng, B., Lawrence, J., Warschauer, M., and andLin, C. H. (2015). Middle school students' writing and feedback in a cloud-based classroom environment. Technol. Knowl. Learn. 20, 201–229. doi: 10.1007/s10758-014-9239-z

CrossRef Full Text | Google Scholar

Keywords: learning, validation studies, adolescent, software validation, psychological assessment

Citation: Gurgel LG, de Oliveira MMC, Joly MCRA and Reppold CT (2017) Learning Potential in Narrative Writing: Measuring the Psychometric Properties of an Assessment Tool. Front. Psychol. 8:719. doi: 10.3389/fpsyg.2017.00719

Received: 23 December 2016; Accepted: 21 April 2017;
Published: 10 May 2017.

Edited by:

Pietro Cipresso, Istituto Auxologico Italiano (IRCCS), Italy

Reviewed by:

Julia H. Chariker, University of Louisville, USA
Pedro Armelim Almiro, University of Coimbra, Portugal

Copyright © 2017 Gurgel, de Oliveira, Joly and Reppold. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Caroline T. Reppold,