Skip to main content

GENERAL COMMENTARY article

Front. Psychol., 05 May 2015
Sec. Cognitive Science
This article is part of the Research Topic Frontiers in the Acquisition of Literacy View all 14 articles

What are the criteria for a good intervention study? Response: “Unrecognized ambiguities in validity of intervention research: an example on explicit phonics and text-centered teaching”

  • 1Institute of Education, Massey University, Auckland, New Zealand
  • 2School of Curriculum and Pedagogy, The University of Auckland, Auckland, New Zealand

A commentary on
Unrecognized ambiguities in validity of intervention research: an example on explicit phonics and text-centered teaching

by Thompson G. B. (2015). Front. Psychol. 5:1535. doi: 10.3389/fpsyg.2014.01535

Thompson (2015) has raised several validity issues about our study (Tse and Nicholson, 2014) while acknowledging that it would score well in terms of Troia's (1999) criteria for “What makes a good study?” A response to the critique is detailed briefly below.

Thompson's first point was that the study lacked evidence about what instruction children received prior to and concurrent with the intervention study. Interactions with teachers and the principals of the schools however indicated that reading instruction was similar from one school to the next. Any differences among schools and classrooms were also controlled for in that participants were randomly assigned to groups thus spreading possible effects of differences in instruction across all groups.

The second point was the absence of justification for the phonics rules taught however the article explained that the taught Anglo-Saxon decoding rules were from Calfee and Patrick's (1995) well-known explanation of Anglo-Saxon letter-sound patterns. The intervention followed their scope and sequence in the study. It is not clear why this might be a validity problem in that the study did reference the source of the phonics rules.

The third point was that participants' vocabulary age was low at 4.8 years compared with chronological age of 6.3 years and thus Big Books may have been inappropriate. Their standard score was 86 which is close to the average range (90–110) and there are studies to support Big Book reading with lower SES children such as these (Nicholson and Whyte, 1992; Valdez-Menchaca and Whitehurst, 1992; Whitehurst et al., 1994). The Big Books were also selected so as to be at the reading level of the children who were being taught and given that their reading level was in the beginner range the language should have been understandable for them.

The fourth point was that the article did not discuss whether children had opportunities to use their decoding skills to process the items of the pre and post-test measures. Although not reported our data did confirm that the combined group scored better on regular words (e.g., went) than irregular (e.g., love). The Bryant Test of Basic Decoding skills also gave opportunities to use decoding skills.

The fifth point was that the phonics group practiced phonics quizzes but the Big Book group did not practice reading of text. This was not completely the case. Children in the Big Book group did get opportunities to practice reading of text through the Big Book lessons. They did three readings of each text and read along with the teacher.

The sixth point was that the orthogonal analysis was not sufficient and needed to compare the performance gains of the combined group with those of the treatment control group (math-only). To do this however risked statistical error so instead of carrying out all possible comparisons among the four groups the decision was to use Helmert contrasts which were pre-planned orthogonal contrasts. This approach offered protection against statistical error (Kwon, 1996; Keppel and Wickens, 2004). As Kuehne (1993) has pointed out, using post-hoc comparisons increases the chance of type 1 error (in the study, to do six post-hoc comparisons across four groups would increase the possibility of type 1 error to 26%). The Helmert contrast procedure is common in other disciplines but less common in education. The way the Helmert contrasts worked in the study was that the control group mean was first compared with the overall mean score for the other three groups. Then the phonics enhanced Big Books group mean was compared with the overall mean for the two remaining groups (Big Book and phonics). Finally the means of the Big Book and phonics groups were compared. It was like peeling an onion. The logic was that if the control group was not better than the mean of the other three groups and if the phonics enhanced group was better than the mean of the combined Big Book and phonics groups, and if there was no difference in the contrast between the Big Book and phonics groups, then it can be inferred that the phonics enhanced group was superior to the other groups. The orthogonal contrast worked just as well as all possible contrasts with less risk of type 1 and 2 error.

The seventh issue was that speed of reading was not reported. Thompson's previous research would suggest a slower reading speed for the phonics enhanced Big Book group but it could counter-wise be argued that they would have gained similar fluency to the Big Book group because they also read Big Books. To answer this question, fluency would be a useful variable for future studies to find out which approach is more effective for fluency.

To conclude, one reviewer commented that the present study could be “a model for how such work might be conducted on a larger scale, which might lead New Zealand and other nations to progress in dealing with the [achievement] gap issue.” Replicating and scaling up the present study will clarify further whether enhancing Big Book reading with explicit phonics brings disadvantaged children closer to their expected reading and spelling age in a short time with only a small adjustment to Big Book instruction.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

Calfee, R. C., and Patrick, C. L. (1995). Teach Our Children Well. Bringing K-12 Education into the 21st Century. Stanford, CA: Stanford University Alumni Association.

Keppel, G., and Wickens, T. D. (2004). Design and Analysis. Upper Saddle River, NJ: Pearson.

Google Scholar

Kuehne, C. C. (1993). “The advantages of using planned comparisons over post hoc tests,” in Paper Presented at the Annual Meeting of the Mid-South Educational Research Association (New Orleans, LA).

Kwon, M. (1996). “The use of planned comparisons in analysis of variance research,” Paper Presented at Annual Meeting of the Southwest Educational Research Association (New Orleans, LA).

Nicholson, T., and Whyte, B. (1992). “Matthew effects in learning new words while listening to stories,” in Literacy Research, Theory, and Practice: Views from Many Perspectives eds C. K. Kinzer and D. J. Leu (Chicago: National reading Conference), 499–503.

Thompson, G. B. (2015). Unrecognized ambiguities in validity of intervention research: an example on explicit phonics and text-centered teaching. Front. Psychol. 5:1535. doi: 10.3389/fpsyg.2014.01535

PubMed Abstract | Full Text | CrossRef Full Text | Google Scholar

Troia, G. A. (1999). Phonological awareness intervention research: a critical review of the experimental methodology. Read. Res. Q. 34, 28–52.

Google Scholar

Tse, L., and Nicholson, T. (2014). The effect of phonics-enhanced Big Book reading on the language and literacy skills of six-year-old pupils of different reading ability attending lower SES schools. Front. Psychol. 5:1222. doi: 10.3389/fpsyg.2014.01222

PubMed Abstract | Full Text | CrossRef Full Text | Google Scholar

Valdez-Menchaca, M. C., and Whitehurst, G. J. (1992). Accelerating language development through picture book reading: a systematic extension to Mexican day-care. Dev. Psychol. 28, 1106–1104.

Google Scholar

Whitehurst, G. J., Arnold, D. S., Epstein, J. N., Angell, A. L., Smith, M., and Fischel, J. E. (1994). A picture book reading intervention in day-care and home for children from low-income families. Dev. Psychol. 30, 679–689.

Google Scholar

Keywords: teaching reading, big books, explicit phonics, text-centered teaching, planned comparisons, post-hoc comparisons, validity, intervention research

Citation: Nicholson T and Tse L (2015) What are the criteria for a good intervention study? Response: “Unrecognized ambiguities in validity of intervention research: an example on explicit phonics and text-centered teaching”. Front. Psychol. 6:508. doi: 10.3389/fpsyg.2015.00508

Received: 05 January 2015; Accepted: 10 April 2015;
Published: 05 May 2015.

Edited by:

Claire Marie Fletcher-Flinn, University of Auckland, New Zealand

Reviewed by:

Brian Byrne, University of New England, Australia

Copyright © 2015 Nicholson and Tse. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Tom Nicholson, t.nicholson@massey.ac.nz

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.