# Confidence intervals and tests are two sides of the same research question

- Business School, Massey University, Palmerston North, New Zealand

A commentary on

The new statistics: why and how

*by Cumming, G. (2014). The new statistics: why and how. Psychol. Sci. 25(1), 7–29. doi: 10.1177/0956797613504966*

Cumming's (2014) article was commissioned by *Psychological Science* to reinforce the journal's 2014 publication guidelines. It exhorts substituting confidence intervals (CIs) for Null Hypothesis Significance Testing (NHST) as a way of increasing the scientific value of psychological research. Cumming's article is somehow biased, hence the aims of my commentary: to balance out the presentation of statistical tests and to fend CIs against misinterpretations. Researchers with an interest in the correct philosophical application of tests and CIs are my target audience.

## Technical Tests

NHST is a philosophical mismatch of three incompatible theories: Fisher's, Neyman-Pearson's, and Bayes's (Gigerenzer, 2004). Technologically, however, it reduces to either of the former two, most often to Fisher's (Cortina and Dunlap, 1997). Few researchers concern themselves with the philosophical underpinnings of NHST and rather use it as a mere technology. Therefore, it is more interesting to discuss Fisher's (1954) or Neyman and Pearson's theories (1933) than NHST. Cumming does so in his book (2012) but not in his article, thus painting an inaccurate picture of the value of data testing in research.

In a nutshell, Fisher's relevant constructs are null hypotheses, levels of significance and *ad hoc p*-values. Neyman-Pearson's relevant constructs are main and alternative hypotheses, long-run errors under repeated sampling (α, β), critical test values, and power (1–β)—*p*-values are not relevant here but can be used as proxies for deciding between hypotheses (Perezgonzalez, 2014).

Cumming's Figure 1 about the dance of p's (alternatively, http://youtu.be/ez4DgdurRPg) is not suitable for representing Fisher's *ad hoc* approach (and, by extension, most NHST projects). It is, however, adequate for representing Neyman-Pearson's repeated sampling approach, a simple count of significant tests irrespective of their *p*-value. As it turns out, Cumming's figure is a textbook example of what to expect given power (Table 1). For example, 48% of tests are significant at α = 0.05 out of 50% expected.

## Fair Comparison

Cumming writes “CIs and *p* values are based on the same statistical theory” (p. 13), which is partly correct. CIs were developed by Neyman (1935) and, thus, CIs and Neyman-Pearson's tests are grounded on the same statistical philosophy: repeated sampling from the same population, errors in the long run (1−CI = β), and assumption of true population parameters (which are unknown in the case of CIs)—*p*-values, however, are part of Fisher's theory.

CIs and Neyman-Pearson's tests are, thus, two sides of the same coin. Tests work on the main hypotheses, with known population parameters but unknown sample probabilities, and calculate point estimates to make decisions about those samples. CIs work on the alternative hypotheses, with known sample interval probabilities but unknown population parameters, and calculate CIs to describe those populations.

To be fair, a comparison of both requires equal ground. At interval level, CIs compare with power, and Cumming's figure reveals that 92% of sample statistics fall within the population CIs (95% are expected) versus the power results presented in Table 1. At point estimate level, means (or a CI bound) compare with *p*-values, and Cumming's figure reveals a well-choreographed dance between those. Namely, CIs are not superior to Neyman-Pearson's tests when properly compared although, as Cumming discussed, CIs are certainly more informative.

## Fending Against Fallacies

The common philosophy underlying CIs and tests implies that they share similar fallacies. Cumming touches on some but does not pre-emptily resolve them for CIs.

Cumming writes, “if *p* reveals truth… ” (p. 12). It is not clear what truth Cumming refers to, most probably about two known fallacies: that *p* is the probability of the main hypothesis being true and, consequently, that 1–*p* is the probability of the alternative hypothesis being true (e.g., Kline, 2004). Similar fallacies equally extend to the power of the alternative hypothesis. Yet accepting the alternative hypothesis does not mean a 1–β probability of it being true, but a probability of capturing 1–β samples pertaining to its population in the long run. The same can be said about CIs (insofar CI = 1–β): they tell something about the data—about their probability of capturing a population parameter in the long run—not about the population—i.e., the observed CI may not actually capture the true parameter.

Another fallacy touched upon is that *p* informs about replication. *P*-values only inform about the *ad hoc* probability of data under the tested hypothesis, thus “(they) have little to do with replication in the usual scientific sense” (Kline, 2004, p. 66). Similarly, CIs do not inform about replicability either. They are a statement of expected frequency under repeated sampling from the same population. The 83% next-experiment replicability reported by Cumming (2014), although interesting, is not part of the frequentist understanding of CIs and seems explainable by the size of the confidence interval.

## Fending Against Mindless use of CIs

Finally, there is the ever present risk that CIs will be used as mindlessly as tests. For one, CIs share the same philosophical background than Neyman-Pearson's tests, yet many researchers take them to mean an *ad hoc* probability of personal confidence (Hoekstra et al., 2014). On the other hand, the inferential value of a CI rests on the assumption that the unknown population parameter has equal chance of being anywhere within that interval. Using a point estimate leads to the wrong conclusion: that such point estimate is more probable than any other in the interval.

## Final Note

Both CIs and tests are useful tools (Gigerenzer, 2004). Neyman's CIs are more descriptive, Fisher's tests find significant results and foster meta-analyses, Neyman-Pearson's tests help with decisions, NHST means trouble. Yet, to keep Psychology at the frontier of science we ought to consider alternative tools in the statistical toolbox, such as exploratory data analysis (Tukey, 1977), effect sizes (Cohen, 1988), meta-analysis (Rosenthal, 1984), cumulative meta-analysis (Braver et al., 2014), and Bayesian applications (Dyjas et al., 2012; Barendse et al., 2014), lest we forget.

## Conflict of Interest Statement

The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

## References

Barendse, M. T., Albers, C. J., Oort, F. J., and Timmerman, M. E. (2014). Measurement bias detection through Bayesian factor analysis. *Front. Psychol*. 5:1087. doi: 10.3389/fpsyg.2014.01087

Pubmed Abstract | Pubmed Full Text | CrossRef Full Text | Google Scholar

Braver, S. L., Thoemmes, F. J., and Rosenthal, R. (2014). Continuously cumulating meta-analysis and replicability. *Perspect. Psychol. Sci*. 9, 333–342. doi: 10.1177/1745691614529796

Cohen, J. (1988). *Statistical Power Analysis for the Behavioural Sciences, 2nd Edn*. New York, NY: Psychology Press.

Cortina, J. M., and Dunlap, W. P. (1997). On the logic and purpose of significance testing. *Psychol. Methods* 2, 161–172.

Cumming, G. (2012). *Understanding the New Statistics. Effect Sizes, Confidence Intervals, and Meta-Analysis*. New York, NY: Routledge.

Cumming, G. (2014). The new statistics: why and how. *Psychol. Sci*. 25, 7–29. doi: 10.1177/0956797613504966

Pubmed Abstract | Pubmed Full Text | CrossRef Full Text | Google Scholar

Dyjas, O., Grasman, R. P. P. P., Wetzels, R., van der Maas, H. L. J., and Wagenmakers, E.-J. (2012). What's in a name: a Bayesian hierarchical analysis of the name-letter effect. *Front. Psychol*. 3:334. doi: 10.3389/fpsyg.2012.00334

Pubmed Abstract | Pubmed Full Text | CrossRef Full Text | Google Scholar

Fisher, R. A. (1954). *Statistical Methods for Research Workers, 12th Edn*. Edinburgh: Oliver and Boyd.

Gigerenzer, G. (2004). Mindless statistics. *J. Soc. Econ*. 33, 587–606. doi: 10.1016/j.socec.2004.09.033

Pubmed Abstract | Pubmed Full Text | CrossRef Full Text | Google Scholar

Hoekstra, R., Morey, R. D., Rouder, J. N., and Wagenmakers, E.-J. (2014). Robust misinterpretation of confidence intervals. *Psychon. Bull. Rev*. 21, 1157–1164. doi: 10.3758/s13423-013-0572-3

Pubmed Abstract | Pubmed Full Text | CrossRef Full Text | Google Scholar

Kline, R. B. (2004). *Beyond Significance Testing: Reforming Data Analysis Methods in Behavioral Research*. Washington, DC: APA Books.

Neyman, J., and Pearson, E. S. (1933). On the problem of the most efficient tests of statistical hypotheses. *Philos. T. R. Soc. A* 231, 289–337.

Perezgonzalez, J. D. (2014). A reconceptualization of significance testing. *Theor. Psychol*. 24, 852–859. doi: 10.1177/0959354314546157

Keywords: confidence interval, CI, significance testing, NHST, Fisher, Neyman-Pearson

Citation: Perezgonzalez JD (2015) Confidence intervals and tests are two sides of the same research question. *Front. Psychol*. **6**:34. doi: 10.3389/fpsyg.2015.00034

Received: 01 December 2014; Accepted: 08 January 2015;

Published online: 28 January 2015.

Edited by:

Prathiba Natesan, University North Texas, USAReviewed by:

Ratna Nandakumar, University of Delaware, USAMei Chang, University of North Texas, USA

Copyright © 2015 Perezgonzalez. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: j.d.perezgonzalez@massey.ac.nz