Impact Factor 2.990 | CiteScore 3.5
More on impact ›


Front. Psychol., 17 July 2018 |

Utility of Alternative Effect Size Statistics and the Development of a Web-Based Calculator: Shiny-AESC

  • Psychology, Louisiana State University, Baton Rouge, LA, United States

Alternative displays of effect size statistics can enhance the understandability and impact of validity evidence in a variety of applied settings. Arguably, the proliferation of alternative effect size statistics has been limited due to the lack of user-friendly tools to create them. Common statistical packages do not readily produce these alternative effect sizes and existing tools are outdated and inaccessible. In this paper, I introduce a free-to-use web-based calculator ( for generating alternative effect size displays from empirical data. This calculator requires no mathematical or programming expertise, and therefore, is ideal for academics and practitioners. I also present results from an empirical study that demonstrates the benefits of alternative effect size displays for enhancing lay people's perceived understandability of validity information and attitudes toward the use of standardized testing for college admissions.

The idea that expectancy tables, or other methods that go beyond simply reporting the correlation coefficient, may more clearly show the value of selection tests is not a new one, but it is an idea that must be regularly rediscovered as test critics continue to focus on variance accounted for” – (Bridgeman et al., 2009)

Effect size statistics are universal in the social and behavioral sciences. In the academic literature, standardized effect size indices—such as the correlation coefficient or coefficient of determination—provide a metric for describing the strength of association between two variables (e.g., standardized test scores and academic performance) or the effect of an intervention (e.g., job training and job performance). Standardized indices serve as a common metric for comparing results across scientific studies. Understanding the magnitude of an effect allows decision makers to make informed decisions in domains such as employee selection, health intervention, and education policy.

Real world decisions, however, are often made by non-academics in context-rich environments. Standardized effect size indices preclude any contextual information and require statistical expertise to interpret them. These indices also tend to obscure the practical impact of a statistical effect. Lawshe and Bolda (1958), for instance, commented that the correlation coefficient does not clearly communicate the “predictive efficiency” of a variable. As a result, lay decision makers often find standardized statistical effect sizes hard to comprehend and ineffective for communicating the practical implications of research findings (Brogden, 1946; Rosenthal and Rubin, 1982b; Soyer and Hogarth, 2012; Highhouse et al., 2017). Given the limitations of standardized effect size statistics, there is considerable need to explore alternatives. Indeed, Mattern et al. (2009) commented that “…the ability to communicate validity research findings in a way that is meaningful to the general public should a be a top concern and priority for researchers” (p. 230).

To date, researchers have introduced several alternatives for translating traditional effect size statistics into more meaningful metrics (Lawshe and Bolda, 1958; Rosenthal, 1991; McGraw and Wong, 1995). Although these alternative metrics were developed to facilitate the interpretation of statistical effect size for non-experts, there are—unfortunately—no easily available tools for non-experts to produce these indices. Commonplace statistical programs, such as SPSS or R, do not readily generate alternative effect size statistics. Existing tools are outdated and inaccessible. Myors (1994), for instance, created a computer program for calculating theoretical expectancy charts but the original paper has never been cited. Similarly, Dunlap (1999) developed a program for calculating Common Language Effect Sizes (CLES) and has only been cited 28 times as of writing. Moreover, both programs were developed in FORTRAN, which is a severely outdated programing language. Some other existing tools require programming experience and do not provide a comprehensive list of common alternative effect size statistics (Table 1).


Table 1. Tools for calculating alternative effect sizes.

The lack of tools for computing these effect sizes may limit its proliferation in applied settings. As the opening quote illustrates, the reliance on traditional effect size indices such as correlations continues to obscure the practical impact of evidence-based selection tests: decades after the introduction of alternative effect size displays. The purpose of this paper is to introduce a web-based application that allows scholars and practitioners to easily generate and visualize a variety of alternative effect size metrics. The calculator does not require existing expertise in statistics or programming. In developing this tool, I hope to enhance the accessibility and visibility of alternative effect size displays in organizational and educational settings. Finally, I present an empirical experiment to illustrate the benefits alternative effect sizes have on the understandability of validity information and judgments toward evidence-based selection methods.


Traditional effect size indices, such as the correlation coefficient, are commonplace in the academic literature. Unfortunately, they are often difficult to understand and not easily translated into real-world outcomes. Moreover, the practical utility of correlations are often obscured: critics of using the SAT as a college admissions test asserted that “the SAT only adds 5.4 percent of variance explained by HSGPA [high school grade point average] alone” (Kidder and Rosner, 2002), even though the same evidence was used to support its utility in college admission decisions (e.g., Kuncel and Hezlett, 2007). Similarly, human resource professionals often judge academic literature as inaccessible and impractical (Terpstra and Rozell, 1998; Gelade, 2006), despite the abundance of validity data provided by academic researchers (Schmidt and Hunter, 1998; Kuncel et al., 2004; Huffcutt, 2011). Effect size information—when communicated effectively—should be easy to understand and elucidate the practical impacts of interventions or relations it aims to represent.

In response to the limitations of traditional effect size statistics, several alternative displays of effect size have been developed: the expectancy chart (Schrader, 1965), the Binomial Effect Size Display (Rosenthal and Rubin, 1982a) and the Common Language Effect Size (McGraw and Wong, 1992). Expectancy charts communicate the relationship between two variables (e.g., ACT score and GPA) by presenting the proportion of the sample with score above a cut-off criterion (e.g., GPA above 3.5) at a given score interval on the predictor (e.g., ACT score between 24 and 26). Similarly, the BESD uses a 2 × 2 matrix to present the linear relation between two variables as the probability of an outcome (e.g., GPA above 3.5) based on one's standing on a dichotomized predictor (e.g., ACT above vs. below 26). The theoretical values in the cells can be calculated with the equations: (0.5+r2) and (0.5-r2) where r is the bivariate sample correlation between the two variables of interest (Rosenthal and Rubin, 1982a). The CLES describes the difference between two groups (e.g., control vs. intervention group) with the probability that a random score from one group will differ from the control group (McGraw and Wong, 1992, also see Improvement Index, Clearinghouse, 2014). CLES can be calculated based on the mean and standard deviations of the groups. The effectiveness of the SAT training program, for example, can be described in CLES as “there is a 60% chance that a score from someone who took the ACT training will be better than someone without training.” Dunlap (1994) extended the original CLES to describe the relationship between two continuous variables where the comparison groups are operationalized as the subset of the sample with predictor scores above vs. below a cut-off. For example, the relationship between ACT scores and college GPA can be described as “there is a 60% chance that a student with an ACT score above the median will have a higher college GPA than a student with an ACT score below the median.”

Empirical vs. Theoretical Effect Sizes

Existing methods of computing alternative effect size metrics rely on computationally transforming an observed traditional effect size (e.g., sample correlation) into a theoretical alternative effect size (e.g., hypothetical expectancy chart). The resulting proportions and probabilities in the alternative effect size are not the actual values in the data, but rather, theoretical population values based on the observed sample correlation. The distinction between empirical and theoretical methods of computing alternative effect sizes is particularly salient in the development of expectancy charts (Lawshe and Bolda, 1958; Lawshe et al., 1958). Tiffin and Vincent (1960) found that hypothetical expectancy charts are appropriate when the sample size is adequately large. Existing calculators also use the hypothetical method for computing alternative effect sizes (e.g., Myors, 1994; Cucina et al., 2017).

Although hypothetical effect sizes may be more representative of population parameters, there are several theoretical and practical drawbacks for validity communication. First, when theoretical proportions are computed from a correlation coefficient, the results do not always reflect the data. Rosenthal (1991), for example, showed that the same correlation coefficient can produce markedly different proportions in the BESD and CLES depending on the cut-off values in the criterion and the sample sizes within criterion ranges. Therefore, it might inaccurate and possibly disingenuous to present stake holders with proportions of expected outcomes when those proportions are not, in fact, actual proportions in the data. Secondly, theoretical effect sizes require the audience to understand the difference between sample vs. population parameters: a distinction that may not resonate with a statistically naïve population. Baldridge et al. (2004) argued that “data should be described in a way that fits with how practitioner would describe the situation being addressed in the study.” Proportions presented “as is” (i.e., empirically derived from the sample) may alleviate an additional barrier to comprehension for non-experts. Third, hypothetical transformations of an effect size assume a linear relationship between two variables. This assumption is particularly relevant in expectancy charts, where a theoretical expectancy chart will necessarily depict a linear change in desired outcome as a function of the predictor. However, researchers have started to discover non-linear trends between predictors and criterion in applied settings [e.g., personality and job performance, (Carter et al., 2014)]. When non-linear trends are expected to exist in the data, theoretical transformations of traditional effect sizes are no longer appropriate. In contrast, empirically based displays may be more informative for observing possible curvilinear relationships. Based on these advantages, an empirical alternative effect size calculator was created. In the next section, I describe the development of a web application aimed at producing alternative effect size displays from empirical data.

Shiny-AESC Web App

The Alternative Effect Size Calculator (Shiny-AESC1 is created using Shiny (, which is a web application framework in the statistical programming language R. The popularity of R has risen considerably over the years (Robinson, 2017). Shiny allows developers to implement statistical procedures from R into user-friendly online applications. Users of Shiny apps do not need existing knowledge of R. Shiny is particularly useful for developing dynamic and interactive visualizations of data (Ellis and Merdian, 2015).

The Shiny-AESC is separated into the two sections: (1) the input panel, which allows users to upload their own data and manipulate various parameters for the effect size calculation; and (2) the output panel, which displays a variety of statistical information related to the variables. The output panel is separated into five tabs: expectancy chart, raw data, traditional statistics, alternative effect sizes, and help. Figures 1, 2 contains screenshots of the web app in the expectancy tab and the effect size tab. The presented web app is fully interactive, and the effect size outputs are dynamically generated based on the user input. In other words, as the user changes the input parameters, the resulting effect size displays are updated in real time.


Figure 1. Screenshot of the Shiny-AESC main screen and expectancy chart output.


Figure 2. Screenshot of alternative effect size displays.

Interactive data visualizations have several benefits for statistical comprehension. Interactivity refers to giving user control over the graphical elements in the visual display (e.g., words, pictures; Mayer and Chandler, 2001), which serves to enhance the engagement of the viewer (Perez and White, 1985; Rieber, 1990; Ancker et al., 2011). Moreover, interactive graphs often contain animations, which are ideal for communicating changes or trends over time or across categories (Morrison et al., 2000). Effectively implemented user-interactivity can reduce the viewer's cognitive load and lead to better learning outcomes (Mayer and Chandler, 2001). For instance, Boucheix and Guignard (2005) found that user-interactivity improved performance on the comprehension of a technical manual. Mayer and Chandler (2001) found that modest amount of interactivity improved students' learning of scientific concepts. The recent technological developments in web-based applications has allowed for more user-interactive platforms to manipulate and visualize data (Tay et al., 2016). Given the benefits of user-interactivity, some researchers have advocated the development of interactive data visualization tools to enhance statistical communication (Ellis and Merdian, 2015).

Demonstration of Shiny-AESC With ACT Data

The Shiny-AESC is used to generate displays of the validity of ACTs as a predictor of college GPA using archival data. The sample bivariate correlation between ACT and GPA in the dataset is 0.295. The expectancy chart was created with criterion cut-off set to a GPA of 3.5. I chose 3.5 because it is—for many universities in the United States—the cut-off value for being on the dean's list. It is notable that the average GPA in the sample was 3.35. Cut-off values for the predictor is also required for calculating the BESD and CLES. By default, the median value is used to ensure symmetry of sample sizes. To be sure, both predictor and criterion cut-offs are free to vary at the discretion of the user. Figure 1 depicts the association between ACT and GPA as an expectancy chart, and Figure 2 contains the BESD and CLES for the same relation.

Empirical Study: Perceptions of Validity Comprehension and ACT Usefulness

To illustrate the psychological benefits of alternative displays of effect sizes, I conducted an experiment that examined the effects of alternative vs. traditional effect size displays on the perceived understandability of validity evidence and people's attitudes toward the use of ACTs as a college admissions test. Standardized testing, including but not limited to ACTs, have received considerable criticism from academics and the public (Sternberg, 2006). Despite a substantial body of evidence in favor of the standardized tests as a valid predictor of college performance, scholars and college administrators often judge these tests as unfair and useless (Sherley, 2007). Many universities and colleges across the countries have started adopting a test-optional policy (Belasco et al., 2015). Proponents of the standardized tests have suggested that their perceived validity may be undermined by the choice of statistic (Mattern et al., 2009). In their report, Mattern et al. (2011) used primarily expectancy charts when presenting the validity of their admissions test; likewise, Kuncel and Hezlett (2007) used a BESD in response to criticisms of the magnitude of cognitive ability tests' predictive validity.

The benefits of alternative effect size statistics have also been shown in controlled experiments. Brooks et al. (2014), for instance, found that people perceived alternative effect sizes as easier to understand than traditional effect size indices (r and r2). Moreover, the authors found that the perceived utility of a training program was greater when its validity was presented with alternative effect sizes. In this experiment, I will present a lay audience with validity of ACTs presented as five different effect size metrics generated from the calculator (r, r2, BESD, CLES, and Expectancy Chart) and ask participants to report their understandability of the validity information and the usefulness of ACTs. The experiment is a conceptual replication and extension of findings from Brooks and colleagues. This study extends the previous study in two ways: (1) in additional to BESD and CLES, I will also examine the effect of the expectancy chart on validity communication; and (2) whereas Brooks and colleagues used theoretically derived effect sizes, this experiment will use empirically calculated effect size. Consistent with previous findings, I expect that:

Hypothesis 1. People will perceive alternative effect sizes (BESD, CLES, and Expectancy Chart) as easier to understand than traditional effect size statistics (r and r2).

Hypothesis 2. The ACT will be judged more favorably as a college admissions test when its validity is presented with alternative effect sizes displays than with traditional effect size statistics.



Adult participants (n = 225) from the United States were recruited using Amazon Mechanical Turk (MTurk). Past research has demonstrated that the MTurk population generalizes well to an adult population and is a valuable platform for conducting experimental research (Buhrmester et al., 2011; Highhouse and Zhang, 2015). Fifty-three participants were dropped for incorrectly responding to our quality check question (“ACT is a test of students' physical abilities”). The final sample had 172 participants (53% male, mean age = 36, 80% Caucasian). Each participant received 75 cents for completing the survey, which took approximately 5 min.


Participants first read a short description of the ACT as a standardized college entrance exam in the United States. Next, they were presented with information regarding the validity of the ACT. Each participant was randomly assigned to one of five conditions, each corresponding to a different type of validity display (correlation; coefficient of determination; CLES; BESD; and Expectancy Chart)2 Finally, participants reported—on 5-point Likert scales—the degree to which they understood the validity information using a four-item measure (e.g., “I understood the information about the relationship between ACT score and college GPA”) and the degree to which they judged the ACT as a useful tool for making college admission decisions using a three-item measure (e.g., “I would endorse the use of ACTs for admission purposes”). Both measures were adapted based on Brooks et al. (2014) and had Cronbach α values of 0.92 and 0.93 respectively. Appendix A contains all the items used in this study. For exploratory purposes, I also measured participants' numeracy using the four-item Berlin Numeracy Test (Cokely et al., 2012) and their self-reported ACT score.


Table 2 contains the means, standard deviations, and intercorrelation of the study's measures. Perceived comprehension and perceived utility of the ACT were moderately correlated. Self-reported ACT score was moderately correlated with scores on the Berlin Numeracy Test. Both scores were positively correlated with perceived comprehension of the statistical evidence, but not with the perceived utility of the ACT. It is notable that people who scored higher on the ACT did not necessarily exhibit better reaction toward the test (e.g., Wright et al., 2012).


Table 2. Means, standard deviations, and correlations of study variables.

The purpose of the study was to examine the effects of traditional (r and r2) and alternative (CLES, BESD, and Expectancy Chart) validity displays on their perceived comprehension and subsequent judgments toward the ACT. Therefore, I collapsed the display conditions into traditional vs. alternative effect size displays for the ease of exposition. Table 3 contains the results of the Welch's t-tests3 for both dependent variables. There was a statistically significant difference in the perceived comprehension of the validity information. As expected, alternative effect sizes were perceived as easier to understand than traditional effect sizes. There was also a statistically significant effect of display type on the perceived usefulness of the ACT. People judged the ACT to be more useful for making college admission decisions when the validity information was presented as an alternative effect size display than a traditional effect size statistic. Parallel Bayesian independent t-tests were also conducted using JASP (JASP Team, 2018; Wagenmakers et al., 2018). Observed Bayes Factors for the alternative hypotheses (BF10) were greater than 100 for both dependent variables. In other words, the alternative hypotheses were 100 times more likely than the null hypotheses. Based on the results, Hypotheses 1 and 2 were fully supported.


Table 3. Independent samples t-test of dependent variables.

For exploratory purposes, I also examined the difference in the dependent variables between each of the display types (Figure 3). There were no noticeable differences in the perceived comprehension nor perceived usefulness of the ACT across the different alternative effect size displays. Although there was a noticeable difference in the perceived comprehension between the correlation and coefficient of determination, this difference was not statistically significant after correcting for multiple comparisons.


Figure 3. Means of perceived comprehension and perceived usefulness across display types. Besd, binomial effect size display; cles, common language effect display; exp, expectancy chart; corr, correlation coefficient; rsquared, coefficient of determination.


In the empirical experiment, people judged the alternative effect size indices as easier to understand than traditional effect sizes. The results are consistent with previous investigations of alternative effect sizes (Brooks et al., 2014). I also extended previous research by examining the expectancy chart, which is frequently used in organizational and education settings. I did not find any differences in the understandability between the expectancy chart and other alternative displays (CLES and BESD). Interestingly, there was a small—but not statistically significant—difference in the perceived comprehension between the correlation and coefficient of determination. One explanation is that the term “correlation” is used more frequently in non-academic settings, whereas “coefficient of determination” and “variance” are both technical concepts that are rarely mentioned outside the context of scientific research. Still, the differences in perceived comprehension did not translate to differences in attitude toward the ACT. Most importantly, participants of this study judged the ACT to be more useful when its validity was presented as one of the alternative effect size displays. The results suggest that the reluctance to use standardized tests in admission settings may be due—in part—to the way in which academics communicate validity information.

To facilitate the generation of alternative effect size statistics using empirical data, I have developed a working web-based interactive alternative effect size calculator ( The calculator will read in raw data and produce a series of traditional (e.g., Pearson's r, r2, Hedges' g) and alternative (e.g., BESD, CLES, Expectancy Chart) effect size displays. At the current stage, the calculator will only produce alternative effect sizes based on two continuous variables. Future developments of the calculator will allow for the calculation of the combined and incremental validity of multiple predictors (e.g., Bridgeman et al., 2004; Krasikova et al., 2018).

Although I have listed several benefits of empirically derived alternative effect sizes, it is important to note its limitations. Specifically, alternative effect sizes derived directly from the data are subject to sampling error and the observed patterns are subject to the influences of cut-off values and number of bins. The number of bins in the expectancy chart, for example, could either exacerbate or mask the observed relations, especially when empirical data is used. Indeed, there are numerous ways in which the disingenuous presentation of data can mislead the receivers of the information (Best, 2001; Parikh, 2014). Therefore, as with any decisions of displaying data, the user must carefully choose meaningful parameters. Future research should also examine differences in the interpretation of hypothetical vs. empirically derived alternative effect sizes, particularly with respect to the perceived believability and relevance of the results. Finally, the nature of the convenient sample may limit the empirical experiment. Future empirical research should examine the effect of validity displays of ACTs on a more invested population (e.g., policy makers, educators, parents).

This research has several practical implications. First, the Shiny-AESC provides practitioners and researchers with an easy-to-use tool for creating alternative effect size displays from empirical data. Practitioners are encouraged to use this tool when communicating validity evidence to relevant stakeholders and organizational decision-makers. Second, researchers are encouraged to include alternative effect size displays in their scientific publications. As shown in this paper and previous research, alternative effect sizes are easier to understand for a non-expert population and are more effective in communicating the practical benefit of the research findings. By presenting research evidence with more accessible metrics, we may finally begin to bridge the gap between academic research and practice in applied areas of psychology.

Ethics Statement

This study was carried out in accordance with the recommendations of the Institutional Review Board at Louisiana State University. The protocol was approved by the Institutional Review Board at Louisiana State University. All subjects gave written informed consent in accordance with the Declaration of Helsinki.

Author Contributions

The author confirms being the sole contributor of this work and approved it for publication.

Conflict of Interest Statement

The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.


1. ^Link to the app (

2. ^Stimulus material can be found on (

3. ^Welch's t-test was used to accommodate unequal group sizes between the traditional (n = 73) and alternative effect size (n = 102) groups.


Ancker, J. S., Weber, E. U., and Kukafka, R. (2011). Effects of game-like interactive graphics on risk perceptions and decisions. Med. Decis. Making 31, 130–142. doi: 10.1177/0272989X10364847

PubMed Abstract | CrossRef Full Text | Google Scholar

Baldridge, D. C., Floyd, S. W., and Markóczy, L. (2004). Are managers from Mars and academicians from venus? Toward an understanding of the relationship between academic quality and practical relevance. Strateg. Manage. J. 25, 1063–1074. doi: 10.1002/smj.406

CrossRef Full Text | Google Scholar

Belasco, A. S., Rosinger, K. O., and Hearn, J. C. (2015). The test-optional movement at America's selective liberal arts colleges. Educ. Eval. Policy Anal. 37, 206–223. doi: 10.3102/0162373714537350

CrossRef Full Text | Google Scholar

Best, J. (2001). Damned Lies and Statistics: Untangling Numbers From the Media, Politicians, and Activists. Berkeley, CA: University of California Press.

Google Scholar

Boucheix, J.-M., and Guignard, H. (2005). What animated illustrations conditions can improve technical document comprehension in young students? Format, signaling and control of the presentation. Eur. J. Psychol. Educ. 20, 369–388. doi: 10.1007/BF03173563

CrossRef Full Text

Bridgeman, B., Burton, N., and Cline, F. (2009). A note on presenting what predictive validity numbers mean. Appl. Measure. Educ. 22, 109–119. doi: 10.1080/08957340902754577

CrossRef Full Text | Google Scholar

Bridgeman, B., Pollack, J., and Burton, N. (2004). Understanding what SAT Reasoning Test™ scores add to high school grades: a straightforward approach. ETS Res. Rep. Ser. 2004, i–20. doi: 10.1002/j.2333-8504.2004.tb01967.x

CrossRef Full Text

Brogden, H. E. (1946). On the interpretation of the correlation coefficient as a measure of predictive efficiency. J. Educ. Psychol. 37, 65–76. doi: 10.1037/h0061548

PubMed Abstract | CrossRef Full Text | Google Scholar

Brooks, M. E., Dalal, D. K., and Nolan, K. P. (2014). Are common language effect sizes easier to understand than traditional effect sizes? J. Appl. Psychol. 99, 332–340. doi: 10.1037/a0034745

PubMed Abstract | CrossRef Full Text | Google Scholar

Buhrmester, M., Kwang, T., and Gosling, S. D. (2011). Amazon's mechanical turk: a new source of inexpensive, yet high-quality, data? Perspect. Psychol. Sci. 6, 3–5. doi: 10.1177/1745691610393980

PubMed Abstract | CrossRef Full Text | Google Scholar

Carter, N. T., Dalal, D. K., Boyce, A. S., O'connell, M. S., Kung, M.-C., and Delgado, K. M. (2014). Uncovering curvilinear relationships between conscientiousness and job performance: how theoretically appropriate measurement makes an empirical difference. J. Appl. Psychol. 99, 564–586. doi: 10.1037/a0034688

PubMed Abstract | CrossRef Full Text | Google Scholar

Clearinghouse, W. W. (2014). Procedures and Standards Handbook (Version 3.0). Washington, DC: US Department of Education.

Cokely, E. T., Galesic, M., Schulz, E., Ghazal, S., and Garcia-Retamero, R. (2012). Measuring risk literacy: the Berlin numeracy test. Judgm. Decis. Mak. 7, 25–47.

Google Scholar

Cucina, J. M., Berger, J. L., and Busciglio, H. H. (2017). Communicating criterion-related validity using expectancy charts: a new approach. Pers. Assess. Decis. 3:1. doi: 10.25035/pad.2017.001

CrossRef Full Text | Google Scholar

Dunlap, W. P. (1994). Generalizing the common language effect size indicator to bivariate normal correlations. Psychol. Bull. 116, 509–511. doi: 10.1037//0033-2909.116.3.509

CrossRef Full Text | Google Scholar

Dunlap, W. P. (1999). A program to compute McGraw and Wong's common language effect size indicator. Behav. Res. Methods Instrum. Comput. 31, 706–709.

PubMed Abstract | Google Scholar

Ellis, D. A., and Merdian, H. L. (2015). Thinking outside the box: developing dynamic data visualizations for psychology with shiny. Front. Psychol. 6:1782. doi: 10.3389/fpsyg.2015.01782

PubMed Abstract | CrossRef Full Text | Google Scholar

Gelade, G. A. (2006). But what does it mean in practice? The Journal of Occupational and Organizational Psychology from a practitioner perspective. J. Occup. Org. Psychol. 79, 153–160. doi: 10.1348/096317905X85638

CrossRef Full Text

Highhouse, S., Brooks, M. E., Nesnidol, S., and Sim, S. (2017). Is a.51 validity coefficient good? Value sensitivity for interview validity. Int. J. Select. Assess. 383–389. doi: 10.1111/ijsa.12192

CrossRef Full Text | Google Scholar

Highhouse, S., and Zhang, D. (2015). The new fruit fly for applied psychological research. Industr. Org. Psychol. 8, 179–183. doi: 10.1017/iop.2015.22

CrossRef Full Text | Google Scholar

Huffcutt, A. I. (2011). An empirical review of the employment interview construct literature. Int. J. Select. Assess. 19, 62–81. doi: 10.1111/j.1468-2389.2010.00535.x

CrossRef Full Text | Google Scholar

JASP Team (2018). JASP (Version 0.8.6)[Computer software]. Available online at:

Kidder, W. C., and Rosner, J. (2002). How the SAT creates built-in-headwinds: an educational and legal analysis of disparate impact. Santa Clara L. Rev. 43:131. Available online at:

Google Scholar

Krasikova, D. V., Le, H., Bachura, E., Krasikova, D. V., Le, H., and Bachura, E. (2018). Common language effect size indicator for multiple toward customer-centric organizational science : a common language effect size indicator for multiple linear regressions and regressions with higher-order terms. J. Appl. Psychol. 103, 659–675. doi: 10.1037/apl0000296

PubMed Abstract | CrossRef Full Text | Google Scholar

Kuncel, N. R., and Hezlett, S. A. (2007). Standardized tests predict graduate students' success. Science 315, 1080–1081. doi: 10.1126/science.1136618

PubMed Abstract | CrossRef Full Text | Google Scholar

Kuncel, N. R., Hezlett, S. A., and Ones, D. S. (2004). Academic performance, career potential, creativity, and job performance: can one construct predict them all? J. Pers. Soc. Psychol. 86, 148–161. doi: 10.1037/0022-3514.86.1.148

PubMed Abstract | CrossRef Full Text | Google Scholar

Lawshe, C. H., and Bolda, R. A. (1958). Expectancy charts: I. Their use and empirical development. Pers. Psychol. 11, 353–365. doi: 10.1111/j.1744-6570.1958.tb00023.x

CrossRef Full Text

Lawshe, C. H., Bolda, R. A., Brune, R. L., and Auclair, G. (1958). Expectancy charts II. Their theoretical development. Pers. Psychol. 11, 545–559. doi: 10.1111/j.1744-6570.1958.tb00040.x

CrossRef Full Text | Google Scholar

Mattern, K. D., Kobrin, J., Patterson, B., Shaw, E., and Camara, W. (2009). “Validity is in the eye of the beholder: Conveying SAT research findings to the general public,” in The Concept of Validity: Revisions, New Directions, and Applications, ed R. W. Lissitz (Charlotte, NC: Information Age Publishing), 213–240.

Mattern, K. D., Shaw, E. J., and Kobrin, J. L. (2011). An alternative presentation of incremental validity discrepant SAT and HSGPA performance. Educ. Psychol. Meas. 71, 638–662. doi: 10.1177/0013164410383563

CrossRef Full Text | Google Scholar

Mayer, R. E., and Chandler, P. (2001). When learning is just a click away: does simple user interaction foster deeper understanding of multimedia messages? J. Educ. Psychol. 93, 390–397. doi: 10.1037/0022-0663.93.2.390

CrossRef Full Text | Google Scholar

McGraw, K. O., and Wong, S. P. (1992). A common language effect size statistic. Psychol. Bull. 111, 361–365. doi: 10.1037/0033-2909.111.2.361

CrossRef Full Text

McGraw, K., and Wong, S. P. (1995). Use of a common language effect size statistic. Am. J. Psychiatry 152:151.

Google Scholar

Morrison, J. B., Tversky, B., and Betrancourt, M. (2000). Animation: does it facilitate learning. AAAI Technical Report, 53–59.

Google Scholar

Myors, B. (1994). A theoretical expectancy tablecalculator. Behav. Res. Methods Instrum. Comput. 26, 467–469. doi: 10.3758/BF03204667

CrossRef Full Text | Google Scholar

Parikh, R. (2014). How to Lie with Data Visualization. Available online at: (Accessed February 28, 2018).

Perez, E. C., and White, M. A. (1985). Student evaluation of motivational and learning attributes of microcomputer software. J. Comput. Based Instr. 12, 39–43.

Google Scholar

Rieber, L. P. (1990). Using computer animated graphics in science instruction with children. J. Educ. Psychol. 82, 135–140. doi: 10.1037/0022-0663.82.1.135

CrossRef Full Text | Google Scholar

Robinson, D. (2017). The Impressive Growth of R - Stack Overflow Blog. Available online at: (Accessed February 27, 2018).

Rosenthal, R. (1991). Effect sizes: Pearson's correlation, its display via the BESD, and alternative indices. Am. Psychol. 46, 1086–1087. doi: 10.1037/0003-066X.46.10.1086

CrossRef Full Text | Google Scholar

Rosenthal, R., and Rubin, D. B. (1982a). A simple, general purpose display of magnitude of experimental effect. J. Educ. Psychol. 74, 166–169. doi: 10.1037/0022-0663.74.2.166

CrossRef Full Text | Google Scholar

Rosenthal, R., and Rubin, D. B. (1982b). Comparing effect sizes of independent studies. Psychol. Bull. 92, 500–504. doi: 10.1037/0033-2909.92.2.500

CrossRef Full Text | Google Scholar

Schmidt, F., and Hunter, J. (1998). The validity and utility of selection methods in personnel psychology: practical and theoretical implications of 85 years of research findings. Psychol. Bull. 124, 262–274. doi: 10.1037/0033-2909.124.2.262

CrossRef Full Text | Google Scholar

Schrader, W. B. (1965). A taxonomy of expactancy tables. J. Educ. Measure. 2, 29–35. doi: 10.1111/j.1745-3984.1965.tb00388.x

CrossRef Full Text | Google Scholar

Sherley, J. L. (2007). The utility of standardized tests. Science 316, 1694–1697.

PubMed Abstract | Google Scholar

Soyer, E., and Hogarth, R. M. (2012). The illusion of predictability: how regression statistics mislead experts. Int. J. Forecast. 28, 695–711. doi: 10.1016/j.ijforecast.2012.02.002

CrossRef Full Text | Google Scholar

Sternberg, R. J. (2006). The rainbow project enhancing the SAT through assessments of analytical practical and creative skills. Intelligence 34, 321–350. doi: 10.1016/j.intell.2006.01.002

CrossRef Full Text | Google Scholar

Tay, L., Parrigon, S., Huang, Q., and LeBreton, J. M. (2016). Graphical descriptives a way to improve data transparency and methodological rigor in psychology. Perspect. Psychol. Sci. 11, 692–701. doi: 10.1177/1745691616663875

PubMed Abstract | CrossRef Full Text | Google Scholar

Terpstra, D. E., and Rozell, E. J. (1998). Human resource executives' perceptions of academic research. J. Bus. Psychol. 13, 19–29.

Google Scholar

Tiffin, J., and Vincent, N. L. (1960). Comparison of empirical and theoretical expectancies. Pers. Psychol. 13, 59–64. doi: 10.1111/j.1744-6570.1960.tb01517.x

CrossRef Full Text | Google Scholar

Wagenmakers, E.-J., Love, J., Marsman, M., Jamil, T., Ly, A., Verhagen, J., et al. (2018). Bayesian inference for psychology. Part II: example applications with JASP. Psychonom. Bull. Rev. 25, 58–76. doi: 10.3758/s13423-017-1323-7

PubMed Abstract | CrossRef Full Text

Wright, C. W., Sablynski, C. J., Manson, T. M., and Oshiro, S. (2012). Why are manhole covers round? A laboratory study of reactions to puzzle interviews. J. Appl. Soc. Psychol. 42, 2834–2857. doi: 10.1111/j.1559-1816.2012.00963.x

CrossRef Full Text | Google Scholar

Appendix A. Items of Dependent Variables

Perceived Comprehension

1. It was easy to understand the information about the effectiveness of the ACT

2. The relationship between ACT and college GPA was clearly communicated

3. I understood the information about the relationship between ACT score and college performance

4. I understood the value of the ACT as a tool for predicting college GPA

Perceived Usefulness

1. There is a strong association between ACT scores and college GPA

2. The ACT is an accurate tool for predicting college GPA

3. The ACT score is a strong predictor of college GP

Keywords: effect size statistics, validity, visual-aids, decision-aids, shiny R, science communication

Citation: Zhang DC (2018) Utility of Alternative Effect Size Statistics and the Development of a Web-Based Calculator: Shiny-AESC. Front. Psychol. 9:1221. doi: 10.3389/fpsyg.2018.01221

Received: 28 May 2018; Accepted: 27 June 2018;
Published: 17 July 2018.

Edited by:

Laura Badenes-Ribera, Universitat de València, Spain

Reviewed by:

Jose D. Perezgonzalez, Massey University Business School, New Zealand
Patrizio E. Tressoldi, Università degli Studi di Padova, Italy

Copyright © 2018 Zhang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Don C. Zhang,