The APP procedure for estimating the Cohen's effect size

Xiangfei Chen (Mathematical Sciences, New Mexico State University, Las Cruces, New Mexico, USA)
David Trafimow (Department of Psychology, New Mexico State University, Las Cruces, New Mexico, USA)
Tonghui Wang (Mathematical Sciences, New Mexico State University, Las Cruces, New Mexico, USA)
Tingting Tong (Mathematical Sciences, New Mexico State University, Las Cruces, New Mexico, USA)
Cong Wang (Mathematics, University of Nebraska Omaha, Omaha, Nebraska, USA)

Asian Journal of Economics and Banking

ISSN: 2615-9821

Article publication date: 21 September 2021

Issue publication date: 26 October 2021

766

Abstract

Purpose

The authors derive the necessary mathematics, provide computer simulations, provide links to free and user-friendly computer programs, and analyze real data sets.

Design/methodology/approach

Cohen's d, which indexes the difference in means in standard deviation units, is the most popular effect size measure in the social sciences and economics. Not surprisingly, researchers have developed statistical procedures for estimating sample sizes needed to have a desirable probability of rejecting the null hypothesis given assumed values for Cohen's d, or for estimating sample sizes needed to have a desirable probability of obtaining a confidence interval of a specified width. However, for researchers interested in using the sample Cohen's d to estimate the population value, these are insufficient. Therefore, it would be useful to have a procedure for obtaining sample sizes needed to be confident that the sample. Cohen's d to be obtained is close to the population parameter the researcher wishes to estimate, an expansion of the a priori procedure (APP). The authors derive the necessary mathematics, provide computer simulations and links to free and user-friendly computer programs, and analyze real data sets for illustration of our main results.

Findings

In this paper, the authors answered the following two questions: The precision question: How close do I want my sample Cohen's d to be to the population value? The confidence question: What probability do I want to have of being within the specified distance?

Originality/value

To the best of the authors’ knowledge, this is the first paper for estimating Cohen's effect size, using the APP method. It is convenient for researchers and practitioners to use the online computing packages.

Keywords

Citation

Chen, X., Trafimow, D., Wang, T., Tong, T. and Wang, C. (2021), "The APP procedure for estimating the Cohen's effect size", Asian Journal of Economics and Banking, Vol. 5 No. 3, pp. 289-306. https://doi.org/10.1108/AJEB-08-2021-0095

Publisher

:

Emerald Publishing Limited

Copyright © 2021, Xiangfei Chen, David Trafimow, Tonghui Wang, Tingting Tong and Cong Wang

License

Published in Asian Journal of Economics and Banking. Published by Emerald Publishing Limited. This article is published under the Creative Commons Attribution (CC BY 4.0) licence. Anyone may reproduce, distribute, translate and create derivative works of this article (for both commercial and non-commercial purposes), subject to full attribution to the original publication and authors. The full terms of this licence may be seen at http://creativecommons.org/licences/by/4.0/legalcode.


1. Introduction

Cohen (e.g. 1988) famously argued that researchers should be concerned not only with whether an effect is present but with the size of the effect too. Cohen discussed a variety of different effect size indices, and other researchers have added to the effect size toolbox. Nevertheless, for typical studies, where economic data for two groups are compared, economic data for two countries are compared, or where an experimental group is compared to a control group, Cohen's d remains by far the most popular effect size index. As will be explained in more detail in the subsequent section, Cohen's d denotes the difference in means divided by the standard deviation. Thus, Cohen's d provides valuable information about how much the means differ in standard deviation units.

One reason that scientists in the social and economic sciences have found Cohen's d useful is that many of the dependent measures these sciences do not have intrinsic meaning. For instance, whereas it might be reasonably clear what a dollar means, the meaning of a unit on an attitude scale might be less clear. If the mean attitude in the experimental condition is 2 and the mean attitude in the control condition is 1, is this a small difference or a large one? By converting the difference in means to Cohen's d, the researcher can gain an idea of the size of the difference in standard deviation units, even when the scale units are not themselves intrinsically meaningful. Another advantage of Cohen's d is that it facilitates comparisons across studies. Even if scale units are different for different studies, thereby rendering them seemingly impossible to compare, researchers still can compare in terms of standard deviation units. Many researchers have taken advantage of this, particularly in meta-analytic research. Despite the popularity of Cohen's d and its obvious usefulness, there remains an important limitation. Specifically, the Cohen's d that a researcher obtains in a particular experiment is a sample statistic. It is not a population value. Typically, researchers are not interested in sample statistics for their own sake, but because they provide useful estimates of population values. Thus, there is an important question that has not been properly addressed: how well does Cohen's d estimate the population effect size? Although researchers have long known how to compute traditional confidence intervals for Cohen's d, traditional confidence intervals do not properly address the question. This is because, for example, although 95% of 95% confidence intervals surround the population parameter, it is not the case that the population parameter has a 95% chance of being within a 95% confidence interval. This last is unknown. In addition, Trafimow and Uhalt (2020) have shown that sample confidence intervals tend to be inaccurate representations of population confidence intervals unless sample sizes are much larger than those typically employed. An alternative way to address the issue is to use the a priori procedure (APP) that has been employed previously in a variety of ways not pertaining to Cohen's d (e.g. Li et al., 2020; Trafimow, 2017, 2019; Trafimow and MacDonald, 2017; Trafimow et al., 2020a; Wang et al., 2020, 2021; Wei et al., 2020). Although the APP uses confidence intervals, it does so in a way that deviates importantly from traditional confidence intervals. To use APP thinking to address Cohen's d, the researcher would ask the two bullet-pointed questions below.

RQ1.

The precision question: How close do I want my sample Cohen's d to be to the population value?

RQ2.

The confidence question: What probability do I want to have of being within the specified distance?

For example, the research might wish to have a 95% probability of obtaining Cohen's d within a tenth of a standard deviation of the population value. The present goal is to determine the sample size the researcher needs to collect to meet the precision and confidence specifications in the contexts of independent and matched samples experimental designs.

This paper is organized as follows. Definitions of Cohen's effect sizes for both populations and samples are given in Section 2, together with properties of noncentral t distributions. In Section 3, the APP methods are applied for estimating population effect size θ in independent case and θD in dependent case. In Section 4, the simulation study, the coverage rate, and real data examples are provided in supporting our main results given in Section 3. Conclusion remarks are given in Sections 5.

2. Preliminaries

Effect size is a statistical concept that measures the strength of the relationship between two variables on a numeric scale. For example, in medical education research studies that compare different educational interventions, effect size is the magnitude of the difference between groups. The absolute effect size is the difference between the average, or mean, outcomes in two different intervention groups. The standard deviation of the effect size is of critical importance, since it indicates how much uncertainty is included in the measurement. For more details and applications, see, Sullivan and Feinn (2012), Schafer and Schwarz (2019), Bhandari (2020).

Cohen's d is one of the most common ways to measure effect size, which is known as the difference of two population means and it is divided by the standard deviation from the data. Mathematically, Cohen's effect size is denoted by:

(2.1)θ=μ1μ2σ,
where μ1 and μ2 are means of two populations, and σ is the standard deviation based on either or both populations.

Cohen's d is defined as the difference between two means divided by a standard deviation for the data obtained from both populations:

(2.2)d=X̅1X̅2S,
where X̅1 and X̅2 are sample means and S, defined by Jacob Cohen, is the pooled standard deviation (for two independent samples)
S=(n11)S12+(n21)S22n1+n22,
and n1, S1, n2, S2 are sample sizes and sample variances of two independent samples, respectively.

Note that confidence intervals of standardized effect sizes, especially Cohen's d, rely on the calculation of confidence intervals of noncentrality parameters. In order to find the minimum sample sizes for estimating the Cohen's effect size θ given in (2.1) by Cohen's d given in (2.2), we need the following definition.

Definition 2.1.

Let Z and U be independent random variables, ZN(λ, 1), the normal distribution with mean λ and standard deviation 1, and Uχm2, the chi-square distribution with m degrees of freedom. The random variable T given by

(2.3) T=ZU/m
is said to have a noncentral t-distribution with m degrees of freedom and the noncentrality parameter λ, denoted by Ttm(λ).

It is easy to obtain the following properties of Ttm(λ) (see Nguyen and Wang, 2008).

  1. The probability density function (pdf) of T is given by

fT(t;λ)=12(m+1)/2Γm2mπ0xm12exp12x+txmλ2dx.
  1. The mean and variance of T are

E(T)=λm2Γ((m1)/2)Γ(m/2),form>1
and
Var(T)=m(1+λ2)m2λ2m2Γ((m1)/2)Γ(m/2)2form>2,
respectively. For convenience, if we use the correction factor J(m) given by
(2.4)J(m)=m2Γ((m1)/2)Γ(m/2),
then the mean and the variance of T are
(2.5)E(T)=J(m)λ,form>1,
and
(2.6)Var(T)=m(1+λ2)m2J(m)λ2,form>2.

We will use the following results in the proofs of our main results to be given in next section.

Proposition 2.1.

Let X11,,X1n1 be a random sample of size n1 from N(μ1, σ2), X21, …, X2n2 be a random sample of size n2 from N(μ2, σ2). Assume that two samples are independent. Let θ and d be the Cohen's effect sizes given in (2.1) and (2.2), respectively. Then

n1n2n1+n2dtn1+n22(λ)withλ=n1n2n1+n2θ.

Proof. From the basic statistics, we know that.

  1. X̅1X̅2N(μ1μ2,(n11+n21)σ2) so that Zn1n2n1+n2dN(λ,1).

  2. (n11)S12σ2χn112 and (n21)S22σ2χn212 so that (n1+n22)S2σ2χn1+n222 as S12 and S22 are independent.

Now by Definition 2.1, T=n1n2n1+n2dtn1+n22(λ), the desired result follows. □

For the dependent case, we have the following result.

Proposition 2.2.

Let (X11,X21),,(X1n,X2n) be a random sample of size n from a bivariate normal population with mean vector μ and covariance matrix Σ, where

μ=μ1μ2andΣ=σ21ρρ1.

Let Di = X1i − X2i, i = 1, …, n and θD=μ1μ2σ and dD=D̅SD be the Cohen's sizes of the population and matched sample, respectively, where

D̅=1ni=1nDiandSD2=1n1i=1n(DiD̅)2
are the mean and variance of Di's, respectively. Then ndDtn1(λD), where the noncentrality parameter λD=n2(1ρ)θD.

Proof. Note that D̅N(μ1μ2,2(1ρ)σ2n) so that ZnD̅2(1ρ)σN(λD,1), where λD=n2(1ρ)θD. Also it is easy to know that

U(n1)SD22(1ρ)σ2χn12,
and D̅ and SD2 are independent. Therefore, by Definition 2.1,
(2.7)T=ZU/(n1)=nD̅2(1ρ)σ(n1)SD22(1ρ)σ2/(n1)=ndDtn1(λD).
so that the desired result follows. □
Remark 2.1.

The graphs of density curves of tm(λ) with different mean values J(m)λ and different degrees of freedom m are given in Figures 1 and 2, respectively. From graphs, we know that (i) density curves are symmetric about mean J(m)λ, which is a function of m, so that the equal tailed confidence intervals should be the best choice, and (ii) density curves tend to N(λ, 1) as m to .

Remark 2.2.

For a given the confidence level c = 1 − α, the c 100% confidence intervals of λ based on T1 with m1 degrees of freedom and T2 with m2 with m1 < m2 degrees of freedom have the following relationship:

(2.8) c=P(|T1λ|tm1,(1c)/2)P(|T2λ|tm1,(1c)/2),
where tm,(1−c)/2 is the critical value of the t-distribution with m degrees of freedom. More details for the confidence interval of Cohen's effect size θ will be given in next section.

3. The APP methods for estimating both θ and θD

In this section, we will apply the APP methods for estimating Cohen's effect size θ in independent samples' case and θD in matched sample case.

3.1 The minimum sample size required for estimating θ

First, we consider two independent samples from two normal populations N(μ1,σ12) and N(μ2,σ22) with equal unknown variances: σ12=σ22=σ2.

Theorem 3.1.

Let X11,,X1n1 be a random sample of size n1 from N(μ1, σ2), X21,,X2n2 be a random sample of size n2 from N(μ2, σ2). Assume that two samples are independent. Let c be the confidence level and f be the precision which satisfies

(3.1) Pfσd1dE(d)fσd1=c,
where E(d) = J(n1 + n2 − 2) θ, d1=X11X21S and
Var(d1)σd12=n1+n22n1+n242+θ2(J(n1+n22)θ)2,

J(n1 + n2 − 2) is the correction factor given in (2.4). Both θ and d are given in (2.1) and (2.2). Let n =  min{n1, n2} and fT(⋅) be the density of t-distribution with degrees of freedom 2(n − 1) and noncentrality parameter λ*=n/2θ. Then the required sample size n can be obtained by solving

(3.2) LUfH*(h)dt=c,
where d*=(X̅1*X̅2*)/S*, X̅1*, S1*2 and X̅2*, S2*2 are sample means and variances of independent samples with same sample size n, respectively. Here S*2=(S1*2+S2*2)/2, H*=[d*E(d*)]/σd*, which has density given by
(3.3) fH*(h)=n2σd*fTn2σd*h+J(2n2)λ*,
where
E(d*)=J(2n2)θ,Var(d*)=σd*2=2nn1n2(1+λ*)2[J(2n2)λ*]2,
(3.4) L=fσd1*σd*,U=L,σd1*2=n1n22+θ2(J(2n2)θ)2.

Proof. By Proposition 2.1, we know that T1=n1n2n1+n2dtn1+n22(λ), where λ=n1n2n1+n2θ. Thus, the mean and variance of d, are given, respectively, by

(3.5)E(d)=En1+n2n1n2T1=J(n1+n22)θ
and
(3.6)Var(d)σd2=n1+n2n1n2n1+n22n1+n241+λ2J(n1+n22)λ2.

Now it is easy to obtain the density of H=dE(d)σd, which is symmetric about 0 and given by

(3.7)fH(t)=Var(T1)fT1Var(T1)t+J(n1+n22)λ,
where fT1 is the density of T1. Note that if we have two independent random samples of sizes n1 and n2, we can construct the 100c% confidence interval for θ based on (3.7). Now, let T2=d12=X1Y12S, which has a noncentral t distribution with n1 + n2 − 2 degrees of freedom and noncentrality parameter θ/2. Thus, the unbiased estimator of θ is 2T2. Note that the variance of d1 is
(3.8)σd12=2V(T2)=n1+n22n1+n242+θ2[J(n1+n22)]2θ2.

Therefore, we can setup the confidence interval for given confidence level c and precision f, which is given in (3.1). Since there are two unknowns n1 and n2, there are no solutions using Equation (3.1) so we need to modify this equation. Let n =  min{n1, n2} so that the degrees of freedom n1 + n2 − 2 ≥ 2n − 2. Suppose that we have two independent samples of size n in both; then, S*2=S12+S222, and the distribution of d*=(X̅1X̅2)/S* is the noncentral t with 2(n − 1) degrees of freedom and the noncentrality parameter λ*=n2θ so that its mean and variance are

E(d*)=J(2n2)θandVar(d*)=σd*2=2nn1n2(1+λ*)2[J(2n2)λ*]2.

Thus, the standardized random variable H*=d*E(d*)S* has the density given in Equation (3.3). Therefore, the required n can be obtained by solving Equation (3.2). Similarly to Remark 2.2, we know that H2(n1),(1c)/2Hn1+n22,(1c)/2, where Hm,(1−c)/2 is the critical value of the distribution H. It is easy to see that

(3.9)c=PH2(n1),(1c)/2H*H2(n1),(1c)/2PH2(n1),(1c)/2HH2(n1),(1c)/2
so that the desired results follows. □
Remark 3.1.

The required n obtained in Theorem 3.1 is unique. Also, if the conditions in Theorem 3.1 are satisfied, we can construct a c × 100% confidence interval for θ=μ1μ2σ

(3.10) d*σd1*fJ(2n2),d*+σd1*fJ(2n2),
where σd1* is given in (3.4) and θ = θ0, which can be obtained from the previous data, otherwise the default θ = 0.

Remark 3.2.

In order to see that Equation (3.9) holds numerically, we provide probabilities of the c = 95% confidence intervals for different sample sizes n1 and n2, which is given in Table 1.

Remark 3.3.

Researchers can access at the following website https://appcohensd.shinyapps.io/independent/ to obtain the required sample size. The input variables are the value of θ0 from the previous data by previous data or θ0 = 0, precision f, and confidence level c. For convenience, the output variable is the required sample size n =  min{n1, n2}. The required sample sizes for different values of precision f = 0.1, 0.15, 0.2, 0.25, confidence levels c = 0.95, 0.90 and θ0 = 0, 0.1, , 1 for independent case are given in Table 2. The relationship between required sample n and parameter θ for different values of precision f is given in Figure 3.

3.2 The minimum sample size of Cohen's d needed for a given sampling precision in matched samples

Theorem 3.2.

Let (X11,X21),,(X1n,X2n) be a random sample of size n from a bivariate normal population with mean vector μ and covariance matrix Σ, where

μ=μ1μ2andΣ=σ21ρρ1.

Denote Di = X1iX2i, i = 1, …, n. Let θD=μ1μ2σ and dD=D̅SD be the Cohen's effect sizes of the population and matched sample, respectively, where

D̅=1ni=1nDiandSD2=1n1i=1n(DiD̅)2
are the mean and variance of Di's, respectively. The density function of ndD is fT, the density noncentral t-distribution with n − 1 degrees of freedom and noncentrality parameter λD=n2(1ρ)θD. Also, the mean and variance of dD are given by
E(dD)=J(n1)λD/nandσdD2=1nn1n31+λD2J(n1)λD2.

Let c be the confidence level and f be the precision which satisfies

(3.11) P(fσdD1dDE(dD)fσdD1)=c,
where dD1=D1SD and
E(dD1)=J(n1)θD2(1ρ),Var(dD1)=σdD12=n1n31+θD22(1ρ)[J(n1)θD]22(1ρ).

Then the required sample size n can be obtained by

(3.12) LUfH(t)dt=c,
where fH is the density of H=[dDE(dD)]/σdD given by
(3.13) fH(h)=nσdDfTnσdDh+J(n1)λD
and
(3.14) L=fσdD1σdDU=L.

Proof. By Proposition 2.2, we know that T=ndDtn1(λD), where λD=n2(1ρ)θD.

It is easy to see that the mean and variance of dD are given by

E(dD)=J(n1)λD/nandσdD2=1nn1n31+λD2J(n1)λD2.

Thus, the density of H is given in Equation (3.13). Now, we know that dD1=D1/SDtn1(λD1), where λD1=θD2(1ρ). Then, the variance of dD1 is

σdD12=n1n31+λD12J(n1)λD12.

From the distribution of H, we can obtain the required sample size n which is given in Equations (3.12) - (3.14) so that the desired result follows. □

Remark 3.4.

The value of n obtained is unique with f. Also, if the conditions in Theorem 3.2 are satisfied, we can construct a c × 100% confidence interval for θD given by

(dDσdD1f)2(1ρ)J(n1),(dD+σdD1f)2(1ρ)J(n1),
where θD=θD0, which can be obtained from the previous data, otherwise the default θD = 0.

Remark 3.5.

Researchers can access at the following website: https://appcohensd.shinyapps.io/matched/

The input variables are the confidence level c, precision f, correlation coefficient ρ and θD=θD0 obtained from previous data information. The default value of θD = 0. The output is the desired sample size n required. Table 3 provides the n for c = 0.90, 0.95 for different values of f and different values of θD0's. Also the relationship between n and θD is given in Figure 4.

4. Simulation and real data examples

In this section we will provide some simulation results and two real data examples to support our main results. Based on M = 100,000 runs, coverage rates of the confidence intervals and the corresponding point estimates of θ in independent case and θD for matched data are given in Tables 4 and 5. From these two tables, we can see that the performances of our APP procedures work very well and the biases are really small.

To evaluate our results, we provide a real data example for independent and match sample cases, respectively.

Example 4.1.

Consider the salary data (in 2011) of all professors from College of Arts and Sciences and College of Engineering (population 1 with size 85) and other colleges (population 2 with size 93) of New Mexico State University, which are publicly available https://riograndefoundation.org/downloads/rgf_pr_nmsu.pdf

The estimated distributions based on the data set are approximately N(8.0868, 1.10992) for population 1 and N(9.5273, 1.41152) for population 2 (with unit $10,000) (see Figures 5 and 7). The Q–Q plots of the data sets are given in Figures 6 and 8, showing that the scatters lie close to the line with no obvious pattern coming away from the line in both Q–Q plots. Now we consider the 95% confidence interval of θ with precision f = 0.25 and default θ = 0. From Table 2, the minimum sample sizes n needed is 60, so we randomly choose samples of size 60 from both populations and obtain the value of Cohen's d is d = −1.1430. Then by Remark 3.1, the 95% confidence interval of θ is [−1.4900, −0.7814], which includes the true value of population effect size θ = −1.1284.

Example 4.2.

For matched case, we consider the data set named “Rugby” from Pakage “PairedDate” in R by Champely (2018). This data set provides the ratings on a continuous ten-point scale of two experts about 93 actions during several rugby union matches. Let D be the difference between ratings of two experts. The histogram and estimated density curve of D are given in Figure 9. From the data, we obtain D̅=0.3011 and SD = 1.4872. The pattern of the points of the scatter plot shown in Figure 10 shows a positive linear relationship between this two variables. After calculation, we get the correlation coefficient is ρ = 0.85. Now we consider the 95% confidence interval of population effect size θD with precision f = 0.25 and default θ = 0. By the shinyApp provided in Section 4 for matched data, the minimum sample sizes n needed is 59. Randomly select a sample with 59 paired data; then by Remark 3.2, we have the 95% confidence interval of θD is [−0.3122, −0.0371], which includes the true value of population effect size −0.1122.

5. Conclusion remarks

Our goal was to derive ways to perform the APP with respect to Cohen's d for independent and matched samples. The present mathematics provide those derivations. In turn, computer simulations support the mathematical derivations. We also provide links to free and user-friendly programs to facilitate researchers performing the APP to determine sample sizes to meet their specifications for precision and confidence. An advantage of the programs is that even researchers who are unsophisticated in mathematics nevertheless can avail themselves of APP advantages.

In addition to the obvious benefit of aiding researchers who wish to compute Cohen's d determine the samples sizes they need the present mathematics provide an additional benefit. Specifically, the famous article in Science by the Open Science Collaboratio (2015) included replications of studies in top psychology journals. They found that the average effect size in the replication cohort of studies was less than half that in original cohort of studies. Thus, effect sizes tend not to replicate across study cohorts. Our suspicion is that one reason for irreproducibility is that sample sizes are too small and traditional power analyses are insufficient because they do not address the precision issue (Trafimow and Myüz, 2019; Trafimow et al., 2020b), though significance testing doubtless plays a role too. The present mathematics, along with the links to computer programs, provide a solution. We hope and expect that researchers who wish to use Cohen's d to index their effect sizes will be better able to determine appropriate sample sizes, and thereby increase reproducibility in the social sciences.

Figures

The density curve of noncentral t-distribution with degrees of freedom 100 and different values of noncentral parameter λ

Figure 1

The density curve of noncentral t-distribution with degrees of freedom 100 and different values of noncentral parameter λ

The density curve of noncentral t-distribution with noncentral parameter λ = 1 and different values of degrees of freedom

Figure 2

The density curve of noncentral t-distribution with noncentral parameter λ = 1 and different values of degrees of freedom

The relationship between required sample n and parameter θ for c = 0.90 and different values of precision f

Figure 3

The relationship between required sample n and parameter θ for c = 0.90 and different values of precision f

The relationship between required sample n and parameter θD for ρ = 0.5, c = 0.90 and different values of precision f

Figure 4

The relationship between required sample n and parameter θD for ρ = 0.5, c = 0.90 and different values of precision f

Histogram and estimated density curve

Figure 5

Histogram and estimated density curve

Q–Q plot

Figure 6

Q–Q plot

Histogram and estimated density curve

Figure 7

Histogram and estimated density curve

Q–Q plot

Figure 8

Q–Q plot

Histogram and estimated density curve

Figure 9

Histogram and estimated density curve

Correlation

Figure 10

Correlation

Probabilities of the c = 95% confidence intervals for different sample sizes n1 and n2

θn1n2dfH2n−2,(1−c)/2P(|H| < H2n−2,(1−c)/2)
010010019813.873660.95
120218 0.9395
150248 0.9267
0.210010019813.873660.9512
120218 0.9408
150248 0.9281
0.510010019814.048870.9528
120218 0.9427
150248 0.9302
0.810010019814.144430.9547
120218 0.9445
150248 0.9322

The desired sample sizes n for f = 0.1, 0.15, 0.2, 0.25, θ = 0, 0.1, , 1 and c = 0.95, 0.9 in the independent case

θ00.10.20.30.40.50.60.70.80.91
f = 0.1c = 0.95382384386389392396405411415423436
c = 0.90268269271273276279283287294300304
f = 0.15c = 0.95169170171173174176180183186191194
c = 0.90118119120121122124125127130133135
f = 0.2c = 0.95949596979899100102104106110
c = 0.906666676768697071737476
f = 0.25c = 0.956060616263636465676870
c = 0.904141424243444445464749

The desired sample sizes n for f = 0.1, 0.15, 0.25, θD = 0, 0.2, 0.5, 0.8 and c = 0.95, 0.9 in the case of ρ = 0.2, 0.5, 0.8

θD00.20.50.8
f = 0.1ρ = 0.2c = 0.95382389418470
c = 0.90267273293330
ρ = 0.5c = 0.95382392436517
c = 0.90267276304363
ρ = 0.8c = 0.95382407514705
c = 0.90267284361490
f = 0.15ρ = 0.2c = 0.95168173186207
c = 0.90117120130144
ρ = 0.5c = 0.95168175198231
c = 0.90117122135163
ρ = 0.8c = 0.95168180234325
c = 0.90117126162216
f = 0.25ρ = 0.2c = 0.9559616776
c = 0.9040424651
ρ = 0.5c = 0.9559627181
c = 0.9040424858
ρ = 0.8c = 0.95596582118
c = 0.9040445877

The corresponding point estimates of θ and the coverage rates when n satisfies the required minimum sample size for f = 0.1, 0.15, 0.2, 0.25, θ = 0, 0.2, 0.5, 0.7, 1 and c = 0.95, 0.9 in the independent case

fθc = 0.95c = 0.90
nθˆCoverage ratenθˆCoverage rate
0.103820.00010.94892680.00010.8971
0.23860.19990.95092710.20040.8976
0.53960.50020.94992790.50070.8988
0.74110.70070.95092870.70120.9008
14361.00070.95093041.00120.8997
0.150169−0.00020.9489118−0.00030.8976
0.21710.20080.94991200.20060.8998
0.51760.50070.95031240.50090.9006
0.71830.70130.95111270.70230.9011
11941.00190.95081351.00310.8987
0.2094−0.00040.9473660.00050.8981
0.2960.20130.948670.19980.8978
0.5990.50090.9497690.50450.898
0.71020.70230.952710.70290.8985
11101.00250.9469761.00510.9007
0.250600.00020.946941−0.00050.8913
0.2610.20090.9472420.20260.8951
0.5630.50280.9469440.50490.898
0.7650.70460.9489450.70580.8974
1701.00540.9522481.00850.8983

The corresponding point estimates of θD and the coverage rates when n satisfies the required minimum sample size for f = 0.1, 0.15, 0.2, 0.25, θ = 0, 0.2, 0.5, 0.7, 1, ρ = 0.2, 0.5, 0.8 and c = 0.95 in the matched case

fθDc = 0.95, ρ = 0.2c = 0.95, ρ = 0.5c = 0.95, ρ = 0.8
nθDˆCoverage ratenθDˆCoverage ratenθDˆCoverage rate
0.103820.00010.9503382−0.00010.94883820.00010.9499
0.23890.20020.94963920.20040.95024070.20060.9501
0.54180.50070.95144360.50070.95055140.50080.9535
0.74470.70080.95064930.70110.95296250.70090.9513
15141.00150.95175941.00110.95398921.00080.9542
0.150168−0.00030.9468168−0.00010.947168−0.00010.9486
0.21730.20130.94991750.20090.94951730.20070.9509
0.51860.50240.94981980.50190.95292070.50140.9544
0.71990.70270.95152220.70240.95442910.70190.9564
12341.00330.95452571.00320.95093911.00180.9517
0.20930.00010.944893−0.00030.94449300.9452
0.2970.20160.9486980.20130.9491020.20140.9511
0.51040.50380.94851100.50390.95091290.50260.9524
0.71150.7040.95461250.70430.95311590.70340.9513
11291.00580.95271541.00440.95752281.00340.956
0.250590.00010.9448590.00010.945159−0.00030.9438
0.2610.20250.9486620.20220.9489650.20190.9491
0.5670.50620.9497710.50570.9514820.50420.9532
0.7730.70710.9526800.70660.95381050.70510.9562
1821.00980.9529971.00750.95631461.00550.956

References

Bhandari, P. (2020), Effect Size in Statistics-Scribbr, available at: https://www.scribbr.com/statistics/effect-size/.

Cohen, J. (1988), Statistical Power Analysis for the Behavioral Sciences, 2nd ed., Erlbaum, Hillsdale, NJ.

Champely, S. (2018), “PairedData: paired data analysis. R pakage version: 1.1.1”, available at: https://CRAN.R-project.org/package=PairedData.

Li, H., Trafimow, D., Wang, T., Wang, C. and Hu, L. (2020), “User-friendly computer programs so econometricians can run the a priori procedure”, Frontiers in Management and Business, Vol. 1 No. 1, pp. 2-6, doi: 10.25082/FMB.2020.01.002.

Nguyen, H.T. and Wang, T. (2008), A Graduate Course in Probability and Statistics, Vol. 2, Tsinghua University Press, Beijing.

Open Science Collaboration (2015), “Estimating the reproducibility of psychological science”, Science, Vol. 349 No. 6251, p. aac4716, doi: 10.1126/science.aac4716.

Schafer, T. and Schwarz, M. (2019), “The meaningfulness of effect sizes in psychological research: differences between sub-disciplines and the impact of potential biases”, Frontiers in Psychology, Vol. 10, p. 813. doi: 10.3389/fpsyg.2019.00813.

Sullivan, G.M. and Feinn, R. (2012), “Using effect size-or why the p value is not enough”, Journal of Graduate Medical Education, Vol. 4 No. 3, pp. 279-282, doi: 10.4300/JGME-D-12-00156.1.

Trafimow, D. (2017), “Using the coefficient of confidence to make the philosophical switch from a posteriori to a priori inferential statistics”, Educational and Psychological Measurement, Vol. 77 No. 5, pp. 831-854, doi: 10.1177/0013164416667977.

Trafimow, D. (2019), “A frequentist alternative to significance testing, p-values, and confidence intervals”, Econometrics, Vol. 7 No. 2, pp. 1-14, available at: https://www.mdpi.com/2225-1146/7/2/26.

Trafimow, D. and MacDonald, J.A. (2017), “Performing inferential statistics prior to data collection”, Educational and Psychological Measurement, Vol. 77 No. 2, pp. 204-219, doi: 10.1177/0013164416659745.

Trafimow, D. and Myüz, H.A. (2019), “The sampling precision of research in five major areas of psychology”, Behavior Research Methods, Vol. 51 No. 5, pp. 2039-2058, doi: 10.3758/s13428-018-1173-x.

Trafimow, D. and Uhalt, J. (2020), “The inaccuracy of sample-based confidence intervals to estimate a priori ones”, Methodology: European Journal of Research Methods for the Behavioral and Social Sciences, Vol. 16 No. 2, pp. 112-126, doi: 10.5964/meth.2807.

Trafimow, D., Wang, C. and Wang, T. (2020a), “Making the a priori procedure (APP) work for differences between means”, Educational and Psychological Measurement, Vol. 80 No. 1, pp. 186-198, doi: 10.1177/0013164419847509.

Trafimow, D.T., Hyman, M.R. and Kostyk, A. (2020b), “The (im)precision of scholarly consumer behavior research”, Journal of Business Research, Vol. 114, pp. 93-101, doi: 10.1016/j.jbusres.2020.04.008.

Wang, C., Wang, T., Trafimow, D. and Talordphop, K. (2020), “Extending the a priori procedure to one-way analysis of variance model with skew normal random effects”, Asian Journal of Economics and Banking, Vol. 4 No. 2, pp. 77-90.

Wang, C., Wang, T., Trafimow, D., Li, H., Hu, L. and Rodriguez, A. (2021), “Extending the A Priori procedure (APP) to address correlation coefficients”, in Ngoc Thach, N., Kreinovich, V. and Trung, N.D. (Eds), Data Science For Financial Econometrics, Springer. doi: 10.1007/978-3-030-48853-610.

Wei, Z., Wang, T., Trafimow, D. and Talordphop, K. (2020), “Extending the a priori procedure to normal Bayes models”, International Journal of Intelligent Technologies and Applied Statistics, Vol. 13 No. 2, pp. 169-183, doi: 10.6148/IJITAS.202006-13(2).0004.

Further reading

Baguley, T. (2009), “Standardized or simple effect size: what should be reported?”, British Journal of Psychology, Vol. 100 No. Pt 3, pp. 603-617.

Bobko, P., Roth, P.L. and Bobko, C. (2001), “Correcting the effect size of d for range restriction and unreliability”, Organizational Research Methods, Vol. 4 No. 1, pp. 46-61.

Cohen, J. (2013), Statistical Power Analysis for the Behavioral Sciences, Elsevier Science, United Kingdom.

Demidenko, E. (2016), “The p-value you can't buy”, The American Statistician, Vol. 70 No. 1, pp. 33-38, doi: 10.1080/00031305.2015.1069760.

Gillett, R. (2003), “The metric comparability of meta-analytic effect-size estimators from factorial designs”, Psychological Methods, Vol. 8 No. 4, pp. 419-433.

Hedges, L.V. (1981), “Distribution theory for Glass's estimator of effect size and related estimators”, Journal of Educational Statistics, Vol. 6 No. 2, p. 107, doi: 10.2307/1164588.

Wang, C., Wang, T., Trafimow, D. and Chen, J. (2019), “Extending a priori procedure to two independent samples under skew normal settings”, Asian Journal of Economics and Banking, Vol. 3 No. 2, pp. 29-40.

Wang, C., Wang, T., Trafimow, D. and Xu, Z. (2021), “A priori procedure (APP) for estimating the scale parameter in gamma populations for known shape parameter”, in Sriboonchitta, S., Kreinovich, V. and Yamaka, W. (Eds), Quantum Computing in Econometrics and Quantum Economics, Springer-Verlag.

Corresponding author

Tonghui Wang can be contacted at: twang@nmsu.edu

Related articles