28 Jan 2014: Kuo CL, Zaykin D (2014) Correction: The Ranking Probability Approach and Its Usage in Design and Analysis of Large-Scale Studies. PLOS ONE 9(1): 10.1371/annotation/6bc48309-0280-4066-b937-e88debd1579c. https://doi.org/10.1371/annotation/6bc48309-0280-4066-b937-e88debd1579c View correction
In experiments with many statistical tests there is need to balance type I and type II error rates while taking multiplicity into account. In the traditional approach, the nominal -level such as 0.05 is adjusted by the number of tests, , i.e., as 0.05/. Assuming that some proportion of tests represent “true signals”, that is, originate from a scenario where the null hypothesis is false, power depends on the number of true signals and the respective distribution of effect sizes. One way to define power is for it to be the probability of making at least one correct rejection at the assumed -level. We advocate an alternative way of establishing how “well-powered” a study is. In our approach, useful for studies with multiple tests, the ranking probability is controlled, defined as the probability of making at least correct rejections while rejecting hypotheses with smallest P-values. The two approaches are statistically related. Probability that the smallest P-value is a true signal (i.e., ) is equal to the power at the level , to an excellent approximation. Ranking probabilities are also related to the false discovery rate and to the Bayesian posterior probability of the null hypothesis. We study properties of our approach when the effect size distribution is replaced for convenience by a single “typical” value taken to be the mean of the underlying distribution. We conclude that its performance is often satisfactory under this simplification; however, substantial imprecision is to be expected when is very large and is small. Precision is largely restored when three values with the respective abundances are used instead of a single typical effect size value.
Citation: Kuo C-L, Zaykin D (2013) The Ranking Probability Approach and Its Usage in Design and Analysis of Large-Scale Studies. PLoS ONE8(12): e83079. https://doi.org/10.1371/journal.pone.0083079
Editor: Chuhsing Kate Hsiao, National Taiwan University, Taiwan
Received: August 19, 2013; Accepted: November 7, 2013; Published: December 20, 2013
This is an open-access article, free of all copyright, and may be freely reproduced, distributed, transmitted, modified, built upon, or otherwise used by anyone for any lawful purpose. The work is made available under the Creative Commons CC0 public domain dedication.
Funding: This research was supported by the Intramural Research Program of the NIH, National Institute of Environmental Health Sciences. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Development of novel statistical methods to accompany technological advances in biological sciences is essential in order to further scientific progress. Human genetics is a prime example of a field where data is currently being produced at an accelerating rate. The sheer volume of data and its complexity create new challenges for statistical science. How to design studies appropriately, and how not to drown in the data while extracting useful signals? Studies now routinely involve multiple testing of statistical hypotheses on a massive scale. Thus, both design and analysis of large-scale experiments must entail considerations of multiplicity.
Design of studies often involves sample size and power calculations. In exploratory studies where many statistical tests are anticipated, it is reasonable to assume that most test statistics and their respective P-values are generated under the condition where the null hypothesis holds. One would also assume that there are multiple real signals. In reality, effect sizes (e.g. odds ratios) for these true signals follow some distribution, although it might be tempting to assume, for simplicity, a single “typical” effect size. In multiple testing contexts, statistical power can be defined as the probability of correctly rejecting at least one (or at least ) null hypotheses. Power defined this way increases with the total number of true signals. Thus, not only the total number of tests, but the number of truly associated loci and the distribution of effect sizes must be taken into account. Figure 1 is an illustrative example that shows a sample (P-value) plot for a simulation experiment with multiple tests. The graph shows ordered of 1000 P-values (Y axis) for a one degree of freedom chi-square statistic, plotted against what is expected under the null hypotheses (), i.e. when P-values would follow the uniform (0,1) distribution. Ten out of 1000 P-values, denoted by red circles, were simulated assuming the alternative hypothesis so that these P-values originated from statistics having a noncentral chi-square distribution. The largest value (4.9), that happened to be a true positive, is the only one significant at the level of 0.05/1000. Yet, among the first ten largest values, there are four true signals. These observations raise some important statistical questions:
red circles– true signals; crosses– false signals.
- What is the probability that the largest value on the graph is a true signal?
- What is the expected number (or proportion) of true signals among, for example, the first ten top results?
We will seek a simple and practical way with a minimum number of assumptions to allow for such computations.
In the simulation experiment to construct Figure 1, ten noncentralities for the ten assumed true signals were sampled from the Gamma(1/2, 15) distribution. Given such information, the questions above can be answered exactly, with the help of simple formulas that will be detailed later. Specifically, probability that the minimum P-value (top red dot) is a true signal is equal to 0.93, and the expected proportion of true signals among the first ten top results is 0.35 (the actual proportion observed in this particular experiment is 4/10). Because the noncentrality is the multiple of sample size (Equation 4), one can also calculate a sample size needed for a specified proportion of true signals to aggregate among a speciffied number of top hits.
For the next motivating example, we considered a genome-wide association study (GWAS) with association trend tests among which there were 75 true signals with the allele frequency 0.15 and the relative risk 1.15. We assumed 3000 cases, 3000 controls, which resulted in the noncentrality 7.8 that governed the distribution of P-values for true signals. Under these assumptions,
- Power to detect any particular associated variant at the -level of is only 0.003.
- Power to detect at least one of the 75 true signals is 0.19, at the -level of .
More revealing are calculations based on the ranks of true signals:
- Probability that the best hit (i.e. the smallest among P-values) is a true signal is 0.5.
- There is 0.95 probability that at least one true signal will be encountered among the 11 smallest P-values.
- Expected proportion of true signals among these 11 smallest P-values is 0.23, i.e. there are about 2–3 true signals expected among the 11 smallest P-values.
Table 1 shows an expanded list for the number of true signals () expected to be encountered among a specific number of the smallest P-values (that would contain both true and false signals, +). These numbers are also shown in Figure 2. The number of real discoveries grows sharply at first; however, a large number of P-values would be required to “cover” a good portion of the 75 true signals. At the same time, the proportion of true signals among top P-values, /(+), quickly drops. There is a good chance to find some true signals at the top of the list of best results, but only a few of the 75 can be found. If we suppose a confirmatory study of top hits, then following up a large number of results would again incur a substantial multiple testing penalty.
The number of true signals (T) expected to be encountered among a specific number of the smallest P-values.
All of the calculations for the presented examples are derived from the ranking probability, i.e. the probability that at least true signals would rank among the top hits, . As these examples illustrate, ranking probability is a concept that allows characterization of how well-powered a study is in a way that is usefully different from the regular power calculation. The ranking approach depends on the same assumed parameters as the statistical power approach, yet it provides additional information about a study. Ranking probabilities are related to the false discovery rate (FDR), in particular, the proportion of true signals among a specified number of top P-values can be expressed via the average of ranking probabilities. We provided approximations to and showed good performance of these approximations under local dependence of P-values, such as dependence due to linkage disequilibrium in GWAS . Here, we specifically address the issue of accuracy in estimating the expected proportion of true signals among top hits of a study. A major difficulty with applications of the approach is its reliance on a parametric distribution used to characterize the effect size distribution . Not only it is difficult to estimate parameters of such a distribution reliably, but the effect of misspecification of that distribution has been unclear. A distribution can be specified erroneously in a multitude of possible ways: not only the parameter values, but the distribution itself might be incorrect. Here we addressed the issue of misspecification by taking a perspective that very little might be guessed well about the effect size distribution, perhaps only its mean or the abundance of a few effect size values, and evaluated performance of the ranking approach under these conditions. With only one or a few specific effect size values, the approach simplifies considerably and can be more readily implemented and applied for design of genetic studies.
Materials and Methods
For ease of exposition, we assume one-degree of freedom chi-squared statistics, commonly found in genetic association studies. P-values are random variables, with the cumulative distribution function (CDF) given by(1)where and denote the CDF of the test statistic under the null and the alternative hypothesis and is the noncentrality parameter. When , we get the uniform distribution, . P-value density for a fixed can be obtained by differentiating :(2)where is the density that corresponds to the cumulative distribution . The inverse of Equation (1) can be used conveniently to generate random P-values under the alternative hypothesis, i.e. assuming a certain noncentrality, as(3)where is a random number from the Uniform(0,1) distribution . The noncentrality parameter can be calculated assuming an effect size, such as odds ratio, and a particular statistic. For example, the usual chi-square statistic for testing log odds ratio () being equal to zero has the noncentrality that is approximately(4)where , i.e. is one half of the harmonic mean of the counts for the table row totals (they are assumed to represent two binomial samples) and is the frequency, pooled across the rows. Recently, Lee and Wray reported simple formulas for computing the noncentrality given various experimental designs .
Equation (1) gives the P-value distribution for a fixed value of the noncentrality, . Different signals have distinct associated noncentrality values. When different tests have approximately the same sample size, as in GWAS, noncentralities can be thought to arise from some distribution, .
For continuously distributed effect sizes, Gamma(,) distribution can be utilized. This distribution provides flexible shapes, and for its shape parameter values 1, the distribution is L-shaped, which corresponds to a biologically realistic assumption that there are many small effects, but a few that are large. The ranking probability calculation makes use of the marginal P-value distribution, averaged over all possible noncentrality values,(5)Similarly, assuming that noncentralities for true signals arise from some distribution, , the marginal density of P-values for true associations is(6)
Effect size distribution may be specified in a tabulated fashion and statistical methods begin to emerge for estimating the effect size distribution, discretized into bins (e.g. ). Each bin is defined by the average effect size, , and the relative abundance of signals that fall into that bin, . The discrete version of the marginal distribution can thus be defined as follows:(7)
When true signals are independent, the CDF of the -th ordered true signal P-value is(8)where Bin denotes the binomial CDF evaluated at successes in trials ( is the number of true signals) with the success probability . We previously derived a precise approximation for the ranking probability . Specifically, ranking probability, as the chance that at least true signals will rank among top smallest P-values is(9)(where = for ). We note that although the ranking probability depends on the total number of true and false signals, as well as on the effect size distribution, only and are the varying indices in a given study. Thus, we omit dependence on other parameters to make the notation succinct. The approximation given by Equation (9) improves as both the rank and the total number of tests increase, but the error for is only around 1–3% at as small as 10 (data not shown). Even though our focus is on large-scale testing, the ranking approach is valid when is small and provides an excellent approximation when and tests are independent. If there is considerable local correlation between P-values, as found in large-scale association studies, the term can be dropped for increased precision. The resulting formula still provides a good approximation, unless the correlation is extreme and is close to one. In a study with tests that behaved as only 6.4% of that number due to extreme correlation (as judged by the distribution of the minimum P-value), the approximation was adequate at 15 .
The main question that we address here is how the ranking approach is affected by the precision in specification of the effect size distribution that is utilized by equations for the marginal P-value distribution (Equations 5, 6). Considering noncentralities ('s) as random, we can rewrite Equation (5) as the expectation,Using the first order Taylor approximation,(10)where . This is also Equation (7) with . Approximations utilizing the mean introduce imprecision, compared to the approaches based on the distribution, but the degree of inaccuracy due to these approximations has been unclear.
Relation between ranking probability, power, and the expected P-value
If we replace by a significance threshold , Equation (1) will define power at the level for a particular signal with the noncentrality . With multiple () independent true signals, power to detect at least one of them isEquation (9) reveals a relation of probability that the smallest P-value is a true signal to power to detect a particular true signal. If , so that ,Thus, probability that a true signal ranks first when results of a multiple testing experiment are sorted by P-value is the average power to detect at least one signal at the level ten times higher than the conventional, Bonferroni-adjusted level, .
The ranking approach can also be related to the “expected P-value” approach of Sackrowitz and Samuel-Cahn  who considered the probability that a single false signal P-value () is smaller than a single true signal P-value (). This probability can be written in terms of the expectation of the true signal P-values:(11)In the context of multiple testing considered in our ranking approach, there are multiple false P-values. Let us denote the -th smallest P-value (without regard whether that is a true or false signal) by . and will denote the -th smallest P-value among false and true signals, respectively. With this notation,(12)Considering a particular true signal and setting ,(13) is the CDF of the -th ordered P-value under the null hypothesis, which for continuous and independent test statistics has the Beta(, ) distribution, where is the number of false signals. When there are only two tests, one of which is a false signal, Equation (12) is equal to Equation (11).
Ranking false discovery rate (rFDR)
The ranking false discovery rate, rFDR() is defined as the proportion of null signals among the top hits. A study-specific rFDR value is(14)where is the -th smallest P-value, and is the posterior probability of the null hypothesis given the -th smallest P-value. These posterior probabilities are unaffected by the fact of selection of the smallest P-values, that is, they do not depend on , but only on the value :(15)The second equality assumes that under , P-values follow the uniform (0,1) distribution. The rFDR can be related to the traditional FDR concept, introduced by Benjamini and Hochberg . In their method, the proportion of false discoveries among discoveries is controlled by the expectation that is taken over multiple-testing experiments with or without rejections. Storey  confined attention to control of the FDR among only those experiments that contained rejections of (“positive FDR”, pFDR). While the pFDR is defined for a fixed P-value threshold, the study-specific rFDR is defined for a fixed rank, . The expected rFDR() is defined as the average across many multiple-testing experiments:(16)(17)Equations (16), (17) show the relation between the posterior probability and the ranking probability. Approximation, rather than equality arises from the fact that Equation (17) assumes that the number true signals, , is fixed, while Equation (16) assumes that each P-value corresponds to a true signal with probability 1-, resulting in the expected number of true signals across multiple-testing experiments being equal to . There is a slight difference in values provided by these equations when is close to one, hence the relation is approximate.
We used simulation experiments to evaluate precision of rFDR estimates. These simulations were used to construct Tables 2, 3, 4. At each simulation, we sampled noncentrality parameters for a specific number of true signals from an L-shaped Gamma(shape = 1, scale = ) distribution. P-values for true signals were simulated using this sample of noncentralities using Equation (3), while P-values for the remaining signals were sampled from the Uniform(0,1) distribution. Next, P-values were sorted, the smallest P-values were taken and the proportion of true signals was recorded. 1-rFDR values were estimated via Equation (17). This describes a single simulation experiment. The 1-rFDR values were averaged across 1000 simulations to produce an estimate as given by Equation (16).
In addition, for each simulation setting, the expected rFDR was estimated via the sum of ranking probabilities, by utilizing (1) the true (continuous) effect size distribution; (2) the distribution reduced to three bins (by utilizing Equation 7 with the bins defined as [0, ], [, ], and [, ]; and their relative sizes () obtained from the integration on the true distribution); (3) the distribution reduced to its mean. The terms in Equation (17) were computed using the approximation given in Equation (9). Note that this computation does not utilize P-values and in that way it is analogous to a classical power computation.
In Figure 1, transformed ordered P-values were plotted against their expected values, computed under . The −log transformation provides better visualization of deviations from the expected line, compared to the plots of non-transformed P-values. A reviewer raised the issue of a possible usage of in place of . This appears to be a common practice due to simplicity of the resulting expression, however, bringing the expectation inside introduces a bias due to Jensen's inequality. This bias causes a tendency for the smallest null P-values to appear above the expected line. The correct expected values can be found as follows. As before, denote -th smallest P-value by . Under , and has the expectationwhere is the base of the logarithm, , , , is an arbitrary scale parameter of the gamma distribution, and is the digamma function.
As shown in Materials and Methods, power calculations, ranking probabilities, Bayesian forms of FDR, and posterior probabilities of depend on the assumed distribution of effect sizes. The noncentrality can be regarded as a standardized effect size. Park et al.  explicitly defined the effect size to be , where is defined by Equation (4). In the following, we will keep sample size in the “effect size” definition, thus by “effect size” we will simply mean .
In simulation experiments, we utilized the ranking false discovery rate, rFDR(), defined as the proportion of null signals among the top hits. We investigated imprecision in estimates of the proportion of true signals (1-rFDR) due to approximations by Equations (7, 10) via simulation experiments as detailed in Materials and Methods: Simulation Experiments. Tables 2, 3, 4 show the results for three simulation settings, where we considered = 100 (the number of true signals) and = , , (total number of P-values). We conducted simulations assuming independence, however, as increases, ranking probabilities computed under the independence assumption become robust in the presence of even extreme correlation among false signals, which are assumed to represent majority of the tests.
In Tables 2, 3, we considered the number of top hits to be = 100 and = 200, correspondingly; and in Table 4, was set to 10. Noncentralities were assumed to arise from an L-shaped Gamma(, ) distribution. In Table 2, the shape was set to = 1 and the scale was varied. In Table 4, both the shape () and the scale () were varied. Smaller shape and scale values result in smaller mean values of the noncentralities (), thus, in lower true values of 1-rFDR.
As expected, 1-rFDR computed by the method that utilizes the actual continuous effect size distribution (via Equation 5) is precise in all scenarios. The three bin approximation performed similarly to the integration method and consistently gave estimates close to the true 1-rFDR values. The error introduced by the simplified analysis, where instead of the integration of the noncentrality distribution, its mean value is utilized, gives acceptable results for relatively large values of top hits ( = 100 and = 200). For = 10, there appears to be considerable inaccuracy, especially when becomes as large as .
Figure 3 shows the expected number of true signals among the first 200 top results, as a function of the number of truly associated signals. Assuming the single value approximation, we postulated the term in Equation (4) to be equal to 0.0024, which could have resulted assuming that typical value of the odds ratio is 1.15 and the allele frequency . Further, we assumed 106 tests and three different sample sizes. The smallest sample size (5000 cases and 5000 controls) allows to detect less than one half of the actual number of true signals. Doubling the sample size makes the number of discovered signals to rise considerably more sharply, as the number of true signals increases. The combined sample (15000 cases and 15000 controls) allows to detect nearly all true signals.
Classical counterparts to ranking and posterior probabilities are statistical power and P-values. Power is calculated for a fixed significance threshold while in our ranking approach it is the number of top hits that is fixed. “Fixed” in this context simply means not-random: it is possible to vary the number of top hits by incrementing starting from one, until a predefined value of rFDR is reached, or until the ranking probability (e.g. ) is at a needed value. Gail et al. provided an alternative approach for estimating the expected proportion of true signals among a fixed number of top hits of a study . Our approach is conceptually the same as that of Gail et al., but differs in the implementation. Our approach utilizes P-values, and can accommodate variety of test statistics. It also provides additional flexibility in the way the effect size distribution is specified. In particular, the distribution is not limited to a normal, in fact, one can avoid usage of any parametric distribution by employing the tabulated effect size distribution (Equation 7). Our approach hinges on expressing the rFDR in terms of the ranking probability introduced in Zaykin and Zhivotovsky  and on the usage of approximations of Kuo and Zaykin . Usage of the tabulated distribution (a typical effect size value, e.g., the distribution mean, is a special case) brings about additional computational simplicity, and we investigated performance of this approach as an approximation to the true continuous effect size distribution.
Ranking probabilities, rFDR, as well as standard power calculations when multiple tests are anticipated, all depend on the effect size distribution and the number of true signals. Specification of these parameters can be challenging. Here, we investigated whether the usage of the mean of the effect size distribution would provide acceptable precision. Wacholder et al.  proposed the “False Positive Report Probability” (FPRP), defined as the proportion of true signals among P-values that are smaller or equal to a fixed threshold . FPRP also utilizes a single “typical” effect size instead of using an assumed effect size distribution. This was criticized as an oversimplification , . But usage of a single value taken to be the mean of the underlying prior distribution can be justified as the first order Taylor approximation to the method that takes into account the entire distribution (Equation 10). Our simulation experiments show that imprecision introduced by this simplification may be tolerable. Usage of the distribution mean instead of the entire distribution is much simpler computationally. From the practitioner's viewpoint, it is relatively easy to specify a “typical” effect size, that is the mean, rather than the distribution. In addition, it might be possible to estimate the mean of the effect size distribution more reliably than the distribution itself.
P-value distribution functions, and , can also be expanded to the second order Taylor series, and in the case of the usual chi-square test statistic these approximations are easily computed:where is the variance of . The expectation operator can be thought of as a summation of approximation values (weighted by the density of at each particular point) across the entire domain of . However, it is known that higher order expansions may work worse than lower order ones away from the point of expansion (in our case, ). Our investigation into quality of these second order approximations did not reveal marked improvements over the first order approximations, especially at tiny P-values, where the second order ones sometimes performed worse (results not shown). Thus, in this paper, we focused on simple first order approximations. Satisfactory performance utilizing the mean of the effect size distribution suggests that our approach is robust against misspecification of the shape of the effect size distribution. However, there is considerable imprecision when is very large and is small. This underlines importance of development of statistical methods for estimating parameters of the effect size distribution. While replacing a continuous effect size distribution by its mean value may result in a loss of precision, collapsing that distribution into three bins performed surprisingly well in all simulations. Correct specification of three typical values (low, medium, high) with the relative abundance of each appears sufficient for computing the proportion of true positives among top hits with high precision and can be recommended for practical use.
R programs to compute the expected number of true (truly associated) signals among top hits for a given sample size are available at https://sites.google.com/site/entrust2013/. The programs can also be used to choose a sample size such that a desired number of true signals would rank among the top hits.
We are grateful to anonymous reviewers and the editor for insightful suggestions that improved this work.
Conceived and designed the experiments: CLK DZ. Performed the experiments: CLK DZ. Analyzed the data: CLK DZ. Wrote the paper: CLK DZ.
- 1. Kuo CL, Zaykin DV (2011) Novel rank-based approaches for discovery and replication in genome-wide association studies. Genetics 189: 329–340.
- 2. Zaykin DV, Zhivotovsky LA (2005) Ranks of genuine associations in whole-genome scans. Genetics 171: 813–823.
- 3. Lee SH, Wray N (2013) Novel genetic analysis for case-control genome-wide association studies: Quantification of power and genomic prediction accuracy. PLoS ONE 8: e71494.
- 4. Park J, Wacholder S, Gail M, Peters U, Jacobs K, et al. (2010) Estimation of effect size distribution from genome-wide association studies and implications for future discoveries. Nat Genet 42: 570–575.
- 5. Sackrowitz H, Samuel-Cahn E (1999) P values as random variables – expected P values. American Statistician 53: 326–331.
- 6. Benjamini Y, Hochberg Y (1995) Controlling the false discovery rate: a practical and powerful approach to multiple testing. J Royal Stat Soc B 57: 289–300.
- 7. Storey J (2002) A direct approach to false discovery rates. J Royal Stat Soc B 64: 479–498.
- 8. Gail MH, Pfeiffer RM, Wheeler W, Pee D (2008) Probability of detecting disease-associated single nucleotide polymorphisms in case-control genome-wide association studies. Biostatistics 9: 201–215.
- 9. Wacholder S, Chanock S, Garcia-Closas M, El Ghormli L, Rothman N (2004) Assessing the probability that a positive report is false: an approach for molecular epidemiology studies. J Natl Canc Inst 96: 434–442.
- 10. Lucke J (2009) A critique of the false-positive report probability. Genet Epidemiol 33: 145–150.
- 11. Wakefield J (2008) Reporting and interpretation in genome-wide association studies. Int J Epidemiol 37: 641–53.