## Figures

## Abstract

Plasmode is a term coined several years ago to describe data sets that are derived from real data but for which some truth is known. Omic techniques, most especially microarray and genomewide association studies, have catalyzed a new zeitgeist of data sharing that is making data and data sets publicly available on an unprecedented scale. Coupling such data resources with a science of plasmode use would allow statistical methodologists to vet proposed techniques empirically (as opposed to only theoretically) and with data that are by definition realistic and representative. We illustrate the technique of empirical statistics by consideration of a common task when analyzing high dimensional data: the simultaneous testing of hundreds or thousands of hypotheses to determine which, if any, show statistical significance warranting follow-on research. The now-common practice of multiple testing in high dimensional experiment (HDE) settings has generated new methods for detecting statistically significant results. Although such methods have heretofore been subject to comparative performance analysis using simulated data, simulating data that realistically reflect data from an actual HDE remains a challenge. We describe a simulation procedure using actual data from an HDE where some truth regarding parameters of interest is known. We use the procedure to compare estimates for the proportion of true null hypotheses, the false discovery rate (FDR), and a local version of FDR obtained from 15 different statistical methods.

## Author Summary

Plasmode is a term used to describe a data set that has been derived from real data but for which some truth is known. Statistical methods that analyze data from high dimensional experiments (HDEs) seek to estimate quantities that are of interest to scientists, such as mean differences in gene expression levels and false discovery rates. The ability of statistical methods to accurately estimate these quantities depends on theoretical derivations or computer simulations. In computer simulations, data for which the true value of a quantity is known are often simulated from statistical models, and the ability of a statistical method to estimate this quantity is evaluated on the simulated data. However, in HDEs there are many possible statistical models to use, and which models appropriately produce data that reflect properties of real data is an open question. We propose the use of plasmodes as one answer to this question. If done carefully, plasmodes can produce data that reflect reality while maintaining the benefits of simulated data. We show one method of generating plasmodes and illustrate their use by comparing the performance of 15 statistical methods for estimating the false discovery rate in data from an HDE.

**Citation: **Gadbury GL, Xiang Q, Yang L, Barnes S, Page GP, Allison DB (2008) Evaluating Statistical Methods Using Plasmode Data Sets in the Age of Massive Public Databases: An Illustration Using False Discovery Rates. PLoS Genet 4(6):
e1000098.
doi:10.1371/journal.pgen.1000098

**Editor: **Greg Gibson, The University of Queensland, Australia

**Received: **December 19, 2007; **Accepted: **May 15, 2008; **Published: ** June 20, 2008

**Copyright: ** © 2008 Gadbury et al. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.

**Funding: **Research supported in part by NIH grants U54CA100949 from the National Cancer Institute (NCI); P30DK056336 from the National Institute Of Diabetes And Digestive And Kidney Diseases (NIDDK); and R01ES09912 from the National Institute of Environmental Health Sciences (NIEHS). The content is solely the responsibility of the authors and does not necessarily represent the official views of the NCI, NIDDK, or NIEHS.

**Competing interests: ** The authors have declared that no competing interests exist.

## Introduction

“Omic” technologies (genomic, proteomic, etc.) have led to high dimensional experiments (HDEs) that simultaneously test thousands of hypotheses. Often these omic experiments are exploratory, and promising discoveries demand follow-up laboratory research. Data from such experiments require new ways of thinking about statistical inference and present new challenges. For example, in microarray experiments an investigator may test thousands of genes aiming to produce a list of promising candidates for differential genetic expression across two or more treatment conditions. The larger the list, the more likely some genes will prove to be false discoveries, i.e. genes not actually affected by the treatment.

Statistical methods often estimate both the proportion of tested genes that are differentially expressed due to a treatment condition and the proportion of false discoveries in a list of genes selected for follow-up research. Because keeping the proportion of false discoveries small ensures that costly follow-on research will yield more fruitful results, investigators should use some statistical method to estimate or control this proportion. However, there is no consensus on which of the many available methods to use [1]. How should an investigator choose?

Although the performance of some statistical methods for analyzing HDE data has been evaluated analytically, many methods are commonly evaluated using computer simulations. An analytical evaluation (i.e., one using mathematical derivations to assess the accuracy of estimates) may require either difficult-to-verify assumptions about a statistical model that generated the data or a resort to asymptotic properties of a method. Moreover, for some methods an analytical evaluation may be mathematically intractable. Although evaluations using computer simulations may overcome the challenge of intractability, most simulation methods still rely on the assumptions inherent in the statistical models that generated the data. Whether these models accurately reflect reality is an open question, as is how to determine appropriate parameters for the model, what realistic “effect sizes” to incorporate in selected tests, as well as if and how to incorporate correlation structure among the many thousands of observations per unit [2].

Plasmode data sets may help overcome the methodological challenges inherent in generating realistic simulated data sets. Catell and Jaspers [3] made early use of the term when they defined a plasmode as “a set of numerical values fitting a mathematico-theoretical model. That it fits the model may be known either because simulated data is produced mathematically to fit the functions, or because we have a real—usually mechanical—situation which we know with certainty must produce data of that kind.” Mehta et al. (p. 946) [2] more concisely refer to a plasmode as “a real data set whose true structure is known.” The plasmodes can accommodate unknown correlation structures among genes, unknown distributions of effects among differentially expressed genes, an unknown null distribution of gene expression data, and other aspects that are difficult to model using theoretical distributions. Not surprisingly, the use of plasmode data sets is gaining traction as a technique of simulating reality-based data from HDEs [4].

A plasmode data set can be constructed by spiking specific mRNAs into a real microarray data set [5]. Evaluating whether a particular method correctly detects the spiked mRNAs provides information about the method's ability to detect gene expression. A plasmode data set can also be constructed by using a current data set as a template for simulating new data sets for which some truth is known. Although in early microarray experiments, sample sizes were too small (often only 2 or 3 arrays per treatment condition) to use as a basis for a population model for simulating data sets, larger HDE data sets have recently become publicly available, making their use feasible for simulation experiments.

In this paper, we propose a technique to simulate plasmode data sets from previously produced data. The source-data experiment was conducted at the Center for Nutrient–Gene Interaction (CNGI, www.uab.edu/cngi), at the University of Alabama at Birmingham. We use a data set from this experiment as a template for producing a plasmode null data set, and we use the distribution of effect sizes from the experiment to select expression levels for differentially expressed genes. The technique is intuitively appealing, relatively straightforward to implement, and can be adapted to HDEs in contexts other than microarray experiments. We illustrate the value of plasmodes by comparing 15 different statistical methods for estimating quantities of interest in a microarray experiment, namely the proportion of true nulls (hereafter denoted π_{0}), the false discovery rate (FDR) [6] and a local version of FDR (LFDR) [7]. This type of analysis enables us, for the first time, to compare key omics research tools according to their performance in data that, by definition, are realistic exemplars of the types of data biologists will encounter. The illustrations given here provide some insight into the relative performance characteristics of the 15 methods in some circumstances, but definitive claims regarding uniform superiority of one method over another would require more extensive evaluations over multiple types of data sets.

## Results

### Simulation Design – Producing the Plasmode Data Sets

Steps for plasmode creation that are described herein are relatively straightforward. First, an HDE data set is obtained that reflects the type of experiment for which statistical methods will be used to estimate quantities of interest. Data from a rat microarray experiment at CNGI were used here. Other organisms might produce data with different structural characteristics and methods may perform differently on such data. The CNGI data were obtained from an experiment that used rats to test the pathways and mechanisms of action of certain phytoestrogens [8],[9]. In brief, rats were divided into two large groups, the first sacrificed at day 21 (typically the day of weaning for rats), the second sacrificed at day 50 (the day, corresponding to late human puberty, when rats are most susceptible to chemically induced breast cancer). Each of these groups was subdivided into smaller groups according to diet. At 21 and 50 days, respectively, the relevant tissues from these rat groups were appropriately processed, and gene expression levels were extracted using GCOS (GeneChip Operating Software). We exported the microarray image (*.CEL) files from GCOS and analyzed them with the Affymetrix Package of Bioconductor/R to extract the MAS 5.0 processed expression intensities. The arrays and data were investigated for outliers using Pearson's correlation, spatial artifacts [10] and a deleted residuals approach [11]. It is important to note that only one normalization method was considered, but the methods could be compared on RMA normalized data as well. In fact, comparisons of methods' performances on data from different normalization techniques could be done using the plasmode technique.

Second, an HDE data set that compares effect of a treatment(s) is analyzed and the vector of effect sizes is saved. The effect size used here was a simple standardized mean difference (i.e., a two sample t-statistics) but any meaningful metric could be used. Plasmodes, in fact, could be used to compare the performance of statistical methods when different statistical tests were used to produce the P-values. We chose two sets of HDE data as templates to represent two distributions of effect sizes and two different null distributions. We refer to the 21-day experiment using the control group (8 arrays) and the treatment group (EGCG supplementation, 10 arrays) as data set 1, and the 50-day experiment using the control group (10 arrays) and the treatment group (Resveratrol supplementation, 10 arrays) as data set 2. There were 31,042 genes on each array, and two sample pooled variance t-tests for differential expression were used to create a distribution of P-values. Histograms of the distributions for both data sets are shown in Figure 1.

P-values were computed from the original data using two sample pooled variance t-tests.

The distribution of P-values for data set 1 shows a stronger signal (i.e., a larger collection of very small P-values) than that for data set 2, suggesting either that more genes are differentially expressed or that those that are expressed have a larger magnitude treatment effect. This second step provided a distribution of effects sizes from each data set.

Next, create the plasmode null data set. For each of the HDE data sets, we created a random division of the control group of microarrays into two sets of equal size. One consideration in doing so is that if some arrays in the control group are ‘different’ from others due to some artifact in the experiment, then the null data set can be sensitive to how the arrays are divided into two sets. Such artifacts can be present in data from actual HDEs, so this issue is not a limitation of plasmode use but rather an attribute of it, that is, plasmodes are designed to reflect actual structure (including artifacts) in a real data set. We obtained the plasmode null data set from data set 1 by dividing the day 21 control group of 8 arrays into two sets of 4, and for data set 2 by dividing the control group of 10 arrays into two sets of 5 arrays. Figure 2 shows the two null distributions of P-values obtained using the two sample t-test on the plasmode null data sets. Both null distributions are, as expected, approximately uniform, but sampling variability allows for some deviation from uniformity.

P-values were computed from two sample pooled variance t-tests.

A proportion 1−π_{0} of effect sizes were then sampled from their respective distributions using a weighted probability sampling technique described in the Methods section. What sampling probabilities are chosen can be a tuning parameter in the plasmode creation procedure. The selected effects were incorporated into the associated null distribution for a randomly selected proportion 1−π_{0} of genes in a manner also described in the Methods section. What proportion of genes is selected may depend upon how many genes in an HDE are expected to be differentially expressed. This may determine whether a proportion equal to 0.01 or 0.5 is chosen to construct a plasmode. Proportions between 0.05 and 0.2 were used here as they are in the range of estimated proportions of differentially expressed genes that we have seen from the many data sets we have analyzed.

Finally, the plasmode data set was analyzed using a selected statistical method. We used two sample t-tests to obtain a plasmode distribution of P-values for each plasmode data set because the methods compared herein all analyze a distribution of P-values from an HDE. P-values were declared statistically significant if smaller than a threshold τ. Box 1 summarizes symbol definitions.

When comparing the 15 statistical methods, we used three values of π_{0} (0.8, 0.9, and 0.95) and two thresholds (τ = 0.01 and 0.001). For each choice of π_{0} and threshold τ, we ran B = 100 simulations. All 15 methods provided estimates of π_{0}, 14 provided estimates of FDR, and 7 provided estimates of LFDR. Because the true values of π_{0} and FDR are known for each plasmode data set, we can compare the accuracy of estimates from the different methods.

### Methods for Estimating FDR

There are two basic strategies for estimating FDR, both predicated on an estimated value for π_{0}, the first using equation (1) below, the second using a mixture model approach. Let *P _{K}* =

*M*/

*K*be the proportion of tests that were declared significant at a given threshold, where

*M*and

*K*were defined with respect to quantities in Table 1. Then one estimate for

*FDR*at this threshold is,(1)The mixture model (usually a two-component mixture) approach uses a model of the form,(2)where

*f*is a density,

*p*represents a P-value,

*f*

_{0}a density of a P-value under the null hypothesis,

*f*

_{1}a density of a P-value under the alternative hypothesis, π

_{0}is interpreted as before, and θ a (possibly vector) parameter of the distribution. Since valid P-values are assumed,

*f*

_{0}is a uniform density. LFDR is defined with respect to this mixture model as,(3)FDR is defined similarly except that the densities in (3) are replaced by the corresponding cumulative distribution functions (CDF), that is,(4)where

*F*

_{1}(τ) is the CDF under the alternative hypothesis, evaluated at a chosen threshold τ. (There are different definitions of FDR and the definition in (4) is, under some conditions, the definition of a positive false discovery rate [12]. However, in cases with a large number of genes many of the variants of FDR are very close [13]).

The methods are listed for quick reference in Table 2. Methods 1–8 use different estimates for π_{0} and, as implemented herein, proceed to estimate FDR using equation (1). Method 9 uses a unique algorithm to estimate LFDR and does not supply an estimate of FDR. Methods 10–15 are based on a mixture model framework and estimate FDR and LFDR using equations (3) and (4) where the model components are estimated using different techniques. All methods were implemented using tuning parameter settings from the respective paper or ones supplied as default values with the code in cases where the code was published online.

### Results of the Statistical Methods Tests

First, to compare their differences, we used the 15 methods to analyze the original two data sets, with data set 1 having a “stronger signal” (i.e., lower estimates of π_{0} and FDR). Estimates of π_{0} from methods 3 through 15 ranged from 0.742 to 0.837 for data set 1 and 0.852 to 0.933 for data set 2. (Methods 1 and 2 are designed to control for rather than estimate FDR and are designed to be conservative; hence, their estimates were much closer to 1.) Results of these analyses can be seen in the Supplementary Tables S1 and S2.

Next, using the two template data sets we constructed plasmode data sets in order to compare the performance of the 15 methods for estimating π_{0} (all methods), FDR (all methods except method 9), and LFDR (methods 9–15). Figures 3 and 4 show some results based on data set 2. More results are available in the Figures S1, S2, S3, S4, S5, and S6.

Two cases are shown representing A. π_{0} = 0.8 and B. π_{0} = 0.9, represented by the horizontal line in the two plots A and B, respectively.

Estimates calculated at two thresholds τ = 0.01 (A and B) and 0.001 (C and D) are shown. For the plots of *FDR* estimates, the horizontal line is the mean of the 100 true values of *FDR* in the plasmodes and the horizontal dashed lines are the minimum and maximum. True values of *LFDR* are not known.

Figure 3 shows the distribution of 100 estimates for π_{0} using data set 2 when the true value of π_{0} is equal to 0.8 and 0.9. Methods 1 and 2 are designed to be conservative (i.e., true values are overestimated). With a few exceptions, the other methods tend to be conservative when π_{0} = 0.8 and liberal (the true value is underestimated) when π_{0} = 0.9. The variability of estimates for π_{0} is similar across methods, but some plots show a slightly larger variability for methods 12 and 15 when π_{0} = 0.9.

Figure 4 shows the distribution of estimates for FDR and LFDR at the two thresholds. The horizontal lines in the plots show the mean (solid line) and the minimum and maximum (dashed lines) of the true FDR value for the 100 simulations. A true value for LFDR is not known in the simulation procedure. The methods tend to be conservative (overestimate FDR) when the threshold τ = 0.01 and are more accurate at the lower threshold. Estimates of FDR are more variable for methods 11, 13, and 14 and estimates for LFDR more variable for methods 13 and 14, with the exception of a few unusual estimates obtained from method 9. The high variability of FDR estimates from method 11 may be due to a “less than optimal” choice of the spanning parameter in a numerical smoother (see also Pounds and Cheng [27]). We did not attempt to tune any of the methods for enhanced performance.

## Discussion

Researchers have been evaluating the performance of the burgeoning number of statistical methods for the analysis of high dimensional omic data, relying on a mixture of mathematical derivations, computer simulations, and sadly, often single dataset illustrations or mere *ipse dixit* assertions. Recognizing that the latter two approaches are simply unacceptable approaches to method validation [2] and that the first two suffer from limitations described earlier, an increasing number of investigators are turning to plasmode datasets for method evaluation [28]. An excellent example is the Affycomp website (http://affycomp.biostat.jhsph.edu/) that allows investigators to compare different microarray normalization methods on datasets of known structure. Other investigators have also recently used plasmode-like approaches which they refer to as ‘data perturbation’ [29],[30], yet it is not clear that these ‘perturbed datasets’ can distinguish true from false positives, suggesting greater need for articulation of principles or standards of plasmode generation.

As more high dimensional experiments with larger sample sizes become available, researchers can use a new kind of simulation experiment to evaluate the performance of statistical analysis methods, relying on actual data from previous experiments as a template for generating new data sets, referred to herein as plasmodes. In theory, the plasmode method outlined here will enable investigators to choose *on an empirical basis* the most appropriate statistical method for their HDEs.

Our results also suggest that large, searchable databases of plasmode data sets would help investigators find existing data sets relevant to their planned experiments. (We have already implemented a similar idea for planning sample size requirements in HDEs [31],[32].) Investigators could then use those data sets to compare and evaluate several analytical methods to determine which best identifies genes affected by the treatment condition. Or, investigators could use the plasmode approach on their own data sets to glean some understanding of how well a statistical method works on their type of data. Our results compare the performance of 15 statistical methods as they process the specific plasmode data sets constructed from the CNGI data. Although identifying one uniformly superior method (if there is one) is difficult within the limitations of this one comparison, our results suggest that certain methods could be sensitive to tuning parameters or different types of data sets. A comparison over multiple types of source data sets with different distributions of effects sizes could add the detail necessary to clearly recommend certain methods over others [1].

Other papers have used simulation studies to compare the performance of methods for estimating π_{0} and FDR (e.g., Hsueh et al. [33]; Nguyen [34]; Nettleton et al. [35]). We compared methods that use the distribution of P-values as was done in Broberg [36] and Yang and Yang [37]. Unlike our plasmode approach, most earlier comparison studies used normal distributions to simulate gene expression data and incorporated dependence using a block diagonal correlation structure as in Allison et al [26].

A key implication and recommendation of our paper is that, as data from the growing number of HDEs is made publicly available, researchers may identify a previous HDE similar to one they are planning or have recently conducted and use data from these experiments to construct plasmode data sets with which to evaluate candidate statistical methods. This will enable investigators to choose the most appropriate method(s) for analyzing their own data and thus to increase the reliability of their research results. In this manner, statistical science (as a discipline that studies the methods of statistics) becomes as much an empirical science as a theoretical one.

## Methods

The quantities in Table 1 are those for a typical microarray experiment. Let *N* = *A*+*B* and *M* = *C*+*D* and note that both *N* and *M* will be known and *K* = *N*+*M*. However, the number of false discoveries is equal to an unknown number *C*. The proportion of false discoveries for this experiment is *C/M*. Benjamini and Hochberg [6] defined *FDR* as, *P*(*M*>0) where *I*_{{M>0}} is an indicator function equal to 1 if *M*>0 and zero otherwise. Storey [12] defined the positive *FDR* as . Since *P*(*M*>0)≥1−(1−τ)* ^{K}*, and since

*K*is usually very large,

*FDR*≈

*pFDR*, so we do not distinguish between

*FDR*and

*pFDR*as the parameter being estimated and simply refer to it as

*FDR*with estimates denoted (and ).

Suppose we identify a template data set corresponding to a two treatment comparison for differential gene expression for *K* genes. Obtain a vector, δ, of effect sizes. One suggestion is the usual t-statistic, where the i^{th} component of δ, is given by(5)where *n _{trt}*,

*n*are number of biological replicates in the treatment and control group, respectively,

_{ctrl}*X̅*

_{i}_{,trt},

*X̅*

_{i}_{,ctrl}are the mean gene expression levels for gene

*i*in treatment and control groups, and , is the usual pooled sample variance for the ith gene, where the two sample variances are given by , . In what follows, we will use this choice for δ

*since it allows for effects to be described by a unitless quantity, i.e., it is scaled by the standard error of the observed mean difference*

_{i}*X̅*

_{i}_{,trt}−

*X̅*

_{i}_{,ctrl}for each gene.

For convenience, assume that *n _{ctrl}* is an even number and divide the control group into two sets of equal size. Requiring

*n*≥4 allows for at least two arrays in each set, thus allowing estimates of variance within each of the two sets. This will be the basis for the plasmode “null” data set. There are ways of making this division. Without loss of generality, assume that the first

_{ctrl}*n*/2 arrays after the division are the plasmode control group and the second

_{ctrl}*n*/2 are the plasmode treatment group. Specify a value of π

_{ctrl}_{0}and specify a threshold, τ, such that a P-value ≤τ is declared evidence of differential expression. Execute the following steps.

- Sample without replacement (1−π
_{0})*K*(rounding down to the nearest integer) from the integers 1,…,*K*. Denote this set as*S*^{*}. This set will denote those genes that will be differentially expressed. - Sample (1−π
_{0})*K*(rounding down) effect sizes without replacement from the vector δ with components given by equation (5), where the ith component is selected with a weighted probability, . Denote this vector as δ^{*}. This will be the set of effect sizes used to differentially express genes. The weighted probability sampling allows for the fact that the original vector δ contains effects for both differentially expressed genes and genes corresponding to true null hypotheses. Thus larger effects are more likely to be selected, but the chance remains for very small effects to be selected as well. The weighted probabilities could be modified to allow for a higher (or lower) probability of large effects being sampled and, as such, could be a tuning adjustment in a plasmode simulation procedure. - For each expression level in the plasmode treatment group and for each gene,
*j*, in the set*S*^{*}, add the amount where*S*is the sample standard deviation for the_{j,ctrl}*j*th gene in the original control group. This is one plasmode data set with a null reference data set obtained within the control group but effect sizes borrowed from the full microarray experiment. - Conduct a statistical test for differentially expressed genes on the plasmode data set and record the distribution of P-values. Determine which genes have P-values ≤τ.
- Note that π
_{0}and the set*S*^{*}are known, so a true value of FDR for this data set is available. This true value will change with each simulated data set since the set*S*^{*}and the vector δ^{*}will be different in each simulation. - Apply a statistical method that estimates π
_{0}, FDR, LFDR and other quantities of interest. Estimates of FDR and LFDR are computed at a preset threshold τ. Some methods compute these estimates at the observed P-values in which case we interpolate the estimates computed at the two nearest P-values above and below τ. - Repeat steps 1–6 B times. Record summary statistics such as the mean, standard deviation, and range of the true FDR over the B plasmodes, and the summary statistics from the estimates obtained from the statistical method that is being evaluated.
- Choose another threshold τ and/or another value of π
_{0}and repeat for a new simulation case.

One can then obtain another data set and repeat the entire process to evaluate a method on a different type of data, perhaps from a different organism having a different null distribution, or a different treatment type giving a different distribution of effect sizes, δ. Alternatively, one might choose to randomly divide the control group again and repeat the entire process. This would help assess how differences in arrays within a group or possible correlation structure might affect results from a method. If some of the arrays in the control group have systematic differences among them (e.g., differences arising from variations in experimental conditions—day, operator, technology, etc.), then the null distribution can be sensitive to the random division of the original control group into the two plasmode groups, particularly if *n _{ctrl}* is small.

**Box 1:** Notation for parameters used in modeling high dimensional data

- π
_{0}= A true proportion of genes for which there is no differential expression. This value is controlled by the experimenter in a simulation study. - 1−π
_{0}= the proportion of genes that are truly differentially expressed. - πˆ
_{0}= An estimate of π_{0}obtained using a statistical method on data from an HDE. - τ = A threshold set by the investigator below which P-values are declared statistically significant.

## Supporting Information

### Figure S1.

Boxplots plasmode simulations dataset 1.

doi:10.1371/journal.pgen.1000098.s001

(0.02 MB PDF)

### Figure S2.

Boxplots plasmode simulations dataset 2.

doi:10.1371/journal.pgen.1000098.s002

(0.02 MB PDF)

## Acknowledgments

The authors thank Deanna Calvert for editing assistance and Vinodh Srinivasasainagendra for graphics assistance.

## Author Contributions

Conceived and designed the experiments: GG DA. Performed the experiments: QX. Analyzed the data: LY. Wrote the paper: GG SB GP DA.

## References

- 1. Allison DB, Cui X, Page GP, Sabripour M (2006) Microarray data analysis: From disarray to consolidation and consensus. Nat Rev Gen 7: 55–65.
- 2. Mehta TS, Tanik M, Allison DB (2004) Towards Sound Epistemological Foundations of Statistical Methods for High-Dimensional Biology. Nat Gen 36: 943–947.
- 3. Cattell RB, Jaspars J (1967) A general plasmode (No. 30-10-5-2) for factor analytic exercises and research. Multivariate Behav Res 67: 1–212.
- 4. Singhal S, Kyvernitis CG, Johnson SW, Kaisera LR, Leibman MN, Albelda SM (2003) Microarray data simulator for improved selection of differentially expressed genes. Cancer Biol Ther 2(4): 383–391.
- 5. Mehta TS, Zakharkin SO, Gadbury GL, Allison DB (2006) Epistemological issues in omics and high-dimensional biology: Give the people what they want. Physiol Genomics 28: 24–32.
- 6. Benjamini Y, Hochberg Y (1995) Controlling the false discovery rate: a practical and powerful approach to multiple testing. J R Stat Soc Ser B 57: 289–300.
- 7. Efron B (2004) Large-scale simultaneous hypothesis testing: The choice of a null hypothesis. J Am Stat Assoc 99: 96–104.
- 8. Whitsett T, Lamartiniere CA (2006) Genistein and resveratrol: mammary cancer chemoprevention and mechanisms of action in the rat. Expert Rev Anticancer Ther 6(12): 1699–706.
- 9. Whitsett T, Carpenter DM, Lamartiniere CA (2006) Resveratrol, but not EGCG, in the diet suppresses DMBA-induced mammary cancer in rats. J Carcinog 5: 15.
- 10. Kim K, Page GP, Beasley TM, Barnes S, Scheirer KE, Allison DB (2006) A proposed metric for assessing the measurement quality of individual microarrays. BMC Bioinformatics 7: 35.
- 11. Persson S, Wei H, Milne J, Page GP, Somerville CR (2005) Identification of genes required for cellulose synthesis by regression analysis of public microarray data sets. Proc Natl Acad Sci USA 102(24): 8633–8638.
- 12. Storey JD (2002) A direct approach to false discovery rates. J R Stat Soc Ser B 64: 479–498.
- 13. Tsai C, Hsueh H, Chen JJ (2003) Estimation of false discovery rates in multiple testing: Application to gene microarray data. Biometrics 59: 1071–1081.
- 14. Benjamini Y, Hochberg Y (2000) On the adaptive control of the false discovery rate in multiple testing with independent statistics. J Ed Behav Stat 25: 60–83.
- 15. Mosig MO, Lipkina E, Khutoreskayaa G, Tchourzynaa E, Sollera M, Friedmanna A (2001) A whole genome scan for quantitative trait loci affecting milk protein percentage in Israeli-Holstein cattle, by means of selective milk DNA pooling in a daughter design, using an adjusted false discovery rate criterion. Genetics 157: 1683–1698.
- 16. Storey JD, Tibshirani R (2003) Statistical significance for genomewide studies. Proc Nat Acad Sci 100: 9440–9445.
- 17. Storey JD, Taylor JE, Siegmund D (2004) Strong control, conservative point estimation and simultaneous conservative consistency of false discovery rates: a unified approach. J R Stat Soc Ser B 66: 187–205.
- 18. Schweder T, Spjøtvoll E (1982) Plots of p-values to evaluate many tests simultaneously. Biometrika 69: 493–502.
- 19. Dalmasso C, Broët P, Moreau T A (2005) A simple procedure for estimating the false discovery rate. Bioinformatics 21: 660–668.
- 20. Langaas M, Lindqvist BH, Ferkingstad E (2005) Estimating the proportion of true null hypotheses, with application to DNA microarray data. J R Stat Soc Ser B 67: 555–572.
- 21. Scheid S, Spang RA (2004) Stochastic downhill search algorithm for estimating the local false discovery rate. IEEE/ACM Trans Compu Biol Bioinform 1: 98–108.
- 22. Pounds S, Morris SW (2003) Estimating the occurrence of false positive and false negative in microarray studies by approximating and partitioning the empirical distribution of p-values. Bioinformatics 19: 1236–1242.
- 23. Pounds S, Cheng C (2004) Improving false discovery rate estimation. Bioinformatics 20: 1737–1745.
- 24. Liao JG, Lin Y, Selvanayagam ZE, Shih WJ (2004) A mixture model for estimating the local false discovery rate in DNA microarray analysis. Bioinformatics 20: 2694–2701.
- 25. Broberg P (2004) A new estimate of the proportion unchanged genes in a microarray experiment. Genome Biology 5: P10.
- 26. Allison DB, Gadbury GL, Heo M, Fernanfez JR, Lee C, Prolla TA, Weindruch RA (2002) A mixture model approach for the analysis of microarray gene expression data. Compu Stat Data Anal 39: 1–20.
- 27. Pounds S, Cheng C (2006) Robust estimation of the false discovery rate. Bioinformatics 22: 1979–1987.
- 28. Mavromatis K, Ivanova N, Barry K, Shapiro H, Goltsman E, et al. (2007) Use of simulated data sets to evaluate the fidelity of metagenomic processing methods. Nat Methods 4(6): 495–500.
- 29. Yu J, Pressoir G, Briggs WH, Bi IV, Yamasaki M, et al. (2006) A unified mixed-model method for association mapping that accounts for multiple levels of relatedness. Nat Genetics 38(2): 203–208.
- 30. Zhao K, Aranzana MJ, Kim S, Lister C, Shindo C, et al. (2007) An arabidopsis example of association mapping in structured samples. PLoS Genet 3(1): e4.
- 31. Gadbury GL, Page GP, Edwards J, Kayo T, Prolla TA, Weindruch R, Permana PA, Mountz JD, Allison DB (2004) Power and sample size estimation in high dimensional biology. Stat Methods Med Res13: 325–338.
- 32. Page GP, Edwards JW, Gadbury GL, Yelisetti P, Wang J, Trivedi P, Allison DB (2006) The PowerAtlas: a power and sample size atlas for microarray experimental design and research. BMC Bioinformatics 7: 84.
- 33. Hsueh H, Chen JJ, Kodell RL (2003) Comparison of methods for estimating the number of true null hypotheses in multiplicity testing. J Biopharm Stat 13: 675–689.
- 34. Nguyen D (2004) On estimating the proportion of true null hypotheses for false discovery rate controlling procedures in exploratory DNA microarray studies. Comp Stat Data Anal 47: 611–637.
- 35. Nettleton D, Hwang GJT, Caldro RA, Wise RP (2006) Estimating the number of true null hypotheses from a histogram of p-values. J Agr Biol Environ Stat 337–356.
- 36. Broberg PA (2005) A comparative review of estimates of the proportion unchanged genes and the false discovery rate. BMC Bioinformatics 6: 199.
- 37. Yang JJ, Yang MC (2006) An improved procedure for gene selection from microarray experiments using false discovery rate criterion. BMC Bioinformatics 7: 15.