Storeys Q ValuesEdit
Storey's q-values are a practical tool for interpreting results in large-scale hypothesis testing, offering an approach that balances discovery with a principled awareness of false positives. Named after statistician John D. Storey and developed with colleagues, including Robert Tibshirani, this framework is particularly influential in fields that generate vast numbers of statistical tests, such as genomics and high-throughput biology. At its core, a q-value is the minimal false discovery rate (FDR) at which a given result would be deemed significant, providing a way to rank findings by their reliability rather than by p-values alone.
Put simply, storey's q-values take the p-values from many tests and translate them into estimates of the proportion of false positives one would accept if a given result were declared significant. This makes it easier to identify true signals in datasets with thousands or millions of tests, where traditional p-value cutoffs can be either too lax or too conservative. The method is designed to work in conjunction with the idea of a false discovery rate false discovery rate and has become a staple in tools for multiple hypothesis testing multiple hypothesis testing.
Origins and definitions
The concept emerged in the early 2000s as researchers sought a more powerful alternative to conventional familywise error rate control in settings with many simultaneous tests. Storey and his collaborators formalized the notion of a q-value as the smallest FDR at which a particular hypothesis test would be called significant. This creates a local measure of significance that adapts to the distribution of p-values across the entire study, rather than applying a single universal threshold to all tests.
- John D. Storey is the primary figure behind the development of q-values, and his work is often cited together with Robert Tibshirani for its impact on genomics and other data-rich disciplines.
- The approach rests on estimating the proportion of true null hypotheses, denoted pi0, among the tested hypotheses. A common way to estimate pi0 uses the distribution of p-values, particularly the tail of the p-value spectrum above a chosen cutoff lambda, where true nulls are expected to behave like a uniform distribution on [0,1] pi0.
- After pi0 is estimated, q-values are computed from the ordered p-values p_(1) ≤ p_(2) ≤ … ≤ p_(m). The basic idea is to adjust each p_(i) by a factor that accounts for the number of tests and the estimated null proportion, and then enforce a monotone nonincreasing sequence so that higher p-values do not end up with smaller q-values. This yields a set of q-values q_(i) that reflect the minimum FDR at which each observed result could be considered significant.
Estimation of pi0 and the lambda parameter
A key practical step in storey's approach is estimating pi0, the proportion of tests for which the null hypothesis is true. Since this quantity is rarely known in advance, practitioners estimate it from the data. A common method involves selecting a threshold lambda in [0,1) and using the proportion of p-values exceeding lambda to infer pi0. The estimate typically takes the form:
pi0_hat(lambda) = (number of p-values > lambda) / (m * (1 - lambda))
where m is the total number of tests. Since pi0 is unknown, analysts often explore several lambda values and either choose a fixed one (for example, lambda around 0.5 to 0.95) or adopt a smoothing rule to stabilize the estimate. The choice of lambda and any smoothing strategy can influence the resulting q-values, especially in studies with a substantial fraction of true alternatives or with complex dependence structures among tests.
Computing q-values in practice
With pi0_hat in hand, q-values are derived from the ordered p-values. A standard procedure is:
- Sort p-values: p_(1) ≤ p_(2) ≤ … ≤ p_(m).
- For each i, compute a provisional value t_i = (pi0_hat * m * p_(i)) / i.
- Enforce monotonicity by setting q_(m) = min(t_m, 1) and, for i from m-1 down to 1, q_(i) = min(t_i, q_(i+1)).
- The resulting q-values q_(i) represent the minimum FDR at which the i-th test would be called significant.
In practice, one reports q-values alongside p-values so that researchers can decide on an acceptable FDR threshold (e.g., q* = 0.05) and declare all tests with q_(i) ≤ q* as discoveries. Software implementations in popular statistical environments, such as the Bioconductor ecosystem, provide ready-to-use tools for estimating pi0, computing q-values, and performing downstream analyses on large datasets Bioconductor and R (programming language).
Advantages and limitations
Advantages
- Increased power: By estimating the null proportion and adapting to the data, storey's q-values often yield more discoveries than rigid p-value cutoffs, especially when many tests are truly null.
- Direct FDR interpretation: A q-value gives a clear statement about the expected proportion of false positives among results declared significant at that threshold.
- Compatibility with high-throughput workflows: The method is well-suited to modern experiments where thousands to millions of hypotheses are tested.
Limitations and cautions
- pi0 estimation sensitivity: If the estimation of pi0 is unstable (e.g., due to peculiar p-value distributions or strong dependence among tests), q-values may be biased.
- Dependence structure: The theoretical guarantees for FDR control under storey-type q-values rely on certain dependence conditions among tests. Widely differing dependencies can affect how well FDR is controlled in practice.
- Interpretation risks: As with any FDR-based metric, q-values describe an expected rate of false discoveries, not the certainty that a particular finding is true. Misinterpretation can lead to overconfidence if the underlying assumptions are not considered.
- Comparison with alternative approaches: Some researchers prefer the local false discovery rate (lfdr) perspective or the original Benjamini–Hochberg procedure in certain settings. Each method has tradeoffs in terms of interpretation and power under various data-generating conditions.
Applications and impact
Storey's q-values have become a standard tool in fields that routinely conduct large-scale hypothesis testing. In genomics, researchers use q-values to prioritize candidate genes or genomic regions in microarray studies, RNA-sequencing analyses, and genome-wide association studies. The approach also appears in proteomics, metabolomics, neuroscience imaging, and other data-intensive disciplines where balancing discovery with error control is crucial.
- In the context of gene expression analysis, q-values help researchers separate true biological signals from noise when thousands of genes are tested for differential expression. See gene expression studies and related methodologies.
- In RNA-Seq and other high-throughput assays, q-values underpin many published results that claim robust findings while acknowledging the risk of false positives.
- Methodological discussions around q-values often feature comparisons with the Benjamini–Hochberg procedure, as both aim to control FDR but with different underlying estimators and assumptions Benjamini–Hochberg procedure.
Controversies and debates
As with any statistical tool designed for broad applicability, storey's q-values have sparked ongoing discussion about best practices. Key points in the discourse include:
- When to rely on pi0 estimation: Critics occasionally argue that pi0 estimates can be unstable in small samples or when the signal fraction is substantial, potentially inflating false positives in some contexts.
- Dependence among tests: The robustness of FDR control under complex dependence patterns is an active area of methodological research. In some datasets, strong positive or negative dependence can affect the advertised error rates.
- Practical interpretation: Some practitioners emphasize that q-values are a guide for discovery rather than a direct measure of a single hypothesis’s truth value, requiring careful communication in publications and policy decisions.
- Alternative perspectives: Proponents of alternative error rate concepts, such as the local false discovery rate (lfdr), argue for frameworks that model the distribution of test statistics more explicitly. The choice between approaches often depends on the scientific question, data characteristics, and tolerance for false positives.
From a pragmatic standpoint, storey's q-values are valued for offering a transparent, data-driven path to prioritize findings in large-scale studies while keeping false positives under explicit control. In many real-world analyses, they complement traditional p-value reporting by providing an interpretable measure of reliability that aligns with the goals of substantial discovery without sacrificing rigor.